File operation

I am beginer in Labview. I have two queries.
1. I want to count number of lines in a file and display it in the front panel. How?
2. How to identify white spaces in a file (tab, linefeed etc.)

Try this in 7.0 format.
Attachments:
lines and whitespace.vi ‏33 KB

Similar Messages

  • Saving master page in search template throws error "UserAgent not available, file operations may not be optimized"

    Hi Folks,
    I was trying to save basic search template master page "seattle.master" after making change to the template.
    I have added just "CompanyName" folder and update the line below in seattle.master.
    Change is this : <SharePoint:CssRegistration Name="Themable/CompanyName/corev15.css" runat="server"/>
    When I save it, and refresh page on browser, it shows "Something went wrong" error.
    ULS says the following error : "UserAgent not Available, file operation may not be optimized"
    Pls let us know if there is a solution.
    Any help Much appreciated !
    Thanks,
    Sal
    

    Hi Salman,
    Thanks for posting this issue, 
    Just remove this below given tag and check out. It might be happened that your control is conflicting with others. 
    Also, browse the below mentioned URL for more details
    http://social.msdn.microsoft.com/Forums/office/en-US/b32d1968-81f1-42cd-8f45-798406896335/how-apply-custom-master-page-to-performance-point-dashboard-useragent-not-available-file?forum=sharepointcustomization
    I hope this is helpful to you. If this works, Please mark it as Answered.
    Regards,
    Dharmendra Singh (MCPD-EA | MCTS)
    Blog : http://sharepoint-community.net/profile/DharmendraSingh

  • ACS v5.2 - Unable to update User integer attributes through File Operations

    Hi,
         I have created some internal users on ACS v5.2 and added some Unsigned Integer attributes for each user. I am trying to do a bulk update of these integer attributes using the File Operation facility. However no matter what number I put on the import template it doesn't get updated and displays a "0" in the user config.
    The import template is validated successfully with no errors and also the string attributes are updated correctly.
    There is a sort of work around of deleting the users and adding them back in with the updated values. But this is not feasible as it would reset their passwords. I have also tried saving the csv file in Open Ofifce instead of Excel
    Has any one else come across this problem?
    (I am unable to see this issue in the Release notes or Bug tool kit although there is a similar issue when updating devices in CSCth68051)

    Hi,
         Thanks for the reply. I have managed to recreate the problem to show you but it is a bit more complicated than I first thought. The problem only occurs when the integer attributes are added after the user is created.
    I created a dummy user. The MTL and TLS attributes were present before the user was added. I then added the XXX and ZZZ attributes afterwards and assigned them default values. The default values show up in the GUI config.
    However when I export the database to a csv file only the values of the MTL and TLS attributes show up in the export file:
    I then downloaded an import template and updated the integer values for TLS,MTL, XXX and ZZZ for the dummy user:
    The file imports successfully with no errors. However, when I display the user config only the MTL and TLS attributes have changed. The XXX and ZZZ attributes have stayed the same.
    I thought it might be because I was assigning a default value of 0 to the new attributes but I assigned ZZZ a default value of 1 and the same thing occurred.

  • USB Drive Resets during file copy, interrupting file operations

    Fresh Arch Linux install, working USB thumb drive (Windows, previous Arch Linux setups, etc.), formatted FAT32.
    Seems to randomly reset, eject, and auto-re-mount during extensive file operations (i.e. copying 4gb worth of 800 files from the USB thumb drive to disk).
    I've tried various USB ports to no avail; no external hub being used.
    Thoughts? Not sure where to begin diagnosing:
    dmesg:
    usb 1-1: configuration #1 chosen from 1 choice
    scsi17 : SCSI emulation for USB Mass Storage devices
    usb-storage: device found at 15
    usb-storage: waiting for device to settle before scanning
    scsi 17:0:0:0: Direct-Access OCZ RALLY2 1100 PQ: 0 ANSI: 0 CCS
    sd 17:0:0:0: Attached scsi generic sg2 type 0
    usb-storage: device scan complete
    sd 17:0:0:0: [sdb] 31326208 512-byte logical blocks: (16.0 GB/14.9 GiB)
    sd 17:0:0:0: [sdb] Write Protect is off
    sd 17:0:0:0: [sdb] Mode Sense: 43 00 00 00
    sd 17:0:0:0: [sdb] Assuming drive cache: write through
    sd 17:0:0:0: [sdb] Assuming drive cache: write through
    sdb: sdb1
    sd 17:0:0:0: [sdb] Assuming drive cache: write through
    sd 17:0:0:0: [sdb] Attached SCSI removable disk
    usb 1-1: reset high speed USB device using ehci_hcd and address 15
    sd 17:0:0:0: [sdb] Unhandled error code
    sd 17:0:0:0: [sdb] Result: hostbyte=0x07 driverbyte=0x00
    end_request: I/O error, dev sdb, sector 7589504
    sd 17:0:0:0: [sdb] Unhandled error code
    sd 17:0:0:0: [sdb] Result: hostbyte=0x07 driverbyte=0x00
    end_request: I/O error, dev sdb, sector 7589744
    sd 17:0:0:0: [sdb] Unhandled error code
    sd 17:0:0:0: [sdb] Result: hostbyte=0x07 driverbyte=0x00
    end_request: I/O error, dev sdb, sector 7589776
    usb 1-1: USB disconnect, address 15
    FAT: bread failed in fat_clusters_flush
    FAT: FAT read failed (blocknr 3805)
    FAT: Directory bread(block 24147944) failed
    FAT: FAT read failed (blocknr 3818)
    <snip lots of repeats>
    FAT: FAT read failed (blocknr 4032)
    FAT: bread failed in fat_clusters_flush
    FAT: bread failed in fat_clusters_flush
    FAT: bread failed in fat_clusters_flush
    usb 1-1: new high speed USB device using ehci_hcd and address 16
    usb 1-1: configuration #1 chosen from 1 choice
    scsi18 : SCSI emulation for USB Mass Storage devices
    usb-storage: device found at 16
    usb-storage: waiting for device to settle before scanning
    scsi 18:0:0:0: Direct-Access OCZ RALLY2 1100 PQ: 0 ANSI: 0 CCS
    sd 18:0:0:0: Attached scsi generic sg2 type 0
    usb-storage: device scan complete
    sd 18:0:0:0: [sdb] 31326208 512-byte logical blocks: (16.0 GB/14.9 GiB)
    sd 18:0:0:0: [sdb] Write Protect is off
    sd 18:0:0:0: [sdb] Mode Sense: 43 00 00 00
    sd 18:0:0:0: [sdb] Assuming drive cache: write through
    sd 18:0:0:0: [sdb] Assuming drive cache: write through
    sdb: sdb1
    sd 18:0:0:0: [sdb] Assuming drive cache: write through
    sd 18:0:0:0: [sdb] Attached SCSI removable disk
    lsusb -v for the drive:
    Bus 001 Device 016: ID 0325:ac02 OCZ Technology Inc ATV Turbo / Rally2 Dual Channel USB 2.0 Flash Drive
    Device Descriptor:
    bLength 18
    bDescriptorType 1
    bcdUSB 2.00
    bDeviceClass 0 (Defined at Interface level)
    bDeviceSubClass 0
    bDeviceProtocol 0
    bMaxPacketSize0 64
    idVendor 0x0325 OCZ Technology Inc
    idProduct 0xac02 ATV Turbo / Rally2 Dual Channel USB 2.0 Flash Drive
    bcdDevice 11.00
    iManufacturer 1 OCZ Technology
    iProduct 2 RALLY2
    iSerial 3 AA04012700275633
    bNumConfigurations 1
    Configuration Descriptor:
    bLength 9
    bDescriptorType 2
    wTotalLength 32
    bNumInterfaces 1
    bConfigurationValue 1
    iConfiguration 0
    bmAttributes 0x80
    (Bus Powered)
    MaxPower 300mA
    Interface Descriptor:
    bLength 9
    bDescriptorType 4
    bInterfaceNumber 0
    bAlternateSetting 0
    bNumEndpoints 2
    bInterfaceClass 8 Mass Storage
    bInterfaceSubClass 6 SCSI
    bInterfaceProtocol 80 Bulk (Zip)
    iInterface 0
    Endpoint Descriptor:
    bLength 7
    bDescriptorType 5
    bEndpointAddress 0x81 EP 1 IN
    bmAttributes 2
    Transfer Type Bulk
    Synch Type None
    Usage Type Data
    wMaxPacketSize 0x0200 1x 512 bytes
    bInterval 255
    Endpoint Descriptor:
    bLength 7
    bDescriptorType 5
    bEndpointAddress 0x02 EP 2 OUT
    bmAttributes 2
    Transfer Type Bulk
    Synch Type None
    Usage Type Data
    wMaxPacketSize 0x0200 1x 512 bytes
    bInterval 255
    Device Qualifier (for other device speed):
    bLength 10
    bDescriptorType 6
    bcdUSB 2.00
    bDeviceClass 0 (Defined at Interface level)
    bDeviceSubClass 0
    bDeviceProtocol 0
    bMaxPacketSize0 64
    bNumConfigurations 1
    Device Status: 0x0000
    (Bus Powered)
    Last edited by lieut_data (2009-12-26 17:39:56)

    Jithin wrote:
    When ever I copy some files or create a file in my external USB drive which is a FAT32 one. file permissions are not preserved.
    From <http://support.apple.com/kb/HT3764>:
    "The MS-DOS (FAT32) file system format does not support permissions, file owners, and groups. Such permissions are synthesized on Mac OS X with some default permissions. Because of this all files will have the same permissions […]"
    So it's not "how it work in mac", but it's a limitation of the FAT32 file system.

  • URGENT: File Adapter List Files operation Issue

    Hi All,
    we are using List files operation in one of the SOA composite which lists all files available in the directory. what we observed files are not listing as for the timestamps.
    is there any property to list all files ascending or descending based on time stamp?. we tried with ListSorter property which is suggested by Oracle,but it works for only INBOUND operations. [http://docs.oracle.com/cd/E23943_01/integration.1111/e10231/adptr_file.htm#BABBIJCJ]
    Any suggestions will be greatly appreciated.

    Hi,
    You can try 2 options:
    1. You would need to capture/collect all the file names, you might have to use BPM and create a separate interface.
    2. You can also pick up those files from the archive directory using FTP and push them using mail adapter.
    Regards,
    Pavan

  • Unable to run a Batch File Operating System Command

    Using XI 3.0, I am unable to run a Batch File Operating System Command After Message Processing.
    My Batch file:
    gpg -se -r BOA3RSKY --armor --passphrase-fd 0 %1 < C:\Progra~1\GNU\GnuPG\gpgin
    My Command Line (ID scenario)
    exec "cmd.exe /c C:\Progra~1\GNU\GnuPG\boagpg.bat %F"
    If I execute
    exec "cmd.exe /c type C:\Progra~1\GNU\GnuPG\boagpg.bat >xis.txt"
    It displays the contents of boagpg.bat file in xis.txt.
    I just don't understand why when I run the batch file, I would expect an %F.asc encrypted file in the same directory as the %F unencrypted file.
    Any ideas?
    or will I need Basis to create commands that will allow me to run GPG from XI Command Line?

    Check this links if its helpful
    http://help.sap.com/saphelp_nw04/helpdata/en/bb/c7423347dc488097ab705f7185c88f/frameset.htm
    /people/sap.user72/blog/2004/01/30/command-line-help-utility
    Check this thread a similar problem
    Process Integration (PI) & SOA Middleware
    Note 841704 - XI File & JDBC Adapter: Operating system command
    http://service.sap.com/sap/support/notes/841704
    Try to see the below links
    /people/michal.krawczyk2/blog/2005/08/17/xi-operation-system-command--error-catching
    OS Command on FTP
    OS command line script - Need help
    FTP - Run OS Command before file processing
    Note: reward points if solution found helpfull
    Regards
    Chandrakanth.k

  • ORA-29283 invalid file operation

    NLSRTL      10.2.0.5.0     Production
    Oracle Database 10g Enterprise Edition      10.2.0.5.0     64bi
    PL/SQL      10.2.0.5.0     Production
    TNS for IBM/AIX RISC System/6000:      10.2.0.5.0     Productio
    I am trying to get the content of a trace file generated for me.
    Because I don't have privileges to log on the server and copy the trace file for me directly with some os user, I am doing the following:
    1. I alter my session trace identifier to easier identify the trace file
    ALTER SESSION SET TRACEFILE_IDENTIFIER = 'Func01';2. I invoke DBMS_MONITOR
    3. I run the procedure I want to monitor.
    4. I disable the monitoring by calling DBMS_MONITOR
    5. At this point I run the following query to identify my trace file:
    select u_dump.value || '/' || instance.value || '_ora_' || v$process.spid || nvl2(v$process.traceid, '_' || v$process.traceid, null ) || '.trc'"Trace File"
    from V$PARAMETER u_dump
    cross join V$PARAMETER instance
    cross join V$PROCESS
    join V$SESSION on v$process.addr = V$SESSION.paddr
    where 1=1
       and u_dump.name = 'user_dump_dest'
       and instance.name = 'instance_name'
       and V$SESSION.audsid=sys_context('userenv','sessionid');It gives me: /ORACLE/MYDB/trace/MYDB_ora_3616822_Func01.trc
    I have created directory in advanced on the path where the traces are stored:
    CREATE OR REPLACE DIRECTORY trace_dir AS '/ORACLE/MYDB/trace/';
    SELECT * FROM dba_directories WHERE directory_name = 'TRACE_DIR';
    Output:
    OWNER DIRECTORY_NAME DIRECTORY_PATH
    SYS     TRACE_DIR      /ORACLE/MYDB/trace/I don't have rights to grant read, write on TRACE_DIR to my user, as I am not logged with SYS.
    I created a table to store in it the lines from the trace file:
    CREATE TABLE tmp_traces_tab
      callnum NUMBER,
      line NUMBER,
      fileline CLOB
    );Then I run the following PL/SQL block to retrieve the content of the trace and store it in the table T:
    DECLARE
      l_file            UTL_FILE.file_type;
      l_location     VARCHAR2 (100) := 'TRACE_DIR';
      l_filename    VARCHAR2 (255) := 'MYDB_ora_3616822_Func01.trc';
      l_text           VARCHAR2 (32767);
      l_line           NUMBER := 1;
      l_call           NUMBER := 1;
    BEGIN
      -- Open file.
      l_file := UTL_FILE.fopen (l_location, l_filename, 'r', 32767);
      -- Read and output first line.
      UTL_FILE.get_line (l_file, l_text, 32767);
      INSERT INTO tmp_traces_tab (callnum, line, fileline) VALUES (l_call, l_line, l_text);
      l_line := l_line + 1;
      BEGIN
        LOOP
          UTL_FILE.get_line (l_file, l_text, 32767);
           INSERT INTO tmp_traces_tab (callnum, line, fileline) VALUES (l_call, l_line, l_text);
           l_line := l_line + 1;
        END LOOP;
      EXCEPTION
        WHEN NO_DATA_FOUND THEN
          NULL;
      END;
      INSERT INTO tmp_traces_tab (callnum, line, fileline) VALUES (l_call, l_line, l_text);
      l_line := l_line + 1;
      UTL_FILE.fclose (l_file);
    END;
    /And when I run the code I get the error: ORA-29283 invalid file operation.
    Is it possible to a role my user to be able to get the content of the trace files in the directory TRACE_DIR without having explicit READ , WRITE privileges on it?
    My user currently has these roles:
    select * from dba_role_privs where grantee = USER;
    Output:
    U1     OPR_ROLE_LOSS_SNAPSHOT_READER     YES     YES
    U1     RESOURCE     NO     YES
    U1     CONNECT     NO     YES
    U1     DBA     NO     YES
    U1     OPR_ROLE_SUPPORT_USER     YES     YESI know that on another db with different user I hit no errors when doing completely the same (of course the program unit I monitor is different).
    That other user with which I have NO issues has these roles:
    select * from dba_role_privs where grantee = USER;
    Output:
    U2    DBA    NO    YES
    U2    EXEC_SYS_PACKAGES_ROLE    NO    YES
    U2    EXECUTE_CATALOG_ROLE    NO    YES
    U2    CONNECT    NO    YES

    Verdi wrote:
    NLSRTL      10.2.0.5.0     Production
    Oracle Database 10g Enterprise Edition      10.2.0.5.0     64bi
    PL/SQL      10.2.0.5.0     Production
    TNS for IBM/AIX RISC System/6000:      10.2.0.5.0     Productio
    And when I run the code I get the error: ORA-29283 invalid file operation.
    Is it possible to a role my user to be able to get the content of the trace files in the directory TRACE_DIR without having explicit READ , WRITE privileges on it?
    My user currently has these roles:
    select * from dba_role_privs where grantee = USER;
    Output:
    U1     OPR_ROLE_LOSS_SNAPSHOT_READER     YES     YES
    U1     RESOURCE     NO     YES
    U1     CONNECT     NO     YES
    U1     DBA     NO     YES
    U1     OPR_ROLE_SUPPORT_USER     YES     YESI know that on another db with different user I hit no errors when doing completely the same (of course the program unit I monitor is different).
    Thanks for posting version alongwith other details.
    TO my knowledge, No you cannot.
    Privileges acquired via a Role are not valid in PL/SQL. You need to have explicit privileges.

  • Laggy file operations windows 7 clients with windows server 2008 R2

    Hope this is posted in the right place..
    Ok so this is driving me insane...
    I have a DNS/file server running Windows Server 2008 R2, and 5 client pc's running windows 7 x64 and x86.
    All the client PC's have extremely laggy file operations when for example right clicking to create a new folder or opening certain files (AutoCAD). Talking about 3-10 second lag for a response.
    However if I do a straight file transfer to test the transfer speed its all normal. Everything is connected @ gigabit network speeds.
    This just seemed to happen one day a few months ago, and has remained ever since. Before that (couple years) it was not doing this...
    Anyone have any ideas, I'm at a loss... I've checked the performance monitor on the server nothing seems abnormal when this happens.
    Thank you,

    Hi, 
    Please test to disable all third party applications on file server to see if that is the cause. Sometimes such kind of issues are caused by antivirus program, firewall etc.
    A quick step is to boot into Clean Boot mode as provided in following article:
    http://support.microsoft.com/kb/929135
    If you have any feedback on our support, please send to [email protected]

  • Filestream Creation Unable to Open Physical File Operating System Error 259

    Hey Everybody,
    I have run out of options supporting a customer that is having an error when creating a database with a file stream.  The error displayed is unable to open physical file operating system error 259 (No more data is available).  We're using a pretty
    standard creation SQL script that we aren't having issues with other customers:
    -- We are going to create our data paths for the filestreams.  
    DECLARE @data_path nvarchar(256);
    SET @data_path = (SELECT SUBSTRING(physical_name, 1, CHARINDEX(N'master.mdf', LOWER(physical_name)) - 1)
                      FROM master.sys.master_files
                      WHERE database_id = 1 AND file_id = 1);
    -- At this point, we should be able to create our database.  
    EXECUTE ('CREATE DATABASE AllTables
    ON PRIMARY
        NAME = AllTables_data
        ,FILENAME = ''' + @data_path + 'AllTables_data.mdf''
        ,SIZE = 10MB
        ,FILEGROWTH = 15%
    FILEGROUP FileStreamAll CONTAINS FILESTREAM DEFAULT
        NAME = FSAllTables
        ,FILENAME = ''' + @data_path + 'AllTablesFS''
    LOG ON
        NAME = AllTables_log
        ,FILENAME = ''' + @data_path + 'AllTables_log.ldf''
        ,SIZE = 5MB
        ,FILEGROWTH = 5MB
    GO
    We are using SQL Server 2014 Express.  File streams were enabled on the installation SQL Server.  The instance was created successfully and we are able to connect to the database through SSMS. The user is using an encrypted Sophos. 
    We have tried the following:
    1. Increasing the permissions of the SQL Server server to have full access to the folders.
    2. Attempted a restore of a blank database and it failed.
    There doesn't seem to be any knowledge base articles on this particular error and I am not sure what else I can do to resolve this.  Thanks in advance for any help!

    Hi Ryan,
    1)SQL Server(any version) can't be installed on encrypted drives. Please see a similar scenario in the following link
    https://ask.sqlservercentral.com/questions/115761/filestream-and-encrypted-drives.html
    2)I don't think there is any problem with permissions on the folder, if the user can create a database in the same folder. Am not too sure. Also see the article by
    Jacob on configuring the FILESTREAM for SQL Server that describes how to configure FILESTREAM access level & creating a FILESTREAM enabled database
    Hope this helps,
    Thanks
    Bhanu 

  • Build a tool for controlling a copy file operation base on its file size in any folder of Window Explorer

    Hi all,
    We are developing a tool which will control a copy file operation in any folder of Window Explorer base on file size limit.
    If these copied files are not satisfied the file size condition (its size is larger than a predefined value), the copy/move operation must not be executed.
    My question is that: "Are there any availability support of Windows API/tool for implementing this feature?"
    BR,
    Mr_Le

    Hi Mr_Le,
    Thank you for posting in the MSDN forum.
    Based on your description, I’m afraid that it is not the correct forum for this issue, since this forum is to discuss:
    Visual Studio WPF/SL Designer, Visual Studio Guidance Automation Toolkit, Developer Documentation and Help System, and Visual Studio Editor.
    To help you find the correct development forum for this issue, please tell me the real project type (Winforms or others) and the real development language (C#, VB or others) you want to use.
    If you are not very sure that which kind of project fulfils this requirement, maybe the language development forum would be better for this issue:
    https://social.msdn.microsoft.com/Forums/vstudio/en-US/home?category=vslanguages&filter=alltypes&sort=lastpostdesc
    If there's any concern, please feel free to let me know.
    Best Regards,
    Jack
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • How to install File Operations components in Data Integrator

    Hello Endeca Forum,
    I'm interested in using the File Operations components in Data Integrator, but getting and installing them is surprisingly obscure.
    I'm talking about...
    http://doc.cloveretl.com/documentation/UserGuide/index.jsp?topic=/com.cloveretl.gui.docs/docs/file-operations.html
    The Clover documentation (http://doc.cloveretl.com/documentation/UserGuide/index.jsp?topic=/com.cloveretl.gui.docs/docs/components.html) states...
    "+Note if you cannot see this component category, navigate to Window → Preferences → CloverETL → Components in Palette and tick both checkboxes next to File Operations+"
    However, when I go to Window --> Preferences --> CloverETL --> Components in Palette, I don't even see File Operations. This tells me that the version of Clover packaged with OEID doesn't have the components installed.
    Going to Help --> Check for Updates also didn't yield anything that looked like the File Operations components.
    I'm hoping someone here is familiar with the necessary steps to get these components installed...
    Thanks,
    Jerome

    I see what you're saying Purvesh, thanks for the clarification.
    Again, it's a little challenging to find this out, but there's a Clover support team member who mentions here...
    http://forum.cloveretl.com/viewtopic.php?f=4&t=6428&view=previous
    ... that the File Operations were introduced in the 3.3.0 M3 release.
    Guess this means we're out of luck, which is unfortunate.
    I'm fully aware that the SystemExecute component is an option for file management, but it's not the best fit for what I need to do. In any case, my question has been answered.
    Thanks!
    Jerome

  • Very, very slow file operations

    I needed to make a specific change to an exif field of some images. So what I did was to move identify the files that I needed to change in Lightroom (they were in various folders on my disk) and move them to other folders so they would be easy to operate on. Then I used exiftool to change the field in these files, and then I thought it would be easy to read in the changed data and move them back to their original folders in Lightroom. What I found included:
    I tried to Synchronize Metadata on the folders after changing the exif field. One folder had about 500 images in it, the other about 2,000. After 3 hours, the progress bar had barely moved, and I aborted.
    When I tried to quit Lightroom it said that it was "writing metadata", even tho the synchronization had changed the external data and it should only have been reading metadata. I quit anyways and rebooted.
    After rebooting, and relaunching Lightroom I looked at the "metadata status" column in Grid view. It was (as always, it seems) completely wrong about which files were up to date.
    I was able to "read metadata" on these 2,500 images in "only" about an hour or two. Note that I didn't ask Lightroom to update the cached images, only to read the metadata. Reading metadata at 1,000-2,000 images an hour seems exceptionally slow.
    Now I am trying to move the files back to their correct folders. I dragged about 1,000 raw images (CR2 + xmp) from one folder to another in Lightroom. Three hours later it is about 50% finished according to the progress bar. I'm going to bed and hope it will finish by the morning. I have to move all the other images as well, but because Lightroom won't update the filter bars in Grid View while moving files, I can't start these moves until Lightroom finishes the first.
    Do other people fine absolutely absurd speeds for file operations in Lightroom? Is there any solution other than quitting and rebooting (which doesn't always work)? I hate to quit in the middle of a synchronize, read metadata, or move files operation even when it seems to be barely chugging along, because I am afraid that when I abort the operation, I will somehow leave a file orphaned, or it's metadata out of sync (which happens to me all the time, but I never know why).
    When I am doing other operations I also find that I can use Lightroom for only a few hours before it gets so slow that I quit and reboot, after which operations speed up considerably. This doesn't seem like it should be necessary for a "modern" program and operating system, but in my experience it is.
    I'm using a Macintosh (4 core i7, 16 GB of RAM), Mavericks latest, Lightroom 5.5, and a RAID disk, so hardware shouldn't be limiting. During this last interminable copy operation, it appears that Lightroom is pegged at 100% of a processor (core).

    Sorry, but I don't think your workaround works. I am moving a bunch of files from one folder to various folders (using Lightroom to discriminate between files based on their exif/iptc values), then I change them using an outside program (exiftool), then I put them back in their original folders. This involves having Lightroom read in the new data in the files, which is what is so slow. Reconnecting, if I understand it, is for when you move a folder of files to another location, and Lightroom needs to be informed of that move.
    As I pointed out, the problem is that Lightroom gets progressively and painfully slower the more you use it. But quitting and restarting can be annoying if you are in the middle of an operation because you can lose context (especially if you restart your computer, which seems to help).
    And I do optimize my database nearly daily when using Lightroom.
    I would spend the time to fully document this problem and report it as a bug, but my experience is that Adobe shows no evidence of actually reading my bug reports, let alone responding to them.

  • Invoke a FTP "Get File" operation throws java.lang.ClassCastException:

    Hi there,
    I am working on a very simple sample project that will pull the file from a ftp site. I create a partner link with FTP adpater with "Get File" operation.
    When I deploy and run the SIMPLE project, it throws the following exception.
    I am wondering if any of you experienced the same problem and how to resolve the issue. Your help will be greatly appreciated.
    Here is the error message:
    <messages><input><Invoke_1_Get_InputVariable><part xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" name="Root-Element"><Root-Element xmlns="http://TargetNamespace.com/ftp_service"/>
    </part></Invoke_1_Get_InputVariable></input><fault><bindingFault xmlns="http://schemas.oracle.com/bpel/extension"><part name="summary"><summary>file:/C:/product/10.1.3.1/OracleAS_1/bpel/domains/default/tmp/.bpel_FTPProcess2_1.0_709914d835a81c407ed668866e080b9e.tmp/ftp_service.wsdl [ Get_ptt::Get(Root-Element) ] - WSIF JCA Execute of operation 'Get' failed due to: Could not instantiate InteractionSpec oracle.tip.adapter.ftp.inbound.FTPActivationSpec due to: oracle.tip.adapter.ftp.inbound.FTPActivationSpec; nested exception is:
         java.lang.ClassCastException: oracle.tip.adapter.ftp.inbound.FTPActivationSpec; nested exception is:
         org.collaxa.thirdparty.apache.wsif.WSIFException: Could not instantiate InteractionSpec oracle.tip.adapter.ftp.inbound.FTPActivationSpec due to: oracle.tip.adapter.ftp.inbound.FTPActivationSpec; nested exception is:
         java.lang.ClassCastException: oracle.tip.adapter.ftp.inbound.FTPActivationSpec</summary>
    </part><part name="detail"><detail>org.collaxa.thirdparty.apache.wsif.WSIFException: Could not instantiate InteractionSpec oracle.tip.adapter.ftp.inbound.FTPActivationSpec due to: oracle.tip.adapter.ftp.inbound.FTPActivationSpec; nested exception is:
         java.lang.ClassCastException: oracle.tip.adapter.ftp.inbound.FTPActivationSpec</detail>
    </part></bindingFault></fault></messages>

    Take a look at the stack trace, it shows you exactly where the error is coming from:
    java.lang.ClassCastException: org.theclass.candidate.view.SearchForm
    org.theclass.candidate.view.CandidateListAction.execute(CandidateListAction.java:41)Line 41 in your CandidateListAction.
    Probably it is failing on a cast of your action form.
    Taking a closer look at your struts-config, you are "chaining" actions.
    Your AddCandidateAction uses the candidateForm, and forwards to CandidateListAction.do
    Your CandidateListAction uses the searchForm.
    That will be the cause of your class cast exception.
    <action path="/Add"
      name="candidateForm"
      type="org.theclass.candidate.view.AddCandidateAction"
      validate ="true"
      input="/jsp/addcandidate.jsp">
      <forward name="success" path="/CandidateList.do"/>
    </action>
    <action path="/CandidateList"
    type="org.theclass.candidate.view.CandidateListAction"
    name="searchForm"
    scope="request" >
    <forward name="failure" path="/jsp/list.jsp"/>
    <forward name="success" path="/jsp/candidatelist.jsp"/>
    </action>Check out: http://struts.apache.org/1.x/faqs/newbie.html#chaining
    Solutions?
    - don't chain actions ;-)
    - use the same action form for both actions (possible?)
    - make the actionForward a "redirect" one. That means a new request, and losing any request parameters/attributes but should prevent this class cast exception.
    Hope this helps,
    evnafets

  • Cisco SSL VPN "The following error occurred while attempting the file operation: Unable to view the contents of the Domain/Workgroup. "

    Hey People, 
    Ive set up an SSL Clientless VPN on the Cisco 2821. Ive set up WINS, and the NBNS entries in the VPN config. When i log onto the VPN , i can access the file servers by typing in their names in the network fi
    le box, but when i click browse network and select the network name i get the following message
    "The following error occurred while attempting the file operation:
    Unable to view the contents of the Domain/Workgroup. "
    Has anyone come accross this before?
    Im using Windows Server 2008R2 for the DC, Windows Server 2003 R2 For WINS and File Sharing. 
    The connection goes WAN->BROADBANDROUTER>CISCO2821
    Any helo would be much greatly appreciated. 
    Thanks in advance! 

    Please see old threads which discuss the same topic -- http://forums.oracle.com/forums/search.jspa?threadID=&q=An+error+occurred+while+attempting+to+establish+an+Applications+File+Server+connection+with+the+node&objID=c3&dateRange=all&userID=&numResults=15&rankBy=10001
    Thanks,
    Hussein

  • How do I the interpret "Disk file operations I/O" wait event?

    I have a large and very busy batch database. All of a sudden the "Disk file operations I/O" wait event is in the top 5 in AWR.
    The manual page isn't very helpful:
    http://download.oracle.com/docs/cd/E11882_01/server.112/e17110/waitevents003.htm#insertedID40
    Disk file operations I/O
    This event is used to wait for disk file operations (for example, open, close, seek, and resize). It is also used for miscellaneous I/O operations such as block dumps and password file accesses.
    So here is my question: What exactly is going on when I see this wait event? Why doesn't it show up as one of the other I/O events? Can I make it go away? Should I make it go away?
    DR

    sb92075 wrote:
    All of a sudden the "Disk file operations I/O" wait event is in the top 5 in AWR.Top wait event
    In EVERY Top Wait Event list, one wait event will ALWAYS be on top as #1; by definition of the list.
    Simply because any item, even #1, appears on this list does not mean this is a problem & needs to be fixed.
    If the Top Wait Event accounts for only 5 seconds out of a 1 hour sample,
    then reducing it to ZERO won't measurably improve overall application performance.
    The actual Time Waited is required to determine if it is a problem or not.It's taking 20% of time in a 15 minute sample. Anything that takes 20% of deserves to be understood....So: What actually causes it?
    DR

  • VNI-2002 File Operation Error

    I am trying to do an export of a sample database i have running
    on Oracle8i on Windows 2000 professional.
    Whenever I schedule a job, it fails with the message 'VNI-2002 :
    File Operation Error".
    What should I do to elliminate this problem?
    thanks a bunch.

    I have done as follows:
    Verify that sufficient permission is granted to createfiles in the $ORACLE_HOME/network/agent directory
    If connecting from a client, drop the node and rediscover
    I also created an OS user (oraJob) with permissions:
    read/write permissions to the $ORACLE_HOME\Net80 or $ORACLE_HOME/network directory and all of its sub directories, read/write permissions to the $TEMP directory, as described in Oracle Books.
    Updated the preferred credentials with this user on the node.
    Assign IMP_FULL_DATABASE to the DBA role
    Assign all necessary privileges to the IMP_FULL_DATABASE role
    These steps cleared up many things but unfortunately I'm still stuck with a VNI-2002 error, although (according to my logic) I have all the necessary permissions to access the jobq etc. There's almost nothing out in the Internet community relating to the VNI-2002. Apart from sites hosting Oracle Books I have only encountered two (this one posted here) and another via yahoo, complaining about it. None has a suitable explanation, that is one that could guide you onto the real cause.
    So for now I still searching...

Maybe you are looking for

  • Adobe Acrobat 9.0 has stopped working

    I have just downloaded and installed adobe acrobat 9.0 on a Windows Server 2008. Installl process went find. When starting the program it eventually wil give me the error "Adobe Acrobat 9.0 has stopped working". I get an option to find a solution on

  • How do I get a large audio memo from my iphone to my computer?

    My itunes is currently down and so my icloud is not working.

  • Oracle Apps Report

    Hi Folks, There is a AR Reports in oracle Reports which is giving the Data perfectly fine but the user is not able to export the Data properly on to a excel sheet . When he is exporting the data on to a excel all the line comes in a single Cell thus

  • APEX Interactive Report Compute Case with multiple conditions to highlight

    My ultimate is to highligt a row in an interactive report based on two conditions. I didn't see a way to use the highligt feature with two conditions. So I thought I would try to create a computation based on the two conditions. Then use that result

  • Where is snap back in Safari 6?

    When on a Google search,  I see they have removed the snap back feature!  Is it somewhere else?  Is there another quick way to get back to the main search after reading multiple pages of a particular site?  I am using OSX 10.7.4.