KDE: Awfully slow file operations

So I don't know what the deal is, but it's been moving ~118mb (two folders with an iconset each) for 3 minutes now, and is only about 30% done. Both source and destination being an ext4 formatted SSD.
I can see a few kio-file processes running and my CPU is capped at 100% on each core. What in seven hells is dolphin doing? A move operation to the same drive should be near instantaneous.
Extracting the same two archives directly to the destination takes a few seconds at most, as expected. Moving the same files via console is as fast as extracting from the 7z archive.

It turns out I forgot that folder is mounted on my HDD, not the SSD, so it was an actual copy to destination -> delete source operation.
Preview is off, and I disabled the indexer (it wasn't producing any CPU activity anyway), still same problem.
Dolphin is by far the slowest file manager according to my tests:
Setup:
copy 13k icons ext4 SSD -> ext4 HDD, destination folder deleted between runs. Each file manager tested 3 times
Results:
dolphin: 1:23
thunar: 1:17
nautilus: 0:26
windows exlorer: 0:14
pcmanfm: 0:10
command line: 0:02
Double the numbers for move (I guess that's because it has to delete the files as well). Here's hoping they can figure out how to speed it up with the KDE5 dolphin port...
edit: added windows explorer for comparison. It does quite well, surprisingly.
Last edited by Soukyuu (2015-02-05 12:30:44)

Similar Messages

  • Very, very slow file operations

    I needed to make a specific change to an exif field of some images. So what I did was to move identify the files that I needed to change in Lightroom (they were in various folders on my disk) and move them to other folders so they would be easy to operate on. Then I used exiftool to change the field in these files, and then I thought it would be easy to read in the changed data and move them back to their original folders in Lightroom. What I found included:
    I tried to Synchronize Metadata on the folders after changing the exif field. One folder had about 500 images in it, the other about 2,000. After 3 hours, the progress bar had barely moved, and I aborted.
    When I tried to quit Lightroom it said that it was "writing metadata", even tho the synchronization had changed the external data and it should only have been reading metadata. I quit anyways and rebooted.
    After rebooting, and relaunching Lightroom I looked at the "metadata status" column in Grid view. It was (as always, it seems) completely wrong about which files were up to date.
    I was able to "read metadata" on these 2,500 images in "only" about an hour or two. Note that I didn't ask Lightroom to update the cached images, only to read the metadata. Reading metadata at 1,000-2,000 images an hour seems exceptionally slow.
    Now I am trying to move the files back to their correct folders. I dragged about 1,000 raw images (CR2 + xmp) from one folder to another in Lightroom. Three hours later it is about 50% finished according to the progress bar. I'm going to bed and hope it will finish by the morning. I have to move all the other images as well, but because Lightroom won't update the filter bars in Grid View while moving files, I can't start these moves until Lightroom finishes the first.
    Do other people fine absolutely absurd speeds for file operations in Lightroom? Is there any solution other than quitting and rebooting (which doesn't always work)? I hate to quit in the middle of a synchronize, read metadata, or move files operation even when it seems to be barely chugging along, because I am afraid that when I abort the operation, I will somehow leave a file orphaned, or it's metadata out of sync (which happens to me all the time, but I never know why).
    When I am doing other operations I also find that I can use Lightroom for only a few hours before it gets so slow that I quit and reboot, after which operations speed up considerably. This doesn't seem like it should be necessary for a "modern" program and operating system, but in my experience it is.
    I'm using a Macintosh (4 core i7, 16 GB of RAM), Mavericks latest, Lightroom 5.5, and a RAID disk, so hardware shouldn't be limiting. During this last interminable copy operation, it appears that Lightroom is pegged at 100% of a processor (core).

    Sorry, but I don't think your workaround works. I am moving a bunch of files from one folder to various folders (using Lightroom to discriminate between files based on their exif/iptc values), then I change them using an outside program (exiftool), then I put them back in their original folders. This involves having Lightroom read in the new data in the files, which is what is so slow. Reconnecting, if I understand it, is for when you move a folder of files to another location, and Lightroom needs to be informed of that move.
    As I pointed out, the problem is that Lightroom gets progressively and painfully slower the more you use it. But quitting and restarting can be annoying if you are in the middle of an operation because you can lose context (especially if you restart your computer, which seems to help).
    And I do optimize my database nearly daily when using Lightroom.
    I would spend the time to fully document this problem and report it as a bug, but my experience is that Adobe shows no evidence of actually reading my bug reports, let alone responding to them.

  • A slow file IO operation was encountered

    Hola a todos, el fin de semana hubo algunos cambios en los equipos de comunicación y ahora tengo este error en Event Viewer:
    EventID 2170 Source MSExchangeRepl
    A slow file IO operation was encountered on file 'I:\Program Files\Microsoft\Exchange Server\V14\Mailbox\DB02\E03005761DA.log' for copy 'DB02\MXSCVMMBOX02'. The observed latency was 8580.605 ms while performing a 'MissingFileNotification'. This may indicate
    an overloaded system or a storage-related problem.
    Este evento se repite continuamente. Tengo 4 servidores Mailbox los cuales pertenecen a un DAG, las unidades de almacenamiento están presentadas por Netapp conectadas vía iSCSI. El uso de red es mínimo en las tarjetas de red y todas están online y accesibles.
    Alguien tiene alguna idea?

    Translation
    Hi all, the weekend were some changes in
    communications equipment and now I have this error
    in Event Viewer:
    EventID 2170 Source MSExchangeRepl
    A slow file IO operation
    was encountered on file 'I:
    \ Program Files \ Microsoft \ Exchange
    Server \ V14 \ Mailbox \
    DB02 \ E03005761DA.log' for copy 'DB02
    \ MXSCVMMBOX02'. The
    Observed was 8580.605 ms
    latency while performing a 'MissingFileNotification'.
    This May Indicate an overloaded
    system or a storage-related
    problem.
    This event repeats continuously. I have 4
    Mailbox servers which belong to a
    DAG, the storage units are made by
    Netapp connected via iSCSI. Network usage
    is minimal network cards and they are all
    online and accessible.
    Anyone have any idea?
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help.

  • Slow File I/O with newer compilers (SPARCompiler 5 and gcc 3)

    Well it's more like slow File I, as I'm only really testing Input, but anyway, here goes...
    I have a simple program that takes a file of about 250mb and simply loops through the file reading it in chunks of 8,192 bytes.
    Basically I'm timing the read speed to try and figure out why there is such a large discrepancy between the older SPARCompiler 3.0.1 : C++ 4.0.1 and the newer 5.0 as well as gcc 3.0 in reading a file. I realize that the old compiler doesn't support STL and that SPARCompiler 5's .h headers basically point to the STL headers, so there is a difference there, but the numbers I get are around 35s with SC 3.0.1 : C++ 4.0.1 to about 90s with SC 5.0 and 76s with gcc 3.0. I understand the numbers are going to vary depending on the test input file size and the machine you're running on, but the ratio's should stay relatively similar if you do run your own tests. That slowdown is pretty much unacceptable in my opinion and I really want the benefits of a new compiler, am I doing something wrong, is this a known issue??
    Command Lines (I did my best to try with different versions of the OS, but Sun licensing makes it hard!):
    With C++ 4.0.1 (tried on Solaris 2.6 and 2.7)
    CC -fast -o IOtest.CC io_read_tst.cpp
    With C++ 5 (used on Solaris 2.5.1 and 2.8)
    CC -O3 -o IOtest.CC5 io_read_tst.cpp
    With gcc 3.0 (tried on Solaris 2.5.1, 2.6, 2.7, and 2.8)
    gcc -O3 -o IOtest.gcc3 io_read_tst.cpp
    I've got the latest patches for the OS and compilers, I've tried compiling on multiple versions of the OS (Solaris 2.5.1, 2.6, 2.7, and 2.8) and always I get similar results, I know I haven't tested every OS possibility, but you have to re-install the OS to test since the compiler is licensed to only one machine, at least for the SUN Compilers anyway! Any help would be greatly, greatly appreciated, code is as follows:
    #include <iostream.h>
    #include <fstream.h>
    #include <time.h>
    int main( )
    char sBuffer[8192];
    time_t startTime, endTime;
    ifstream TestFile;
    TestFile.open( "/data4/tax101/taxtab" ); // Change this to a large file local to your system (~250mb)
    if( TestFile.fail() )
    return( EXIT_FAILURE );
    time( &startTime );
    while( !TestFile.eof() )
    TestFile.read( sBuffer, 8192 );
    time( &endTime );
    cout << endl << "Read Time: " << endTime - startTime << "s\n";
    return( EXIT_SUCCESS );
    Thanks,
    Lance Beddawi
    [email protected]

    Alright, here is my code reposted:
    // Warning.java
    // Reads student data from a text file and writes data to another text file.
    import java.util.;
    import java.io.;
    public class Warning
    // // Reads student data (name, semester hours, quality points) from a
    // text file, computes the GPA, then writes data to another file
    // if the student is placed on academic warning.
    public static void main (String[] args)
    int creditHrs; // number of semester hours earned
    double qualityPts; // number of quality points earned
    double gpa; // grade point (quality point) average
    Scanner scan=null;
    PrintWriter outFile=null;
    String name, inputName = "students.txt";
    String outputName = "warning.txt";
    try
    // Set up Scanner to input file
    scan=new Scanner(new FileInputStream(inputName));
    // Set up the output file stream
    outFile = new PrintWriter(new FileWriter(outputName));
    // Print a header to the output file
    outFile.println ();
    outFile.println ("Students on Academic Warning");
    outFile.println ();
    // Process the input file, one token at a time
    while (scan.hasNext())
    // Get the credit hours and quality points and
    // determine if the student is on warning. If so,
    // write the student data to the output file.
    name=scan.next();
    creditHrs=scan.nextInt();
    qualityPts=scan.nextDouble();
    gpa=qualityPts/creditHrs;
    if ((gpa < 1.5 && creditHrs < 30) || (gpa < 1.75 && creditHrs < 60) || (gpa < 2.0 && creditHrs >= 60))
    outFile.print(name " ");
    outFile.print(creditHrs " ");
    outFile.print(qualityPts " ");
    outFile.print(gpa);
    //Add a catch for each of the specified exceptions, and in each case
    //give as specific a message as you can
    catch (FileNotFoundException e)
    System.out.println("The file " inputName " was not found.");
    catch (IOException e)
    System.out.println("The I/O operation failed and " outputName + " could not be created.");
    catch (InputMismatchException e)
    System.out.println("The input information was not of the right type.");

  • Disk file operations I/O waits

    Hi,
    I have a production database running 11.2.0.3 on a Windows server.
    It has been running for about a year without problems, but today we experiences big performance problems.
    When I take a look at the AWR report we have "DIsk file operations I/O" waits for 30% of the DBTime.
    Checking the AWR for last week we had no such waits.
    The disks are on SAN. and the SAN guys can not see any problems so far.
    Any suggestions? Why do we suddenly have these kind of waits??
    Thanks!
    Brg
    jio

    954496 wrote:
    Hi,
    I have a production database running 11.2.0.3 on a Windows server.
    It has been running for about a year without problems, but today we experiences big performance problems.
    Quantify "big"?
    Is every SQL slow?
    Even manual SQL issued outside of the application?
    Consider providing us details as enumerated in URL below.
    How do I ask a question on the forums?
    Search for "related documents" as Altavista or Google do it?
    Handle:     954496
    Status Level:     Newbie
    Registered:     Aug 23, 2012
    Total Posts:     12
    Total Questions:     4 (4 unresolved)
    Why do you waste your time & our time here; since you NEVER get answers to your questions?

  • Extremely slow file copying over airport extreme network

    Hi,
    I have an Airport Extreme Base Station (10/100) network, also connected to the internet. When copying files from my MacBook Pro to a USB drive connected to the base station, OR when copying files to another computer (a windows machine) the throughput maxes out at about 700 KB/sec which is 10 times slower than when copying to the USB drive connected directly to my MacBook Pro. Although I have updated everything to the latest and greatest versions, I still have very slow file copy speeds. It's always been this way. I've read dozens of threads for potential fixes but none of them have helped. I am using Activity Monitor to measure the throughput and also timing the copy of a 10GB file. At these speeds, that file takes 4 hours. Useless.
    My base station is within a few feet. I've tried different channels, 2.4GHz, 5GHz, getting rid of WPA, you name it. At first I thought it might be some issue with the USB drive connected to the base station, but copying to another computer on my network is the same. It is very frustrating.
    Any insights would be greatly appreciated.

    Thanks. I thought about buying the TC but I'm worried that if my network is operating slowly, which it appears to be, that I'm not going to get the throughput. I also noticed that my windows machine which is on the same network accesses the internet faster as well. I'm wondering if it is something with my MacBook.
    BTW - I went to school in Pittsburgh. Great great town.

  • Saving master page in search template throws error "UserAgent not available, file operations may not be optimized"

    Hi Folks,
    I was trying to save basic search template master page "seattle.master" after making change to the template.
    I have added just "CompanyName" folder and update the line below in seattle.master.
    Change is this : <SharePoint:CssRegistration Name="Themable/CompanyName/corev15.css" runat="server"/>
    When I save it, and refresh page on browser, it shows "Something went wrong" error.
    ULS says the following error : "UserAgent not Available, file operation may not be optimized"
    Pls let us know if there is a solution.
    Any help Much appreciated !
    Thanks,
    Sal
    

    Hi Salman,
    Thanks for posting this issue, 
    Just remove this below given tag and check out. It might be happened that your control is conflicting with others. 
    Also, browse the below mentioned URL for more details
    http://social.msdn.microsoft.com/Forums/office/en-US/b32d1968-81f1-42cd-8f45-798406896335/how-apply-custom-master-page-to-performance-point-dashboard-useragent-not-available-file?forum=sharepointcustomization
    I hope this is helpful to you. If this works, Please mark it as Answered.
    Regards,
    Dharmendra Singh (MCPD-EA | MCTS)
    Blog : http://sharepoint-community.net/profile/DharmendraSingh

  • ACS v5.2 - Unable to update User integer attributes through File Operations

    Hi,
         I have created some internal users on ACS v5.2 and added some Unsigned Integer attributes for each user. I am trying to do a bulk update of these integer attributes using the File Operation facility. However no matter what number I put on the import template it doesn't get updated and displays a "0" in the user config.
    The import template is validated successfully with no errors and also the string attributes are updated correctly.
    There is a sort of work around of deleting the users and adding them back in with the updated values. But this is not feasible as it would reset their passwords. I have also tried saving the csv file in Open Ofifce instead of Excel
    Has any one else come across this problem?
    (I am unable to see this issue in the Release notes or Bug tool kit although there is a similar issue when updating devices in CSCth68051)

    Hi,
         Thanks for the reply. I have managed to recreate the problem to show you but it is a bit more complicated than I first thought. The problem only occurs when the integer attributes are added after the user is created.
    I created a dummy user. The MTL and TLS attributes were present before the user was added. I then added the XXX and ZZZ attributes afterwards and assigned them default values. The default values show up in the GUI config.
    However when I export the database to a csv file only the values of the MTL and TLS attributes show up in the export file:
    I then downloaded an import template and updated the integer values for TLS,MTL, XXX and ZZZ for the dummy user:
    The file imports successfully with no errors. However, when I display the user config only the MTL and TLS attributes have changed. The XXX and ZZZ attributes have stayed the same.
    I thought it might be because I was assigning a default value of 0 to the new attributes but I assigned ZZZ a default value of 1 and the same thing occurred.

  • Slow Files Copy File Server DFS Namespace

    I have two file servers running on VM both servers are on different physical servers.
    Both connect with dfs namespace.
    The problem part is both servers never have same copy speed.
    Sometime very slow files copy about 1MBps on FS01 and fast copy 12MBps on FS02.
    Sometime fast on FS01 and slow on FS02.
    Sometime both of them slow..
    So as usual I reboot the servers. Doesn't work.
    Then I reboot the DC01 also doesn't work. There is another brother DC02.
    After I reboot DC02, one of the FS become normal and another FS still slow.
    FS01 and FS02 randomly. They never get faster speed together.
    Users never complain slow FS because 1MBps is acceptable for them to open word excel etc.,.
    The HUGE problem is I don't have backup when the slow FS days.
    The problem since two weeks I'm giving up fixing it myself and need help from you expert guys.
    Thanks!
    DC01, DC02, FS01, FS02 (Win 2012 and All VMs)

    Hi,
    Since the slow copy is also occurred when you tried the direct copy from both shared folder, you could enable the disk write cache on the destination server to check the results.
    HOW TO: Manually Enable/Disable Disk Write Caching
    http://support.microsoft.com/kb/259716
    Windows 2008 R2 - large file copy uses all available memory and then tranfer rate decreases dramatically (20x)
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/3f8a80fd-914b-4fe7-8c93-b06787b03662/windows-2008-r2-large-file-copy-uses-all-available-memory-and-then-tranfer-rate-decreases?forum=winservergen
    You could also refer to the FAQ article to troubleshoot the slow copy issue:
    [Forum FAQ] Troubleshooting Network File Copy Slowness
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/7bd9978c-69b4-42bf-90cd-fc7541ccb663/forum-faq-troubleshooting-network-file-copy-slowness?forum=winserverPN
    Best Regards,
    Mandy 
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • USB Drive Resets during file copy, interrupting file operations

    Fresh Arch Linux install, working USB thumb drive (Windows, previous Arch Linux setups, etc.), formatted FAT32.
    Seems to randomly reset, eject, and auto-re-mount during extensive file operations (i.e. copying 4gb worth of 800 files from the USB thumb drive to disk).
    I've tried various USB ports to no avail; no external hub being used.
    Thoughts? Not sure where to begin diagnosing:
    dmesg:
    usb 1-1: configuration #1 chosen from 1 choice
    scsi17 : SCSI emulation for USB Mass Storage devices
    usb-storage: device found at 15
    usb-storage: waiting for device to settle before scanning
    scsi 17:0:0:0: Direct-Access OCZ RALLY2 1100 PQ: 0 ANSI: 0 CCS
    sd 17:0:0:0: Attached scsi generic sg2 type 0
    usb-storage: device scan complete
    sd 17:0:0:0: [sdb] 31326208 512-byte logical blocks: (16.0 GB/14.9 GiB)
    sd 17:0:0:0: [sdb] Write Protect is off
    sd 17:0:0:0: [sdb] Mode Sense: 43 00 00 00
    sd 17:0:0:0: [sdb] Assuming drive cache: write through
    sd 17:0:0:0: [sdb] Assuming drive cache: write through
    sdb: sdb1
    sd 17:0:0:0: [sdb] Assuming drive cache: write through
    sd 17:0:0:0: [sdb] Attached SCSI removable disk
    usb 1-1: reset high speed USB device using ehci_hcd and address 15
    sd 17:0:0:0: [sdb] Unhandled error code
    sd 17:0:0:0: [sdb] Result: hostbyte=0x07 driverbyte=0x00
    end_request: I/O error, dev sdb, sector 7589504
    sd 17:0:0:0: [sdb] Unhandled error code
    sd 17:0:0:0: [sdb] Result: hostbyte=0x07 driverbyte=0x00
    end_request: I/O error, dev sdb, sector 7589744
    sd 17:0:0:0: [sdb] Unhandled error code
    sd 17:0:0:0: [sdb] Result: hostbyte=0x07 driverbyte=0x00
    end_request: I/O error, dev sdb, sector 7589776
    usb 1-1: USB disconnect, address 15
    FAT: bread failed in fat_clusters_flush
    FAT: FAT read failed (blocknr 3805)
    FAT: Directory bread(block 24147944) failed
    FAT: FAT read failed (blocknr 3818)
    <snip lots of repeats>
    FAT: FAT read failed (blocknr 4032)
    FAT: bread failed in fat_clusters_flush
    FAT: bread failed in fat_clusters_flush
    FAT: bread failed in fat_clusters_flush
    usb 1-1: new high speed USB device using ehci_hcd and address 16
    usb 1-1: configuration #1 chosen from 1 choice
    scsi18 : SCSI emulation for USB Mass Storage devices
    usb-storage: device found at 16
    usb-storage: waiting for device to settle before scanning
    scsi 18:0:0:0: Direct-Access OCZ RALLY2 1100 PQ: 0 ANSI: 0 CCS
    sd 18:0:0:0: Attached scsi generic sg2 type 0
    usb-storage: device scan complete
    sd 18:0:0:0: [sdb] 31326208 512-byte logical blocks: (16.0 GB/14.9 GiB)
    sd 18:0:0:0: [sdb] Write Protect is off
    sd 18:0:0:0: [sdb] Mode Sense: 43 00 00 00
    sd 18:0:0:0: [sdb] Assuming drive cache: write through
    sd 18:0:0:0: [sdb] Assuming drive cache: write through
    sdb: sdb1
    sd 18:0:0:0: [sdb] Assuming drive cache: write through
    sd 18:0:0:0: [sdb] Attached SCSI removable disk
    lsusb -v for the drive:
    Bus 001 Device 016: ID 0325:ac02 OCZ Technology Inc ATV Turbo / Rally2 Dual Channel USB 2.0 Flash Drive
    Device Descriptor:
    bLength 18
    bDescriptorType 1
    bcdUSB 2.00
    bDeviceClass 0 (Defined at Interface level)
    bDeviceSubClass 0
    bDeviceProtocol 0
    bMaxPacketSize0 64
    idVendor 0x0325 OCZ Technology Inc
    idProduct 0xac02 ATV Turbo / Rally2 Dual Channel USB 2.0 Flash Drive
    bcdDevice 11.00
    iManufacturer 1 OCZ Technology
    iProduct 2 RALLY2
    iSerial 3 AA04012700275633
    bNumConfigurations 1
    Configuration Descriptor:
    bLength 9
    bDescriptorType 2
    wTotalLength 32
    bNumInterfaces 1
    bConfigurationValue 1
    iConfiguration 0
    bmAttributes 0x80
    (Bus Powered)
    MaxPower 300mA
    Interface Descriptor:
    bLength 9
    bDescriptorType 4
    bInterfaceNumber 0
    bAlternateSetting 0
    bNumEndpoints 2
    bInterfaceClass 8 Mass Storage
    bInterfaceSubClass 6 SCSI
    bInterfaceProtocol 80 Bulk (Zip)
    iInterface 0
    Endpoint Descriptor:
    bLength 7
    bDescriptorType 5
    bEndpointAddress 0x81 EP 1 IN
    bmAttributes 2
    Transfer Type Bulk
    Synch Type None
    Usage Type Data
    wMaxPacketSize 0x0200 1x 512 bytes
    bInterval 255
    Endpoint Descriptor:
    bLength 7
    bDescriptorType 5
    bEndpointAddress 0x02 EP 2 OUT
    bmAttributes 2
    Transfer Type Bulk
    Synch Type None
    Usage Type Data
    wMaxPacketSize 0x0200 1x 512 bytes
    bInterval 255
    Device Qualifier (for other device speed):
    bLength 10
    bDescriptorType 6
    bcdUSB 2.00
    bDeviceClass 0 (Defined at Interface level)
    bDeviceSubClass 0
    bDeviceProtocol 0
    bMaxPacketSize0 64
    bNumConfigurations 1
    Device Status: 0x0000
    (Bus Powered)
    Last edited by lieut_data (2009-12-26 17:39:56)

    Jithin wrote:
    When ever I copy some files or create a file in my external USB drive which is a FAT32 one. file permissions are not preserved.
    From <http://support.apple.com/kb/HT3764>:
    "The MS-DOS (FAT32) file system format does not support permissions, file owners, and groups. Such permissions are synthesized on Mac OS X with some default permissions. Because of this all files will have the same permissions […]"
    So it's not "how it work in mac", but it's a limitation of the FAT32 file system.

  • URGENT: File Adapter List Files operation Issue

    Hi All,
    we are using List files operation in one of the SOA composite which lists all files available in the directory. what we observed files are not listing as for the timestamps.
    is there any property to list all files ascending or descending based on time stamp?. we tried with ListSorter property which is suggested by Oracle,but it works for only INBOUND operations. [http://docs.oracle.com/cd/E23943_01/integration.1111/e10231/adptr_file.htm#BABBIJCJ]
    Any suggestions will be greatly appreciated.

    Hi,
    You can try 2 options:
    1. You would need to capture/collect all the file names, you might have to use BPM and create a separate interface.
    2. You can also pick up those files from the archive directory using FTP and push them using mail adapter.
    Regards,
    Pavan

  • Unable to run a Batch File Operating System Command

    Using XI 3.0, I am unable to run a Batch File Operating System Command After Message Processing.
    My Batch file:
    gpg -se -r BOA3RSKY --armor --passphrase-fd 0 %1 < C:\Progra~1\GNU\GnuPG\gpgin
    My Command Line (ID scenario)
    exec "cmd.exe /c C:\Progra~1\GNU\GnuPG\boagpg.bat %F"
    If I execute
    exec "cmd.exe /c type C:\Progra~1\GNU\GnuPG\boagpg.bat >xis.txt"
    It displays the contents of boagpg.bat file in xis.txt.
    I just don't understand why when I run the batch file, I would expect an %F.asc encrypted file in the same directory as the %F unencrypted file.
    Any ideas?
    or will I need Basis to create commands that will allow me to run GPG from XI Command Line?

    Check this links if its helpful
    http://help.sap.com/saphelp_nw04/helpdata/en/bb/c7423347dc488097ab705f7185c88f/frameset.htm
    /people/sap.user72/blog/2004/01/30/command-line-help-utility
    Check this thread a similar problem
    Process Integration (PI) & SOA Middleware
    Note 841704 - XI File & JDBC Adapter: Operating system command
    http://service.sap.com/sap/support/notes/841704
    Try to see the below links
    /people/michal.krawczyk2/blog/2005/08/17/xi-operation-system-command--error-catching
    OS Command on FTP
    OS command line script - Need help
    FTP - Run OS Command before file processing
    Note: reward points if solution found helpfull
    Regards
    Chandrakanth.k

  • ORA-29283 invalid file operation

    NLSRTL      10.2.0.5.0     Production
    Oracle Database 10g Enterprise Edition      10.2.0.5.0     64bi
    PL/SQL      10.2.0.5.0     Production
    TNS for IBM/AIX RISC System/6000:      10.2.0.5.0     Productio
    I am trying to get the content of a trace file generated for me.
    Because I don't have privileges to log on the server and copy the trace file for me directly with some os user, I am doing the following:
    1. I alter my session trace identifier to easier identify the trace file
    ALTER SESSION SET TRACEFILE_IDENTIFIER = 'Func01';2. I invoke DBMS_MONITOR
    3. I run the procedure I want to monitor.
    4. I disable the monitoring by calling DBMS_MONITOR
    5. At this point I run the following query to identify my trace file:
    select u_dump.value || '/' || instance.value || '_ora_' || v$process.spid || nvl2(v$process.traceid, '_' || v$process.traceid, null ) || '.trc'"Trace File"
    from V$PARAMETER u_dump
    cross join V$PARAMETER instance
    cross join V$PROCESS
    join V$SESSION on v$process.addr = V$SESSION.paddr
    where 1=1
       and u_dump.name = 'user_dump_dest'
       and instance.name = 'instance_name'
       and V$SESSION.audsid=sys_context('userenv','sessionid');It gives me: /ORACLE/MYDB/trace/MYDB_ora_3616822_Func01.trc
    I have created directory in advanced on the path where the traces are stored:
    CREATE OR REPLACE DIRECTORY trace_dir AS '/ORACLE/MYDB/trace/';
    SELECT * FROM dba_directories WHERE directory_name = 'TRACE_DIR';
    Output:
    OWNER DIRECTORY_NAME DIRECTORY_PATH
    SYS     TRACE_DIR      /ORACLE/MYDB/trace/I don't have rights to grant read, write on TRACE_DIR to my user, as I am not logged with SYS.
    I created a table to store in it the lines from the trace file:
    CREATE TABLE tmp_traces_tab
      callnum NUMBER,
      line NUMBER,
      fileline CLOB
    );Then I run the following PL/SQL block to retrieve the content of the trace and store it in the table T:
    DECLARE
      l_file            UTL_FILE.file_type;
      l_location     VARCHAR2 (100) := 'TRACE_DIR';
      l_filename    VARCHAR2 (255) := 'MYDB_ora_3616822_Func01.trc';
      l_text           VARCHAR2 (32767);
      l_line           NUMBER := 1;
      l_call           NUMBER := 1;
    BEGIN
      -- Open file.
      l_file := UTL_FILE.fopen (l_location, l_filename, 'r', 32767);
      -- Read and output first line.
      UTL_FILE.get_line (l_file, l_text, 32767);
      INSERT INTO tmp_traces_tab (callnum, line, fileline) VALUES (l_call, l_line, l_text);
      l_line := l_line + 1;
      BEGIN
        LOOP
          UTL_FILE.get_line (l_file, l_text, 32767);
           INSERT INTO tmp_traces_tab (callnum, line, fileline) VALUES (l_call, l_line, l_text);
           l_line := l_line + 1;
        END LOOP;
      EXCEPTION
        WHEN NO_DATA_FOUND THEN
          NULL;
      END;
      INSERT INTO tmp_traces_tab (callnum, line, fileline) VALUES (l_call, l_line, l_text);
      l_line := l_line + 1;
      UTL_FILE.fclose (l_file);
    END;
    /And when I run the code I get the error: ORA-29283 invalid file operation.
    Is it possible to a role my user to be able to get the content of the trace files in the directory TRACE_DIR without having explicit READ , WRITE privileges on it?
    My user currently has these roles:
    select * from dba_role_privs where grantee = USER;
    Output:
    U1     OPR_ROLE_LOSS_SNAPSHOT_READER     YES     YES
    U1     RESOURCE     NO     YES
    U1     CONNECT     NO     YES
    U1     DBA     NO     YES
    U1     OPR_ROLE_SUPPORT_USER     YES     YESI know that on another db with different user I hit no errors when doing completely the same (of course the program unit I monitor is different).
    That other user with which I have NO issues has these roles:
    select * from dba_role_privs where grantee = USER;
    Output:
    U2    DBA    NO    YES
    U2    EXEC_SYS_PACKAGES_ROLE    NO    YES
    U2    EXECUTE_CATALOG_ROLE    NO    YES
    U2    CONNECT    NO    YES

    Verdi wrote:
    NLSRTL      10.2.0.5.0     Production
    Oracle Database 10g Enterprise Edition      10.2.0.5.0     64bi
    PL/SQL      10.2.0.5.0     Production
    TNS for IBM/AIX RISC System/6000:      10.2.0.5.0     Productio
    And when I run the code I get the error: ORA-29283 invalid file operation.
    Is it possible to a role my user to be able to get the content of the trace files in the directory TRACE_DIR without having explicit READ , WRITE privileges on it?
    My user currently has these roles:
    select * from dba_role_privs where grantee = USER;
    Output:
    U1     OPR_ROLE_LOSS_SNAPSHOT_READER     YES     YES
    U1     RESOURCE     NO     YES
    U1     CONNECT     NO     YES
    U1     DBA     NO     YES
    U1     OPR_ROLE_SUPPORT_USER     YES     YESI know that on another db with different user I hit no errors when doing completely the same (of course the program unit I monitor is different).
    Thanks for posting version alongwith other details.
    TO my knowledge, No you cannot.
    Privileges acquired via a Role are not valid in PL/SQL. You need to have explicit privileges.

  • Laggy file operations windows 7 clients with windows server 2008 R2

    Hope this is posted in the right place..
    Ok so this is driving me insane...
    I have a DNS/file server running Windows Server 2008 R2, and 5 client pc's running windows 7 x64 and x86.
    All the client PC's have extremely laggy file operations when for example right clicking to create a new folder or opening certain files (AutoCAD). Talking about 3-10 second lag for a response.
    However if I do a straight file transfer to test the transfer speed its all normal. Everything is connected @ gigabit network speeds.
    This just seemed to happen one day a few months ago, and has remained ever since. Before that (couple years) it was not doing this...
    Anyone have any ideas, I'm at a loss... I've checked the performance monitor on the server nothing seems abnormal when this happens.
    Thank you,

    Hi, 
    Please test to disable all third party applications on file server to see if that is the cause. Sometimes such kind of issues are caused by antivirus program, firewall etc.
    A quick step is to boot into Clean Boot mode as provided in following article:
    http://support.microsoft.com/kb/929135
    If you have any feedback on our support, please send to [email protected]

  • Filestream Creation Unable to Open Physical File Operating System Error 259

    Hey Everybody,
    I have run out of options supporting a customer that is having an error when creating a database with a file stream.  The error displayed is unable to open physical file operating system error 259 (No more data is available).  We're using a pretty
    standard creation SQL script that we aren't having issues with other customers:
    -- We are going to create our data paths for the filestreams.  
    DECLARE @data_path nvarchar(256);
    SET @data_path = (SELECT SUBSTRING(physical_name, 1, CHARINDEX(N'master.mdf', LOWER(physical_name)) - 1)
                      FROM master.sys.master_files
                      WHERE database_id = 1 AND file_id = 1);
    -- At this point, we should be able to create our database.  
    EXECUTE ('CREATE DATABASE AllTables
    ON PRIMARY
        NAME = AllTables_data
        ,FILENAME = ''' + @data_path + 'AllTables_data.mdf''
        ,SIZE = 10MB
        ,FILEGROWTH = 15%
    FILEGROUP FileStreamAll CONTAINS FILESTREAM DEFAULT
        NAME = FSAllTables
        ,FILENAME = ''' + @data_path + 'AllTablesFS''
    LOG ON
        NAME = AllTables_log
        ,FILENAME = ''' + @data_path + 'AllTables_log.ldf''
        ,SIZE = 5MB
        ,FILEGROWTH = 5MB
    GO
    We are using SQL Server 2014 Express.  File streams were enabled on the installation SQL Server.  The instance was created successfully and we are able to connect to the database through SSMS. The user is using an encrypted Sophos. 
    We have tried the following:
    1. Increasing the permissions of the SQL Server server to have full access to the folders.
    2. Attempted a restore of a blank database and it failed.
    There doesn't seem to be any knowledge base articles on this particular error and I am not sure what else I can do to resolve this.  Thanks in advance for any help!

    Hi Ryan,
    1)SQL Server(any version) can't be installed on encrypted drives. Please see a similar scenario in the following link
    https://ask.sqlservercentral.com/questions/115761/filestream-and-encrypted-drives.html
    2)I don't think there is any problem with permissions on the folder, if the user can create a database in the same folder. Am not too sure. Also see the article by
    Jacob on configuring the FILESTREAM for SQL Server that describes how to configure FILESTREAM access level & creating a FILESTREAM enabled database
    Hope this helps,
    Thanks
    Bhanu 

Maybe you are looking for

  • Can not display BLOB in a REPORT Portal 3.0.9

    I had created a table witha blob field. I want to display this field in a report to display the documento from the aplicaction. What can i do ?

  • Illustrator CS6 hangs opening existing files

    I have been having a problem with Illustrator CS6: Whenever I try to open up an existing file, Illustrator hangs. If I leave it long enough it will open the file. Any click causes a 30 second (at least) delay. I can work in it as long as I'm patient

  • Aironet 1300

    Hi, Does the Aironet 1300 series works as repeater ? And does it works both as bridge and access point at the same time ? Thank you

  • Java app on 64 bit Windows 2008

    I need help finding the solution to a problem that I'm having with an java application that I'm running on 64 bit Windows 2008 Server for a P to V, physical to virtual server migration. My application has been running fine on Windows 2003 32 bit OS.

  • Help with ichat + i didnt respond?

    Date/Time: 2007-01-04 18:09:36.041 -0700 OS Version: 10.4.8 (Build 8L2127) Report Version: 4 iChat Connection Log: AVChat started with ID 0. 0xe8de8d0: State change from AVChatNoState to AVChatStateWaiting. bonjourolivia: State change from AVChatNoSt