AS400 - File adapter not archiving File and Member

Hi Gents,
Specific problem for the archiving of file/members on the AS400 system.    When the file adapter is archiving the AS400 member to an AS400 archive file it is leaving an empty file structure in the origin directory.    So at the end of the day we have a lot of AS400 Files with no members.
AS400 Technical folks are looking for the FTP 'script' that is being executed on the AS400 to pick up the file and archive it.  
Has anyone else encountered this AS400 problem, were you able to solve it?
Also how can I see the exact commands being executed on the AS400 from XI?
Thanks in advance and regards.
Mark

Extend the trace file to Debug in Visual Admin, and re run:
 Services
 Log Configurator
 Locations
 Choose the location (Java package) and the tracing level (severity)
(for example, com.sap.aii.adapter.file)
 Root Location > com > sap > aii > File
 Set the severity and choose Copy Severity to Subtree.
Check the log file which will give you the details.
Don't forget to switch this back afterwards.
There are also quite a few OSS Notes relating to problems with Adapter Framework and AS/400...

Similar Messages

  • ESB FTP Adapter Not Archiving to FTP Server

    I have a FTP adapter in which I have pointed a logical archive directory to a directory on the FTP server. However, the file is not getting archived to that folder. If I point the logical directory to a shared folder on the server it archives the file. Can I not archive a file on the FTP server? I have checked permissions on the FTP folder and they are correct.
    Thanks.

    I found the solution to this issue. There is a property named UseRemoteArchive that needs to be set in the WSDL in the <jca:operation> element. See below for details.
    <jca:operation
    FileType="ascii"
    LogicalDirectory="incomingDirectory"
    ActivationSpec="oracle.tip.adapter.ftp.inbound.FTPActivationSpec"
    LogicalArchiveDirectory="archiveDirectory"
    DeleteFile="true"
    IncludeFiles="*\.xml"
    PollingFrequency="100"
    MinimumAge="0"
    OpaqueSchema="false"
    UseRemoteArchive="true" >
    </jca:operation>

  • Writing File and File Member to AS400 using Access Classes

    Hi! All,
    I am in a fix here and would sincerely appreciate any help I can get. I need to write a pogram that will allow me to write a file and many FileMembers to the File in the AS400.
    I have the following method which I thought would help me do it, but I get exceptions.
    public static void writeFileOnAS400(String pFileName, String pFileMemberName, String pContents ) throws Exception
    AS400 as400 = new AS400();
    String as400FileName = "/QSYS.LIB/MYLIB.LIB/" + pFileName + ".FILE";
    String as400MemberName = pFileMemberName + ".MBR";
    // Create a file object that represents the file
    SequentialFile newFile = new SequentialFile(as400, as400FileName + "/" + as400MemberName);
    // Create the file
    newFile.create(pContents.getBytes().length, "*DATA", "Data file created with " + pContents.getBytes().length + "byte record." );
    // Open the file for writing only.
    // Note: The record format for the file
    // has already been set by create()
    newFile.open(AS400File.WRITE_ONLY, 0, AS400File.COMMIT_LOCK_LEVEL_NONE);
    // Write a record to the file. Because the record
    // format was set on the create(), getRecordFormat()
    // can be called to get a record properly formatted
    // for this file.
    Record writeRec = newFile.getRecordFormat().getNewRecord();
    writeRec.setField(0, pContents);
    newFile.write(writeRec);
    // Close the file since I am done using it
    newFile.close();
    // Disconnect since I am done using
    // record-level access
    as400.disconnectService(AS400.RECORDACCESS);
    This works the first time for the 1st Member(ie; when te file or Member is not yet created). When I want to add a new Member to the same file, by passing in the same fileName, but a different fileMemberName, it does not work. The string pContents is what I need to write to the Member.
    Anyone Know about the AS400?
    Thanks

    The whole class is as shown below
    import com.ibm.as400.access.*;
    public class AS400Utility
    public static void writeFileOnAS400(String pFileName, String pFileMemberName, String pContents ) throws Exception
    AS400 as400 = new AS400();
    String as400FileName = "/QSYS.LIB/IQFILES.LIB/" + pFileName + ".FILE";
    String as400MemberName = pFileMemberName + ".MBR";
    // Create a file object that represents the file
    SequentialFile newFile = new SequentialFile(as400, as400FileName + "/" + as400MemberName);
    // Create the file
    newFile.create(pContents.getBytes().length, "*DATA", "Data file created with " + pContents.getBytes().length + "byte record." );
    // Open the file for writing only.
    // Note: The record format for the file
    // has already been set by create()
    newFile.open(AS400File.WRITE_ONLY, 0, AS400File.COMMIT_LOCK_LEVEL_NONE);
    // Write a record to the file. Because the record
    // format was set on the create(), getRecordFormat()
    // can be called to get a record properly formatted
    // for this file.
    Record writeRec = newFile.getRecordFormat().getNewRecord();
    writeRec.setField(0, pContents);
    newFile.write(writeRec);
    // Close the file since I am done using it
    newFile.close();
    // Disconnect since I am done using
    // record-level access
    as400.disconnectService(AS400.RECORDACCESS);
    public static void main(String args[])
    System.out.println("Writing file to AS400...");
    try
    writeFileOnAS400_2("BPL0000004", "M001", "<SampleXML>The sample file 1</SampleXML>");
    System.out.println("Finished writing file 1");
    writeFileOnAS400_2("BPL0000004", "M002", "<SampleXML>The sample file 2</SampleXML>");
    catch(Exception e)
    e.printStackTrace();
    System.out.println("Completed file writing to AS400...");
    }

  • JCBC Adapter and File adapter not processing messages

    Hi
    I noticed that messages are being delivered to Adapter engine and the same are visible in Runtime Workbench with status "to be delivered". But, JDBC Adapter and File Adapter not processing these messages.
    Any idea where I can find the problem?
    I was able to re-deliver successfully via the JDBC Adapter using the MessagingSystem GUI using XISUPER user.
    Regards
    Chandu

    Hi,
    1.Status: TO_BE_DELIVERED
    Which means that the message was successfully delivered from Integration Server point of view and it states that the messages is initially handed over to the Messaging System.
    TO_BE_DELIVERED occurs while the message is put into the Messaging System receive queue.
    Regards
    Agasthuri Doss

  • My right click is not working. the context menu does not come up and i can no longer drag files/folders. I have checked the mouse and track pad in system preferences and the correct boxes are all checked (with a support member). Please can you help

    My right click is not working and producing the context menu or ability to drag files and folders. I just spoke with a phone support person and we checked that all the mouse/trackpad options for right click were correct and checked.
    Please can you help with this.
    It has only just happened in the last few days.
    Many thanks

    You might want to update your profile so that we can see what model iMac you have and what version of OSX you're running.
    For this question, it might help to know if you're dealing with a trackpad, Magic Mouse, or something else?
    (Until your issue's resolved, you can Control-Left Click to get the right click functionality).

  • Took backup of data files and controlfile but not archive file db is 11g xe

    took backup of data files and controlfile but not redo file db is 11g xe r2..was trying complete shut backup,..missed redo..
    anyway to restore the db back..
    I had failed to backup redo log file..
    db was shut...properly
    any info highly appreciated.
    cheers
    Edited by: zycoz100 on Feb 27, 2013 4:42 AM

    A controlfile create is a fairly trivial task, for either a cold backup or hot backup scenario. Well, its an easy task after you've done the process a few times.
    If you have a healthy oracle instance running any place else, preferably at 11gR2, or even a 10g instance can work as an example, get a system connection and do an `alter database backup controlfile to trace;` and the instance will create a new trace file with all SQL commands needed to rebuild the controlfile in the instance trace directory.
    Have a look at the trace file, a `show parameter diag` reveals the diagnostic_dest parameter, look for the "trace" directory under that folder, for the latest *trc file. In 10g its `show parameter dump` and the trace file goes straight to the user_dump_dest directory, no digging required.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Need Help-SOA 11g File Adapter unable to delete input file and its crashing

    Hi All
    Please find the details below:
    1. We have created a simple SOA composite to Read file from an input directory, archive the file in an archive directory using Inbound File Adapter Read
    and then use Outbound File Adapter Write to move the file to a output directory.
    2. File Adapter needs to delete the file after successful read/retrieval.
    3. We are using the "Use Trigger File" for invoking the file adapter. This is a new feature in SOA 11g
    4. Also we are using the option of reading the file as an attachment as we are not doing any transformation in the composite
    Issue Details_
    1. When the trigger file is put in the input directory for the first time, the File Adapter reads the file, archives it and moves it to the output directory
    2. However it does not delete the input file from the input directory and raises Fatal Exception mentioned below:
    [*2011-01-12T16:55:48.639+05:30] [soa_server1] [WARNING] [] [oracle.soa.adapter] [tid: weblogic.work.j2ee.J2EEWorkManager$WorkWithListener@19c243d]*
    [userId: <anonymous>] [ecid: 0000IptyLrL9_aY5TrL6ic1DBOS_000009,0] [APP: soa-infra] File Adapter FileAdapterTriggerFilePOC PostProcessor::
    Delete failed, the operation will be retried for max of [0] times
    [2011-01-12T16:55:48.639+05:30] [soa_server1] [WARNING] [] [oracle.soa.adapter] [tid: weblogic.work.j2ee.J2EEWorkManager$WorkWithListener@19c243d]
    [userId: <anonymous>] [ecid: 0000IptyLrL9_aY5TrL6ic1DBOS_000009,0] [APP: soa-infra] File Adapter FileAdapterTriggerFilePOC [[
    BINDING.JCA-11042
    File deletion failed.
    File deletion failed.
    File : C:\Dibya\AttachmentTest\InputDir\TestFile3.txt could not be deleted.
    Delete the file and restart server. Contact oracle support if error is not fixable.
    If any one has faced similar issues, kindly provide pointers on how to resolve it.
    Regards,
    Dibya

    Hi,
    Using the file adapter, you can poll from multilple locations...
    Keep the following property in your .jca file
    <property name="DirectorySeparator" value="," />
    While giving the path in File Adapter configuration, keep comma and give the next location....then the file will be picked up from the locations you gave....
    Hope this helps...
    Thanks,
    N

  • Sender File Adapter  not able to get file via FTP from an external system

    Hi,
    I have a Sender File Adapter using FTP which polls a server every 10 seconds looking for a file ".xml".  The adapter will connect and pick up the waiting file, but sometimes the adapter not able to get file via FTP from an external system. The file is on the server, but despite the polling, the adapter does not pick up it. There is no return error.
    The connect mode is Permanently and the transfer mode is Binary.
    Can you help me?
    Thanks.
    T.

    Hi Teresa,
    does this case appears with a file size higher the others?
    Personally, I do not recommend to have a polling under 1min.
    10s could be too just, if the file is not small and/or if the file server is far (several kilometers) to where the Adapter Engine is, and/or this file server is beside a firewall. Do not forget also that, with previous points, you could have like a conflict, betwen the 1st polling and the 2nd polling, if in less 10s PI adapter has not enough time: to connect to the FTP server (several Km), find the file (perhaps among 100 other files in the directory), go back to the file to Adapter Engine server (several Km again), convert your file to XML message (with a complex Content Conversion), then check the XML structure, and when all that is done, return back to FTP server (several Km), to archive the source file... Well It's my understanding of a FTP exchange with FILE adapter...
    If you really need a 10s polling, that's mean your have a pseudo real time interface. In that case, in my mind file should not be used! but we should use a Web service, JDBC JMS (any exchange without a physical object like a file). For me, it's a non-sens to have a file in that case.
    regards.
    mickael

  • FTP adapter --- Not picking up files

    Hi,
    I have set up an FTP adapter in archive mode to read files from FTP server and process it while archiving a copy of it. I have set up Poll interval at 40 secs and retry interval also set up. It works great the first time . Processes the file and archives it. But when the same files comes in again in the FTP server, I see the foll. message in the adapter monitor "Error: Stopped due to error: com.sap.aii.adapter.file.ftp.FTPEx: inputdat.txt: No such file or directory: No such file or directory."..The file exists in the FTP server but its not processed..I am not sure what I am missing...
    -Teresa

    Hi Michal,
      I have timestamp set up for archive directory..So every run will create an archive file with timestamp..But the input filename remains the same..Archive functionality would remove the file inthe FTP server..So when the file comes again into the FTP server, the FTP adapter doesn't trigger automatically. Everytime I had to activate the IB for FTP adapter to trigger the process. I know this is not how XI-FTP adapter should work..Our scenario is we would receive weekly files into the FTP adpater with the same name. We need to archive each weekly file and also send it for further processing..I can't be activating the process manually ....I know something is going wrong but can't figure it out...:(..Any help is appreciated...:)
    -Teresa

  • FTP File Adapter: more Archive Directory

    Hello everybody,
    i have some interfaces whit FTP-XI-Idoc scenarion.
    we decided to save files after elaboration in different directory depending on month (yy-mm).
    Example:
    /Rxarchivio/0901
    /Rxarchivio/0902
    I see this link
    https://wiki.sdn.sap.com/wiki/display/XI/SAPXIFileAdapterOSCommandLine+Feature
    very helpfull but i don't know how to parameterize variable of year and month.
    i try with this but don't work:
    SET CC=%DATE:6,4%%DATE:3,2%
    open ftp
    VCE01\selzsapd
    dpas_zles
    cd /Rxarchivio
    mkdir %CC%
    mget
    cd /Rxarchivio/%CC%
    mput
    quit
    any suggestion?

    Hi,
    I m not taking about FTP server. Your PI is on Windows or Unix.
    Id it is on UNIX the put the UNIX script to PI file system and give the path.
    FTP adapter pick the file from FTP location and the pick the UNIX script from PI FIle system and run script ..archive the files and then send message for further processing.
    Thanks
    Aman

  • Faults generated by Oracle file adapter not reaching BPEL process

    Hi,
    I have a process that uses the Oracle file adapter to receive the contents of a file, and am suffering from a number of problems when it is not able to parse the source file.
    The process is initiated by a receive activity associated with the file adapter partnerline that has the create instance flag ticked.
    One problem is that if the file adapter encounters an error when reading the file, there is no record of anything happening from the BPEL Console. I presume this is because no instance of the process has been created as the error was encountered before invoking the process. The only record I have of a problem ever occuring is in the logfiles.
    To make matters worse, if you select the Delete files after successful retrieval, the source file is deleted even though an exception has been encountered. If you specify the Archive processed files options, it does not archive the file.
    So to summarise the problems:
    * there is no record of an attempt to read, or of a faulty read occurring when trying to read the file from the BPEL console when a faulty read does occurs.
    * no instance of the process is ever created, so it is not possible to implement a fault handler to send out a notification of the problem.
    * when a problem is encountered, the source file is deleted and not archived.
    I assume that this is a bug, so can anyone advise if there is a patch to address these problems?
    Thanks
    Toby

    Ok, so now I'm using a rejection handler to invoke a BPEL process, and have defined the inputVariable type to be rej:RejectionMessage and imported RejectionMessage.wsdl.
    The process is being invoked successfully, and I can see that the processes input contains data.
    So I am writing the opaque MessagePayload to a file successfully using a file adapter, and can see the file content that is written is as I would expect (it matches the file used as input that would have caused the fault - although the content is truncated).
    Now I want to modify the file adapter so that the file's filename is set to the RejectionId attribute of the BPEL process. However, I'm receiving the following message:
    xpath variable/expression expression "/ns2:RejectedMessage/@ns2:RejectionId" is empty at line 50, when attempting reading/copying it.
    Please make sure the variable/expression result "/ns2:RejectedMessage/@ns2:RejectionId" is not empty.
    I can see from the console that the RejectionId is in fact populated, and the copy from the RejectionMessage's MessagePayload is completed ok.
    Does anyone know what's going on here?
    Thanks
    Toby

  • Changing of the timestamp in sender file adapter in archive mode

    Hi,
    I have a requirement where in I have to archive a file with timestamp different from that generated by XI.
    Please let me know if this can be done and if so how can we handle the changes to be made to the timestamp in the sender adapter in archive mode.
    regards,
    Srinivas.

    Srinivas,
    Option 1) Create a bat file..to run the perl script you call..
    Perl script..
    #!/usr/bin/perl -w
    print("Starter that you want to change: ");
    chomp($badex = <STDIN>);
    print("Starter that you want added: ");
    chomp($goodex = <STDIN>);
    foreach $file (<$badex*>){
        @fields = split(/$badex/,$file);
        $goodfile = ("$goodex" . "$fields[1]");
        rename("$file","$goodfile");
    Run that on the os
    That should fix it.........
    Option 2) On your local Machine create a java file..add this code to it
    public class Utils
         public static int Randomizer(){
              int randomInt = 0;
              randomInt = (int) (Math.random()*1000);
              return randomInt;
    public static void main(String[] args)
         Randomizer();
    save and compile..
    Create a bat file to add the number returned from the random to your targetFilename
    so it would be something like..
    mv oldFileName Newfilename+randomizer... and also get this command written to a file..helpful later on.........
    Hope that helps
    Regards
    Ravi Raman
    PS:Dont forget the points if helpful

  • I can only backup the datafiles but not the control files and SPFiles

    Hello, all,
    I am using the Oracle 10g on a linux server which runs redhat 9. I want to backup the whole database to the disk, but not the flash recovery area. It is /rawdata. I will show how I configure the backup part in EM. Is there anything wrong?
    1. the ARCHIVELOG is on
    2. under Administration, in the All Initialization Parameters, I remove the string value of “log_archive_dest_1”. After that, I go to the SPFile and remove the string value of “log_archive_dest_1”
    3. configure backup settings and policies. The Disk Backup Location field is set to /rawdata. Select Image Copy for Disk Backup type. Test Disk backup is successful.
    4. go to Policy page. Select “Automatically backup the control file and server parameter file (SPFile) with every backup and database structural change.) input /rawdata into the filed: Autobackup Disk Location
    Select “Optimize the whole database backup by skipping unchanged files such as read-only and offline datafiles that have been backed up”
    Select “Retain backups that are necessary for a recovery to any time whin the specified number of days (point-in-time recovery)
    Accept the default of 31days
    5. go to Schedul Backup
    6. select customized from the backup Strategy drop-down menu. Select the Whole Database
    7. in the Options page, select “Full backup” in the Backup Type section. Select “Online Backup” in the Backup Mode section. Select “back up all archived logs on disk” in the Advanced section.
    8. in the Setting page, select “Disk”
    9. in the Schedule page, accept the default job name and description. Select Immediately to execute the job immediately.
    10. just submit the job.
    After several minutes, the job is failed while the Prebackup and Postbackup are succeeded. The log says:
    Recovery Manager: Release 10.1.0.2.0 - Production
    Copyright (c) 1995, 2004, Oracle. All rights reserved.
    RMAN>
    connected to target database: ORCL10G (DBID=770753237)
    using target database controlfile instead of recovery catalog
    RMAN>
    Starting backup at 11-OCT-06
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: sid=268 devtype=DISK
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00001 name=+DATADG/orcl10g/system01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-SYSTEM_FNO-1_68hvin50 tag=BACKUP_ORCL10G_000_101106111844 recid=108 stamp=603544772
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:45
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00003 name=+DATADG/orcl10g/sysaux01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-SYSAUX_FNO-3_69hvin6d tag=BACKUP_ORCL10G_000_101106111844 recid=109 stamp=603544813
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:35
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00005 name=+DATADG/orcl10g/example01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-EXAMPLE_FNO-5_6ahvin7g tag=BACKUP_ORCL10G_000_101106111844 recid=110 stamp=603544829
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:15
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00002 name=+DATADG/orcl10g/undotbs01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-UNDOTBS1_FNO-2_6bhvin7v tag=BACKUP_ORCL10G_000_101106111844 recid=111 stamp=603544835
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:08
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00004 name=+DATADG/orcl10g/users01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-USERS_FNO-4_6chvin87 tag=BACKUP_ORCL10G_000_101106111844 recid=112 stamp=603544841
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:03
    Finished backup at 11-OCT-06
    Starting Control File and SPFILE Autobackup at 11-OCT-06
    piece handle=/rawdata/c-770753237-20061011-05 comment=NONE
    Finished Control File and SPFILE Autobackup at 11-OCT-06
    RMAN>
    Starting backup at 11-OCT-06
    current log archived
    using channel ORA_DISK_1
    archived log /rawdata/arch_D-ORCL10G_id-770753237_S-1062_T-1_A-586368151_cqhvf270 not found or out of sync with catalog
    trying alternate file for archivelog thread 1, sequence 1062
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of backup command at 10/11/2006 11:20:54
    RMAN-06059: expected archived log not found, lost of archived log compromises recoverability
    ORA-19625: error identifying file /rawdata/arch_D-ORCL10G_id-770753237_S-1062_T-1_A-586368151_puhkv41f
    ORA-27037: unable to obtain file status
    Linux Error: 2: No such file or directory
    Additional information: 3
    RMAN>
    Recovery Manager complete.
    It seems the system tries to find a file “arch_D-ORCL10G_id-770753237_S-1062_T-1_A-586368151_cqhvf270” in /rawdata. I just can not understand.
    What should I do? Any advice is highly appreciated!
    Qian

    the string value in the field of 'log_archive_dest_1" under "All Initialization parameters" is "LOCATION=+DATADG/orcl10g/". The same value was in the field under SPFile.
    I reset these values and tried to backup the whole database. The same error appeared. The log says:
    Recovery Manager: Release 10.1.0.2.0 - Production
    Copyright (c) 1995, 2004, Oracle. All rights reserved.
    RMAN>
    connected to target database: ORCL10G (DBID=770753237)
    using target database controlfile instead of recovery catalog
    RMAN>
    Starting backup at 11-OCT-06
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: sid=245 devtype=DISK
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00001 name=+DATADG/orcl10g/system01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-SYSTEM_FNO-1_6ghvj9rd tag=BACKUP_ORCL10G_000_101106043753 recid=114 stamp=603563922
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:45
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00003 name=+DATADG/orcl10g/sysaux01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-SYSAUX_FNO-3_6hhvj9sq tag=BACKUP_ORCL10G_000_101106043753 recid=115 stamp=603563963
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:35
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00005 name=+DATADG/orcl10g/example01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-EXAMPLE_FNO-5_6ihvj9tu tag=BACKUP_ORCL10G_000_101106043753 recid=116 stamp=603563977
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:15
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00002 name=+DATADG/orcl10g/undotbs01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-UNDOTBS1_FNO-2_6jhvj9ud tag=BACKUP_ORCL10G_000_101106043753 recid=117 stamp=603563984
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:07
    channel ORA_DISK_1: starting datafile copy
    input datafile fno=00004 name=+DATADG/orcl10g/users01.dbf
    output filename=/rawdata/data_D-ORCL10G_I-770753237_TS-USERS_FNO-4_6khvj9uk tag=BACKUP_ORCL10G_000_101106043753 recid=118 stamp=603563990
    channel ORA_DISK_1: datafile copy complete, elapsed time: 00:00:03
    Finished backup at 11-OCT-06
    Starting Control File and SPFILE Autobackup at 11-OCT-06
    piece handle=/rawdata/c-770753237-20061011-07 comment=NONE
    Finished Control File and SPFILE Autobackup at 11-OCT-06
    RMAN>
    Starting backup at 11-OCT-06
    current log archived
    using channel ORA_DISK_1
    archived log /rawdata/arch_D-ORCL10G_id-770753237_S-1062_T-1_A-586368151_cqhvf270 not found or out of sync with catalog
    trying alternate file for archivelog thread 1, sequence 1062
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of backup command at 10/11/2006 16:40:05
    RMAN-06059: expected archived log not found, lost of archived log compromises recoverability
    ORA-19625: error identifying file /rawdata/arch_D-ORCL10G_id-770753237_S-1062_T-1_A-586368151_puhkv41f
    ORA-27037: unable to obtain file status
    Linux Error: 2: No such file or directory
    Additional information: 3
    RMAN>
    Recovery Manager complete.
    Why the system keeps trying to find files under /rawdata? It is the destination, not the source.
    Any help is appreciated!
    Qian
    Message was edited by:
    QianChen

  • How do I share a link to a file and have the other person download the file without having to add them explicitly as a member, like Dropbox does?

    With Dropbox, I right click on a file within my designated Dropbox folder and choose share Dropbox link.  This copies a unique URL that I can send to anyone to then download that file and that file only.  I do not have to add people to my access list to allow the download.
    How do I do this with Oracle Documents?  I have one off large files that I prefer not to send via email, however, I do not want to add everybody in the company as a member to my access list.

    There is an option called Public Links which allows you to do this. Select your item and click Share. This throws a pop-up dialog. On the left side of the dialog you may have 3 options - Members, Member Links, Public Links. If the Public Links option is not there, then your service administrator has disallowed them from use. The Public Links option is a controversial one in the enterprise space. DropBox's original offering had no qualms with offering public links because the content was guaranteed to be the property of the dropbox user. However, at enterprise level, the content belongs to the owning enterprise, so service administrators may elect not to allow this feature. This is down to your organisation.
    If you see the Public Links option, click on it and the right side of the popup dialog will change. You can specify:
    a name for each link you create (so you can keep track of which link is which),
    Link Permissions, only the following are available for obvious reasons - Viewer, Downloader, Contributor
    Link Expiration Date - a date after which the link will not work
    Access Code - a pass code which you can divulge to expected recipients by alternative means of your own choosing. This access code guarantees a measure of security on a link which could otherwise be used by anybody
    You can create several public links for the same item, and each of those can have the above settings, so you could have one link for downloaders and a separate link for contributors, for example. Each link can then be sent, either by copy/paste or through the email button to the right of the link. You can also modify the link using the pen icon, so you can unexpire it or reset its expiry date. You can also delete a link using the trash button next to it.
    I hope this explains the public link feature.
    Regards,
    Frank.

  • Receiver File Adapter Not Executing OS Command

    Hello All!
    We are currently running PI 7.1 and wish to use a Receiver File Adapter to FTP a file to a mainframe and override the FTP system defaults.
    Tried using the 'quote site recfm=fb lrecl=80' OS command but discovered that this would not work as the Adapter Engine is NOT on the mainframe.
    Regrouped a bit and made a second attempt which involved writing a Unix script to ftp a file and override rthe system defaults.   The idea was to use the File Adapter to FTP the data to the server, invoke the script and get the file to the mainframe with the proper attributes.    The script when executed on the server works perfectly.   
    However, when the IDoc comes thru PI, the File Adapter ftp's the file ok but the script is never executed.
    Any ideas why this is so or any suggestions?    
    Thank you.

    Hi Ken,
    Script only executes at Network File sharing (NFS) only. It dont work at FTP level.
    So you can better write the file on NFS and then within the script to connect to ftp and do your processing.
    Regards,
    ---Satish

Maybe you are looking for

  • How to pass '&' as a parameter in IDOC Status avoiding it to be replaced

    Hi,    This is the issue I have. Say an IDOC got status 51 when trying to post it. One of the messages to be appended in table EDIDS is: Message Class   Message Number    Text ZZ                      001                       This is parmeter1  & and

  • Oracle VM 3.1.1 Manager - OVM Servers in "Starting" status

    Hi, today I've downloaded Oracle VM Manager 3.1.1(Upgrade only) from e-delivery, and upgraded my previous version of OVMM 3.0.3 without any problems. My servers are still running on OVM Server 3.0.3 version. Usually, my practice for upgrading was: -

  • JAXB2 marshaller throws Validation Exception

    Hello All.. I am trying to marshall a JAXB Object using JAXB2 marshaller and validate it against an xsd. During this process of validation, the parser when reads the xsd fails with the parse error. The schema xsd is a working xsd from a webservice pr

  • Quality of Pictures looks terrible to Facebook and Email

    OK, I have a Pearl and everything is great, except when I post pictures to Facebook and email them to people. Now posting to Flickr they look fine but sending them to Facebook and email like a Gmail account the pics come up looking slightly dithered

  • PB G4 12'' - No/Dead Display

    So, I've got this Powerbook that I'm trying to get working. The display is completely dead, as far as I can see. It's not cracked, but nothing appears on it when I turn on the computer. The backlight does not turn on either. It gives the comforting M