Outbound file adapter for multiple record write

am using jdev 10.1.3.1
and SOA suite 10.1.3.1
I am trying to read a table and write the records to a csv file.
The bpel process I have constructed is :
receive > assign input value to table select parameter > select records from table > assign table select output values to file input value > invoke file adapter to write a csv file > assign a literal to process out value.
when i execute this process it is showing me an error on the file adapter invoke activity.
the error is :
</part></Invoke_2_Write_InputVariable></input><fault><bindingFault xmlns="http://schemas.oracle.com/bpel/extension"><part name="code"><code>null</code>
</part><part name="summary"><summary>file:/C:/product/10.1.3.1/OracleAS_1/bpel/domains/default/tmp/.bpel_TableReader_1.0_594707f7f819a3a4a2002c98afbd42fa.tmp/WritePersonFile.wsdl [ Write_ptt::Write(People) ] - WSIF JCA Execute of operation 'Write' failed due to: Translation Error.
Error while translating message to native format.
; nested exception is:
     ORABPEL-11017
Translation Error.
Error while translating message to native format.
Check the error stack and fix the cause of the error. Contact oracle support if error is not fixable.
</summary>
any help would be appreciated.
thanks

Hi,
In your case, to write any data to a file, you need to define/use a schema when you are using a file adapter.
Cheers,
Abhi

Similar Messages

  • Can the File Adapter handle Multiple Record Layouts in COBOL?

    I am trying to parse flat files that have a record type field in the first four characters on each record. As a test I tried to use the following record definitions:
    01 CONTRACT-HEADER.
    05 RECORD-TYPE PIC X(4).
    05 TRANSACTION-TYPE PIC X.
    05 CONTRACT-NUMBER PIC X(32).
    05 CONTRACT-RELEASE-NBR PIC X(30).
    05 CONTRACT-FILLER PIC X(2025).
    01 CONTRACT-LINEITEM.
    05 RECORD-TYPE PIC X(4).
    05 TRANSACTION-TYPE PIC X.
    05 LINE-FILLER PIC X(1664).
    I also tried smaller records with simple test data. When I press the "test" button in the native file adapter the XML is never generated if I have more than one record type as above. Suppose that the contract-header record type always has the string "X001" in it, while the contract-lineitem always has "X015" in the record-type. I chose Us-ascii and chose $eol as a delimiter. The file on my windows computer has carriage return line feed after each record. I know that the fixed length converter works, but because the records really have 50 fields on each record type and are long, I don't think I can actually use that one in one sitting. Has anyone ever gotten two different record lenght cobol definitions to work? Do I have to hand edit the XML to specify the delimiter? I don't see any delimter specified in the generated XML before I attempt the test.

    Rahul,
    I recommned you should use "|" as "," can be embedded in a text string. I guess you can pass a directive to use to deliver "|" delimited file :). If you want to handle both then you might have to write a shell script or bat to pass delimiter as an argument which will edit and replace your control file.
    Regards

  • Bapi_po_create is not creating Purchase Orders for multiple records in file

    Hi All.
    iam trying to create contracts and Purchase Orders  In me21n,me31k .
    here iam using bdc for contract creation against services and using bapi_po_create for PO Creations.
    in this process i could create contracts and POs for the first record in the file but for second record bapi_po_create couldnt create POs and the return table in bapi says
    1.document contains no items.
    2.no services or limits have been maintained.
    wil be waiitng for  r great answer.
    bye.
    regards.
    seeta.

    Hi Seeta Ram,
    Did you pass the table PO_ITEM_SCHEDULES to BAPI_PO_CREATE with the coresponding Item numbers for each item in the table PO_ITEMS?
    Regards,
    Vitz.

  • Soa Suite: How to limit Outbound File Adapter retry to write to specified location.

    Hi all,
    I have created a file adapter that tries to write files to the specified location.
    If location is wrong, I was expecting to get an error.
    But none is generated.
    It seems that the service keeps on trying to write to the specified location until it finds it (which is never to happen).
    Please tell if I am wrong at this.
    I want my service to stop retrying after a period of time or after a number of retries or whatever is happening.
    I want to get error if location is wrongly specified.
    Regards and thanks in advanced,
    ShuklaG

    Hi,
    Place the receive (or invoke) in a Scope and place a OnAlarm on the scope.
    Set the OnAlarm to a max timeout. When the OnAlarm is reached the sequence of that OnAlarm will catch the transaction and you can throw an Error.
    You can then catch or retrow the error.
    Below some images that show the concept on a receive (download to view).
    - ScopeWithOnAlarm
    - OnAlarmSettings
    Cheers,
    Robert van Mölken
    Senior Oracle Integration Specialist

  • OutBound File Adapter Issue

    Hi,
    I need some help in configuring file adapters:
    Process: I have a inbound adapter where i am feeding 4000 - 5000 rows of data from a flat file. I use these records to query a database and get response. I have a outbound file adapter where I write database response to a file.
    Problem: I have set the, write new file to 100 records. But if there are 3950 records going to outbound file adapter, It is only writing 3900 records in the output file directory.
    I need to write all records to the file. Please help me achieve this.
    Thank You
    Anil

    Hi,
    the batching criteria (100) is not met.
    You could improve the bahaviour by setting "Elapsed Time Exceeds" parameter, but still, there are some probs because critaria is only evaluated on invocation (anyone has an answer to that?!):
    from Doc:
    Elapsed Time Exceeds: Specify a time which, when exceeded, causes a new outgoing file to be created.
    Note:
    The Elapsed Time Exceeds batching criteria is evaluated and a new outgoing file is created, only when an invocation happens.
    For example, if you specify that elapsed time exceeds 15 seconds, then the first message that is received is not written out, even after 15 seconds, as batching conditions are not valid. If a second message is received, then batching conditions become valid for the first one, and an output file is created when the elapsed time exceeds 15 seconds.

  • Append functionality in Outbound File Adapter

    Hi all,
    I have a question regarding append functionality in outbound file adapter.
    When I append files for example I get a XML file with different XML files in one file, but is it also possible to collect and bundle the files and get a same result when u use BPM?
    Do I have to add some parameters?
    Kind regards,
    Kamran Vakili

    It is not possible, when you append XML files.
    The result XML file is not valid, as you have multiple declarations and root tags.
    The append mode for the file adapter should be used only for flat files.
    Regards
    Stefan

  • TargetFilename in the Outbound File Adapter

    Hi all,
    I have created an interface that sends projects form SAP PS (by tRFC) to a file directory (using an outbound file adapter). Each project is created as a individual file. All this works well, but the filenames created for the files are not very intuitive for the users looking at the files at the receiver point since there are quite a lot of similar files only distinguished from each other by a timestamp (usualy only separated by milliseconds).
    Is there a way to make the file adapter append data from one XML tag in the file name? I.e:
    file.TargetFilename='<project></project>'.txt
    file.writeMode=addTimeStamp
    giving i.e a file '1234567_timestamp.txt'
    This way I can include i.e. the project number + timestamp as filename making it user friendly for the people using the files.
    Any suggestions will be appreciated.
    Best Regards
    Odd Hilt

    HI
    Read the file without using key field. Just define the FCC parameters Header and Items PI will pick the file.
    DT_Source
       Record
         Header 0..1
           PO_Number String 0..1
         Items  0..unbounded
           Items String 0..1
    Do the FCC
    Header.fieldnames PO_Number
    other config for Header
    Items.fieldnames Items
    other config for items
    Thanks
    Gaurav
    Edited by: Gaurav Bhargava on Nov 12, 2008 10:27 AM
    Edited by: Gaurav Bhargava on Nov 12, 2008 10:27 AM

  • Can we use single DB Adapter for multiple tables

    Hi,
    I have 3 tables. ONE is the Master table and other 2 tables having Primary and foreign key relationships with the Master Table.
    Data is enter in the Tables 2 & 3 (based on Master Table Primary key)
    I have to use a DB Adapter in BPEL Process which will be in Transaction (what type of Transaction is Best in such a Scenario).
    - How do i implement it in BPEL.
    - Can i use a Single DB Adapter for Multiple Tables ? (*Views* should not be created for tables 1, 2 & 3),
    as we want to avoid using multiple DB Adapters, as in future Tables may increase or decrease...
    - while in Transaction if Table 2 fails in Inserting / Updating, it should roll back on error and also if Table 3 fails in Inserting / Updating, it should roll back table 2 and Table 3.
    - how access to concurrent users is maintained / managed in such a scenarios.
    But everything should be done using Single DB Adapter without creating a separate view.
    Is this Possible in BPEL..
    If not, How do we design, Implementation such a BPEL Process in Transaction with a best in Performance (both BPEL Process & DB Resources)
    thanks in advance
    thanks & regards,
    anvv sharma

    Hi,
    Thanks for your quick response.
    I have few quires.
    If we use PL/SQL, how do we implement transaction and Exception Handling appropriately based on the Error / Exception in BPEL.
    If the Table 2 or Table 3 gives out Error / Exception, i have to handle it separately and intimate accordingly.
    Suppose, if Table 2 got Error / Exception - Error Handling would be 'Roll back due to error in Table2'
    Thanks & regads,

  • Dynamic file names in outbound file adapter

    I have to configure a flow in XI system wich takes an input file from one machine and after mapping process leaves the output file in target machine with the same name of input file.
    Name of input file won't be always the same (*.dat) so i cannot hardcode this name in outbound file adapter.
    Does anyone know how to transfer input filename from inbound file adapter to XI message and then to outbound file adapter?
    I know there is an option to create dynamic file names in file adapter but i think it isn't enought to solve my problem.

    I have to configure a flow in XI system wich takes an input file from one machine and after mapping process leaves the output file in target machine with the same name of input file.
    Name of input file won't be always the same (*.dat) so i cannot hardcode this name in outbound file adapter.
    Does anyone know how to transfer input filename from inbound file adapter to XI message and then to outbound file adapter?
    I know there is an option to create dynamic file names in file adapter but i think it isn't enought to solve my problem.

  • Get the filename after an outbound file adapter operation

    Hello,
    I am calling an outbound file adapter operation in BPEL (10.1.3.3) - using a filename mask with a sequence (i.e. po_%SEQ%.xml). I would like to provide the filename created at run-time (i.e. po_123.xml) to my BPEL process so I can send this in an e-mail. Is this possible?
    Thanks in advance!
    Rod
    Message was edited by: Rod - fixed a typo
    Rod Tunnels

    Well, I took a different approach - which looks like it will work. Instead of letting the adapter decide the filename - I'm generating that in BPEL and passing that to the adapter using the header variable. So ... nevermind. ;-)

  • Receiver file adapter posts multiples for singel source files

    Hi,
    I am using wild character in my input source ftp adapter but my receiver file adapter is creating multiple duplicate files. I am not sure why this happens as QOS is set to "Exactly Once" with archive processing mode. I tried with delete mode and the same error occurs. We had no problems testing it in our Dev environment but in QA its acting so strangely...Any advise is appreciated..
    -Teresa

    Hi,
    2. Homany files you posted at oubound directory?
    1.             You said duplicate files at receiver side, is the files having time stamp? if so those are not duplicate files.
    Thanks,
    Satish.

  • N:1 multimapping for multiple records in every file

    Hi ,
    I am merging 2 files into a single file. First file has employee personal data and second file has employee salary data.I have created BPM and used correlation on EmpNumber to merge both the files. In BPM I am using fork step to recieve two files then use Transform step to merge the files and finally send dtep to send the output file.
    If both the input files have single employee record my scenario works fine. However my requirenment is file one will have 10 employee records . File two will have salary details of the same 10 employees . I have to  merge both the files and create 10 final records of employees .
    Here i am not able to use correlation as i dont have unique key in the file.
    Can any one tell me how to achive this?
    Regards,
    Shabari

    Please check the below blog
    http://www.sdn.sap.com/irj/scn/weblogs?blog=/pub/wlg/10526
    Refer the blog for the BPM design. You can make changes such that you include a fork(with two branches)inside the loop.  But without correlation how will you merge similar employee records.

  • TcDataSet for Multiple Records

    Hi,
    I am trying to write one custom adapter to fetch some values from OIM database using sql query
    That sql query returns multiple records.
    I am using tcdataset APIs to get values. However I am able to write the code when it returns 1 value
    What is the method of tcDataSet can be used when it returns multiple rows.
    Pls suggest.It would be great if you could send some used code snippet.
    Thanks!!

    JRS wrote:
    Hi,
    I am trying to write one custom adapter to fetch some values from OIM database using sql query
    That sql query returns multiple records.
    I am using tcdataset APIs to get values. However I am able to write the code when it returns 1 value
    What is the method of tcDataSet can be used when it returns multiple rows.
    Pls suggest.It would be great if you could send some used code snippet.
    Thanks!!
    tcDataSet dataSet = new tcDataSet();
    dataSet.setQuery(this.oDataProvider, strCheckQuery);
    dataSet.executeQuery();
    for(int i=0;dataSet.getRowCount() i++){
         dataset.gotoRow(i);
    }

  • Use of LIKE in where clause of select statement for multiple records

    Hi Experts,
    I have a account number field which is uploaded from a file. Now this account numbers uploaded does not match fully with sap table account numbers but it contains all of the numbers provided in the file mostly in the upright positions.
    For example in file we have account number as 2ARS1 while in sap table the value is 002ARS1.
    And i want to fetch data from sap table based on account number uploaded. So, i am trying to use LIKE with for all entries but its not working as mentioned below but LIKE is not working with FOR ALL ENTRIES.
    data : begin of t_dda occurs 0,
            dda(19) type c,
           end of t_dda.
    data : begin of t_bukrs occurs 0,
            bukrs type t012k-bukrs,
           end of t_bukrs.
    data : dda type t012k-bankn,
           w_dda type t012k-bankn.
    CONCATENATE '%'
                             '2ARS1'
                     INTO  W_DDA.
    MOVE W_DDA TO T_DDA-DDA.
    APPEND T_DDA.
    CLEAR T_DDA.
    free t_bukrs.
    SELECT BUKRS
      FROM T012K
      into TABLE t_bukrs
        for all entries in t_dda
    WHERE BANKN like t_dda-dda.
    Can anybody suggest what should i use to get the data for multiple account numbers using one select statement only instead on using SELECT UP TO 1 ROWS in LOOP....ENDLOOP ?
    Thanks in advance,
    Akash

    Hi,
    yes, For All entries won't work for LIKE with '%  '.
    I think the other alternative is go for Native SQL by writing sub-query
    sample code is here:
    data: begin of i_mara occurs 0,
              matnr like mara-matnr,
              matkl like mara-matkl,
           end of i_mara.
    exec sql.
    select matnr, matkl from mara where matnr in (select matnr from marc) and matnr like '%ma' into :i_mara
    endexec.
    loop at i_mara.
    write:/ i_mara-matnr, i_mara-matkl.
    endloop.
    hope u got it.
    regards
    Mahesh
    Edited by: Mahesh Reddy on Jan 21, 2009 2:32 PM

  • Physical Path in File Adapter for Windows Server

    Hi There,
    I am trying to read a file from a directory on a windows server and write it to the same server in a different directory.
    In the file adapter I gave the value for physical path as below:
    "\\10.xx.34.xxx\Share\input".
    However I am getting below message in the debug:
    "Value specified for input Physical/Logical Directory is not a directory or is not readable."
    Hence my question is what should be the value in the physical path for windows server.
    Thanks in Advance
    Krishna
    Edited by: user452458 on Jul 19, 2010 11:30 AM

    Thanks for your reply again.
    Actually they are not on the same file system.
    I have successfully used FTP Adapter to connect to a UNIX server.
    However is it possible to use FTP Adapter for Windows Server. If so how should be the corresponding JNDI defined.
    I am getting below error when I use FTP Adapter for Windows Server:
    Unable to send file to server. [Caused by: A remote host refused an attempted connect operation.]
    ; nested exception is:
         ORABPEL-11429
    Error sending file to FTP Server.
    Unable to send file to server. [Caused by: A remote host refused an attempted connect operation.]
    Please ensure 1. Specified remote output Dir has write permission 2. Output filename has not exceeded the max chararters allowed by the OS and 3. Remote File System has enough space.
    Please Advise.
    Thanks
    Krishna

Maybe you are looking for