XML to Idoc- XML size more than 200 MB

Hi Experts,
I have an interface with source xml file size more than 200MB. I need to map it to Idoc and send it to ERP. I am planning to change wsdl of idoc(changing occurence 1.1 to 0.99999) to send multiple idocs in one message.
I want to split XML file somehow and send it. what are the different options i have to achieve that ?  I am short of ideas.
I know if we have flat file,  we can use recordset per message in sender file adapter.
OR
Cleint is willing to send it as a SOAP request, i do not want to use SOAP for such a big file.
Regards
Inder

Hi Inder,
As per sap recommendation we would be able to handle 100 MB, you need to tune your server by increasing the [arametersso that you would be able to handle the messages with big payload.
By default the parameter  icm/HTTP/max_request_size_KB will be 10240 which can handle 100MB of file size.if you increase the parameter value by tuning ur system you can process a file bigger than that..
Please refer to the below links for reference....
[link1|http://help.sap.com/saphelp_nw04s/helpdata/en/58/108b02102344069e4a31758bc2c810/content.htm]
[link2|http://help.sap.com/saphelp_nwpi71/helpdata/de/95/1528d8ca4648869ec3ceafc975101c/content.htm]
as per the above suggestions the best practice is to sand as a multiple idocs splitting into chunks.
Cheers!!!!
Naveen.

Similar Messages

  • Handling xml message of size more than 100mb in SAP PI 7.1

    Dear Experts,
    Is it possible for PI to pick-up and process a XML message of size more than 100 MB in PI 7.1 EHP-1?
    If yes, can you please let me know how to handle it?
    Thank  you.

    Hi Saravana,
    it is not a best practice to more than 100mb..
    you can increase below parameters and so that you would be able to process for the best..
    u2022     UME Parameters :  May be we need to look into the pool size and poolmax wait parameters - UME recommended parameters (like: poolmaxsize=50, poolmaxwait=60000)
    u2022     Tuning Parameters:  May be we need to look/define the Message Size Limit u201Clike: EO_MSG_SIZE_LIMIT = 0000100u201D under tuning category
    u2022     ICM Parameters: May be we need to consider ICM parameters (ex: icm/conn_timeout = 900000. icm/HTTP/max_request_size_KB = 2097152)
    Thanks and Regards,
    Naveen

  • Load and Read XML file size more than 4GB

    Hi All
    My environment is Oracle 10.2.0.4 on Solaris and I have processes to work with XML file as below detail by PL/SQL
    1. I read XML file over HTTP port into XMLTYPE column in table.
    2. I read value no.1 from table and extract to insert into another table
    On test db, everything is work but I got below error when I use production XML file
         ORA-31186: Document contains too many nodes
    Current XML size about 100MB but the procedure must support XML file size more than 4GB in the future.
    Belows are some part of my code for your info.
    1. Read XML by line into variable and insert into table
    LOOP
    UTL_HTTP.read_text(http_resp, v_resptext, 32767);
    DBMS_LOB.writeappend (v_clob, LENGTH(v_resptext), v_resptext);
        END LOOP;
        INSERT INTO XMLTAB VALUES (XMLTYPE(v_clob));
    2. Read cell value from XML column and extract to insert into another table
    DECLARE
    CURSOR c_xml IS
    (SELECT  trim(y.cvalue)
    FROM XMLTAB xt,
    XMLTable('/Table/Rows/Cells/Cell' PASSING xt.XMLDoc
    COLUMNS
    cvalue
    VARCHAR(50)
    PATH '/') y;
        BEGIN
    OPEN c_xml;
    FETCH c_xml INTO v_TempValue;
    <Generate insert statement into another table>
    EXIT WHEN c_xml%NOTFOUND;
    CLOSE c_xml;
        END
    And one more problem is performance issue when XML file is big, first step to load XML content to XMLTYPE column slowly.
    Could you please suggest any solution to read large XML file and improve performance?
    Thank you in advance.
    Hiko      

    See Mark Drake's (Product Manager Oracle XMLDB, Oracle US) response in this old post: ORA-31167: 64k size limit for XML node
    The "in a future release" reference, means that this boundary 64K / node issue, was lifted in 11g and onwards...
    So first of all, if not only due to performance improvements, I would strongly suggest to upgrade to a database version which is supported by Oracle, see My Oracle Support... In short Oracle 10.2.x was in extended support up to summer 2013, if I am not mistaken and is currently not supported anymore...
    If you are able to able to upgrade, please use the much, much more performing XMLType Securefile Binary XML storage option, instead of the XMLType (Basicfile) CLOB storage option.
    HTH

  • How to change the size of a simple brush by tilt by more than 200%?

    Hi,
    how to change the size of a simple brush with tilt by more than 200%?
    Greetings!
    Daniel
    Ps: sorry for saying this, but your forum is surprisingly userunfriendy designed for a design oriented company.

    Hi,
    How is it going? Are the suggestions provided by the above replied helpful? If you need further help regarding this issue, please don’t hesitate to let us know.
    Besides, for scripts, in order to get better help, we can ask for suggestions in the following scripting forums.
    The Official Scripting Guys Forum
    http://social.technet.microsoft.com/Forums/scriptcenter/en-US/home?forum=ITCG
    Windows PowerShell
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/home?forum=winserverpowershell
    TechNet Subscriber Support
    If you are TechNet Subscription user and have any feedback on our support quality, please send your feedback here.
    Best regards,             
    Frank Shen

  • CONVERT_OTFSPOOLJOB_2_PDF is not working with more than 200 pages

    HI all,
    i am trying to print out Inspection Results (OTFs generated by QGA3 Transaction) via the CONVERT_OTFSPOOLJOB_2_PDF funtion module.
    But if the Inspection Results have more than 200 pages the FM only creates a PDF without content.
    The PDF is there but it has 0 bytes.
    This is how i fill the FM:
    CALL FUNCTION 'CONVERT_OTFSPOOLJOB_2_PDF'
             EXPORTING
                  src_spoolid              = i_spool
                  no_dialog                = 'X'
             IMPORTING
                  pdf_bytecount            = wl_numbytes
                  pdf_spoolid              = wl_pdfspoolid
                  btc_jobname              = wl_jobname
                  btc_jobcount             = wl_jobcount
             TABLES
                  pdf                      = tl_pdf
             EXCEPTIONS
                  err_no_otf_spooljob      = 1
                  err_no_spooljob          = 2
                  err_no_permission        = 3
                  err_conv_not_possible    = 4
                  err_bad_dstdevice        = 5
                  user_cancelled           = 6
                  err_spoolerror           = 7
                  err_temseerror           = 8
                  err_btcjob_open_failed   = 9
                  err_btcjob_submit_failed = 10
                  err_btcjob_close_failed  = 11.
    Does anyone have an idea ?
    Thanks and Regards
    Volker

    That's because the function reverts to background processing when the page size is over 99 pages (see the code).  I posted how to handle this a while back but basically you can either use RSPO_RETURN_SPOOLJOB directly or you can continue to use your current function but you have to poll for the completion of the job (BP_JOB_STATUS_GET) and then get the resulting spool in RAW format with RSPO_RETURN_SPOOLJOB then convert it:
    FIELD-SYMBOLS: <fs_line> TYPE x.
    DATA: gv_pdf TYPE xstring.
    CALL FUNCTION 'RSPO_RETURN_SPOOLJOB'
      EXPORTING
        rqident              = p_spono
        desired_type         = 'RAW'
      TABLES
        buffer               = gt_buffer
      EXCEPTIONS
        no_such_job          = 1
        job_contains_no_data = 2
        selection_empty      = 3
        no_permission        = 4
        can_not_access       = 5
        read_error           = 6
        type_no_match        = 7
        OTHERS               = 8.
    LOOP AT gt_buffer INTO gs_buffer.
      ASSIGN gs_buffer TO <fs_line> CASTING TYPE x.
      CONCATENATE gv_pdf <fs_line> INTO gv_pdf IN BYTE MODE.
    ENDLOOP.

  • Case structure with more than 200 cases, slower lab view app

    Problem Description :
    i am currently making a vi which compose of a simple case structures, a case structure with more than 200 cases, what i noticed is that the more i added cases the slower the labview reacts, it takes time for me to route wire on each of the cases.
    i dont know if there are any setting in the labview environment that i can increase the buffer size? i dont know... but its getting slow (only on the
    vi with more cases in a case structure)
    mytestautomation.com
    ...unleashed the power, explore and share ideas on power supply testing
    nissanskyline.org
    ...your alternative nissan skyline information site

    Can you clarify what you mean by "slow". Does it run more slowly or does it feel sluggish during editing?
    Could you attach the actual VI?
     Overall, I agree with the notion of local variable overuse. For example, you read the same local variable twice in the same case. Is there a need for this? Is there a chance that it varies between the two reads? Typically, you would place the terminal in front of the case structure and wire to all the locations where you need the value. This ensures that the data is consistent. Still, this should not cause any slowdown problem.
    (On a sidenote, the code in the upper left corner is inefficient. You convert the same data to lower case with each iteration, so you might want to place this before the loop and do it only once (well, if you are lucky, LabVIEW will fold the operation, but still....). Also your "index array" operation can be replaced by autoidenxing of the 2D array on the outer loop boundary.)
    LabVIEW Champion . Do more with less code and in less time .

  • How to display more than 200 rows in the table?

    Hi Experts,
    Is that possible to display more than 200 rows in a Table.
    when i query from table, it has 1000 rows, i want to display all the 1000 rows in the Table.
    when i query, the values are displaying up to 201 rows only,
    when i click next on 200 - 201 it throws Exception says
    *"Query has exceeded 200 rows. Potentially more rows exist, please restrict your query."*
    let me know how to display all the 1000 rows in a table without Exception.
    Any idea will be highly appreciated.
    Thanks
    Aswath

    The number of rows retrieved is controlled by profile option "FND: View Object Max Fetch Size". I believe 200 is the default value. Pl see these MOS Docs
    386402.1- Query Has Exceeded 200 Rows
    275876.1 - Oracle Application Framework Profile Options Release 11i (11.5.10)
    HTH
    Srini

  • Can someone confirm that "Sent" and "Trash" can hold more than 200 email?

    Can some confirm that Mail on iPad/iPhone
    can hold more than 200 email in the "Sent" and "Trash" folders, and that
    setting "Show 200 Most Recent Msgs" affects only the size of "Inbox" ?
    Thanks in advance.
    (The ISP offers only a POP server; no IMAP or MS Exchange.)

    Check the bit rate for your songs. If 200 songs use up 7GB, they must be in a large bit rate format.
    In iTunes, highlight a song and click on File/Get Info/Summary and check the Kind, Size and Bit Rate. Apple's figures for the number of songs on an iPod is based on a four minute song at 128Kbps.
    I have a song in three different formats  in my iTunes Library. Notice the difference is Size for the different formats. The song is 3 minutes 55 seconds long.
    Kind is AAC audio file. Size is 7.5MB. Bit Rate is 256Kbps
    Kind is Apple Losslesss audio file. Size is 20.8MB. Bit Rate is 738Kbps.
    Kind is AAC audio file. Size is 3.8MB. Bit Rate is 128Kbps
    Notice the enormous difference. I believe the iTunes Store now sells MP3s in 256KBps format, wheras they used to be in 128Kbps MP3 format. For a given Bit Rate, the MP3 format is about the same size as Apple's AAC format.
    So you will need to check several songs, because there will be a variety, I'm sure.

  • How to change the owner of a GPO using command line/script? - More than 200!

    I have more than 200 GPOs on the AD DOmain
    Half of the GPos were created by sveeral different people or are owned by the domain admins
    How can I change all of then, to a new owner, using the more "quick and easy" method?
    there a tool/utility for that? (like icacls.exe/subinacl.exe?)
    There is a VBS function or PS cmdlet?

    Hi,
    How is it going? Are the suggestions provided by the above replied helpful? If you need further help regarding this issue, please don’t hesitate to let us know.
    Besides, for scripts, in order to get better help, we can ask for suggestions in the following scripting forums.
    The Official Scripting Guys Forum
    http://social.technet.microsoft.com/Forums/scriptcenter/en-US/home?forum=ITCG
    Windows PowerShell
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/home?forum=winserverpowershell
    TechNet Subscriber Support
    If you are TechNet Subscription user and have any feedback on our support quality, please send your feedback here.
    Best regards,             
    Frank Shen

  • Outofmemory while creating a new object of file with size more than 100 MB

    I have created an application which generates a report by getting the data from our archived files (.zip file).By the time, the application is reaching a file with size more than 100 mb, it is running out fo memory while creating the object of that particular file. Can some one help me by tellin if there id way to resolve this issue?
    Thanks in advance

    If you're getting OutOfMemoryError, the simplest thing to try is to give the VM more memory at startup. For Sun's VM, I believe the default is 64 MB. You can increase this by using the -X args that control the heap size. For example: java -Xms128m -Xmx256m ... etc. ... This says start with 128 MB of heap, and allow it to grow up to 256 MB.
    One thing to consider, though, is do you need that much stuff in memory at once? A more intelligent approach might be to unpack the archive to the file system, and then read a file at a time or a line at a time or a whatever at a time is appropriate for the processing you need to do.

  • HT204406 itunes crashes my computer (windows 7) while trying to upload my music. Its happened multiple times now and I have no songs which are more than 200 mb. Dont understand what is breaking it

    My itunes keeps crashing while trying to upload music ti icloud using match. I have checked my collection and there are no songs more than 200 mb. The problem is getting so bad that now my OS (windows 7) is crashing. Dont understand what is breaking itunes....

    My itunes keeps crashing while trying to upload music ti icloud using match. I have checked my collection and there are no songs more than 200 mb. The problem is getting so bad that now my OS (windows 7) is crashing. Dont understand what is breaking itunes....

  • How to send a mail from SAP to Users more than 200

    Hi
    some one can help i have one scenario in my   company we have to send a mail to our user more than 200 to inform price changes once in a month or two month
    Thanks in advance
    Best Regards,
    MH

    Hi Mohammed
    As you want to send to around 200 customers then maintain a condition record with the combination of Output medium 7 in the Output medium field
    Second Option is Go to VA02 -> Edit -> Editor to reach the SAPscript Editor and then come back to the header output data and then The system sends the electronic mail at the time you have specified in the timing data..
    Regards
    Srinath

  • Couldn't Upload file with size more than 2K

    Hi,
    i am developing an webapplication where in i have to upload a file.we are using hibernate for data storing and retrieval.problem comes when i am uploading a file with size more than 2k bytes.if the file size lesser than 2k it works fine.i found that its a bug with thin jdbc driver version 9.but i am using 10g jdbc driver.still problem exists. i am not suppose to use oci drivers.
    working on
    OS: windows Xp.
    Apps :weblogic8.1
    DB: oracle 9i
    if anyone has solution plz mailme at [email protected]

    I'm not sure where the issue would be. Are you saying that you are using a 9i driver to access a 10g database? If so, download the newer driver and add it to your WEB-INF/lib directory. Remove the existing driver.
    If, on the other hand, you are using a 10g driver for a 9i database, I would not expect problems. However, you could always download the older driver and try it out.
    Or am I missing something?
    - Saish

  • How to transfer my iphone video to computer,the video siza more than 1.5G

    how to transfer my iphone4 video to computer. the video siza more than 1.5G

    Follow the instructions here or here, and then sync the iPhone with the iTunes library containing the content.
    (112570)

  • SP01: More than 200 downloads of SP01 - great! Please give feedback here:

    Hello,
    2005A SP01 PL02 has been downloaded by more than 200 different SAP Business One partners. That's great and proves the high demand for allowing early hands-on across the partner community!
    However, feedback is rare. Usually this means that no major issues occured and that the information given (e.g. in the documentation of DB and UI changes + potential compatibility issues) is OK as well.
    That would be great!
    Could you confirm - or reject that interpretation?
    Thanks in advance,
    Frank

    I have been studying SBO 2005 A SP1 for a while. I have come to terms with the License Server issue.
    However, I now have a new problem.
    All the add-ons (for different databases) cannot start anymore.
    When I tried Outlook Integration, I will a series of error messages, including
    1. Resource error (no title)
    2. Initialization failed (no title)
    I am given a choice to
    a. Restart Add-on
    b. Log off the company
    c. Continue without the Add-on
    I chose the last option and I get a dialog box that says
    BO_Outlook.exe - Application Error (Title)
    The instruction at "0x016cf850" referenced memory at "0x02de2f48". The memory could not be "read".
    Click on OK to terminate the program.
    When I tried XL Reporter, the error messages are slightly diferent.
    1. Unable to connect to DI. Resource Error (title = XL Reporter)
    I was given the System Message with the 3 options. I chose 3. No further messages.
    Any advice? These add-on modules were working fine.

Maybe you are looking for