JDBC to file scenario - Time limit exceeded

Hi All,
I am getting timelimit exceeded error in the inbound queues when I am processing the interface JDBC-XI-file. I have tried many options like registering the queues, changed the integration engine configuraion parameters for timeout , checked the memory space. When less data is being sent like 50kb then it is executing successfully. With large data the messages are getting stuck in the queues and then giving the timeout error.
Also when I  am checking in sm50 it showing that some rfc is going on hold.
Please suggest .
Thanks,
Aparna

Thanks for the reply Pragati.
What will be the UDF code for RFC lookup in case of passing the data in Queue. I am currently using the below code for gettng single value.
Also is it possible to get multiple output values which can be mapped to different target fields. Currently I am using different rfc lookups for different target fields. So if I can combine them into single rfc lookup them this may effect the performance.
   //write your code here
String content = "";
MappingTrace importanttrace;
importanttrace = container.getTrace();
// Create document builder to create DOM XML document
DocumentBuilderFactory factory = DocumentBuilderFactory.newInstance();
DocumentBuilder builder = null;
factory.setNamespaceAware(false);
factory.setValidating(false);
try {
        // Create XML document using document builder
        builder = factory.newDocumentBuilder();
} catch (Exception e) {
       importanttrace.addWarning("Error creating DocumentBuilder - " + e.getMessage());
        return null;
// filling the string with our RFC-XML (with values)
String m ="<?xml version=\"1.0\" encoding=\"UTF-8\"?><ns1:ZFI_GET_COCODE xmlns:ns1=\"urn:sap-com:document:sap:rfc:functions\"><COMPANYID>" + a + "</COMPANYID></ns1:ZFI_GET_COCODE>";
RfcAccessor accessor = null;
ByteArrayOutputStream out = null;
Payload result = null;
try
// 1. Determine a channel (Business system, Communication channel)
Channel channel = LookupService.getChannel("BS_SAP_DEC210","CC_RECEIVER_RFC_test");
// 2. Get a RFC accessor for a channel.
accessor = LookupService.getRfcAccessor(channel);
// 3. Create a xml input stream representing the function module request message.
InputStream inputStream = new ByteArrayInputStream(m.getBytes());
// 4. Create xml payload
XmlPayload payload = LookupService.getXmlPayload(inputStream);
// 5. Execute lookup
result = accessor.call(payload);
/*InputStream in = result.getContent();
out = new ByteArrayOutputStream(1024);
byte[] buffer = new byte[1024];
for (int read = in.read(buffer); read > 0; read = in.read(buffer)) {
out.write(buffer, 0, read);
content = out.toString();
if (result == null) {
                importanttrace.addWarning("result of RFC call is null");
catch(LookupException e)
importanttrace.addWarning("Error while lookup " + e.getMessage() );
Document docResponse = null;
InputStream in = result.getContent();
String res = "";
try {
        docResponse = builder.parse(in);
        if (docResponse == null) {
                importanttrace.addWarning("docResponse is null");
res =  docResponse.getElementsByTagName("COMPANY_CODE").item(0).getFirstChild().getNodeValue();
if (res == null) {
                importanttrace.addWarning("res is null");
catch (Exception e) {
      importanttrace.addWarning("Error when parsing RFC Response - " + e.getMessage());
try {
        // Free resources, close the accessor..
        if (accessor != null) {
                try {
                        accessor.close();
                } catch (LookupException e) {
                        importanttrace.addWarning( "Error while closing accessor " + e.getMessage());
} catch (Exception e) {
        importanttrace.addWarning("Result value not found in DOM - " + e);
// return the result obtained above
return res;
Thanks.

Similar Messages

  • Time Limit Exceeded. JDBC Adapter

    Hello guys,
    we have problem in an interface in production.
    The interface is asynchronous and is File -> PI -> JDBC
    Some messages have sometimes an error in PI monitorization, and we can see the next error:
    <?xml version="1.0" encoding="UTF-8" standalone="yes" ?>
    - <!--  Receiver Grouping
      -->
    - <SAP:Error SOAP:mustUnderstand="1" xmlns:SAP="http://sap.com/xi/XI/Message/30" xmlns:SOAP="http://schemas.xmlsoap.org/soap/envelope/">
      <SAP:Category>XIServer</SAP:Category>
      <SAP:Code area="INTERNAL">SYSTEM_DUMP</SAP:Code>
      <SAP:P1 />
      <SAP:P2 />
      <SAP:P3 />
      <SAP:P4 />
      <SAP:AdditionalText />
      <SAP:Stack>XXXX : XXXX___0000 : Time limit exceeded.</SAP:Stack>
      <SAP:Retry>M</SAP:Retry>
      </SAP:Error>
    It seems there is an error when sending to the JDBC adapter? The concurrency of the adpter is set to 1 and we send 7 messages to the same adapter and the messages are executed more or less at the same time.
    Could we have to do some PI configuration about queues? or maybe is about the concurrency of the adapter?
    Many thanks in advance.
    Regards,
    Xavier.

    You might need to increase MAXIMUM CONCURRENCY > 1 and set poolWaitingTime > 0 as per SAP Note 831162. By doing this will alow JDBC Receiver to process concurrency the request from Integration Engine.
    or you also can try to increase "Recv.maxConsumers" for JDBC Adapter. The current value is 5
    In addition above, since more messages are concurrently happening more or less at the same time, I would recommend also to use this parameter setting.... Go to advanced tab in the jdbc receiver channel.
    check the option "Disconnect from Database After Processing Each message".  This would help if the connection is not immediately return to the other thread after serving the first thread.  Plus it is a best practice to follow up.

  • Time Limit Exceeded in File - java- IDOC

    Hello,
    I have an interface which reads a text file in XI and uses a java mapping to produce IDOCs in an R3 system synchronously, the interface was running fine for more than 2 years . Since the text file is larger than 20 MB, we are splitting into small text files (2mb each) for easier processing. While processing, each file will hold the outbound queue (SMQ2) until it gets completed. after 2 years of using this interface, it started giving errors in the queue (Time Limit Exceeded) for files larger than 1mb.
    Any hint ??

    Swarup,
    increasing the RFC adapter time out from the visual admin has solve the problem,
    thank you for that.
    but what I was looking for is a way to find out what could be the root cause, since it was running 2 years with (300000 ms time out) without a problem. Now I must increase this time out which mean my system is not performing well!
    Agasthuri,
    as I mentioned, it was running fine with 2mb, now even 1mb cannot be proccessed,
    Thanks again.

  • PI 7.0: IDOCs struck in IDX5 with error "Time Limit Exceeded".

    Hi All,
    We have a File to IDOC scenario in PI 7.0. After mapping the IDOCs are posted from PI to ECC System.
    On a normal day this interface works good, yesterday we received a huge file which resulted in the creation of about 25000 IDOCs from one single file. The mapping went fine, however the IDOCs created were not posted into ECC System. When we monitor the IDOCs using transaction code IDX5 in PI system, we found the error message as "Time limit exceeded", the user shown was "PIAFUSER". To overcome this error, we increased the time limit of PIAFUSER from default to about 1500 seconds.
    Now, I want to push these IDOCs from PI into ECC System. Could you please let us know, how to push these IDOCs sitting in PI system to ECC?
    We do not want to reprocess the file from the beginning. Please let us know if it is possible to push the IDOCs without processing the file? If yes, how to reprocess?
    Thanks in advance.
    Regards,
    Manohar Dubbaka.

    Hi,
    the help documentation is as follows:
    Check the tRFC Status  
    Use
    tRFC calls which transfer IDocs use the function module IDOC_INBOUND_ASYNCHRONOUS at reception (before release 4.0: INBOUND_IDOC_PROCESS).
    If an IDoc in the sending system has been passed to tRFC (IDoc status "03"), but has not yet been input in the receiving system, this means that the tRFC call has not yet been executed.
    Activities
    To check the status of the tRFC calls, choose Tools ® IDoc Interface/ALE ® Administration ® Monitoring ® Troubleshooting ® RFC Queue (SM58) and specify any additional selection criteria.
    The program RSARFCEX restarts unsuccessful tRFC calls.
    You cannot choose the option is being executed in background processing.
    Best Regards,
    Erik Hubers

  • Message getting stuck in XBQO queue - Time limit exceeded

    Hi All,
    We have a BPM scenario in our project (on PI 7.0 SP18), where bundle of PEXR2002 Payment IDocs are received as a single flat file. This file is then consumed by the BPM, to split the message into multiple payments using Java Mapping.
    However, when we get an IDoc file of size greater than 5 MB (more than 500 IDocs), the message gets stuck in XBQO queue and eventually giving a SYSFAIL with the message "Time limit exceeded". Could you please let us know if you have encountered a similar issue and are aware of a possible solution.
    Any pointers to this will be really appreciated.
    Thanks & Regards,
    ROSIE SASIDHARAN

    H Rosie,
    1)  Goto SXMB_ADM-> Integration Engine Configuration->Parameter  EO_MSG_SIZE_LIMIT->possible values 0 - 2,097,151 (KB)
    The parameter EO_MSG_SIZE_LIMIT enables serial processing of messages of a particular size. This applies for messages with the quality of service Exactly Once (EO). If the message is larger than the parameter value, the message is processed in a separate queue.
    2)  Goto SXMB_ADM-> Integration Engine Configuration->Parameter  HTTP_TIMEOUT->possible values n Seconds, where n is a whole number.
    The parameter Specifies the timeout for HTTP connections (time between two data packages at line level). This value overrides the system profile parameter icm/server_port_n (for example, icm/server_port_0 : PROT=HTTP, PORT=50044, TIMEOUT=900). If you do not set the parameter HTTP_TIMEOUT or if you set the parameter to 0, then the setting for the system profile parameter is used.
      See SAP Note 335162  for sysfail issue....
      Hope these will help u....
    Regds,
    Pinangshuk.

  • Time limit exceeded in inbound qRfc

    Hi all!
    Please I need your help with this timeout issue. I have the following scenario:
    FTP --> XI (Java mapping) --> R3 (Idoc)
    Where from that Java mapping I do some RFC's to R3 in order to make conversions and validations.
    When I test this interface in development environment with a 1Mb flat file, it took 5 minutes to process. However, i did the same test in QA environment, it took pretty much more than 5 minutes and I got this "time limit exceeded" in the qRfc inbound queue (sxmb_moni t-code).
    I was told that the hardware beneath development and QA are different but theorically, the later should be faster.
    Anyone has any suggestion to resolve this? I will really appreciate it.
    Best regards!
    Vanesa.

    Hi,
    have a look on this doc called "HowTou2026 Investigate Timeouts in Synchronous XI/PI Scenario"
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/c059d583-a551-2c10-e095-eb5d95e03747
    You will certainly find your answer.
    Regards.
    Mickael

  • IDoc on outbound side from XI - time limit exceeded error

    Hi,
    I have a FIle to IDoc scenario and I'm creating there a lot of idocs (20000) in a single push. I'm getting an error "time limit exceeded" on outbound side with red flag in SXMB_MONI. How to increase this limit parameter? Where to do it? - PI? or R/3?
    Thank you,
    Olian

    Hi,
    Check thios below thread..
    Re: XI timeout error
    /people/michal.krawczyk2/blog/2006/06/08/xi-timeouts-timeouts-timeouts
    Regards,
    Srini

  • Time limit Exceeded problem

    Hi ,
    I have a file to idoc scenario .
    I have a file which generates many idocs (size atmost 5 mb)
    I am aware sap recommends to limit the message the file size to atmost 5mb.
    Do they mean file size of the payload ...because size increases in xi also .
    I just want to process such a file quickly .
    For that I am employing queue priortization at both sender and reciever .
    and would idoc packaging help at the receiver .
    Besides I am getting 'time limit exceeded in sm58' although I am seeing my idocs in idx5.
    Here are some of the things I have checked ..
    1) Parameters in SXMB_ADM - Integration Eningine configuration ---Specific configuration ...tried various tuning parameters ...but it didn't work..
    2) In partner profile in recieving CRM system made chagend the setting from trigger immediately ..to background Task ...
    (I am not certain how far that is relevant .
    If it is a problem with CRM system's ability to accept mass idocs (because a small number is acceptible ...under 1mb file size ) then seetings should in crm so that large number idocs get succesfully posted .
    I have posted this issue several times ...Kindlyhelp
    KIndly what parameters i should tweak in xi or that matter recieving crm system to take care of this problem.
    Kindly help

    HI,
    Large size of file processing
    In general, an extra sizing for XI memory consumption is not required. The total memory of the SAP Web Application Server
    should be sufficient except in the case of large messages (>1MB).
    To determine the memory consumption for processing large messages, you can use the following rules of thumb:
    Allocate 3 MB per process (for example, the number of parallel messages per second may be an indicator)
    Allocate 4 kB per 1kB of message size in the asynchronous case or 9 kB per 1kB message size in the synchronous case
    Example: asynchronous concurrent processing of 10 messages with a size of 1MB requires 70 MB of memory
    (3MB + 4 * 1MB) * 10 = 70 MB With mapping or content-based routing where an internal representation of the message payload
    may be necessary, the memory requirements can be much higher (possibly exceeding 20 kBytes per 1kByte
    message, depending on the type of mapping).
    The size of the largest message thus depends mainly on the size of the available main memory. On a normal 32Bit operating system, there is an upper boundary of approximately 1.5 to 2 GByte per process, limiting the respective largest message size.
    please check these links..
    /community [original link is broken]:///people/michal.krawczyk2/blog/2006/06/08/xi-timeouts-timeouts-timeouts
    Input Flat File Size Determination
    /people/shabarish.vijayakumar/blog/2006/04/03/xi-in-the-role-of-a-ftp
    data packet size  - load from flat file
    How to upload a file of very huge size on to server.
    Regards
    Chilla

  • Time limit Exceeded in SM58

    Hi ,
    I have a file to ido scenario.The file size is  suppose to be big (atmost 5mb).I am getting the following error in SM58
    Time limit Exceeded .
    I am able to see my idocs in idx5.
    Kindly help

    HI,
    Check the below links may be useful
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/43e3b690-0201-0010-c582-deedc80fa6bb
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f078394a-4469-2910-c4bf-853c75674694
    Regards
    Chilla

  • Short dump "Time limit exceeded" when searching for Business Transactions

    Hello Experts,
    We migrated from SAP CRM 5.2 to SAP CRM 7.0. After migration, our business transaction search (quotation, sales order, service order, contract etc) ends with the short dump "Time limit exceeded" in class CL_CRM_REPORT_ACC_DYNAMIC, method DATABASE_ACCESS. The select query is triggered from line 5 of this method.
    Number of Records:
    CRMD_ORDERADM_H: 5,115,675
    CRMD_ORDER_INDEX: 74,615,914
    We have done these so far, but the performance is still either poor or times out.
    1. DB team checked the ORACLE parameters and confirmed they are fine. They also checked the health of indices in table CRMD_ORDER_INDEX and indices are healthy
    2. Created additional indices on CRMD_ORDERADM_H and CRMD_ORDER_INDEX. After the creation of indices, some of the searches(without any criteria) work. But it takes more than a minute to fetch 1 or 2 records
    3. An ST05 trace confirmed that the selection on CRMD_ORDER_INDEX takes the most time. It takes about 103 seconds to fetch 2 records (max hits + 1)
    4. If we specify search parameters, say for example a date or status, then again we get a short dump with the message "Time limit exceeded".
    5. Observed that only if a matching index is available for the WHERE clause, the results are returned (albeit slowly). In the absence of an index, we get the dump.
    6. Searched for notes and there are no notes that could help us.
    Any idea what is causing this issue and what we can do to resolve this?
    Regards,
    Bala

    Hi Michael,
    Thanks. Yes we considered the note 1527039. None of the three scenarios mentioned in the note helped us. But we ran CRM_INDEX_REBUILD to check if the table CRMD_ORDER_INDEX had a problem. That did not help us either.
    The business users told us that they mostly search using the date fields or Object ID. We did not have any problem with search by Object ID. So we created additional indices to support search using the date fields.
    Regards,
    Bala

  • Time Limit exceeded Error while updating huge number of records in MARC

    Hi experts,
    I have a interface requirement in which third party system will send a big file say.. 3 to 4MB file into SAP. in proxy we
    used BAPI BAPI_MATERIAL_SAVEDATA to save the material/plant data. Now, because of huge amount of data the SAP Queues are
    getting blocked and causing the time limit exceeded issues. As the BAPI can update single material at time, it will be called as many materials
    as we want to update.
    Below is the part of code in my proxy
    Call the BAPI update the safety stock Value.
        CALL FUNCTION 'BAPI_MATERIAL_SAVEDATA'
          EXPORTING
            headdata                    = gs_headdata
            CLIENTDATA                  =
            CLIENTDATAX                 =
           plantdata                   = gs_plantdata
           plantdatax                  = gs_plantdatax
           IMPORTING
              return              = ls_return.
        IF ls_return-type <> 'S'.
          CALL FUNCTION 'BAPI_TRANSACTION_ROLLBACK'.
          MOVE ls_return-message TO lv_message.
    Populate the error table and process next record.
          CALL METHOD me->populate_error
            EXPORTING
              message = lv_message.
          CONTINUE.
        ENDIF.
    Can any one please let me know what could be the best possible approach for this issue.
    Thanks in Advance,
    Jitender
    Hi experts,
    I have a interface requirement in which third party system will send a big file say.. 3 to 4MB file into SAP. in proxy we
    used BAPI BAPI_MATERIAL_SAVEDATA to save the material/plant data. Now, because of huge amount of data the SAP Queues are
    getting blocked and causing the time limit exceeded issues. As the BAPI can update single material at time, it will be called as many materials
    as we want to update.
    Below is the part of code in my proxy
    Call the BAPI update the safety stock Value.
        CALL FUNCTION 'BAPI_MATERIAL_SAVEDATA'
          EXPORTING
            headdata                    = gs_headdata
            CLIENTDATA                  =
            CLIENTDATAX                 =
           plantdata                   = gs_plantdata
           plantdatax                  = gs_plantdatax
           IMPORTING
              return              = ls_return.
        IF ls_return-type <> 'S'.
          CALL FUNCTION 'BAPI_TRANSACTION_ROLLBACK'.
          MOVE ls_return-message TO lv_message.
    Populate the error table and process next record.
          CALL METHOD me->populate_error
            EXPORTING
              message = lv_message.
          CONTINUE.
        ENDIF.
    Can any one please let me know what could be the best possible approach for this issue.
    Thanks in Advance,
    Jitender

    Hi Raju,
    Use the following routine to get fiscal year/period using calday.
    *Data definition:
    DATA: l_Arg1 TYPE RSFISCPER ,
          l_Arg2 TYPE RSFO_DATE ,
          l_Arg3 TYPE T009B-PERIV .
    *Calculation:
    l_Arg2  = TRAN_STRUCTURE-POST_DATE. (<b> This is the date that u have to give</b>)
    l_Arg3  = 'V3'.
    CALL METHOD CL_RSAR_FUNCTION=>DATE_FISCPER(
      EXPORTING I_DATE = l_Arg2
                I_PER = l_Arg3
      IMPORTING E_FISCPER = l_Arg1  ).
    RESULT = l_Arg1 .
    Hope it will sove ur problem....!
    Please Assign points.......
    Best Regards,
    SG

  • JDBC-XI-FILE scenario. How to extract data from more than one table in JDBC

    Hi,
    I was asked a question like in JDBC-XI-FILE scenario........ How to extract data from more than one tables (i.e from JDBC system) ?? What is the logic to do the same ??
    I am not sure whether this is a valid question..........but any help in this regards is highly appreciated.
    Regards
    Kumar

    HI,
    Yes it can be possible ,please see the following links
    JDBC  Receiver with Oracle Stored Procedures
    configuring jdbc adapter with multiple tables
    RFC -> XI -> JDBC Scenario Updating Multiple Tables
    /people/alessandro.berta/blog/2005/10/04/save-time-with-generalized-jdbc-datatypes
    JDBC Adapter multiple Selects
    https://www.sdn.sap.com/irj/sdn/advancedsearch?query=jdbc%20with%20multiple%20tables&cat=sdn_all
    Regards
    Chilla..

  • TRFC error "time limit exceeded"

    Hi Prashant,
    No reply to my below thread...
    Hi Prashant,
    We are facing this issue quite often as i stated in my previous threads.
    As you mentioned some steps i have already followed all the steps so that i can furnish the jog log and tRFC details for reference long back.
    This issue i have posted one month back with full details and what we temporarily follow to execute this element successfully.
    Number of times i have stated that i need to know the root cause and permanent solution to resolve this issue as the log clearly states that it is due to struck LUWs(Source system).
    Even after executing the LUWs manually the status is same (Request still running and the status is in yellow color).
    I have no idea why this is happening to this element particularly as we have sufficient background jobs.
    we need change some settings like increasing or decreasing data package size or something else to resolve the issue permanently?
    For u i am giving the details once again
    Data flow:Standard DS-->PSA--->Data Target(DSO)
    In process monitor screen the request is in yellow color. NO clear error message s defined here.under update 0 record updated and missing message with yellow color except this the status against each log is green.
    Job log:Job is finished TRFCSSTATE=SYSFAIL message
    Trfcs:time limit exceeded
    What i follow to resolve the issue:Make the request green and manually update from PSA to Data target and the job gets completed successfully.
    Can you please tell me how to follow in this scenario to resolve the issue as i waiting for the same for long time now.
    And till now i didn't get any clue and what ever i have investigated i am getting replies till that point and no further update beyond this
    with regards,
    musai

    Hi,
    You have mentioned that already you have checked for LUWs, so the problem is not there now.
    In source system, go to we02 and check for idoc of type RSRQST & RSINFO. If any of them are in yellow status, take them to BD87 and process them. If the idoc processed is of RSRQST type, it would now create the job in source system for carrying out dataload. If it was of RSINFO type, it would finish the dataload in SAP BI side as well.
    If any in red, then check the reason.

  • Runtime error(Time limit exceeds)after executing select query

    Dear experts, whenever i executing the select query in this zprogram i am getting runtime error that time limit exceeds.i am using inner join and into table.after that also i am geetting error. how can i resolve it??
    SELECT LIKP~VBELN LIKP~WADAT_IST LIKP~VEHICLE_NO LIKP~TRNAME
              LIKP~VEHI_TYPE LIKP~LR_NO LIKP~ANZPK LIKP~W_BILL_NO
              LIKP~SEALNO1                                       " Seal NO1
              LIKP~SEALNO2                                       " Seal NO2
              LIPS~LFIMG
              VBRP~VBELN VBRP~VGBEL VBRP~MATNR VBRP~AUBEL VBRP~FKIMG
              VBAK~AUART
              VBRK~FKART VBRK~KNUMV VBRK~FKSTO
              FROM LIKP INNER JOIN LIPS ON LIKP~VBELN EQ LIPS~VBELN
                        INNER JOIN VBRP ON LIKP~VBELN EQ VBRP~VGBEL
                        INNER JOIN VBAK ON VBRP~AUBEL EQ VBAK~VBELN
                        INNER JOIN VBRK ON VBRP~VBELN EQ VBRK~VBELN
              INTO TABLE  I_FINAL_TEMP
              WHERE LIKP~VSTEL = '5100' AND
                 LIKP~WADAT_IST IN S_WADAT  AND
                    VBRP~AUBEL IN S_AUBEL AND
                    VBAK~AUART IN ('ZJOB','ZOR') AND
                    VBRK~FKART IN S_FKART AND
    *               VBRK~FKART IN ('ZF8','ZF2','ZS1') AND
                    VBRK~FKSTO NE 'X'.
    When I am debugging the select query.the cursor will not go to next step.after 15-20 minutes i am getting runtime error(time limit exceeds).
    how can i resolve it for that scenario??

    Looks like whole SD flow you trying to fetch in single query
    First you check the database statistic of these table are upto date in system ( Check with basis team )
    if this query was working fine earlier.
    Most of table involved are huge volume tables which queried with any primary key
    Any secondary index on created for LIKP on VSTEL WADET ?
    My suggestion would be split the selection queries and make use of primary or existing secondary index to fetch the desired result if possible. For testing purpose split the queries and find which is taking more time and which needs index by taking squel trace in ST05.
    Also take ST05 trace of this query in debugger ( New debugger -> special tool -> trace > ST05/SE30)

  • Tolerance Key LD: Blanket purchase order time limit exceeded

    Hi All,
    Can anybody tell me how the Invoice Block Tolerance Key <b>LD: Blanket purchase order time limit exceeded</b> works.
    As per my understanding, we need to maintain no days in the Tolreance Key LD , so that it checks for the Contract Validity & the Invoice Date with the tolerance limit days & block the invoice if it exceeds.
    But, it's asking for the value in the LD tolerance limit setting !!! Hope it should be no of days !!! Kindly let me know.
    Also let me know if you can explain me the settings &  steps to run this scenario.
    Thanks!
    Nanda

    Steve,
    Can you answer Nanda's question? I haven't really used this LD tolerance previously. I tried this on my sandbox system and can see the cause for his confusion although I am unable to explain it.
    Thanks,
    H Narayan

Maybe you are looking for