Exceeded time limit in running   CALL FUNCTION 'BAPI_COSTCENTERGROUP_CREATE

Hello all,
Please may I ask your help in creating cost center group.
I used BAPI COSTCENTERGROUP_CREATE, everything is ok except that when I run the program it dumps because the time it process is exceed the time limit. It's my first time to work with BAPI.
This is urgent.
Please help and advice me what to do.
I'll give the highest points.
THANK YOU SO MUCH. God bless.

Hi chenara ,
I did'nt  use this bapi before but this is what i would do :
1. put  a break-point just before enter the bapi and try to see
   what is the problem , ( look at what you send to the bapi ,
  go with F5 inside and try to see where it is stuck ) ,
  i know it is hard to debug sap bapi but sometimes it helps.
2. go to the basis team , they can help you to monitor
   what take the most time.
i hope this could help you.
Lilach Menashe.

Similar Messages

  • Check if A/R summary has exceeded time limit:exceeded by days/hours 42 14 ?

    When I place an sale order with VA01, there is a warnning:
    Check if A/R summary has exceeded time limit:exceeded by days/hours 42 14
    What does it mean?

    Hi
    Check wheather you have done credit management or not. So check in OVA8 wheather you have maintained data or not. Also check in FD32 wheather you have maintained A/R summary for that customer.or not .
    Regards
    Srinath

  • Setting time limit on running report

    Does anyone know if you can set a time limit on a report? I want a report to be canceled if it runs for more than 30 minutes. I see that you have the option in the Queue Manager when creating a job, but I want it to happen for my web-based reports...

    You can count the numer of times something occurs and use that as a condition to stop things.  How you go about stopping things depends on how your animation is set up.  If it is a straight timeline animation, you would just use a stop(); command.

  • Setting a time limit for outgoing calls SPA3102

    I need to put a time limit, say, 5 min, on outgoing calls with a certain prefix (are code). Is it possible with SPA3102 dialplan or otherwise? Thanks.

    Based on my best knowledge - no, it's not possible with just SPA3102 only.
    It is switch (PBX) job to terminate the call.

  • Time limit between running a job?

    A little background - I'm trying to create a trigger that logs all logins to a table. While testing, I noticed that if I deleted the table the trigger was trying to call then I could not log on because the trigger would fail. This would cause me to have
    to start SQL in single-user mode with -f parameter, and then use SQLCMD to connect then delete the trigger.
    I wanted to fix the trigger so that this would not happen. Here's what I did...
    Created a table msdb.MySchema.LogonHistory.
    and a trigger...
    CREATE TRIGGER [Tr_ServerLogon]
    ON ALL SERVER WITH EXECUTE AS 'sa'
    FOR LOGON
    AS
    BEGIN
    IF OBJECT_ID('msdb.MySchema.LogonHistory', 'U') IS NOT NULL
        BEGIN
            INSERT INTO msdb.MySchema.LogonHistory
            (DBUser, AppName, LogonTime, SystemUser)
            VALUES(USER, APP_NAME(), GETDATE(), ORIGINAL_LOGIN())
        END
    END
    This worked perfectly. If I deleted the table people could still log in.
    I wanted to expand to add an ELSE statement to send an e-mail if the table didn't exist.Problem is, I received an e-mail for EVERY connection. Simply logging in to Management Studio adds 5-6 rows to my LogonHistory table. Here's the ELSE statement I added:
    ELSE
        BEGIN
            EXEC msdb.dbo.sp_send_dbmail @profile_name = "DBMail", @recipients = '[email protected]', @subject = 'Your table doesnt exist', @body='Error'
        END
    I was thinking instead of executing sp_send_dbmail I could a job with that in it. The question is can I limit the job to run only once every hour? Or is there something else I could do within the trigger? I was thinking I could check send_request_date in
    sysmail_mailitems where subject = 'Your table doesnt exist' and compare the send_request_date to the current GETDATE() and only e-mail if it's greater than 1 hour difference.
    Just want to make sure there's not something easier/obvious that I'm missing.

    Yep.. You can configure the job to be executed once every hour. For that just go to job properties -> Schedules tab. Create a new schedule as below
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • Pages Exceed Time Limit

    Greetings -
    I have a quick question about logging pages that take more
    than xx seconds to load as set in CF Administrator. Can somebody
    intepret exactly what that means? That is, is that the time it
    takes for the server to send the data or is it the time it takes
    for the user to actually finish downloading the file.
    I guess what I'm really asking is can that be affected by the
    user's connection? If a user has a slow dial up connection, could
    that cause a page to be logged as taking longer than the limit set
    in the CF administrator?
    TIA,
    joe

    Greetings -
    I have a quick question about logging pages that take more
    than xx seconds to load as set in CF Administrator. Can somebody
    intepret exactly what that means? That is, is that the time it
    takes for the server to send the data or is it the time it takes
    for the user to actually finish downloading the file.
    I guess what I'm really asking is can that be affected by the
    user's connection? If a user has a slow dial up connection, could
    that cause a page to be logged as taking longer than the limit set
    in the CF administrator?
    TIA,
    joe

  • Error while running query "time limit exceeding"

    while running a query getting error "time limit exceeding".plz help.

    hi devi,
    use the following links
    queries taking long time to run
    Query taking too long
    with hopes
    Raja Singh

  • Time Limit exceeded while running in RSA3

    Hi BW Experts,
      I am trying to pull 1 lakh data from CRM to BI System.
    Before scheduling, i am trying to execute in RSA3.I am getting the error message as "Time Limit Exceeded".
    Pls suggest, why it is happening like this.
    Thanks in advance.
    Thanks,
    Ram.

    Hi,
                because huge data with in the stipulated time it is not executing and showing the all records ,so it is better to go any selection option by each document type or some else then u can add all the documents ,anyway in bw side we r running this job in background so no problem.if u want see all records at a  time then u can discuss with ur basis people to extend the time for that.
    Thanks & Regards
    sathish

  • I used the iTunes match workaround and now want to start over, but I am being told I have exceeded the limit even though I completely cleared the match cloud, mindful of the 1000-delete-at-a-time limit....what should I do?

    I have over 25k songs in iTunes. I originally tried the "create a second library" solution. I reduced to less than 25k songs and successfully ran iTunes match. Then I discovered the limitations of the second library solution regarding adding more music, managing playlists, and turning back on my main library and being greeted with all sorts of problems.
    So, I created a blank library, ran iTunes match, and its showed me all of my files in the cloud that had been matched or uploaded. I deleted all of the files, mindful of the 1000 file limit per delete. I have confirmed I have now deleted them all. An iTunes match update says I have no files in the cloud. Perfect.
    I reduced my library again, this time using the superior method of just changing the files I do not want to match to voice memos. That is done. I have less than 25 k songs.
    I waited 2 days and ran match again. Yikes. It did not let me match or upload anything, and every music files's cloud status just says "exceeded limit," despite the fact that my cloud is devoid of music and I have no access to any music via match.
    I am in a limbo where I have nothing in the cloud from match, but I can't add anything because match says I exceeded my limit.
    I emailed customer support, and got back a useless email linking me to a very general article about iTunes match. I called customer service and they said this was something the iTunes store people would need to address, and they can only be contacted by email. Sigh . . . .
    Any thoughts?
    Thanks

    Have you confirmed that you successfull purged iTunes Match by also looking on an iOS device?  If so, keep in mind that Apple's servers may be experiencing a heavy load right now.  They just added about 19 countries to the service and I've read a few accounts this morning that suggests all's not running perfectly right now.

  • Time Limit exceeded Error while updating huge number of records in MARC

    Hi experts,
    I have a interface requirement in which third party system will send a big file say.. 3 to 4MB file into SAP. in proxy we
    used BAPI BAPI_MATERIAL_SAVEDATA to save the material/plant data. Now, because of huge amount of data the SAP Queues are
    getting blocked and causing the time limit exceeded issues. As the BAPI can update single material at time, it will be called as many materials
    as we want to update.
    Below is the part of code in my proxy
    Call the BAPI update the safety stock Value.
        CALL FUNCTION 'BAPI_MATERIAL_SAVEDATA'
          EXPORTING
            headdata                    = gs_headdata
            CLIENTDATA                  =
            CLIENTDATAX                 =
           plantdata                   = gs_plantdata
           plantdatax                  = gs_plantdatax
           IMPORTING
              return              = ls_return.
        IF ls_return-type <> 'S'.
          CALL FUNCTION 'BAPI_TRANSACTION_ROLLBACK'.
          MOVE ls_return-message TO lv_message.
    Populate the error table and process next record.
          CALL METHOD me->populate_error
            EXPORTING
              message = lv_message.
          CONTINUE.
        ENDIF.
    Can any one please let me know what could be the best possible approach for this issue.
    Thanks in Advance,
    Jitender
    Hi experts,
    I have a interface requirement in which third party system will send a big file say.. 3 to 4MB file into SAP. in proxy we
    used BAPI BAPI_MATERIAL_SAVEDATA to save the material/plant data. Now, because of huge amount of data the SAP Queues are
    getting blocked and causing the time limit exceeded issues. As the BAPI can update single material at time, it will be called as many materials
    as we want to update.
    Below is the part of code in my proxy
    Call the BAPI update the safety stock Value.
        CALL FUNCTION 'BAPI_MATERIAL_SAVEDATA'
          EXPORTING
            headdata                    = gs_headdata
            CLIENTDATA                  =
            CLIENTDATAX                 =
           plantdata                   = gs_plantdata
           plantdatax                  = gs_plantdatax
           IMPORTING
              return              = ls_return.
        IF ls_return-type <> 'S'.
          CALL FUNCTION 'BAPI_TRANSACTION_ROLLBACK'.
          MOVE ls_return-message TO lv_message.
    Populate the error table and process next record.
          CALL METHOD me->populate_error
            EXPORTING
              message = lv_message.
          CONTINUE.
        ENDIF.
    Can any one please let me know what could be the best possible approach for this issue.
    Thanks in Advance,
    Jitender

    Hi Raju,
    Use the following routine to get fiscal year/period using calday.
    *Data definition:
    DATA: l_Arg1 TYPE RSFISCPER ,
          l_Arg2 TYPE RSFO_DATE ,
          l_Arg3 TYPE T009B-PERIV .
    *Calculation:
    l_Arg2  = TRAN_STRUCTURE-POST_DATE. (<b> This is the date that u have to give</b>)
    l_Arg3  = 'V3'.
    CALL METHOD CL_RSAR_FUNCTION=>DATE_FISCPER(
      EXPORTING I_DATE = l_Arg2
                I_PER = l_Arg3
      IMPORTING E_FISCPER = l_Arg1  ).
    RESULT = l_Arg1 .
    Hope it will sove ur problem....!
    Please Assign points.......
    Best Regards,
    SG

  • PI 7.0: IDOCs struck in IDX5 with error "Time Limit Exceeded".

    Hi All,
    We have a File to IDOC scenario in PI 7.0. After mapping the IDOCs are posted from PI to ECC System.
    On a normal day this interface works good, yesterday we received a huge file which resulted in the creation of about 25000 IDOCs from one single file. The mapping went fine, however the IDOCs created were not posted into ECC System. When we monitor the IDOCs using transaction code IDX5 in PI system, we found the error message as "Time limit exceeded", the user shown was "PIAFUSER". To overcome this error, we increased the time limit of PIAFUSER from default to about 1500 seconds.
    Now, I want to push these IDOCs from PI into ECC System. Could you please let us know, how to push these IDOCs sitting in PI system to ECC?
    We do not want to reprocess the file from the beginning. Please let us know if it is possible to push the IDOCs without processing the file? If yes, how to reprocess?
    Thanks in advance.
    Regards,
    Manohar Dubbaka.

    Hi,
    the help documentation is as follows:
    Check the tRFC Status  
    Use
    tRFC calls which transfer IDocs use the function module IDOC_INBOUND_ASYNCHRONOUS at reception (before release 4.0: INBOUND_IDOC_PROCESS).
    If an IDoc in the sending system has been passed to tRFC (IDoc status "03"), but has not yet been input in the receiving system, this means that the tRFC call has not yet been executed.
    Activities
    To check the status of the tRFC calls, choose Tools ® IDoc Interface/ALE ® Administration ® Monitoring ® Troubleshooting ® RFC Queue (SM58) and specify any additional selection criteria.
    The program RSARFCEX restarts unsuccessful tRFC calls.
    You cannot choose the option is being executed in background processing.
    Best Regards,
    Erik Hubers

  • Program terminated: Time limit exceeded, ABAP performance, max_wprun_time

    Hi,
    I am running an ABAP program, and I get the following short dump:
    Time limit exceeded. The program has exceeded the maximum permitted runtime and has therefore been terminated. After a certain time, the program terminates to free the work processfor other users who are waiting. This is to stop work processes being blocked for too long by
    - Endless loops (DO, WHILE, ...),
    - Database acceses with large result sets,
    - Database accesses without an apporpriate index (full table scan)
    - database accesses producing an excessively large result set,
    The maximum runtime of a program is set by the profile parameter "rdisp/max_wprun_time". The current setting is 10000 seconds. After this, the system gives the program a second chance. During the first half (>= 10000 seconds), a call that is blocking the work process (such as a long-running SQLstatement) can occur. While the statement is being processed, the database layer will not allow it to be interrupted. However, to stop the program terminating immediately after the statement has been successfully processed, the system gives it another 10000 seconds. Hence the maximum runtime of a program is at least twice the value of the system profile parameter "rdisp/max_wprun_time".
    Last error logged in SAP kernel
    Component............ "NI (network interface)"
    Place................ "SAP-Dispatcher ok1a11cs_P06_00 on host ok1a11e0"
    Version.............. 34
    Error code........... "-6"
    Error text........... "connection to partner broken"
    Description.......... "NiPRead"
    System call.......... "recv"
    Module............... "niuxi.c"
    Line................. 1186
    Long-running programs should be started as background jobs. If this is not possible, you can increase the value of the system profile parameter "rdisp/max_wprun_time".
    Program cannot be started as a background job. We have now identified two options to solve the problem:
    - Increase the value of the system profile parameter "rdisp/max_wprun_time"
    - Improve the performance of the following SELECT statement in the program:
    SELECT ps_psp_pnr ebeln ebelp zekkn sakto FROM ekkn
    INTO CORRESPONDING FIELDS OF TABLE i_ekkn
    FOR ALL ENTRIES IN p_lt_proj
    WHERE ps_psp_pnr = p_lt_proj-pspnr
    AND ps_psp_pnr > 0.
    In EKKN we have 200 000 entries.
    Is there any other options we could try?
    Regards,
    Jarmo

    Thanks for your help, this problem seems to be quite challenging...
    In EKKN we have 200 000 entries. 199 999 entries have value of 00000000 in column ps_psp_pnr, and only one has a value which identifies a WBS element.
    I believe the problem is that there isn't any WBS element in PRPS which has the value of 00000000. I guess that is the reason why EKKN is read sequantially.
    I also tried this one, but it doesn't help at all. Before the SELECT statement is executed, there are 594 entries in internal table p_lt_proj_sel:
      DATA p_lt_proj_sel LIKE p_lt_proj OCCURS 0 WITH HEADER LINE.
      p_lt_proj_sel[] = p_lt_proj[].
      DELETE p_lt_proj_sel WHERE pspnr = 0.
      SORT p_lt_proj_sel by pspnr.
      SELECT ps_psp_pnr ebeln ebelp zekkn sakto FROM ekkn
      INTO CORRESPONDING FIELDS OF TABLE i_ekkn
      FOR ALL ENTRIES IN p_lt_proj_sel
      WHERE ps_psp_pnr = p_lt_proj_sel-pspnr.
    I also checked that the index P in EKKN is active.
    Can I somehow force the optimizer to use the index?
    Regards,
    Jarmo

  • IDoc Tracking: TIME LIMIT EXCEEDED

    Hello!
    We have about 15 different R/3 platforms connected to XI using the IDoc adapter. If I use the IDoc Tracking functionality in IDX5 it usually is working fine and the IDoc number in the receiving system and IDoc status are returned within seconds.
    However, for one R/3 platform it was always taking about 30 minutes until the IDoc number and IDoc status were returned. The status bar displays that the IDOC_DATE_TIME_GET function module is executed.
    Now it runs for over one hour until it stops with a TIME LIMIT EXCEEDED error message and no IDoc numbers and status is returned any more. What needs to be corrected to speed up IDoc tracking for this particular system?
    Regards, Tanja

    Not sure if this would help, but this link speaks of how to bypass time limit exceeded
    http://www.erpgenie.com/abaptips/content/view/490/62/

  • Mails to 1 particular user is getting bounced - Command time limit exceeded

    I've about 100 mail clients. Since yesterday morning, I've had one user whose incoming emails have been getting bounced. I've run a "mailbfr -m user" w/o success. The server is current on updates and has been restarted. All other users are functioning correctly. Notes:
    -IMAP account
    -This user has a 4GB Sent folder. All other folders are under 2GB.
    -This user has recently starting syncing Notes from a Blackberry.
    THE SMTP LOG:
    Oct 7 08:26:52 myserver postfix/pipe[488]: D1BCE11CE230: to=<[email protected]>, relay=cyrus, delay=1000, delays=0.01/0/0/1000, dsn=5.3.0, status=bounced (Command time limit exceeded: "/usr/bin/cyrus/bin/deliver")
    A PORTION OF THE BOUNCE EMAIL SENT TO SENDER:
    This is the mail system at host myserver.com.
    I'm sorry to have to inform you that your message could not
    be delivered to one or more recipients. It's attached below.
    For further assistance, please send mail to postmaster.
    If you do so, please include this problem report. You can
    delete your own text from the attached returned message.
    The mail system
    <[email protected]>: Command time limit exceeded:
    "/usr/bin/cyrus/bin/deliver"
    Reporting-MTA: dns; myserver.com
    X-Postfix-Queue-ID: AD79711CE1DD
    X-Postfix-Sender: rfc822; [email protected]
    Arrival-Date: Thu, 7 Oct 2010 08:07:57 -0700 (PDT)
    Final-Recipient: rfc822; [email protected]
    Original-Recipient: rfc822;[email protected]
    Action: failed
    Status: 5.3.0
    Diagnostic-Code: x-unix; internal software error
    Any help?

    Solved. This person was in-process of moving Blackberry notes to the imap account. There was an interruption during the transfer phase and the server's Notes folder became bad somehow. When Mailbfr hit that particular folder, it stopped the rebuild process.
    I removed the Notes folder and ran Mailbfr again. The Notes folder was recreated and the rebuild function completed.
    Mail delivery has been restored to this user.

  • How to run the Function module Parallelly at a time.

    Hi all,
    LOOP AT ITAB.
    call function 'Z_SAMPLE'
    ENDLOOP.
    Here if ITAB is having 4 records, Function module Z_SAMPLE will be runned 4 times one by one .
    My requirement is If ITAB is having 4 record ,Function module should be runned 4 times parallely(parallel process) .
    Please tell me how i can do this  Can i run that statement 4 times by removing the loop or
    Can i Create 4 different jobs and run that jobs at a time if yes please tell me how to creat the jobs and run parallelly.
    Please suggest me the solution
    Please do ask  me if question is not clear
    Thanks in Advance
    Ajay

    Is it possible ?
    DATA : f1 TYPE string VALUE '1',
           f2 TYPE string VALUE '2',
           f3 TYPE string VALUE '3',
           f4 TYPE string VALUE '4'.
    DATA : f1o TYPE string ,
           f2o TYPE string ,
           f3o TYPE string,
           f4o TYPE string.
    PERFORM call_fm USING f1 CHANGING f1o.
    PERFORM call_fm USING f2 CHANGING f2o.
    PERFORM call_fm USING f3 CHANGING f3o.
    PERFORM call_fm USING f4 CHANGING f4o.
    WRITE f1o.
    WRITE f2o.
    WRITE f3o.
    WRITE f4o.
    *&      Form  CALL_FM
    *       text
    *      -->P_F1  text
    FORM call_fm  USING f1 CHANGING f1o.
    * FUNCTION Z_TEST_MULTI.
    *"*"Local Interface:
    *"  IMPORTING
    *"     VALUE(A) TYPE  STRING OPTIONAL
    *"  EXPORTING
    *"     VALUE(B) TYPE  STRING
    *b = a.
    *ENDFUNCTION.
      CALL FUNCTION 'Z_TEST_MULTI'
        EXPORTING
          a = f1
        IMPORTING
          b = f1o.
    ENDFORM.                    " CALL_FM

Maybe you are looking for