Issue in FM-IDOC_INPUT_DEBITOR_MDM-Taking more processing time

Hi Friends,
        The Customer master idocs are taking lot of time to process in production. Currently one idoc needs 8 minutes to process.  This is because of the select in the DEBMAS where in the program searches for the ADRMAS idoc for the same customer.I am using FM "IDOC_INPUT_DEBITOR_MDM", Is there any way to reduce the processing Time.
Thanks in Advance ,
Regards,
Veera

Thanks Gareth for the reply. This is standard FM and after the Call transaction, SAP code reads the  BDCMSGCOLL. If it does not found ans S in that, it fails the IDOC to 51 status.  There is no Z code in this.
I would like to know what config changes can supress the success message . I tried doing BDC thoguth SM35 and there also found that no message is coming when the Customer is created. This would mean that there is something worng in srrting rather than FM and BDC which is causing the BDC failure in Std FM.
Regds
Harish

Similar Messages

  • Oracle internal queries taking more CPU time

    Hi,
    Following are queries taking more CPU time. I got this from awr report. Can anyone tell me why these queries are running? Is it a oracle enterprise manager query? should I use
    emctl stop dbconsole to stop this?
    DECLARE job BINARY_INTEGER := :job; next_date DATE := :mydate; broken BOOLEAN := FALSE; BEGIN EMD_MAINTENANCE.EXECUTE_EM_DBMS_JOB_PROCS(); :mydate := next_date; IF broken THEN :b := 1; ELSE :b := 0; END IF; END;
    SELECT COUNT(*) FROM MGMT_METRIC_DEPENDENCY_DETAILS DEP, MGMT_SEVERITY SEV WHERE DEP.TARGET_GUID = :B5 AND DEP.METRIC_GUID = :B4 AND DEP.KEY_VALUE = :B3 AND DEP.EDEP_TARGET_GUID = SEV.TARGET_GUID AND DEP.EDEP_METRIC_GUID = SEV.METRIC_GUID AND DEP.DEP_KEY_VALUE = SEV.KEY_VALUE AND SEV.COLLECTION_TIMESTAMP BETWEEN :B2 AND :B1
    SELECT CURRENT_STATUS FROM MGMT_CURRENT_AVAILABILITY WHERE TARGET_GUID = :B1
    Thanks in advance
    With Regards
    boobathi.P

    Hi,
    maybe this document will help if you are using 10g:
    SQL run by SYSMAN consuming a lot of resources on OMS with 800+ targets [ID 330383.1]
    https://support.oracle.com/CSP/main/article?cmd=show&type=NOT&doctype=PROBLEM&id=330383.1
    there you'll find Cause and Solution too:
    SYSMAN Job and Queries are Taking Up High CPU (DB Console) [ID 1288301.1]
    https://support.oracle.com/CSP/main/article?cmd=show&type=NOT&doctype=PROBLEM&id=1288301.1
    For databases above version 11.1 there are paches available.
    Best,
    Michael T. Z.

  • Dbms_stats.cleanup_stats_db_proc taking more CPU time.

    Hi All,
    Oracle 11g R2(11.2.0.3) EE , on Linux machine.
    Oracle auto optimizer stats collection enabled in our environment. In that dbms_stats.cleanup_stats_db_proc() taking more CPU time (more than 24 hours).
    Could you please suggest any solution.
    Thanks in advance.
    BR,
    Surya

    suresh.ratnaji wrote:
    NAME                                 TYPE        VALUE
    _optimizer_cost_based_transformation string      OFF
    filesystemio_options                 string      asynch
    object_cache_optimal_size            integer     102400
    optimizer_dynamic_sampling           integer     2
    optimizer_features_enable            string      10.2.0.4
    optimizer_index_caching              integer     0
    optimizer_index_cost_adj             integer     100
    optimizer_mode                       string      choose
    optimizer_secure_view_merging        boolean     TRUE
    plsql_optimize_level                 integer     2
    please let me know why it taking more time in INDEX RANGE SCAN compare to the full table scan?Suresh,
    Any particular reason why you have a non-default value for a hidden parameter, optimizercost_based_transformation ?
    On my 10.2.0.1 database, its default value is "linear". What happens when you reset the value of the hidden parameter to default?

  • Issue with background job--taking more time

    Hi,
    We have a custom program which runs as the background job. It runs every 2 hours.
    Itu2019s taking more time than expected on ECC6 SR2 & SR3 on Oracle 10.2.0.4. We found that it taking more time while executing native SQL on DBA_EXTENTS. When we tried to fetch less number of records from DBA_EXTENTS, it works fine.
    But we need the program to fetch all the records.
    But it works fine on ECC5 on 10.2.0.2 & 10.2.0.4.
    Here is the SQL statement:
    EXEC SQL PERFORMING SAP_GET_EXT_PERF.
      SELECT OWNER, SEGMENT_NAME, PARTITION_NAME,
             SEGMENT_TYPE, TABLESPACE_NAME,
             EXTENT_ID, FILE_ID, BLOCK_ID, BYTES
       FROM SYS.DBA_EXTENTS
       WHERE OWNER LIKE 'SAP%'
       INTO
       : EXTENTS_TBL-OWNER, :EXTENTS_TBL-SEGMENT_NAME,
       :EXTENTS_TBL-PARTITION_NAME,
       :EXTENTS_TBL-SEGMENT_TYPE , :EXTENTS_TBL-TABLESPACE_NAME,
       :EXTENTS_TBL-EXTENT_ID, :EXTENTS_TBL-FILE_ID,
       :EXTENTS_TBL-BLOCK_ID, :EXTENTS_TBL-BYTES
    ENDEXEC.
    Can somebody suggest what has to be done?
    Has something changed in SAP7 (wrt background job etc) or do we need to fine tune the SQL statement?
    Regards,
    Vivdha

    Hi,
    there was an issue with LMT's but that was fixed  in 10.2.0.4
    besides missing system statistics:
    But WHY do you collect every 2 hours this information? The dba_extents view is based on really heavy used system tables.
    Normally , you would start queries of this type against dba_extents ie. to identify corrupt blocks and such:
    SELECT  owner , segment_name , segment_type
            FROM  dba_extents
           WHERE  file_id = &AFN
             AND  &BLOCKNO BETWEEN block_id AND block_id + blocks -1
    Not sure what you want to achieve with it.
    There are monitoring tools (OEM ?) around that may cover your needs.
    Bye
    yk

  • Consistent gets are reduced by 50% but the query taking more elapsed time.

    Hi All,
    While tuning a application my consistent gets are reduced by 50% but the query is still taking the same time.
    In a Warehouse env data is coming from With clause that is having some 40000 rows .Then these 40K rows are joined to a table that is having 2 Billion records having indexes on primary key and date column(bitmap)indexes .
    It is using Hash Joining method.

    Try forcing a hash join, if possible:
    http://dba-oracle.com/tips_oracle_hash_joins.htm
    Can you post the plans?
    http://www.dba-oracle.com/plsql/t_plsql_plans.htm

  • Query taking more response time

    Hi Masters,
    Today i am tuning One Query
    The Following points are the Info regarding my query.
    Table1- has 100 million records.
    Table2 - has 7000 records.
    I am Using table2 as Driving table.
    I used the Ordered, index hint, use_nl hint and all
    My Query:
    select
    max(t1.date)
    from table2 t2*, table1 t1*
    where t2.col1=12354
    and
    t1.col2=t2.col2 -- Is the  Join b/w the two tables.
    and t1.col3=0
    Regarding Index Details:
    There is a composite index on table1 (col2,col3).
    Also there is an index on table1(col1).
    The Query using the Right Index Only.
    I found out from the explain Plan.
    Then what is the Problem with the above Query.
    Do i need to add any more Hints to work better?
    One more Info man,
    When i tried to retrieve only the t1.date column not as max(t.date), it will respond in subsequent mill seconds.
    select
    t1.date
    from table2 t2, table1 t1
    where t2.col1=12354
    and
    t1.col2=t2.col2 -- Is the Join b/w the two tables.
    and t1.col3=0
    And the result contains, 257000 records.
    For finding the max value only it will take several hundred times more!!
    Try to give some useful feedback!!
    Edited by: user10376641 on Aug 21, 2009 1:32 AM

    If i am harsh....sorry for that.
    even after Sybrand's post, you asked for "anyother feedback". so i thought you had not gone through the link completely.
    but good to hear that you are analyzing the things. please post your findings, so that we can help you out.
    once again, sorry for my words.

  • Process SERVER0 its taking High CPU Time in XI

    Hi,
    We installed XI 3.0 develpoment server ,Process <b>SERVER0</b> its taking more CPU time.
    We are using AS/400 OS & DB2 datbase.Can any one tell me the reason & Solution for this.
    Thanks & Regards,
    Gopinath.

    hi,
    Actually the user XIRWBUSER its an RFC user but its running on many dialog process.I think high CPU time due to this user only.
    using 5 work processes and 3/4 of the available CPU for an extended amount of time.
    Total Total DB
    Job or CPU Sync Async CPU
    Task User Number Thread Pty Util I/O I/O Util
    WP11 D6464 485368 00000010 20 54.0 25 37 27.8
    WP05 X4242 498642 00000098 20 49.6 5 0 36.7
    WP04 X4242 498641 00000014 20 48.8 1 0 39.1
    WP02 X4242 498639 0000025A 20 47.1 2 0 37.8
    WP06 X4242 498643 000001E6 20 43.7 0 0 38.3
    WP00 X4242 498637 00000014 20 11.1 502 194 2.9
    pls can any one help me
    Regards,
    Gopinath.

  • How to reduce processing time  in case of jms to idoc scenario

    hi all,
    i am doing jms to idoc scenario.
    in this i'm using xslt and graphical mapping.
    while perfroming E2E testing the messages are taking lot of time to get processed.
    please tell me as how can i reduce this processing time or in short how can i overcome this problem.this is a critical issue in production.
    regards
    madhu

    Hi,
    1. To reduce the processing time for JMS -> IDOC scenario, try to get the Average processing time taken by JMS adapter, IE, IDOC adapter for the messages that are processed so far by going thro the Adapter Audit logs, SXMB_MONI race and Time in WE02 in R/3 system. Calculate this for 5 - 10 messages at different time to get the clear idea of which step is consuming more time. Then it will be helpful to focus on tuning that specific step.  If you are using BPM, check the processing time taken by PE.
    2. Check the Inbound and Outbound queues used by this interface, if required try to configure Prioritized Queue with more time allocated for this interface in Integration Engine.
    3. Check the load on the Integration Server when this Interface messages are getting processed, to understand/ isolate the interfaces which are consuming more processing time and system resources.
    4. If this is not in Production try to perform the Mapping/ transformation in graphical mapping itself, it will be helpful to minimise use of XSLT mapping to get better performance.
    5. Check if the Archive and Delete jobs are running successfully in the Integration Server, since this will free up the occupied resources in the system.
    6. If possible post seperate Forum questions for issues in different Interface scenarios, it will be easy to understand the problem for all of us.
    Hope this will be helpful,
    Regards,
    S.Santhosh Kumar

  • More response time - STMS_QA

    HI,
    The transaction code STMS_QA is taking more time to execute ( nealry 2 mins). while analysing in the STAD it is showing it due to more processing time. Can some one help how to solve this issue.
    Thansks,
    Prashanth

    Hi Deepak,
    Thanks for you reply. Below are the answers for your questions.
    1) Yes trans directory is mounted and currently in Development system.
    2)  All the TR's.
    3)  We have enough space in trans directory.
    Could you please suggest the better solution to solve this issue ASAP?
    Thanks,
    Prashant Raja.

  • How to include processing time in the  LogFormat in httpd.conf of OAS

    Hi,
    I want to add processing time parameter in the LogFormat of the HTTP requests in Application server logs file.
    i tried to edit the httpd.conf file which orginally is as below
    LogFormat "%h %l %u %t \"%r\" \"%{Cache-Control}i\" CC \"%{Cache-Control}o\" %>s %b" common
    to
    LogFormat "%h %l %u %t *%T* \"%r\" \"%{Cache-Control}i\" CC \"%{Cache-Control}o\" %>s %b" common
    But doing so, did not add the Processing time in the GET and POST requests.Although same works for Apache HTTP server.
    I want to know how do i add the Processing time attribute in the LOGFormat.
    By Default the LogFormat is Common, what are the other possibilities fpr LogFormat.
    Please revert.
    Regards
    Ak

    Hi,
    1. To reduce the processing time for JMS -> IDOC scenario, try to get the Average processing time taken by JMS adapter, IE, IDOC adapter for the messages that are processed so far by going thro the Adapter Audit logs, SXMB_MONI race and Time in WE02 in R/3 system. Calculate this for 5 - 10 messages at different time to get the clear idea of which step is consuming more time. Then it will be helpful to focus on tuning that specific step.  If you are using BPM, check the processing time taken by PE.
    2. Check the Inbound and Outbound queues used by this interface, if required try to configure Prioritized Queue with more time allocated for this interface in Integration Engine.
    3. Check the load on the Integration Server when this Interface messages are getting processed, to understand/ isolate the interfaces which are consuming more processing time and system resources.
    4. If this is not in Production try to perform the Mapping/ transformation in graphical mapping itself, it will be helpful to minimise use of XSLT mapping to get better performance.
    5. Check if the Archive and Delete jobs are running successfully in the Integration Server, since this will free up the occupied resources in the system.
    6. If possible post seperate Forum questions for issues in different Interface scenarios, it will be easy to understand the problem for all of us.
    Hope this will be helpful,
    Regards,
    S.Santhosh Kumar

  • Huge Processing Time in XI

    Hi All,
    We have an interface running in production server which is taking huge processing time.
    It is a File to IDoc scenario wherein the file size is about 1MB and it gets converted to something between 4-5MB after Content Conversion. Also, this issue does not happen for all the messages. A  few of the messages, say 2 out of 10 messages take this huge time to process (from 2 hours u2013 4 hours sometimes) and the rest of them get processed in few minutes (max. 10mins).
    I am just confused with this random behavior of this interface. Has anyone come across anything related to this?
    Please help with your inputs.
    Many Thanks,
    Regards,
    Lakshmi.

    Hi,
         Please check if for the instances where the interface is taking a huge time to process, any other interface is running simultaneously and taking up the system resources, due to which this particular message is getting delayed.
    Also, for file with content conversion, you can consider splitting up the file in the sender communication channel (parameter RecordSets per message) to reduce the message size.
    Also, in transaction code SXMB_ADM, Integration Engine Configuration, you can maintain parameter
    EO_MSG_SIZE_LIMIT of category TUNING to process large messages in series. This will ensure that large messages are processed in a seperate queue and do not affect the execution of other interfaces.
    You can refer to the below link for sizing and performance tunning
    http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/2016a0b1-1780-2b10-97bd-be3ac62214c7
    Regards

  • CDP Performance Issue-- Taking more time fetch data

    Hi,
    I'm working on Stellent 7.5.1.
    For one of the portlet in portal its taking more time to fetch data. Please can some one help me to solve this issue.. So that performance can be improved.. Please kindly provide me solution.. This is my code for fetching data from the server....
    public void getManager(final HashMap binderMap)
    throws VistaInvalidInputException, VistaDataNotFoundException,
    DataException, ServiceException, VistaTemplateException
         String collectionID =
    getStringLocal(VistaFolderConstants.FOLDER_ID_KEY);
         long firstStartTime = System.currentTimeMillis();
    HashMap resultSetMap = null;
    String isNonRecursive = getStringLocal(VistaFolderConstants
    .ISNONRECURSIVE_KEY);
    if (isNonRecursive != null
    && isNonRecursive.equalsIgnoreCase(
    VistaContentFetchHelperConstants.STRING_TRUE))
    VistaLibraryContentFetchManager libraryContentFetchManager =
    new VistaLibraryContentFetchManager(
    binderMap);
    SystemUtils.trace(
    VistaContentFetchHelperConstants.CONTENT_FETCH_TRACE,
    "The input Parameters for Content Fetch = "
    + binderMap);
              resultSetMap = libraryContentFetchManager
    .getFolderContentItems(m_workspace);
    * used to add the resultset to the binder.
    addResultSetToBinder(resultSetMap , true);
    else
         long startTime = System.currentTimeMillis();
    * isStandard is used to decide the call is for Standard or
    * Extended.
    SystemUtils.trace(
    VistaContentFetchHelperConstants.CONTENT_FETCH_TRACE,
    "The input Parameters for Content Fetch = "
    + binderMap);
    String isStandard = getTemplateInformation(binderMap);
    long endTimeTemplate = System.currentTimeMillis();
    binderMap.put(VistaFolderConstants.IS_STANDARD,
    isStandard);
    long endTimebinderMap = System.currentTimeMillis();
    VistaContentFetchManager contentFetchManager
    = new VistaContentFetchManager(binderMap);
    long endTimeFetchManager = System.currentTimeMillis();
    resultSetMap = contentFetchManager
    .getAllFolderContentItems(m_workspace);
    long endTimeresultSetMap = System.currentTimeMillis();
    * used to add the resultset and the total no of content items
    * to the binder.
    addResultSetToBinder(resultSetMap , false);
    long endTime = System.currentTimeMillis();
    if (perfLogEnable.equalsIgnoreCase("true"))
         Log.info("Time taken to execute " +
                   "getTemplateInformation=" +
                   (endTimeTemplate - startTime)+
                   "ms binderMap=" +
                   (endTimebinderMap - startTime)+
                   "ms contentFetchManager=" +
                   (endTimeFetchManager - startTime)+
                   "ms resultSetMap=" +
                   (endTimeresultSetMap - startTime) +
                   "ms getManager:getAllFolderContentItems = " +
                   (endTime - startTime) +
                   "ms overallTime=" +
                   (endTime - firstStartTime) +
                   "ms folderID =" +
                   collectionID);
    Edited by: 838623 on Feb 22, 2011 1:43 AM

    Hi.
    The Select statment accessing MSEG Table is Slow Many a times.
    To Improve the performance of  MSEG.
    1.Check for the proper notes in the Service Market Place if you are working for CIN version.
    2.Index the MSEG table
    2.Check and limit the Columns in the Select statment .
    Possible Way.
    SELECT MBLNR MJAHR ZEILE BWART MATNR WERKS LIFNR MENGE MEINS
    EBELN EBELP LGORT SMBLN BUKRS GSBER INSMK XAUTO
    FROM MSEG
    INTO CORRESPONDING FIELDS OF TABLE ITAB
    WHERE WERKS EQ P_WERKS AND
    MBLNR IN S_MBLNR AND
    BWART EQ '105' .
    Delete itab where itab EQ '5002361303'
    Delete itab where itab EQ  '5003501080' 
    Delete itab where itab EQ  '5002996300'
    Delete itab where itab EQ '5002996407'
    Delete itab where itab EQ '5003587026'
    Delete itab where itab EQ  '5003587026'
    Delete itab where itab EQ  '5003493186'
    Delete itab where itab EQ  '5002720583'
    Delete itab where itab EQ '5002928122'
    Delete itab where itab EQ '5002628263'.
    Select
    Regards
    Bala.M
    Edited by: Bala Malvatu on Feb 7, 2008 9:18 PM

  • Post Goods Issue (VL06O) - taking more time approximate 30 to 45 minutes

    Dear Sir,
    While doing post goods issue against delivery document system is taking lots of time, this issue is very very urgent can any one resolved or provide suitable solution for solving this issue.
    We creates every day approximate 160 sales order / delivery and goods issue against the same by using transaction code VL06O system is taking more time for PGI.
    Kindly provide suitable solution for the same.
    Regards,
    Vijay Sanguri

    Hi
    See Note 113048 - Collective note on delivery monitor and search notes related with performance.
    Do a trace with tcode ST05 (look for help from a basis consultant) and search the bottleneck. Search possible sources of performance problems in userexits, enhancements and so on.
    I hope this helps you
    Regards
    Eduardo

  • Re : BSP App. taking More Time for Processing

    Hi,
       We are using WAS 620 , created a bsp application to create trips thru bdc's.
       In our development system its going fine but in our quality system its taking more that 10 mins for posting.
    When we check thru SM04 the icm session and icm req session are there for long time.
       Is there any setting to do with icm for executing bdc's/posting programs.
    Looking  for help from u all to solve this issue....
    Thanks and Regards,
    K Vijayasekar.

    Check out this weblog to get you started:
    /people/mark.finnern/blog/2003/09/24/bsp-in-depth-confusion-between-stateless-stateful-and-authentication

  • HT201274 My iphone4 is taking more than 5 hours to erase all the data still it is under process , still how much more time i have to wait for my mobile to ON

    My iphone4 is taking more than 5 hours to erase all the data still it is under process , still how much more time i have to wait for my mobile to ON ?

    I'm having this EXACT same problem with my iPhone 4, and I have the same computer stats (I have a Samsung Series 7)

Maybe you are looking for

  • Impact of Running COGI program for every two HRS...

    Hi, As per requirement in my business scenario, we have lot of materials with negative stocks currently. Now for physical inventory purpose, customer wants us to remove these negative stocks and want us to run COGI program every 2 hrs. As we are goin

  • Setting up an audio interface with logic 9

    hey im still pretty new to this. i have a powerbook G4 with logic pro 9. i just got a steinberg CI1 usb audio interface which i was told would work with logic. (http://www.steinberg.net/en/products/hardware/ci1/ci1_start.html) im recording vocals and

  • Deleting ALV Variant

    Hi, how can I let user to delete dislpay variant of alv grid. Thank you. Marian

  • Financial reporting objects

    Hi , When I change the database connection of a report , the text objects , in the report , show some random digits. The database connections for the report and for the templates get changed but for the objects it didn't work . If I re-insert the tex

  • I keep getting adobe flash errors and safari stops working, any suggestions?

    I am nothing that when any website has adobe flash in it, I get the spinning wheel and have to shut down Safari.  I have seen three Adobe Error messages in the last day and a half.  Any suggestions?