Troubleshooting the lockwaits for parallel processing jobs

Hi Experts,
I am facing difficulty tracing the job which is interfering with a business critical job.
The job in discussion is using parallel processing and the other jobs running at the time are also using the same.
If I see a lockwait for some process which may be a dailog process or update process spawned by these jobs, I am having difficulty knowing which one holding the lock and which one is waiting.
So, Is there any way we could identify the dailog or update processes which are used for parallel processing for a particular backgorund job.
Please help me as this is business critical and we have a high visibility in this area.
Any suggestions will be appreciated.......
Regards
Raj

Hi Raj,
First of all, please indicate if you are using SAP Business One.  If yes, then you need to check those locks under SQL Management Studio.
Thanks,
Gordon

Similar Messages

  • Framework for parallel processing

    Hello,
    Can anyone help me by giving some links or examples as to how to use the framework for parallel processing ?? (FPP)
    My scenario is that I have a report that handles huge volumes of data and process it. But at present my report is not able to handle huge volume and it either dumps or hangs. So I would want to use the parallel processing framework which would then can handle huge volumes and also reduce the time for execution.
    Thanks and Regards,
    Sandhya

    Is there anyone can provide some comments?

  • SAP job not using all dialog processes that are available for parallel processing

    He Experts,
    The customer is running a job which is not using all the dialog processes that are available for parallel processing. It appears to use up the parallel processes (60) for the first 4-5 minutes of the job and then maxes out about 3-5 processes for the remainder of the job.
    How do I analyze the job to find out the issue from a Basis perspective?
    Thanks,
    Zahra

    Hi Daniel,
    Thanks for replying!
    I don't believe its a standard job.
    I was thinking of starting a trace using ST05 before the job. What do you think?
    Thanks,
    Zahra

  • Job fail with Timeout for parallel process (for SID Gener.): 006000

    Hello all,
    Im getting below error and not able to find any issue with Basis side. Please anyone help on this!
    Job started
    Data package has already been activated successfully (will be skipped)
    Process started
    Process started
    Process started
    Process started
    Process started
    Import from cluster of the data package to be activated () failed
    Process 000001 returned with errors
    Process 000002 returned with errors
    Process 000003 returned with errors
    Process 000004 returned with errors
    Background process BCTL_4XU7J1JPLOHYI3Y5RYKD420UL terminated due to missing confirmation
    Process 000006 returned with errors
    Data pkgs 000001; Added records 1-; Changed records 0; Deleted records 0
    Log for activation request ODSR_4XUG2LVXX3DH4L1WT3LUFN125 data package 000001...000001
    Errors occured when carrying out activation
    Analyze errors and activate again, if necessary
    Activation of M records from DataStore object CRACO20A terminated
    Activation is running: Data target CRACO20A, from 1,732,955 to 1,732,955
    Overlapping check with archived data areas for InfoProvider CRACO20A
    Data to be activated successfully checked against archiving objects
    Parallel processes (for Activation); 000005
    Timeout for parallel process (for Activation): 006000
    Package size (for Activation): 100000
    Task handling (for Activation): Backgr Process
    Server group (for Activation): No Server Group Configured
    Parallel processes (for SID Gener.); 000002
    Timeout for parallel process (for SID Gener.): 006000
    Package size (for SID Gener.): 100000
    Task handling (for SID Gener.): Backgr Process
    Server group (for SID Gener.): No Server Group Configured
    Activation started (process is running under user *****)
    Not all data fields were updated in mode "overwrite"
    Data package has already been activated successfully (will be skipped)
    Process started
    Process started
    Process started
    Process started
    Process started
    Import from cluster of the data package to be activated () failed
    Process 000001 returned with errors
    Process 000002 returned with errors
    Process 000003 returned with errors
    Process 000004 returned with errors
    Errors occured when carrying out activation
    Analyze errors and activate again, if necessary
    Activation of M records from DataStore object CRACO20A terminated
    Report RSODSACT1 ended with errors
    Job cancelled after system exception ERROR_MESSAGE

    Thanks for the link TSharma I will try that today.
    UPDATE:
    I ran a non-parallel Data Pump and just let it run overnight. This time it finished after 9 hours.  In this run I set the STATUS=300 parameter in the PARFILE which basically echos STATUS updates to standard out every 300 seconds (5 minutes).
    And as before after 2 hours it finished 99% of the export and just spit out WAITING status for the last 7 hours until it finished.  The remaining TABLES it exported (a few hundred) were all very small or ZERO rows.  There clearly is something going on that is not normal.  I've done this expdp before on clones of this database and it usually takes about 2-2.5 hours to finish.
    The database is about 415 Gigabytes in size.
    I will update what the TRACE finds and I'm also opening a case with MOS.

  • Regenerate Variants for parallel processing in FI-CA

    Hi everybody,
    does anybody know a possibility to regenerate variants for parallel processing objects (GPART for example) in FI-CA automically in a job? I can start the program RFKKDI05 with a variant but I can not tell the program to automatically regenerate the existing variants.
    Any help appreciated.
    Cheers,
    Markus

    Hello Markus,
    You'll likely need to create a custom program to accomplish this task.  The program may call FM FKK_DI_VARIANT_REFRESH.  This will allow you to specify the object and variant which needs to be regenerated.
    Best regards,
    Matt

  • Regenerate Variants for parallel processing in PSCD

    Hi everybody,
    does anybody know a possibility to regenerate variants for parallel processing objects (GPART for example) in PSCD automically in a job? I can start the program RFKKDI05 with a variant but I can not tell the program to automatically regenerate the existing variants.
    Any help appreciated.
    Cheers,
    Markus

    Hi, This can be done at the time of Payment run, select the technical setting tab and then maintain variants. See to it that the intervals are all equal. Alternatively this can also be done through IMG.
    IMG Menu Path: Financial Accounting > Contract Accounts Receivables and Payable> Basic Functions> Postings and Documents > Basic Settings > Maintain Documents Number Ranges.
    Create a minimum of 10 num range id in this..
    I hope this helps.
    Rgds
    SJ

  • Unable to clone File Adapter receiver channel for parallel processing

    Hi Experts,
    I am using variable substitution for File - RFC - File with out BPM scenario(using request response, oneway bean).
    While i placed the file in the sender FTP folder, the file didnt get picked up, but in communication channel monitoring, i am getting error ' Unable to clone File Adapter receiver channel for parallel processing'.
    Can anybody provide me suggestions to solve this error.
    Note : without variable substitution , the interface is working good.
    Is it due to, i am trying the source structure field in response file adapter?

    Hi,
    In your CC, do you use some additional paramaters ?
    like these one of point 47/48 of [Oss note 821267 - FAQ: XI 3.0 / PI 7.0/ PI 7.1 File Adapter|https://service.sap.com/sap/support/notes/821267]
    Maybe there is conflict with a parallel connexion and the bean used to do asynch-synch bridge...
    Mickael

  • Java Proxy Generation not working - Support for Parallel Processing

    Hi Everyone,
    As per SAP Note 1230721 - Java Proxy Generation - Support for Parallel Processing, when we generate a java proxy from an interface we are supposed to get 2 archives (one for serial processing and another suffixed with "PARALLEL" for parallel processing of jaav proxies in JPR).
    https://websmp230.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=1230721
    We are on the correct patch level as per the Note, however when we generate java proxy from the IR for an outbound interface, it genrates only 1 zip archive (whose name we ourselves provide in the craete new archive section). This does not enable the parallel processsing of the messages in JPR.
    Could you please help me in this issue, and guide as to how archives can be generated for parallel processing.
    Thanks & Regards,
    Rosie Sasidharan.

    Hi,
    Thanks a lot for your reply, Prateek.
    I have already checked SAP Note 1142580 - "Java Proxy is not processing messages in parallel" where they ask to modify the ejb-jar.xml. However, on performing the change in ejb-jar.xml and while building the EAR, I get the following error:
    Error! The state of the source cache is INCONSISTENT for at least one of the request DCs. The build might produce incorrect results.
    Then, on going through the SAP Note 1142580 again, I realised that the SAP Note 1230721 also should be looked onto which will be needed for generating the Java proxy from Message Interfaces in IR for parallel processing.
    Kindly help me if any of you have worked on such a scenario.
    Thanks in advance,
    Regards,
    Rosie Sasidharan.

  • No destination is currently free for parallel processing - SAPRCK10

    Hello ABAPers,
    We have scheduled SAP standard program "SAPRCK10" in background processing at our production server on daily basis.
    It was worked fine on last week before but its now giving a error says "No destination is currently free for parallel processing ".
    Could you please guide us how to resolve this issue ?
    Thanks,
    vthilagaraj

    Dear François Henrotte,
    Could you please let me know how to increase the number of processe in RZ04 ?
    Thanks,
    vthilagaraj

  • Message 'Please check the destinations for parallel MRP'.

    Hi,
    While I was running MD01, I encountered the following message 'Please check the destinations for parallel MRP'.
    Please advise.
    Thank you in advance.

    Hi,
    In MD01 initial screen indiacate parallel processing indicator.
    In SPRO> PRODUCTION > MRP >MASTER DATA > Define parallel processing in MRP. In that give your server name.
    Regards,
    Dharma

  • Urgent:How to get the Sqltext for Oracle Sumbitted Jobs in 10g

    Dear All,
    Could you help me out in sorting the below problem.
    I use to get the current running sql's with following below query in 9i.
    SELECT A.SID,B.HASH_VALUE, OSUSER, USERNAME, SQL_TEXT
    FROM V$SESSION A, V$SQLTEXT B
    WHERE B.HASH_VALUE = A.SQL_HASH_VALUE
    AND USERNAME LIKE upper('%SCHEMA%')
    ORDER BY B.HASH_VALUE, B.PIECE;
    This is will work in 10g also for user-triggered sqls,stored procedures etc.
    But when oracle submits job i'm not able find which qurery is running.
    Seems For oracle jobs in 10g for V$session contain column SQL_HASH_VALUE as Zero and hence i'm not able find the any sql's running.
    Could you please any of you help me out how to get the Sqltext for Orcle submited Jobs in 10g.
    Please revert asap as this is very urgent for me.
    Thanks in Advance
    Anil.

    Have you tried to query WF_ITEMS? -- http://etrm.oracle.com/pls/et1211d9/etrm_pnav.show_object?c_name=WF_ITEMS&c_owner=APPLSYS&c_type=TABLE
    bde_wf_item.sql - Runtime Data of a Single Workflow Item [ID 187071.1]
    Thanks,
    Hussein

  • The Prerequisites for releasing process order

    Hi all,
    What are the Prerequisites for releasing process order?
    Can we issue the material against process order, if it is on CRTD status?

    Hi,
    Prerequiste for REL of Process Order -
    Process order should have CRTD status.
    Other with IMG setting, you can have order status directly REL once it is created also. See Prod Sch Profile setting.
    You cannot do goods posting (GI and GR) unless order is REL status. Not possible with CRTD status.
    Srini

  • What could be the reason for Crawl process to take long time or get in to a hung state.

    Hi All,
    What could be the reason for Crawl process to take long time or get in to a hung state? Is it something also related to the DB Server resources crunch? Does this lead to Index file corruption?
    What should be the process to be followed when index file is corrupted? How do we come to know about that?
    Thanks in Advance.

    "The crawl time depends on what you are crawling -- the number of items, the size of the items, the location of the items. If you have a lot of content that needs to be crawled, it will take much time".
    http://social.msdn.microsoft.com/Forums/sharepoint/en-US/f4cad578-f3bc-4822-b660-47ad27ce094a/sharepoint-2007-crawl-taking-long-time-to-complete?forum=sharepointgeneralprevious
    "The only clean and recommended way to recover from an index corruption is to completely rebuild the index on all the servers in the farm."
    http://blogs.technet.com/b/victorbutuza/archive/2008/11/11/event-id-4138-an-index-corruption-was-detected-in-component-shadowmerge-in-catalog-portal-content.aspx
    Whenever search index file got corrupted it will got the details to Event logs
    http://technet.microsoft.com/en-us/library/ff468695%28v=office.14%29.aspx
    My Blog- http://www.sharepoint-journey.com|
    If a post answers your question, please click Mark As Answer on that post and Vote as Helpful

  • Issue in completing the block step for parallel processing

    Hi,
    i have created a workflow where in i have used a block step to send workitems to multiple agents.  I have used parallel processing in block step. Number of agents are determined in the runtime. Lets say i have two items in my multiple line container( two agent id). Now inside block statement i have put user decesion step. So at the same time workitems goes to the two approver for approval. When both approver take the decesion after that also command is not coming out of the block step. I want the command to be out of block step after this and goes to the next step of workflow.
    Please suggest any helpful solution for it.
    Regards,
    Smit Shah

    I think theremust be a binding problem , the binding must be some thing like the below
    &USERID[&_WF_PARFOREACH_INDEX&]& ----->&_USERID_LINE&
    of the block step from WF container to Block Conatiner. Because when I checked in my system it is behaving as you want., I also include one Decision step inside the block, and then hard coded the userid values int the table USERID and made the above binding and it ia working fine and in the Decision Agent I mentiond the EXPRESSION and assigned the value &_USERID_LINE&

  • Completing the block step for parallel processing

    Hello ,
    I"m  using a  ParForEach  block   .
    It works fine and  2   approves get their decision task  respectively ,
    I  want to end all parrallel  processes if one of the   approvers select the reject  option  .
    I  thought it could be done using the  End condition Option .
    I have defined an end condition    in  the tab  Parallel  Procesing  ,  the condition  is  an  element wf_result is set to REJECT
    was defined.
      But   nothing resulted and althogh i check  the container and the  element is correctly   assigned after one of the approvers
      rejects  the  other is still  sitting there.
    i  ran  the condiion evalate button  on  this step  with no   success .
    Please Advise

    Hi Rina
    In EHP4, there is the following workflows for Compensation:
    04000022 ECM_SUBMIT
    04000024 ECM_REJECT
    04000025 ECM_CREATE
    together with class CL_HRECM00_WORKFLOW_EVENT.
    There, the parallel processing is used.
    Best Regards
    Raquel

Maybe you are looking for