Load Distribution Versus Parallel Jobs processing

Hello,
I would request feedback.
In Industry Solutions - PS, for mass activity - Automatic Clearing run, there is an Technical Settings tab.
This tab has Load distribution settings and an value that can be defined for Number of Jobs.
If I put an value greater than 1, so many number of jobs gets executed in the background.
I would request clarification on the following :
1. What is its use / benefits of putting defining an automatic load distribution ?
2. Does this setting equals the Config settings for IMG - Finnancial Accounting- Contract AR - Technical Settings - Prepare Mass.
In here, we can set parallel jobs that can be initiated by providing an maximum number of jobs under the job control panel.
Does this setting override Load distribution settings or vice-versa  OR IS this an different setting ?
3. What is the difference between Load distribution versus parallel Jobs ?
Any feedback, most welcome and appreciated.
I am functional consultant and hence unable to distinguish between the two.
I am not sure if this query needs to be posted here, but I have also posted the same in IS fourm too.
Regards
Bala
email : [email protected]

Hi Bala,
In any Mass Activity you will find Technical Settings tab. There you will find Parallel Processing Object and Load Distribution. In Parallel Processing Object, you can select object according to Input depending upon which you will able to divide the job. For example, GPART is used when job have to divide depending upon Business partner. In Maintain Variants you have to create Variants.You have to give Variant name and value in either Interval length or Number of Intervals. Interval length will decide what is the maximum number of object will be processed in a single part. Number of Interval is the number which tells job will be divided in how many parallel process.
For example, let you r running a parallel process for 1000 Business Partner.You choose GPART as object.Now if you put 200 in interval length it will divide job in 5 parallel process.If you put 10 in number of interval it will divide the job in 10 parallel process each of 100 Business partner.
After creating variant, you use it in technical setting.
Let u already define variants which divide the job in 30 but you want maximum number of parallel process will be executed at a time is 6.Then put 6 in Number of jobs field in automatic load distribution.
Hope this can able to clear your question. Still if have any doubt feel free to ask me.
Thanks and regards,
Jyotishankar Dutta

Similar Messages

  • Query to Report on Parallel Jobs Running

    Morning!
    I would like to get a query that reports on my parallel jobs.
    For each minute that a procedure is running I would like to know what stages are running.
    I log the whole procedure in a table called run_details and the start and end of each stage in a table called incident.
    I'm running Oracle 9i
    Here is some sample data based on 2 threads Expected output at the bottom
    SQL>CREATE TABLE run_details
      2  (run_details_key  NUMBER(10)
      3  ,start_time       DATE
      4  ,end_time         DATE
      5  ,description      VARCHAR2(50)
      6  );
    SQL>CREATE TABLE incident
      2  (run_details_key NUMBER(10)
      3  ,stage           VARCHAR2(20)
      4  ,severity        VARCHAR2(20)
      5  ,time_stamp      DATE
      6  );
    SQL>INSERT INTO run_details
      2  VALUES (1
      3         ,TO_DATE('08/10/2007 08:00','DD/MM/YYYY HH24:MI')
      4         ,TO_DATE('08/10/2007 08:10','DD/MM/YYYY HH24:MI')
      5         ,'Test'
      6         );
    SQL>INSERT INTO incident
      2  VALUES (1
      3         ,'Stage1'
      4         ,'START'
      5         ,TO_DATE('08/10/2007 08:00','DD/MM/YYYY HH24:MI')
      6         );
    SQL>INSERT INTO incident
      2  VALUES (1
      3         ,'Stage1'
      4         ,'END'
      5         ,TO_DATE('08/10/2007 08:08:53','DD/MM/YYYY HH24:MI:SS')
      6         );
    SQL>INSERT INTO incident
      2  VALUES (1
      3         ,'Stage2'
      4         ,'START'
      5         ,TO_DATE('08/10/2007 08:00','DD/MM/YYYY HH24:MI')
      6         );
    SQL>INSERT INTO incident
      2  VALUES (1
      3         ,'Stage2'
      4         ,'END'
      5         ,TO_DATE('08/10/2007 08:04:23','DD/MM/YYYY HH24:MI:SS')
      6         );
    SQL>INSERT INTO incident
      2  VALUES (1
      3         ,'Stage3'
      4         ,'START'
      5         ,TO_DATE('08/10/2007 08:04:24','DD/MM/YYYY HH24:MI:SS')
      6         );
    SQL>INSERT INTO incident
      2  VALUES (1
      3         ,'Stage3'
      4         ,'END'
      5         ,TO_DATE('08/10/2007 08:10','DD/MM/YYYY HH24:MI')
      6         );
    SQL>select * from incident;
    RUN_DETAILS_KEY STAGE      SEVERITY   TIME_STAMP
                  1 Stage1     START      08/10/2007 08:00:00
                  1 Stage1     END        08/10/2007 08:08:53
                  1 Stage2     START      08/10/2007 08:00:00
                  1 Stage2     END        08/10/2007 08:04:23
                  1 Stage3     START      08/10/2007 08:04:24
                  1 Stage3     END        08/10/2007 08:10:00  So stages 1 and 2 run in parallel from 08:00, then at 08:04:23 stage 2 stops and a second later stage 3 starts.
    set some variables
    SQL>define start_time = null
    SQL>col start_time new_value start_time
    SQL>define end_time = null
    SQL>col end_time new_value end_time
    SQL>
    SQL>SELECT start_time-(1/(24*60)) start_time
      2        ,end_time
      3  FROM   run_details
      4  WHERE  run_details_key =  1;
    START_TIME          END_TIME
    08/10/2007 07:59:00 08/10/2007 08:10:00Get every minute that the process is running for:
    SQL>WITH t AS (SELECT TRUNC(TO_DATE('&start_time','dd/mm/yyyy hh24:mi:ss'),'MI') + rownum/24/60 tm
      2             FROM   dual
      3             CONNECT BY ROWNUM <= (TO_DATE('&end_time','dd/mm/yyyy hh24:mi:ss')
      4                                   -TO_DATE('&start_time','dd/mm/yyyy hh24:mi:ss')
      5                                  )*24*60
      6            )
      7  SELECT tm
      8  FROM t;
    old   1: WITH t AS (SELECT TRUNC(TO_DATE('&start_time','dd/mm/yyyy hh24:mi:ss'),'MI') + rownum/24/60 tm
    new   1: WITH t AS (SELECT TRUNC(TO_DATE('08/10/2007 07:59:00','dd/mm/yyyy hh24:mi:ss'),'MI') + rownum/24/60 tm
    old   3:            CONNECT BY ROWNUM <= (TO_DATE('&end_time','dd/mm/yyyy hh24:mi:ss')
    new   3:            CONNECT BY ROWNUM <= (TO_DATE('08/10/2007 08:10:00','dd/mm/yyyy hh24:mi:ss')
    old   4:                                -TO_DATE('&start_time','dd/mm/yyyy hh24:mi:ss')
    new   4:                                -TO_DATE('08/10/2007 07:59:00','dd/mm/yyyy hh24:mi:ss')
    TM
    08/10/2007 08:00:00
    08/10/2007 08:01:00
    08/10/2007 08:02:00
    08/10/2007 08:03:00
    08/10/2007 08:04:00
    08/10/2007 08:05:00
    08/10/2007 08:06:00
    08/10/2007 08:07:00
    08/10/2007 08:08:00
    08/10/2007 08:09:00
    08/10/2007 08:10:00
    11 rows selected.Get stage, start & end times and duration
    SQL>SELECT ai1.stage
      2        ,ai1.time_stamp start_time
      3        ,ai2.time_stamp end_time
      4        ,SUBSTR(numtodsinterval(ai2.time_stamp-ai1.time_stamp, 'DAY'), 12, 8) duration
      5  FROM   dw2.incident ai1
      6  JOIN   dw2.incident ai2
      7         ON ai1.run_details_key = ai2.run_details_key
      8         AND ai1.stage = ai2.stage
      9  WHERE ai1.severity = 'START'
    10  AND ai2.severity = 'END'
    11  AND ai1.run_details_key  = 1
    12  ORDER BY ai1.time_stamp
    13  /
    STAGE      START_TIME          END_TIME            DURATION
    Stage1     08/10/2007 08:00:00 08/10/2007 08:08:53 00:08:52
    Stage2     08/10/2007 08:00:00 08/10/2007 08:04:23 00:04:22
    Stage3     08/10/2007 08:04:24 08/10/2007 08:10:00 00:05:36Then combine both (or do something else) to get this:
    TM                  THREAD_1 THREAD_2
    08/10/2007 08:00:00 Stage1   Stage2
    08/10/2007 08:01:00 Stage1   Stage2
    08/10/2007 08:02:00 Stage1   Stage2
    08/10/2007 08:03:00 Stage1   Stage2
    08/10/2007 08:04:00 Stage1   Stage2
    08/10/2007 08:05:00 Stage1   Stage3
    08/10/2007 08:06:00 Stage1   Stage3
    08/10/2007 08:07:00 Stage1   Stage3
    08/10/2007 08:08:00 Stage1   Stage3
    08/10/2007 08:09:00          Stage3
    08/10/2007 08:10:00          Stage3Ideally I'd like this to work for n-threads, as I want this to run on different environments that have different numbers of CPUs.
    Thank you for your time.

    > Ideally I'd like this to work for n-threads, as I want this to run on
    different environments that have different numbers of CPUs.
    The number of CPUs are not always a good indication of the processing load that a platform can take - especially when the processing load involves a lot of I/O.
    You can have 99% CPU idle time with a 1000 active processes... as that idle time is in fact CPU time spend waiting on I/O completion. Courtesy of a severely strained I/O channel that is the bottleneck.
    Another factor is memory (resources). You for example have 4 CPUs with 8GB physical memory.. where a single process (typically a Java VM for a complex process) grabs a huge amount of memory. Assuming that 4 threads/CPU or 1 threads/CPU can be a severe overestimate due to the amount of memory needed. Getting this wrong leads in turn to excessive virtual memory paging and reduces the platform's performance drastically.
    CPU alone is a very poor choice when deciding on the platform's capacity to run parallel processes.

  • How to increase parallel jobs for TA SGEN

    Hi all,
    we did an SGEN load for an AIX ApplServer and a LinuxAppServer to know, which load has a longer runtime.
    We saw, that the linux server has more than twice than the AIX server ... But we also noticed in joblog, that
    AIX had 37 parallel jobs for sgen und linux only 20 ...
    I don't know, where to configure, that linux also uses 37 parallel jobs for sgen ...
    I found an notice in help.sap, that perhaps TA rz12 could help me, but how !?
    Kind regards
    Andreas

    Hi,
    SGEN uses the configured quota for asynchronous RFC. Do the following:
    RZ12 => select group "parallel_generators" (probably the group already exists, otherwise create it and assign to the instance(s) you wish to participate in this group)
    For each instance that is a member of the parallel_generators group:
    - Double-click on the entry
    - The relevant quota are "Max. Number of WPs Used (%)" and "Minimum Number of Free WPs". The first is treated as a percentage of the number of DIA processes. The second indicates how many DIA processes must be kept free.
    Save the changes you make.
    Example: assume that an instance has 20 DIA processes and you configure "Max. Number of WPs Used (%)" = 50 and "Minimum Number of Free WPs" = 5. This means that SGEN will start 10 parallel jobs (50% of 20 processes), unless so many DIA processes are busy on other activities that the number of free DIA would fall below 5.
    Hope this helps,
    Mark

  • Hierarchies Job Failing  The job process could not communicate with the dat

    Hi Experts,
    We have a group of hierarchies that run as a separate job on the DS schedules. The problem is this when we schedule the job to run during the production loads it fails but when we run immediately after it fails it runs completely fine. So it basically means that if i run it manually it runs but when its scheduled to run with the production job it fails. Now the interesting thing is If i schedule the job to run anytime after or before the production jobs are done. It works fine.
    The error i get is
    The job process could not communicate with the data flow <XXXXXX> process. For details, see previously logged
                                                               error <50406>.
    Now this XXXXX DF has only Horizontal Flatenning and it does not run as separate process because if i have it has separate process it fails with an EOF . So i removed the run as separate process and changes the DF to use in memory .
    Any Suggestion on this problem...

    Thanks Mike.. I was hoping its a memory issue but the thing i don't understand is when the job is scheduled to run with the production job it fails. when i manually run the job during the production job it runs, this kinda baffles me.
    DS 3.2 (Verison 12.2.0.0)
    OS: GNU/LINUX
    DF Cache Setting :- In Memory
    processor       : 0
    vendor_id       : GenuineIntel
    cpu family      : 6
    model           : 26
    model name      : Intel(R) Xeon(R) CPU           X5670  @ 2.93GHz
    stepping        : 4
    cpu MHz         : 2933.437
    cache size      : 12288 KB
    fpu             : yes
    fpu_exception   : yes
    cpuid level     : 11
    wp              : yes
    flags           : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss syscall nx rdtscp lm constant_tsc ida nonstop_tsc arat pni ssse3 cx16 sse4_1 sse4_2 popcnt lahf_lm
    bogomips        : 5866.87
    clflush size    : 64
    cache_alignment : 64
    address sizes   : 40 bits physical, 48 bits virtual
    power management: [8]
    processor       : 1
    vendor_id       : GenuineIntel
    cpu family      : 6
    model           : 26
    model name      : Intel(R) Xeon(R) CPU           X5670  @ 2.93GHz
    stepping        : 4
    cpu MHz         : 2933.437
    cache size      : 12288 KB
    fpu             : yes
    fpu_exception   : yes
    cpuid level     : 11
    wp              : yes
    flags           : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss syscall nx rdtscp lm constant_tsc ida nonstop_tsc arat pni ssse3 cx16 sse4_1 sse4_2 popcnt lahf_lm
    bogomips        : 5866.87
    clflush size    : 64
    cache_alignment : 64
    address sizes   : 40 bits physical, 48 bits virtual
    power management: [8]
    Thanks for your help

  • Sm59 & load distribution

    Hi Friends,
    I have a question about sm59. Now, I am creating r3 connections from sm59. Lets say the connection name as 'TEST'. In this connection I don't select load distribution. And when I am calling the bapis;
    call funciton 'BAPI1'
    destination TEST
    call function 'BAPI_TRANSACTION_COMMIT'
    destinaton 'TEST'
    I see that If I don't call BAPI_TRANSACTION_COMMIT BAPI1's change operations is not implemented. And If I call BAPI_TRANSACTION_COMMIT, BAPI1's change operations are done. So, I think that 'different function calls via same destination are using same session !!' Maybe this assumption is wrong, If wrong please correct.
    On the other hand, I want to use load distribution in R3 connection. My real question is in this point. As known, in load distriution selection, the balancer distributes the client to different servers. I need to call two function(BAPI1 and its commit). And there is a possibility;
    while calling BAPI1, server1 is used
    while calling COMMIT BAPI server2 is used.
    Like a this possibility, my BAPI1's change operations will be done? I mean that different server will use the same session?
    And can you give document information, about load balancing in sm59.
    Thanks.

    Hi Huseyin,
    Its really weared situation...really speaking if you call two BAPI's separately it cannot be a part of same process...
    ...Anyway, I think correct solution would be to create another z wrapper function module and inside z-wrapper function module you can call both BAPI's...As we know its is manadatory to call Commit Work after calling BAPI's in order to reflect changes.
    One more solution you can call Inbound Proxy and you can call both BAPIs there.
    Nilesh

  • How to restore system RAM memory when i run DI batch job process

    How to restore system RAM memory when i run DI batch job process
    I use SAP-B12005SP53 and i develop with Business Partner Object

    HI ALL
    TAKE CARE: Nothing was working in VB6, but in .NET it sets the default value !
    from help of VS.
    When you assign Nothing to an object variable, it no longer refers to any object instance. If the variable had previously referred to an instance, setting it to Nothing does not terminate the instance itself. The instance is terminated, and the memory and system resources associated with it are released, only after the garbage collector (GC) detects that there are no active references remaining.
    Garbage Collector distroyes the objects at the end of the sub of function, not in the loop.
    So use local variables when you loading the 7000 items. For me working fine and i do not have a memory problems. I am loading 50000 items with prices from EDI file, and memory usage stops at 120MB.
    1st solution  create a sub or a function to populate the data, and run this for every items by as single rutin,
      sub x
        Dim obj As New Object()
            Try
    ... put here your update code
            Catch ex As Exception
            Finally
                GC.Collect()
            End Try
      end sub
    This free up the memory when Garbage
    2nd solution: use inner loop definition
      sub x
    Try
                Do While True
                    Dim obj As New Object()
                Loop
            Catch ex As Exception
            Finally
                GC.Collect()
            End Try
      end sub
    This free up the memory inside the loop.
    I recommend to use Friend Class keyword for your addon declaration and do not use modules.
    Modules and shared classes will be free up the memory when your program terminates execution.
    For more information see MSDN help http://msdn.microsoft.com/en-us/library/7ee5a7s1(VS.80).aspx
    and http://msdn.microsoft.com/en-us/library/0x9tb07z(VS.80).aspx
    Regards,
    J
    Edited by: János Nagy on Oct 15, 2009 2:17 PM

  • Getting incorrect value in CCMS: Load Distribution

    Dear Experts,
    I am Using Ecc 6.0 with db2 udb(DB6) on Aix Os.
    I have configured load balancing on my sap system.
    But when i am going to see average time of load distribution it appears like this
    Instance            Resp.time(ms)
    PRD_02             1,480
    PRD_01                  1-
    PRD_00               2,978
    You can see that value for PRD_01 is (1-).
    how to correct this value...
    <removed_by_moderator>
    Do NOT offer rewards*
    Thanks in advance,
    Dhiraj
    Edited by: Juan Reyes on Jun 1, 2010 12:03 PM

    > PRD_02             1,480
    > PRD_01                  1-
    > PRD_00               2,978
    - Is that instance included in the load balancing (SMLG)?
    - Is that instance configured to have dialog processes?
    Markus

  • Parallel & sequential processing

    Hi Experts,
                       i have little bit confusion regarding sequential and parallel processing ,when to go for sequential and parallel processing,what is the advantage ?
    please help me,i am new to odi.
    Regards
    ksbabu

    ksbabu wrote:
    Hi Experts,
                       i have little bit confusion regarding sequential and parallel processing ,when to go for sequential and parallel processing,what is the advantage ?
    please help me,i am new to odi.
    Regards
    ksbabu
    The advantage ? In simple terms, say you have 3 steps that take 10 minutes, in serial :
    Step 1 -> Step 2 -> step 3
    10 + 10 + 10 = 30 Minutes to complete.
    Now lets say you can run these steps in Parallel at the same time
    10
    10
    10
    They will all be finished in 10 minutes, you have saved 20 minutes running in Parallel compared to Serial. In a data warehouse, we can commonly load Dimensions in parallel (providing there are no dependancies) , then we might load the fact tables in parallel also (again , assuming no dependancies).
    Load Plans allow you to orchestrate Parallel / Serial calls to interfaces and Package scenarios, you can also run steps in Parallel in Packages using asynchronous scenario calls.
    Hope this helps.

  • How does failover and load distribution work in Operations Manager resource pools?

    Hello everyone,
    I have a couple of questions in regard to failover and load distribution in resource pools.
    What is the criteria for an x-plat or network device to initiate a failover?
    How is load distributed in a resource pool - Is it round-robin, performance based or something completely different?
    In case more than two management servers are in the same pool, and a management server becomes unavailable, how is load distributed among the remaining resource pool members?
    Any help is appreciated.
    Thx,
    Claus

    Hi,
    Resource Pools are a collection of Health Services working together to manage instances assigned to the pool. 
    Workflows targeted to the instances are loaded by the Health Service in the Resource Pool that ends up managing that instance. 
    If one of the Health Services in the Resource pool were to fail, the other Health Services in the pool will pick up the work that the failed member was running. 
    And I would like to share this article with you. Hope it helps.
    http://www.systemcentercentral.com/how-does-the-failover-process-work-in-opsmgr-2012-scom-sysctr/
    Niki Han
    TechNet Community Support

  • Error on load: System.IO.IOException: The process cannot access the file : error in event viewer when users want to view documents from this third party deployed scan solution

    Error on load: System.IO.IOException: The process cannot access the file
    '\\server1\SCANSHARED\.pdf' because it is being used by another process.
       at System.IO.__Error.WinIOError(Int32 errorCode, String maybeFullPath)
       at System.IO.FileStream.Init(String path, FileMode mode, FileAccess access, Int32 rights, Boolean useRights, FileShare share, Int32 bufferSize, FileOptions options, SECURITY_ATTRIBUTES secAttrs, String msgPath, Boolean bFromProxy)
       at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, FileOptions options, String msgPath, Boolean bFromProxy)
       at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share)
       at System.IO.File.WriteAllBytes(String path, Byte[] bytes)
       at abc.Scan.Layouts.ICC.Scan.View.Page_Load(Object sender, EventArgs e)
    I faced this  error in event viewer  when users want to view documents from this third party deployed scan solution
    here I have two WFS servers  and they configured with load balancing in F5 .
    when I enable both servers in F5 I receive this error messages in 2nd server,
    when users want to view documents
    adil

    Do you have antiVirus installed on the sharepoint servers?
    These folders may have to be excluded from antivirus scanning when you use file-level antivirus software in SharePoint. If these folders are not excluded, you may see unexpected behavior. For example, you may receive "access denied" error messages when files
    are uploaded.
    Please follow this KB and exclude the folders from Scanning.
    http://support.microsoft.com/kb/952167
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

  • Job processing

    Dear Experts,
    My client has received one processing job for the 1st time. How to map the scenario in Sap.
    the customer will provide plain bottle for printing in plastic boxes to be returned after printing in the same box.
    My idea is as follows.
    1>Create the material plain bottle as sales order stock non-valuated as no excise is involved.
    2>Create finished material with a new material type
    3>Create sales order
    3> Create BOM and routing for finished material assigning plain bottle as component.
    4> Create Production order with a new order type with respect to the sales order
    5> Create condition type in sales where only service tax and education cess are involved.
    6> a new gl will also needs to be created i.e Job processing charges
    Please confirm whether I am on the right track and what configuration needs to be maintained in FI, MM, CO and SD.
    Please explain...
    Best Regards,
    Divraj

    Hi Ajay,
    I need suggestion for below mentioned problem.
    We will be receiving suppose 10000 bottle for printing as job work. There is 1% breakage allowance....
    Now if 100 bottles are broken we need to return to the customer those 100 bottles in broken condition.
    How to map this.
    What we have opined.......
    As we have created plain bottle as non-valuated. We will receive the same through 501 E mvt. type i.e without PO against SO in a different storage location.
    Instead of creating a seperate material code for broken bottle we will generate the same plain bottle through 531 movt type on order itself in a separate storage location. Then do 502 movement return to customer.
    Please suggest the best method...
    Regards,
    Divraj

  • Distribution Manager failed to process package triggers an update of the packages

    Hi all,
    we have randomly following issue:
    SCCM 2012 Sp1 CAS, takes a snapshot of a Windows update package in the morning (not sure how the frequency of this check is done). If for any reason it fails, SCCM automatically redistributes the package to all sites. This happened this morning again for 5
    Windows updates packages. You understand that this means GB sent to all Secondary sites (66) with an useles amount of data sent out.
    From the Status messages I see
    Information Milestone RC0 06.11.2014 07:12:11 SCHVSGGSC600.rccad.net SMS_DISTRIBUTION_MANAGER 2300 Distribution Manager is beginning to process package "SUP-2014.09" (package ID = RC00017B).
    Then lot of updates lists with comment taking a snapshot and finally
    Error Milestone RC0 06.11.2014 07:12:29 SCHVSGGSC600.rccad.net SMS_DISTRIBUTION_MANAGER 2302 Distribution Manager failed to process package "SUP-2014.09" (package ID = RC00017B).    Possible cause: Distribution manager does not have access to either the package source directory or the distribution point.  Solution: Verify that distribution manager can access the package source directory/distribution point.    Possible cause: The package source directory contains files with long file names and the total length of the path exceeds the maximum length supported by the operating system.  Solution: Reduce the number of folders defined for the package, shorten the filename, or consider bundling the files using a compression utility.    Possible cause: There is not enough disk space available on the site server computer or the distribution point.  Solution: Verify that there is enough free disk space available on the site server computer and on the distribution point.    Possible cause: The package source directory contains files that might be in use by an active process.  Solution: Close any processes that maybe using files in the source directory.  If this failure persists, create an alternate copy of the source directory and update the package source to point to it.
    This triggers immediately an update of all DPs
    Information Milestone RC0 06.11.2014 07:43:52 SCHVSGGSC600.rccad.net SMS_DISTRIBUTION_MANAGER 2304 Distribution Manager is retrying to distribute package "RC00017B".    Wait to see if the package is successfully distributed on the retry.
    Any idea
    How this can be avoided, since nobody changed the package and we suppose it was a temp connection issue between the CAS and the package repository server
    If this check can be set up to once a week for instance or even less?
    Thanks,
    Marco

    Hi Daniel,
    thanks for the prompt answer. Actually I saw it yesterday at least for 1 package (the last one). The error is generate by SQL killing a task 
    Adding these contents to the package RC00010D version 347.
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:23
    7796 (0x1E74)
    Sleep 30 minutes... SMS_DISTRIBUTION_MANAGER
    06.11.2014 07:12:23 3652 (0x0E44)
    *** [40001][1205][Microsoft][SQL Server Native Client 11.0][SQL Server]Transaction (Process ID 152) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction.
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:25
    5460 (0x1554)
    STATMSG: ID=2302 SEV=E LEV=M SOURCE="SMS Server" COMP="SMS_DISTRIBUTION_MANAGER" SYS=SCHVSGGSC600.rccad.net SITE=RC0 PID=2144 TID=5460 GMTDATE=jeu. nov. 06 06:12:25.422 2014 ISTR0="SUP-2012.Q4" ISTR1="RC000068" ISTR2=""
    ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=1 AID0=400 AVAL0="RC000068"
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:25
    5460 (0x1554)
    Failed to process package RC000068 after 0 retries, will retry 100 more times
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:25
    5460 (0x1554)
    Exiting package processing thread. SMS_DISTRIBUTION_MANAGER
    06.11.2014 07:12:25 5460 (0x1554)
    Used 4 out of 5 allowed processing threads.
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:26
    3300 (0x0CE4)
    Starting package processing thread, thread ID = 0x894 (2196)
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:27
    3300 (0x0CE4)
    Used all 5 allowed processing threads, won't process any more packages.
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:27
    3300 (0x0CE4)
    Sleep 1828 seconds... SMS_DISTRIBUTION_MANAGER
    06.11.2014 07:12:27 3300 (0x0CE4)
    STATMSG: ID=2300 SEV=I LEV=M SOURCE="SMS Server" COMP="SMS_DISTRIBUTION_MANAGER" SYS=SCHVSGGSC600.rccad.net SITE=RC0 PID=2144 TID=2196 GMTDATE=jeu. nov. 06 06:12:27.716 2014 ISTR0="SUP-2014.M05" ISTR1="RC00011D" ISTR2=""
    ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=1 AID0=400 AVAL0="RC00011D"
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:27
    2196 (0x0894)
    Start updating the package RC00011D... SMS_DISTRIBUTION_MANAGER
    06.11.2014 07:12:27 2196 (0x0894)
    CDistributionSrcSQL::UpdateAvailableVersion PackageID=RC00011D, Version=14, Status=2300
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:27
    2196 (0x0894)
    *** [40001][1205][Microsoft][SQL Server Native Client 11.0][SQL Server]Transaction (Process ID 154) was deadlocked on lock resources with another process and has been chosen as the deadlock victim. Rerun the transaction.
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:27
    1424 (0x0590)
    Taking package snapshot for package RC00011D
    SMS_DISTRIBUTION_MANAGER 06.11.2014 07:12:27
    2196 (0x0894)
    Now as I mentioned it was probably an SQL (more likely) or networ issue, however the question was more about how to avoid that. Finally the packages were not changed and if possible I would avoid any automatic SCCM action on packages unless an operator manually
    triggers them.
    I didn't see any error today, so it should not be a configuration issue.
    Is this possible?

  • How to synchronise the job processing from the program.

    How to synbchronize the job processing when we are creating jobs from the program.
    Thanks,
    Ramana.

    hi,
    yes you can do this..
    try this
    create a two screen suppose 9000 and 9001..
    then right click on your program name...
    create a TCODE say TONE..
    in this give the screen number 9000..
    now again right click on the program name
    create a TCODE say Tsecond
    in this give the screen number 9001...
    hope this will help you..
    Regards
    Ritesh J

  • Disabling the cost distribution button on the Process Purchase Order Screen

    This is probably a simple question, but we have been unable to find the place to disable cost distribution on the SRM Process Purchase Order screen for limit orders ONLY.  Basically, we just want this functionality to not be available to users.  Can anyone direct us to the correct configuration for this?  Or do we need to edit the BADI?
    Thanks  in Advance!

    Hello,
    You can hide this field by creating a screen variant through SHD0 transaction. I have tested here and I
    could hide it (GV_DIST_IND in Program SAPLBBP_PDH_ACC screen 1000).    
    In the screen variant you should mark the output of GV_DIST_IND field.
    However, this option would hide this button for all documents.
    Kind regards,
    Ricardo

  • Problem in table locking while running the parallel jobs (deadlock?)

    Hi,
    I am trying to delete the entried from a custom table. If I schedule the parallel jobs, I am getting the following dump while deleting the entries from table. This is happening for custom table ( in CRM system).
    Exception : DBIF_RSQL_SQL_ERROR
    And complaining that, deadlock occured. I am not sure what it is..
    Could you please help me out, how can we solve this issue?
    Thanks,
    Sandeep

    Hello Sandeep ,
    I would suggest you Record based  Lock Objects ,
    take a look at below link...to get a general idea..!
    <link removed>
    Hope it helps,
    Thanks and Regards,
    Edited by: Suhas Saha on Aug 19, 2011 11:27 AM

Maybe you are looking for

  • Solution for AAM Updater Error : U44M1I210

    After getting Creative Cloud a couple of days ago then running into Updater Error U44M1l210 over and over and making the mistake of calling tech support (which was no help - "just reset it / download from the site" - even though resetting does nothin

  • Same external drive can only be detected on one of my 2 macs

    I have this usb external drive in exfat format, it can be recognized on my late 2013 macbook pro, but won't on my late 2008, when i plug it into my late 2008 mac, it's won't even show in Disk Utility. Is there a way to fix this without losing data?

  • Best backup program for iPhone 3GS??

    I am looking for the BEST backup program for the above - not only a program which makes backups of the programs, but ALO the contents of the programs - of everything! Any recommendations, thanks!

  • Post automatically tick in Trading consumption GL account

    Hi, I am having third party sales scenario in my project. After creating sales order it is creating PR and at the creation of PO, System is picking the trading GL account from OBYC. But it throwing error that automatic posting is not allowed and only

  • CS5 Master Collection: So much installs, even when I choose not to.

    Hello everyone! I have installed Adobe CS5 for the mac, but have been very choosey in my install options. During the install, I only chose Photoshop, After Effects and OnLocation to install, as well as unchecking ALL of the other options. Low and beh