CKMLCP - Process Parallel

I  have process of the transfering between plants. I have processes of transfer between plants. There is some problem in the determination of multilevel run with the option of parallel processing?
The idea is to gain in performance
Thanks,

Dear  Young,
I  have no problems. I just want to know if there is any problem. If the sistem does not consider some kind of movement when I do parallel processing.
I think there is any problem. But I want to share more opinions
Thanks

Similar Messages

  • Internal error occured when processing parallel RFC

    Dear Experts,
    I have created one FM and calling that FM as prallel process in a report program.
    CALL FUNCTION 'ZHR_TIME_EVAL_PARALLEL_PROCESS'
          STARTING NEW TASK first_task
          DESTINATION IN GROUP srv_grp
          EXPORTING
            p_job     = sub_job
          TABLES
            rspar_tab = rspar_tab.
    In development system its working fine. In quality it throwing a dump saying "Internal error occured when processing parallel RFC" . Error mesage number - BT 857.
    Any body came across this problem.Please help me to resolve.
    Thanks in advance

    Hi chen,
    I have executed the function modules.Its throwing the floowing exceptions.
    SPBT_INITIALIZE     is throwing the exception INVALID_GROUP_NAME  Message number  BT  - 831.
    Function modules SPBT_GET_CURR_RESOURCE_INFO , SPBT_FIND_FREE_SERVER, SPBT_GET_CURR_RESOURCE_INFO are throwing the exception PBT_ENV_NOT_INITIALIZED_YET  message number BT - 838 .
    How to resolve this problem.
    Thanks in advance.

  • Parallel process in process chains

    Hi All,
    I have created a process chain with two parallel jobs (for forecast run) in APO DP. whenever i'm checking it, it is showing me that <b>"Too many parallel processes for chosen server"</b>
    In the diagnosis, the message is "On the server  you have chosen, there are only 1 batch processes available. The process chain has been designed in such as way that 2 processes must be processed parallel".
    Do I need to maintain any settings.Please let me know.
    Thank you.
    Regards,
    Raj

    Hi Raj,
    What it is complaining about is that the number of batch processes(background processes) available on the system are only 1 and you have given it 2 background processes..now if the system only has 1 bckgnd process,, how will it run 2 processes ) ??
    You can check the number of background processes in SM50 and then modify the number of background processes in RZ10(start up profile)......Ask your basis guy to increase/check that???
    Thanks
    Abhi

  • Parallel Processing at the time of confirmation

    Hi Gurus,
    During confirmation of production orders / collective orders (T.Code u2013 CO11 / CO12 / COHV) or through REM Backflush done in Repetitive Manufacturing (T.Code - MFBF), material components get locked, for processing when multiple users are transacting either confirmation of above orders or MM-IM transactions in parallel.
    Eg. Goods Receipt for components in stores cannot be made during confirmation by planning production scheduler and vice versa.
    As a result, transaction of Goods Receipt making cannot be executed getting error message u201CUser lockedu201D.
    Please suggest us how parallel processing can be carried out on same materials simultaneously like working on CO11, CO12, CO1P and MIGO for the same material. System should allow processing parallel with minimum execution time.
    Any SAP notes available for this?
    Any one faced such issue and resolved this, please share.
    Thanks in advance,
    Regards,
    Srini

    Hi Srini,
    Normally this should not be a big problem, since it takes just a short time to process automatic goods movements related to production order confirmation (backflush GI, and GR).
    But if in your reality it is, then SAP offers a decoupling concept, which results in posponing those movements, and perform them en-mass at a more convenient time when it less disturbs (like after normal office hours). Look in the link for full details:
    http://help.sap.com/saphelp_erp60_sp/helpdata/en/b1/c04d42439a11d189410000e829fbbd/content.htm
    Regards,
    Mario

  • Warning Error message while scheduling the Process Chain

    Dear All,
    I have created the meta Chain, which contains 7 Local Chains. I have tested this chain in BI Development and Quality. It was working fine. I have moved this Meta Chain to production. While i am trying to schedule this chain, i got the warning message.
    Too many parallel processes for chosen server BICI
    Message no. RSPC118
    Diagnosis
    On the server TVSBICI you have chosen, there are only 0 batch processes available. The process chain has been designed in such as way that 3 processes must be processed parallel.
    System Response
    Only 0 processes can be processed parallel. Note also that to be able to start sub-chains, AT LEAST TWO free batch processes must be available. Otherwise a lock situation may arise.
    For more information, see SAP Note 621400.
    Procedure
    Reduce the number of parallel processes in the chain or include sub- chains. This warning applies to all relevant processes.
    Procedure for System Administration
    Increase the number of batch processes available. You should note however that it makes no sense to set up more batch processes than there are CPUs on server BICI.
    Pls help on this issue.
    Thanks,
    Siva.
    Edited by: Siva Kumar on Jun 15, 2009 9:16 AM

    I have Ignored this warning and now process Chain is working fine.
    Thanks,
    Siva.

  • Error while converting BPMN to BPEL in Business Process Architect

    Hey,
    I am using Business Process Architect 11gR1 where I have modeled my process. Now that I have to move this model to BPM, I am using the option SOA -> share blueprint with IT in BPA. A dialog appears that asks to validate the model. I confim with "Yes" and it shows the conversion progress bar but after the while an error appears (a java exception). A report opens in my browser that has the following message
    Validation of a service-oriented BPMN model in Oracle BPA Suite
    Structure rules for BPMN in Oracle BPA Suite
    Rules for a service-oriented BPMN model in Oracle BPA Suite
    String not found: RULEKIND_NAMEd89733d0-d605-11db-0505-000fb0c4ad32
    The java exception is "Error while creating report:
    report: oraclebpmn2bpel, line 632
    Type error: cannot find function getErrNo in object.java.lang.IllegalArgumentException. (oraclebpmn2bpel#632)
    can anyone please help me as to why this happens and how it can be sorted out?
    Edited by: user10303310 on Dec 13, 2011 3:10 AM

    Hi Julika,
    The error message might not be related to the activities itself. I assume your model is not well structured. Consider that the following points aren't disregarded:
    * Never "jump" out of a particular branch into a different part of the model or into another branch
    * All functions/events have only one incoming/outgoing connection
    * Process parallel flows should be specified by splitting and joining AND/XOR rules, or they should contain either one splitting AND/XOR rule only for which there is no other connection between their paths, or one joining AND/XOR rule only that is met by all connections.
    Let me know if it did not help to fix your problem.
    Best regards,
    Danilo

  • Error when processing dimension using XMLA script

    I am hopeful that someone has an answer for this...
    I am in the process of deploying SQLServer 2014 and have been running in parallel with my production environment for 2 weeks with no issues.  Yesterday morning, my primary data load job failed while executing a process dimension script.  The script
    does a process update on multiple dimensions, but fails on the product dimension.  This is the script:
    <Batch xmlns="http://schemas.microsoft.com/analysisservices/2003/engine">
    <Parallel>
    <Process xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ddl2="http://schemas.microsoft.com/analysisservices/2003/engine/2" xmlns:ddl2_2="http://schemas.microsoft.com/analysisservices/2003/engine/2/2">
    <Object>
    <DatabaseID>WestLakeSales</DatabaseID>
    <DimensionID>Dim Product</DimensionID>
    </Object>
    <Type>ProcessUpdate</Type>
    <WriteBackTableCreation>UseExisting</WriteBackTableCreation>
    </Process>
    </Parallel>
    </Batch>
    It processes the attributes and starts processing the associated measure groups, then fails with a return code that I cannot find any information on:
    <return xmlns="urn:schemas-microsoft-com:xml-analysis">
    <results xmlns="http://schemas.microsoft.com/analysisservices/2003/xmla-multipleresults">
    <root xmlns="urn:schemas-microsoft-com:xml-analysis:empty">
    <Exception xmlns="urn:schemas-microsoft-com:xml-analysis:exception" />
    <Messages xmlns="urn:schemas-microsoft-com:xml-analysis:exception">
    <Error ErrorCode="3238133769" Description="" Source="Microsoft SQL Server 2014 Analysis Services" HelpFile="" />
    </Messages>
    </root>
    </results>
    </return>
    At the end of the job log it just shows this:
    Finished processing the 'Inventory Weekly 2009 08' partition.
    Processing of the 'Inventory Weekly 2009 06' partition has started.
    Finished processing the 'Inventory Weekly 2009 06' partition.
    Processing of the 'Inventory Weekly 2010 02' partition has started.
    The job completed with failure.
    The job completed with failure.
    The job completed with failure.
    Execution complete
    When I run process update from Studio it processes all of the attributes and measure group partitions, then throws a single 'error' result with no description or message.  It appears that all partitions process successfully.  See the screen shot:
    This script ran fine up until yesterday.  There were no system changes.  The same script is currently running fine on the production SQLServer 2008 instance.
    Anyone see anything like this happening in SSAS 2014?  Any tips on further trouble shooting?

    Hi Bendare2,
    Thank you for sharing your solution.
    Regards,
    Simon Hou
    TechNet Community Support

  • Material Ledger Close (CKMLCP) -problem executing during the month

    Due to our tight closing schedule, we are attempting to pre-run the Material Ledger close (CKMLCP) process and Value Flow Monitor (CKMVFM) on the last 2 days of the month in order to catch any issues before the first workday of the close, when it is too late to do many of the normal processes that would correct the problems in our system.  The issue we are having is that some of our production plants are 24/7 and we are not getting good results in our test run due to new transactions that get posted while we are running the CKMLCP steps, resetting the material statuses.  For example between the Single level and multi level calculations for a given material, a goods movement might get posted that resets the material status from "single level calculated" back to "quantities and values entered", which leads to errors when the multilevel calculation tries to roll up the WAC variances.
    Has anyone encountered this issue, and how did you get around it?  Currently, we can only think of locking people out of the system while we are running the pre-checks, but we know the users will balk at that suggestion. 
    Thanks,
    Burns Thomas

    Hi Thomas,
    you might try to use the alternative valuation run as a simulation environment.
    [http://help.sap.com/saphelp_erp60/helpdata/en/d2/cb532d455611d189710000e8322d00/frameset.htm]
    an alterantive valuation run for just one month will create a copy of the ML data in which you will have no problem with material status. the calculation logic for closing is the same, also the reporting, like CKMVFM. Therefore, If you are successfull in closing the AVR that is a good indcator for not having further problems for your real data.
    best regards,
                         Udo

  • Inf reg process chains design

    Hi All,
    My process chain is done succesfully.. I followed as
    Main Chain
    Start->Delete Data (3ods)->Load data(3ods)->Activate(3ods)-> And(all success ful)->2local process chains(series to and) -> Materdata Load(parallel to and)->Attribute change run
    Local Process chain 1: ( Data loads from ods to cube)
    Start->delete index(cube)-> load data(6 info packages)
    Local Process Chain 2: (Data loads from Flat File to Cube)
    Start->load Data(another 6 Info Packages)->and(condition)-> delete Index(cube).
    Now tell me in design why am getting too many parallel processes for chosen sever
    MSg NO:RSPC118
    why these warnings are comming? give some light on this
    regards
    ashwin

    Hi ashwin,
    Too many parallel processes for chosen server
    Message no. RSPC118
    Diagnosis
    On the server you have chosen, there are only 4 batch processes available. The process chain has been designed in such as way that 5 processes must be processed parallel.
    System Response
    Only 4 processes can be processed parallel. Note also that to be able to start sub-chains, AT LEAST TWO free batch processes must be available. Otherwise a lock situation may arise.
    For more information, see SAP Note 621400.
    Procedure
    Reduce the number of parallel processes in the chain or include sub- chains. This warning applies to all relevant processes.
    Procedure for System Administration
    Increase the number of batch processes available. You should note however that it makes no sense to set up more batch processes than there are CPUs on server .
    Suggestions:
    You may created your Process Chain with 5 parrllel lines or more..You don't free Background Process Available in the server where it is running. I think you have only 4 Background Process Available.
    code
    Start Variant
    Variant 1 Variant 2 Variant 3 Variant 4 Variant 5
    Variant6 variant 7 Variant 8 Variant 9 variant 10
    You can replace it as..
    Code
    Start Variant
    Variant 1 Variant 2
    Variant3 Variant 4
    Variant 5 Variant6
    variant 7 Variant 8
    Variant 9 variant 10
    do's:
    1.You can change your Process Chain Model
    2.You can increase background Process on your System.
    3.You can keep it like this if you are working Development. Your Production will have Extra processes.
    Hope this helps,
    Regards
    CSM Reddy

  • Background process BCTL_4CEDE3M8RTW5WPMQ01YVZ4QYE terminated due to missing

    Hi,
    I am getting following error when i am activating ODS (Bi 7.0)
    Background process BCTL_4CEDE3M8RTW5WPMQ01YVZ4QYE terminated due to missing confirmation.
    I tried reloading and activating it,
    i tried with RSRV adjusts.
    I tried changing package size.
    I tried reducing parallel executions.
    Still not able to activate. Can some one help me.

    Here are the short dump details
    Short text of error message:
    Invalid PBT server group name: PARALLEL_GENERATORS
    Long text of error message:
    Diagnosis
         The system tried to initialize the environment for processing
         parallel RFCs using the function module SPBT_INITIALIZE. The server
         group name specified was found to be invalid. The server group name
         describes a group of servers on which the parallel RFCs should be
         processed.
    System Response
    Procedure
         You can use Transaction RZ12 to determine which PBT server groups
         are configured in your system. You should only use one of the
         server group names listed there to process parallel RFCs.
    But i m activating serially with value 1 in option.

  • Parallel & alternative sequence-Routing

    Dear all,
    Can any one please send the document for how to process parallel & alternative sequence.
    I tried but during confirmation its not showing the operation which i created in sequence.
    Kindly let me know.
    Regards
    Rajasekaran

    See the help file
    http://help.sap.com/saphelp_47x200/helpdata/en/7e/d42611455911d189400000e8323c4f/frameset.htm
    it is very clear in thsi web page

  • Ora 01033- Oracle initialization or shutdown in process

    Hi all
    I am getting this error when I try to connect to SQL......I have restart the serveice of orcl but the result is same here
    Is any way to get out from this error?
    Thanks in advance
    Shayan
    [email protected]

    HI sory for late actually i am working on server for this problem and still the result is same
    Here is alert_ora file last line
    Tue Jul 24 12:22:59 2007
    DEBUG: Replaying xcb 0x6be55a24, pmd 0x6b834010 for failed op 8
    Doing block recovery for file 2 block 934
    No block recovery was needed
    Tue Jul 24 12:38:02 2007
    alter database mount
    Tue Jul 24 12:38:02 2007
    ORA-1100 signalled during: alter database mount
    Tue Jul 24 12:38:13 2007
    alter database open
    Tue Jul 24 12:38:13 2007
    ORA-1531 signalled during: alter database open...
    Tue Jul 24 12:38:50 2007
    Shutting down instance (abort)
    License high water mark = 3
    Instance terminated by USER, pid = 2856
    Tue Jul 24 12:39:07 2007
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 2
    KCCDEBUG_LEVEL = 0
    Using LOG_ARCHIVE_DEST_10 parameter default value as USE_DB_RECOVERY_FILE_DEST
    Autotune of undo retention is turned on.
    Dynamic strands is set to TRUE
    Running with 2 shared and 18 private strand(s). Zero-copy redo is FALSE
    IMODE=BR
    ILAT =18
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    Starting up ORACLE RDBMS Version: 10.1.0.2.0.
    System parameters with non-default values:
    processes = 150
    shared_pool_size = 83886080
    large_pool_size = 8388608
    java_pool_size = 50331648
    control_files = D:\ORACLE\PRODUCT\10.1.0\ORADATA\ORCL\CONTROL01.CTL, D:\ORACLE\PRODUCT\10.1.0\ORADATA\ORCL\CONTROL02.CTL, D:\ORACLE\PRODUCT\10.1.0\ORADATA\ORCL\CONTROL03.CTL
    db_block_size = 8192
    db_cache_size = 25165824
    compatible = 10.1.0.2.0
    db_file_multiblock_read_count= 16
    db_recovery_file_dest = d:\oracle\product\10.1.0\flash_recovery_area
    db_recovery_file_dest_size= 2147483648
    undo_management = AUTO
    undo_tablespace = UNDOTBS1
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=orclXDB)
    job_queue_processes = 10
    background_dump_dest = D:\ORACLE\PRODUCT\10.1.0\ADMIN\ORCL\BDUMP
    user_dump_dest = D:\ORACLE\PRODUCT\10.1.0\ADMIN\ORCL\UDUMP
    core_dump_dest = D:\ORACLE\PRODUCT\10.1.0\ADMIN\ORCL\CDUMP
    sort_area_size = 65536
    db_name = orcl
    open_cursors = 300
    pga_aggregate_target = 25165824
    PMON started with pid=3, OS id=3140
    MMAN started with pid=4, OS id=3348
    DBW0 started with pid=5, OS id=2644
    LGWR started with pid=6, OS id=2828
    CKPT started with pid=7, OS id=3120
    SMON started with pid=9, OS id=3036
    RECO started with pid=10, OS id=3532
    CJQ0 started with pid=11, OS id=3724
    Tue Jul 24 12:39:08 2007
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    starting up 1 shared server(s) ...
    Tue Jul 24 12:39:17 2007
    alter database mount
    Tue Jul 24 12:39:17 2007
    Controlfile identified with block size 16384
    Tue Jul 24 12:39:21 2007
    Setting recovery target incarnation to 2
    Tue Jul 24 12:39:21 2007
    Successful mount of redo thread 1, with mount id 1156582901
    Tue Jul 24 12:39:21 2007
    Database mounted in Exclusive Mode.
    Completed: alter database mount
    Tue Jul 24 12:39:30 2007
    alter database open
    Tue Jul 24 12:39:30 2007
    Beginning crash recovery of 1 threads
    attempting to start a parallel recovery with 2 processes
    parallel recovery started with 2 processes
    Tue Jul 24 12:39:30 2007
    Started first pass scan
    Tue Jul 24 12:39:30 2007
    Completed first pass scan
    62 redo blocks read, 9 data blocks need recovery
    Tue Jul 24 12:39:30 2007
    Started redo application at
    Thread 1: logseq 581, block 3, scn 0.0
    Recovery of Online Redo Log: Thread 1 Group 1 Seq 581 Reading mem 0
    Mem# 0 errs 0: D:\ORACLE\PRODUCT\10.1.0\ORADATA\ORCL\REDO01.LOG
    Tue Jul 24 12:39:30 2007
    Completed redo application
    Tue Jul 24 12:39:30 2007
    Completed crash recovery at
    Thread 1: logseq 581, block 65, scn 0.3471104
    9 data blocks read, 9 data blocks written, 62 redo blocks read
    Tue Jul 24 12:39:31 2007
    Thread 1 advanced to log sequence 582
    Maximum redo generation record size = 120832 bytes
    Maximum redo generation change vector size = 116476 bytes
    Private_strands 7 at log switch
    Thread 1 opened at log sequence 582
    Current log# 2 seq# 582 mem# 0: D:\ORACLE\PRODUCT\10.1.0\ORADATA\ORCL\REDO02.LOG
    Successful open of redo thread 1
    Tue Jul 24 12:39:31 2007
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Tue Jul 24 12:39:31 2007
    SMON: enabling cache recovery
    Tue Jul 24 12:39:31 2007
    Successfully onlined Undo Tablespace 1.
    Tue Jul 24 12:39:31 2007
    SMON: enabling tx recovery
    Tue Jul 24 12:39:31 2007
    Database Characterset is WE8MSWIN1252
    Tue Jul 24 12:39:31 2007
    Published database character set on system events channel
    Tue Jul 24 12:39:31 2007
    All processes have switched to database character set
    Tue Jul 24 12:39:32 2007
    Errors in file d:\oracle\product\10.1.0\admin\orcl\udump\orcl_ora_3340.trc:
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []
    DEBUG: Replaying xcb 0x6be82ea4, pmd 0x6b834010 for failed op 8
    Doing block recovery for file 2 block 934
    No block recovery was needed
    Tue Jul 24 12:41:18 2007
    Errors in file d:\oracle\product\10.1.0\admin\orcl\udump\orcl_ora_3340.trc:
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []
    Tue Jul 24 12:41:20 2007
    DEBUG: Replaying xcb 0x6be82ea4, pmd 0x6b834010 for failed op 8
    Doing block recovery for file 2 block 934
    No block recovery was needed
    Tue Jul 24 12:41:20 2007
    Errors in file d:\oracle\product\10.1.0\admin\orcl\udump\orcl_ora_3340.trc:
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []
    Tue Jul 24 12:41:22 2007
    Errors in file d:\oracle\product\10.1.0\admin\orcl\udump\orcl_ora_3340.trc:
    ORA-00603: ORACLE server session terminated by fatal error
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []
    ORA-00600: internal error code, arguments: [4193], [976], [980], [], [], [], [], []

  • MAIN DIFFERENCES BETWEEN PARALLEL AND SEQUENTAIL PRCESSING???

    HI PALS,
    I WANT THE COMPLETE DIFFERENCES BETWEEN PARALLEL AND SEQUENTIAL PROCESSING!
    IN THE CONTEXT OF RFC.

    Hi
    Parallel Processing
    To achieve a balanced distribution of the system load, you can use destination additions to execute function modules in parallel tasks in any application server or in a predefined application server group of an SAP system.
    Parallel-processing is implemented with a special variant of asynchonous RFC. Itu2019s important that you use only the correct variant for your own parallel processing applications: the CALL FUNCTION STARTING NEW TASK DESTINATION IN GROUP keyword. Using other variants of asynchronous RFC circumvents the built-in safeguards in the correct keyword, and can bring your system to its knees
    Details are discussed in the following subsections:
    ·        Prerequisites for Parallel Processing
    ·        Function Modules and ABAP Keywords for Parallel Processing
    ·        Managing Resources in Parallel Processing
    Prerequisites for Parallel Processing
    Before you implement parallel processing, make sure that your application and your SAP system meet these requirements:
    ·        Logically-independent units of work:
    The data processing task that is to be carried out in parallel must be logically independent of other instances of the task. That is, the task can be carried out without reference to other records from the same data set that are also being processed in parallel, and the task is not dependent upon the results of others of the parallel operations. For example, parallel processing is not suitable for data that must be sequentially processed or in which the processing of one data item is dependent upon the processing of another item of the data.
    By definition, there is no guarantee that data will be processed in a particular order in parallel processing or that a particular result will be available at a given point in processing. 
    ·        ABAP requirements:
    ¡        The function module that you call must be marked as externally callable. This attribute is specified in the Remote function call supported field in the function module definition (transaction SE37).
    ¡        The called function module may not include a function call to the destination u201CBACK.u201D
    ¡        The calling program should not change to a new internal session after making an asynchronous RFC call. That is, you should not use SUBMIT or CALL TRANSACTION in such a report after using CALL FUNCTION STARTING NEW TASK.  
    ¡        You cannot use the CALL FUNCTION STARTING NEW TASK DESTINATION IN GROUP keyword to start external programs. 
    ·        System resources: 
    In order to process tasks from parallel jobs, a server in your SAP system must have at least 3 dialog work processes. It must also meet the workload criteria of the parallel processing system: Dispatcher queue less than 10% full, at least one dialog work process free for processing tasks from the parallel job.
    Function Modules and ABAP Keywords for Parallel Processing
    You can implement parallel processing in your applications by using the following function modules and ABAP keywords:
    ·        SPBT_INITIALIZE: Optional function module. 
    Use to determine the availability of resources for parallel processing. 
    You can do the following:
    ¡        check that the parallel processing group that you have specified is correct.
    ¡        find out how many work processes are available so that you can more efficiently size the packets of data that are to be processed in your data.
    ·        CALL FUNCTION Remotefunction STARTING NEW TASK Taskname DESTINATION IN GROUP:
    With this ABAP statement, you are telling the SAP system to process function module calls in parallel. Typically, youu2019ll place this keyword in a loop in which you divide up the data that is to be processed into work packets. You can pass the data that is to be processed in the form of an internal table (EXPORT, TABLE arguments). The keyword implements parallel processing by dispatching asynchronous RFC calls to the servers that are available in the RFC server group specified for the processing.
    Note that your RFC calls with CALL FUNCTION are processed in work processes of type DIALOG. The DIALOG limit on processing of one dialog step (by default 300 seconds, system profile parameter rdisp/max_wprun_time) applies to these RFC calls. Keep this limit in mind when you divide up data for parallel processing calls. 
    ·        SPBT_GET_PP_DESTINATION: Optional function module. 
    Call immediately after the CALL FUNCTION keyword to get the name of the server on which the parallel processing task will be run. 
    ·        SPBT_DO_NOT_USE_SERVER: Optional function module. 
    Excludes a particular server from further use for processing parallel processing tasks. Use in conjunction with SPBT_GET_PP_DESTINATION if you determine that a particular server is not available for parallel processing (for example, COMMUNICATION FAILURE exception if a server becomes unavailable).
    ·        WAIT: ABAP keyword
    WAIT UNTIL
    Required if you wish to wait for all of the asynchronous parallel tasks created with CALL FUNCTION to return. This is normally a requirement for orderly background processing. May be used only if the CALL FUNCTION includes the PERFORMING ON RETURN addition.
    ·        RECEIVE: ABAP keyword
    RECEIVE RESULTS FROM FUNCTION Remotefunction
    Required if you wish to receive the results of the processing of an asynchronous RFC. RECEIVE retrieves IMPORT and TABLE parameters as well as messages and return codes.
    Managing Resources in Parallel Processing
    You use the following destination additions to perform parallel execution of function modules (asynchronous calls) in the SAP system:
    In a predefined group of application servers:
    CALL FUNCTION Remotefunction STARTING NEW TASK Taskname
    DESTINATION IN GROUP Groupname
    In all currently available and active application servers:
    CALL FUNCTION Remotefunction STARTING NEW TASK Taskname
    DESTINATION IN GROUP DEFAULT
    Sequential Processing
    In the following cases, the system chooses sequential (non-parallel) processing:
    ●      In table RSADMIN, entry QUERY_MAX_WP_DIAG has value (column value) 1.
    ●      The entire query consists of one sub-access only.
    ●      The query is running in a batch process.
    ●      The query was started from the query monitor (transaction RSRT) using various debug options (for example, SQL query display, execution plan display). See, Dividing a MultiProvider Query into Sub-Queries.
    ●      The query requests non-cumulative key figures.
    ●      Insufficient dialog processes are available when the query is executed. These are required for parallel processing.
    ●      The query is configured for non-parallel processing.
    ●      You want to save the result of the query in a file or a table.
    In Release SAP NetWeaver 7.0, the system can efficiently manage the large intermediate results produced by parallel processing. In previous releases, the system terminated when it reached a particular intermediate result size and proceeded to read data sequentially. This is no longer the case. Therefore, the RSADMIN parameter that was used in previous releases for reading a MultiProvider sequentially is no longer used.
    Reward If Helpfull,
    Naresh

  • NULL IS NOT NULL filter issue

    Can someone explain why 'Y' = 'N' is not working with PARALLEL Plan? i.e. With the filter like 'Y' = 'N' specified and if PQ is used , it does not return instantly. In fact it reads the entire table.
    Here is the test case.. Goal is to execute only one of the SQL joined by union all. I have included 'Y' = 'N' in both SQLs for the test purpose.
    DB Version is 10.2.0.4
    Create table test_tbl_01 nologging as select do.* from dba_objects do , dba_objects d1 where rownum < 22000001;
    Create table test_tbl_02 nologging as select do.* from dba_objects do , dba_objects d1 where rownum < 22000001;
    execute DBMS_STATS.GATHER_TABLE_STATS('SCOTT', 'TEST_TBL_01');
    execute DBMS_STATS.GATHER_TABLE_STATS('SCOTT', 'TEST_TBL_02');
    *Serial path with 2 table join*
    SQL> select
      2    /* parallel(t1,2 ) parallel(t2,2) */
      3    t1.*
      4    from test_tbl_01 t1 ,test_tbl_02 t2
      5    where t1.object_name = t2.object_name
      6    and  'Y' = 'N'
      7    and  t1.object_type = 'TABLE'
      8    union all
      9    select
    10    /* parallel(t1,2 ) parallel(t2,2) */
    11    t1.*
    12    from test_tbl_01 t1 ,test_tbl_02 t2
    13    where t1.object_name = t2.object_name
    14    and  'Y' = 'N'
    15  /
    no rows selected
    Elapsed: 00:00:00.01
    Execution Plan
    Plan hash value: 3500703583
    | Id  | Operation            | Name        | Rows  | Bytes |TempSpc| Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT     |             |     2 |   168 |       |     0   (0)|          |
    |   1 |  UNION-ALL           |             |       |       |       |            |          |
    |*  2 |   FILTER             |             |       |       |       |            |          |
    |*  3 |    HASH JOIN         |             |   660G|    50T|   449M|  6242K (99)| 24:16:38 |
    |*  4 |     TABLE ACCESS FULL| TEST_TBL_01 |  5477K|   386M|       | 41261   (2)| 00:09:38 |
    |   5 |     TABLE ACCESS FULL| TEST_TBL_02 |    22M|   212M|       | 40933   (2)| 00:09:34 |
    |*  6 |   FILTER             |             |       |       |       |            |          |
    |*  7 |    HASH JOIN         |             |  2640G|   201T|   467M|    24M(100)| 95:54:53 |
    |   8 |     TABLE ACCESS FULL| TEST_TBL_02 |    22M|   212M|       | 40933   (2)| 00:09:34 |
    |   9 |     TABLE ACCESS FULL| TEST_TBL_01 |    21M|  1546M|       | 41373   (3)| 00:09:40 |
    Predicate Information (identified by operation id):
       2 - filter(NULL IS NOT NULL)
       3 - access("T1"."OBJECT_NAME"="T2"."OBJECT_NAME")
       4 - filter("T1"."OBJECT_TYPE"='TABLE')
       6 - filter(NULL IS NOT NULL)
       7 - access("T1"."OBJECT_NAME"="T2"."OBJECT_NAME")
    Statistics
              1  recursive calls
              0  db block gets
              0  consistent gets
              0  physical reads
              0  redo size
            567  bytes sent via SQL*Net to client
            232  bytes received via SQL*Net from client
              1  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
              0  rows processed
    *Parallel path with 2 table join*
    SQL> select
      2    /*+ parallel(t1,2 ) parallel(t2,2) */
      3    t1.*
      4    from test_tbl_01 t1 ,test_tbl_02 t2
      5    where t1.object_name = t2.object_name
      6    and  'Y' = 'N'
      7    and  t1.object_type = 'TABLE'
      8    union all
      9    select
    10    /*+ parallel(t1,2 ) parallel(t2,2) */
    11    t1.*
    12    from test_tbl_01 t1 ,test_tbl_02 t2
    13    where t1.object_name = t2.object_name
    14    and  'Y' = 'N'
    15  /
    no rows selected
    Elapsed: 00:01:09.34
    Execution Plan
    Plan hash value: 1557722279
    | Id  | Operation                   | Name        | Rows  | Bytes |TempSpc| Cost (%CPU)| Time     |    TQ  |IN-OUT| PQ Distrib |
    |   0 | SELECT STATEMENT            |             |     2 |   168 |       |     0   (0)|          |     |         |            |
    |   1 |  PX COORDINATOR             |             |       |       |       |            |          |     |         |            |
    |   2 |   PX SEND QC (RANDOM)       | :TQ10004    |       |       |       |            |          |  Q1,04 | P->S | QC (RAND)  |
    |   3 |    BUFFER SORT              |             |     2 |   168 |       |            |          |  Q1,04 | PCWP |            |
    |   4 |     UNION-ALL               |             |       |       |       |            |          |  Q1,04 | PCWP |            |
    |*  5 |      FILTER                 |             |       |       |       |            |          |  Q1,04 | PCWC |            |
    |*  6 |       HASH JOIN             |             |   660G|    50T|   224M|  3465K (99)| 13:28:42 |  Q1,04 | PCWP |            |
    |   7 |        PX JOIN FILTER CREATE| :BF0000     |  5477K|   386M|       | 22861   (2)| 00:05:21 |  Q1,04 | PCWP |            |
    |   8 |         PX RECEIVE          |             |  5477K|   386M|       | 22861   (2)| 00:05:21 |  Q1,04 | PCWP |            |
    |   9 |          PX SEND HASH       | :TQ10000    |  5477K|   386M|       | 22861   (2)| 00:05:21 |  Q1,00 | P->P | HASH       |
    |  10 |           PX BLOCK ITERATOR |             |  5477K|   386M|       | 22861   (2)| 00:05:21 |  Q1,00 | PCWC |            |
    |* 11 |            TABLE ACCESS FULL| TEST_TBL_01 |  5477K|   386M|       | 22861   (2)| 00:05:21 |  Q1,00 | PCWP |            |
    |  12 |        PX RECEIVE           |             |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,04 | PCWP |            |
    |  13 |         PX SEND HASH        | :TQ10001    |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,01 | P->P | HASH       |
    |  14 |          PX JOIN FILTER USE | :BF0000     |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,01 | PCWP |            |
    |  15 |           PX BLOCK ITERATOR |             |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,01 | PCWC |            |
    |  16 |            TABLE ACCESS FULL| TEST_TBL_02 |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,01 | PCWP |            |
    |* 17 |      FILTER                 |             |       |       |       |            |          |  Q1,04 | PCWC |            |
    |* 18 |       HASH JOIN             |             |  2640G|   201T|   233M|    13M(100)| 53:15:52 |  Q1,04 | PCWP |            |
    |  19 |        PX RECEIVE           |             |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,04 | PCWP |            |
    |  20 |         PX SEND HASH        | :TQ10002    |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,02 | P->P | HASH       |
    |  21 |          PX BLOCK ITERATOR  |             |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,02 | PCWC |            |
    |  22 |           TABLE ACCESS FULL | TEST_TBL_02 |    22M|   212M|       | 22679   (1)| 00:05:18 |  Q1,02 | PCWP |            |
    |  23 |        PX RECEIVE           |             |    21M|  1546M|       | 22924   (2)| 00:05:21 |  Q1,04 | PCWP |            |
    |  24 |         PX SEND HASH        | :TQ10003    |    21M|  1546M|       | 22924   (2)| 00:05:21 |  Q1,03 | P->P | HASH       |
    |  25 |          PX BLOCK ITERATOR  |             |    21M|  1546M|       | 22924   (2)| 00:05:21 |  Q1,03 | PCWC |            |
    |  26 |           TABLE ACCESS FULL | TEST_TBL_01 |    21M|  1546M|       | 22924   (2)| 00:05:21 |  Q1,03 | PCWP |            |
    Predicate Information (identified by operation id):
       5 - filter(NULL IS NOT NULL)
       6 - access("T1"."OBJECT_NAME"="T2"."OBJECT_NAME")
      11 - filter("T1"."OBJECT_TYPE"='TABLE')
      17 - filter(NULL IS NOT NULL)
      18 - access("T1"."OBJECT_NAME"="T2"."OBJECT_NAME")
    Statistics
           1617  recursive calls
              3  db block gets
         488929  consistent gets
         493407  physical reads
            636  redo size
            567  bytes sent via SQL*Net to client
            232  bytes received via SQL*Net from client
              1  SQL*Net roundtrips to/from client
              6  sorts (memory)
              0  sorts (disk)
              0  rows processedHowever single table with UNION ALL and PQ works..
    *NO Joins (i.e. Single Table with PQ )  , Issue does not show-up.*
    _*SERIAL PLAN with one Table*_
    SQL> select
      2    /* parallel(t1,2 )   */
      3    t1.*
      4    from test_tbl_01 t1
      5    where 'Y' = 'N'
      6    and  t1.object_type = 'TABLE'
      7    union all
      8    select
      9    /* parallel(t1,2 )   */
    10    t1.*
    11    from test_tbl_01 t1
    12    where 'Y' = 'N'
    13  /
    no rows selected
    Elapsed: 00:00:00.01
    Execution Plan
    Plan hash value: 2870519681
    | Id  | Operation           | Name        | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT    |             |     2 |   148 |     0   (0)|          |
    |   1 |  UNION-ALL          |             |       |       |            |          |
    |*  2 |   FILTER            |             |       |       |            |          |
    |*  3 |    TABLE ACCESS FULL| TEST_TBL_01 |  5477K|   386M| 41261   (2)| 00:09:38 |
    |*  4 |   FILTER            |             |       |       |            |          |
    |   5 |    TABLE ACCESS FULL| TEST_TBL_01 |    21M|  1546M| 41373   (3)| 00:09:40 |
    Predicate Information (identified by operation id):
       2 - filter(NULL IS NOT NULL)
       3 - filter("T1"."OBJECT_TYPE"='TABLE')
       4 - filter(NULL IS NOT NULL)
    Statistics
              0  recursive calls
              0  db block gets
              0  consistent gets
              0  physical reads
              0  redo size
            567  bytes sent via SQL*Net to client
            232  bytes received via SQL*Net from client
              1  SQL*Net roundtrips to/from client
              0  sorts (memory)
              0  sorts (disk)
              0  rows processed
    _*PARALLEL PLAN with one Table*_
    SQL> select
      2    /*+ parallel(t1,2 )      */
      3    t1.*
      4    from test_tbl_01 t1
      5    where 'Y' = 'N'
      6    and  t1.object_type = 'TABLE'
      7    union all
      8    select
      9    /*+ parallel(t1,2 )      */
    10    t1.*
    11    from test_tbl_01 t1
    12    where 'Y' = 'N'
    13  /
    no rows selected
    Elapsed: 00:00:00.09
    Execution Plan
    Plan hash value: 3114025180
    | Id  | Operation              | Name        | Rows  | Bytes | Cost (%CPU)| Time     |    TQ  |IN-OUT| PQ Distrib |
    |   0 | SELECT STATEMENT       |             |     2 |   148 |     0   (0)|          |        |      |            |
    |   1 |  PX COORDINATOR        |             |       |       |            |          |        |      |            |
    |   2 |   PX SEND QC (RANDOM)  | :TQ10000    |       |       |            |          |  Q1,00 | P->S | QC (RAND)  |
    |   3 |    UNION-ALL           |             |       |       |            |          |  Q1,00 | PCWP |            |
    |*  4 |     FILTER             |             |       |       |            |          |  Q1,00 | PCWC |            |
    |   5 |      PX BLOCK ITERATOR |             |  5477K|   386M| 22861   (2)| 00:05:21 |  Q1,00 | PCWC |            |
    |*  6 |       TABLE ACCESS FULL| TEST_TBL_01 |  5477K|   386M| 22861   (2)| 00:05:21 |  Q1,00 | PCWP |            |
    |*  7 |     FILTER             |             |       |       |            |          |  Q1,00 | PCWC |            |
    |   8 |      PX BLOCK ITERATOR |             |    21M|  1546M| 22924   (2)| 00:05:21 |  Q1,00 | PCWC |            |
    |   9 |       TABLE ACCESS FULL| TEST_TBL_01 |    21M|  1546M| 22924   (2)| 00:05:21 |  Q1,00 | PCWP |            |
    Predicate Information (identified by operation id):
       4 - filter(NULL IS NOT NULL)
       6 - filter("T1"."OBJECT_TYPE"='TABLE')
       7 - filter(NULL IS NOT NULL)
    Statistics
             28  recursive calls
              3  db block gets
              7  consistent gets
              0  physical reads
            628  redo size
            567  bytes sent via SQL*Net to client
            232  bytes received via SQL*Net from client
              1  SQL*Net roundtrips to/from client
              2  sorts (memory)
              0  sorts (disk)
              0  rows processed

    The same behvious appears in 11.1.0.6, and you don't need such a large data set to prove the point. The paralllel distribution may change to a broadcast with a smaller data set, but I demonstrated the effect when my two tables simply selected 30,000 rows each from all_objects.
    I think you should pass this to Oracle Corp. as a bug - using the smaller data set.
    The problem seems to be the way that Oracle combines multiple lines of a plan into groups of operations (as in PCWC, PCWC, PCWP). It looks like this particularly example has managed to fold the FILTER into a group in such a way that Oracle has lost track of the fact that it is a 'pre-emptng - i.e. always false' filter rather than an ordinary data filter; consequently the filter doesn't apply until after the hash join starts running.
    In my example (which did a broadcast distribution) I could see that Oracle read the entire first table, then started to read the second table, but stopped after one row of the second table, because my plan allowed the join and filter to be applied immediately after the first row from the second table. And I think Oracle decided that the filter was alway going to be false at that moment - so stopped running the second tablescan. You've used a hash/hash distribriution, which has resulted in both scans completing because the slaves in each layer don't talk to each other.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    "Science is more than a body of knowledge; it is a way of thinking"
    Carl Sagan

  • I have some dts cd's that I want to import into itunes on a windows pc and then play back on my apple tv 3 (i think its 3, about 9 months old) to listen to through my onkyo amp via hdmi as dts, when I play back I just get white noise.

    I have some dts cd's that I want to import into itunes on a windows pc and then play back on my apple tv 3 (i think its 3, about 9 months old) to listen to through my onkyo amp via hdmi as dts, when I play back I just get white noise. I have seen lots of old posts which go on about converting files but am wondering wherher at the current time with the latest versions of itunes and apple tv whether this should by now be a virtual play out of the box solution.
    I have tried connecting my atv to the amp by optical cable but that didnt produce anything better and am assuming that a hdmi cable should be able to reproduce everything an optical cable does.
    Is this something I should be able to achieve without having to recode the files?
    Tracks are imported to itunes as apple lossless, have tried setting the atv audio settings to auto and 16 bit, neither works yet I was told some time ago by an apple rep that the atv shoud be able to handle any type of audio as it just passes it to the amp to decode. I am not stupid but not that technical when it comes to audio and bitrates etc so any explaination would be appreciated.
    Thanks in advance for any help

    Willy, the procdedure you describe works perfectly for AppleTV G1. In fact, I frequently play DTS music discs ripped to my iTunes library as Apple Lossless (ALAC) and synched to the ATV. What happens is that, as far as iTunes and the ATV are concerned, a conventional stereo 16/44 audio file is being played; however my reciever detects the DTS data stream, and instead of outputting static, it decodes the 5.1 channels correctly. This process parallels the way DTS discs were designed to be played back: a conventional CD player is connceted, via digital out, to a DTS-aware reciever/pre-amp and the two channels of static are recognized as 6 channels of DTS-encoded audio.
    However this process will only work if the player (CD, ATV, PC, etc.) sends a "bit-perfect" 16 bit/44.1 khz data stream to the decoding device. In the case of ATV G2 and G3, as has been discussed elsewhere, the unit resmaples all audio - including 16/44 PCM (Redbook CD)  - to 16/48. While this is not a big deal for conventional audio CDs, MP3s, and AAC files, it mangles the fragile DTS datastream and renders it undecodable. The resulting static that you hear is just like the static you would hear if you played a DTS CD on a non-DTS capable system.
    Sadly, I'm not aware of any easy way around this. You could use a program like Foobar (and the DTS plugin) to convert your DTS CDs to 6-ch .wav or .flac file. From there you could transcode the file to AC-3 and then mux it into a video container that ATV supports. You may need add a dummy video track, for iTunes/AppleTV to be ok with the file, though I'm not sure. This may be a lot of work.
    Sorry for the bad news. I wish that the iTunes/ATV ecosystem had better multi-channel audio support.
    Steven

Maybe you are looking for