WM TABLE PROCESS FLOW

hi friends,
Can anybody tell me the WM Table process Flow???
thanks

Hi,
As far as tables concerned:
http://www.sap-topjobs.com/SAPTABLEREF%5B1%5D.doc
Some important t-codes:
reports
LX02 (~MB52 in IM)
LS24 (~MMBE)
LT22 (~MB51 in IM)
WH activity monitor
LL01
TR
LB10 - list of TRs
LB01/02/03 (create/change/display)
TO
create:
LT01 (LT21 display, you cannot change in standard system)
LT10
(from LB10 you can also process TRs)
LT03 (for delivery)
cancel
LT15
confirm
LT22 (e.g.)
Post change
LU04 - list
LU01/02/03
LQ01/02/03
Inventory
LI01N - create document & activate (LI02N-create, LI03N-change, LX22-overview)
LI11N - enter count result
LI20 - clear difference in WM
LI21 - clear difference in IM
Reward points if helpful,
Regards,
Archit

Similar Messages

  • Locking source tables during process flow

    Hi All,
    I have a process flow which calls multiple mappings and stored procedures. I pull data from source located in different database and insert data into current database. Since, this migration is delta migration, I want to make sure that during migration, no user should add data into source till migration is successfully completed.
    I had called a procedure which locks individual tables, but as soon as next mapping is completed the lock gets released, since each mappings commits data. I want that the lock should be active till all mappings gets completed.
    Regards,
    Danish

    Commit will release all locks. Therefor, no matter how you lock the table, the lock will release when you commit the map. An alternative might be to capture a frozen image of the data into a local GTT created with the on commit preserve rows option.

  • Cannot deploy process flow package specification from file

    Hi,
    we are using owb 11gr2 on linux. We successfully managed to deploy mappings and tables into a specification file, and then deploy from that file into the target database (RAC). But when we try to deploy a process flow package we run into an error:
    A process flow package can be deployed to a specification file. But a deployment from that specification file fails with the error: " OMB05623: Cannot deploy specification from file. Exception follows. nulll"
    ### SETUP
    set CONN_DESIGN "xxx"
    set CONN_TARGET_DESIGN "yyy"
    set CONN_TARGET_RT "xxx'"
    set DIR "D:\\tmp"
    set RELEASE "V0_2"
    set file "$DIR\\$RELEASE"
    set plan "FILE_DEPLOY_$RELEASE"
    ### Deployment via control center: works!
    OMBCONNECT $CONN_DESIGN
    OMBCONNECT CONTROL_CENTER $CONN_TARGET_RT
    OMBCC '/KLINGEL_DWH'
    OMBCREATE TRANSIENT DEPLOYMENT_ACTION_PLAN '$plan'
    OMBALTER DEPLOYMENT_ACTION_PLAN '$plan' ADD ACTION 'PF_TEST' SET PROPERTIES (OPERATION) VALUES ('REPLACE') SET REFERENCE PROCESS_FLOW_PACKAGE '/KLINGEL_DWH/WF_DWH/PF_TEST'
    OMBDEPLOY DEPLOYMENT_ACTION_PLAN '$plan'
    OMBDISCONNECT CONTROL_CENTER
    OMBCOMMIT
    OMBDISCONNECT
    ### Deployment via specificatin file: fails!
    OMBCONNECT $CONN_DESIGN
    OMBCONNECT CONTROL_CENTER $CONN_DESIGN
    OMBCC '/KLINGEL_DWH'
    OMBCREATE TRANSIENT DEPLOYMENT_ACTION_PLAN '$plan'
    OMBALTER DEPLOYMENT_ACTION_PLAN '$plan' ADD ACTION 'PF_TEST' SET PROPERTIES (OPERATION) VALUES ('REPLACE') SET REFERENCE PROCESS_FLOW_PACKAGE '/KLINGEL_DWH/WF_DWH/PF_TEST'
    OMBDEPLOY DEPLOYMENT_ACTION_PLAN '$plan' AS SPECIFICATION TO '$file.xml'
    OMBDROP DEPLOYMENT_ACTION_PLAN '$plan'
    OMBDISCONNECT CONTROL_CENTER
    OMBCOMMIT
    OMBDISCONNECT
    OMBCONNECT CONTROL_CENTER $CONN_TARGET_RT
    OMBDEPLOY SPECIFICATION FROM '$file.xml'
    ### OMB05623: Deployment von Spezifikation aus Datei nicht möglich. Exception folgt. null
    ###OMB05623: Cannot deploy specification from file. Exception follows. nullIf we use the operation DROP it works. But REPLACE and CREATE do not work.
    We get the error regardless whether the code is executed as an expert or on the command line at a windows client.
    Does anyone knows a workaround?
    Deployment via a specification file is a must due to license and architecture requirements.
    Thanks,
    Carsten.

    Its woked fine for me for OWB 10.2.0.4 .... some steps are missing from code like
    when you are connecting again it should be
    OMBDISCONNECT
    OMBCONNECT CONTROL_CENTER $CONN_TARGET_RT
    OMBDISCONNECT
    OMBCC '/KLINGEL_DWH'
    OMBCONNECT CONTROL_CENTER $CONN_TARGET_RTI think you can try the same on different system .
    Cheers
    Nawneet

  • To check whether a process flow is valid or not

    Hi,
    I want to check whether a OWB process flow is valid or not inside a PL/SQL block. Can anyone let me know about any repository table or view where I can check the process flow status.
    Regards,
    Poulose

    Use display schedule job option from the context menu of the start variant of the process chain.
    When u chick this.. u will get the available job on this start process.. There u can see whether the process chain in schedule mode or not?
    If u can see multiple schedule then u can delete the older ones which are not needed and then u can maintain only one schedule job.
    If u click twice.. then there are chances that u might end up with the 2 schedule job with the same time frame.. .. so check the same with the steps mentioned..... if u find two jobs then u need to delete one.. if not there might be chance that ur PC might get trigger twice.. and one trigger will gets failed. and the other one will gets executed..
    Thanks
    I hope this is clear to u..
    Assign points if this helps

  • Experts VS Process Flows on OWB 10g Release 2 (Paris)

    I was looking at the the new features list for OWB 10g Release 2 (Paris), and I notices the new feature called Experts. Does anyone have any insight into when an Expert would be used instead of a Process Flow? It doesn't look like Process Flows are going away, so I was trying to get a grasp of the use for Experts.
    From Here:
    http://www.oracle.com/technology/products/warehouse/htdocs/owb10gr2%20new%20features.htm
    Experts. Experts are solutions that enable advanced users to design solutions that simplify routine or complex tasks that end users perform in Warehouse Builder. You can design these solutions in the Expert Editor which resembles the Process Flow Editor and shares many of the same commands and navigational tools.

    Hi,
    Just to start with the Experts capabilities, they can be used for designing schema objects.......I tried with an expert for sampling the source files, creating External Tables on top of it....validating and deploying..........itz very interesting to work with.........I am unable to see the practical use of this for end users.....
    where as we all know, the PF can only be for executing the deployed objects.....
    Mahesh

  • Error while deploying process flow

    hi All,
    i am getting below error while deploying 1 process flow.
    Description : Runtime User : OWBPC Started : 2011-06-02 19:18:06.0
    Name Action Status Log
    None Error ORA-00942: table or view does not exist
    None Error RPE-02053: The Oracle Workflow version cannot be retrieved. This may have been caused by an incomplete or incorrect installation of Oracle Workflow or the schema may not contain an Oracle Workflow repository. Please check the location details, reinstall Oracle Workflow, if necessary and try again
    i added only 1 mapping in process flow and mapping containing both tables in taret schema.
    kindly tell me how to fix it.?
    rgds,
    Pc

    hi..thnx for replying..
    i have already installed owf manager from below location in my pc..
    D:\app\pc\product\11.1.0\db_1\owb\wf\install
    and installation was successful..then too I am getting same error while deploying process flow.. im using oracle 11g.. direct me now plz..
    Edited by: PC on Jun 3, 2011 3:48 PM

  • Reg: SQL select Query in BPEL process flow

    <p>
    Hi,
    I am suppose to execute a SQL select query (in BPEL Process flow) as mention below in JDeveloper using Database adapter.
    </p>
    <p>
    SELECT LENGTH, WIDTH, HEIGHT, WEIGHT,
    </p>
    <p>
    LENGTH*WIDTH* HEIGHT AS ITEM_CUBE
    </p>
    <p>
    FROM CUBE
    </p>
    <p>
    WHERE ITEM= &lt;xyz&gt;
    </p>
    <p>
    AND OBJECT= (SELECT CASE_NAME FROM CUBE_SUPPLIER WHERE ITEM=&lt;xyz&gt; AND SUPP_IND = &lsquo;Y')
    <strong>Now my question is:
    1.</strong> What does this "*" refer to in the query and how can I retrieve the value of LENGTH*WIDTH* HEIGHT from the query where LENGTH,WIDTH and HEIGHT are the individual field in the table.
    2.What does this " AS" refer to? If " ITEM_CUBE " is the alies for the table name "ITEM" to retrieve the value, then query shoud be evaluated as
    </p>
    <p>
    SELECT LENGTH, WIDTH, HEIGHT, WEIGHT,
    </p>
    <p>
    LENGTH*WIDTH* HEIGHT AS ITEM_CUBE
    </p>
    <p>
    FROM CUBE
    </p>
    <p>
    WHERE <strong>ITEM_CUBE.ITEM</strong>= &lt;xyz&gt;
    </p>
    <p>
    AND <strong>ITEM_CUBE.OBJECT</strong>= (SELECT CASE_NAME FROM CUBE_SUPPLIER WHERE ITEM=&lt;xyz&gt; AND SUPP_IND = &lsquo;Y')
    Is my assumption correct?
    Please suggest asap.
    Thanks...
    </p>
    <p>
    </p>

    Hi
    Thank for your reply!
    I have a nested select query which performs on two different table as shown below:
    <p>
    SELECT LENGTH, WIDTH, HEIGHT, WEIGHT,
    </p>
    <p>
    LENGTH*WIDTH* HEIGHT AS ITEM_CUBE
    </p>
    <p>
    FROM CUBE
    </p>
    <p>
    WHERE ITEM= &lt;abc&gt;
    </p>
    <p>
    AND OBJECT= (SELECT NAME FROM SUPPLIER WHERE ITEM=&lt;Item&gt; AND SUPP_IND = &lsquo;Y')
    I am using DB adapter of Oracle JDeveloper in BPEL process flow, where I can able to select only one master table in DB adapter say SUPPLIER and its attributes at a time.But as per my requirment I need to select both the table (CUBE and SUPPLIER) in a single adapter to execute my query.
    It can be achievable by using two DB adapter , One to execute the nested query and another to execute the main qyery considering value of nested query as a parameter.But I want to achieve it by using a single one.
    Am I correct with my concept?
    Please suggest how to get it ?
    </p>
    Edited by: user10259700 on Oct 23, 2008 12:17 AM

  • Process Flow and Workflow

    Hi experts, does a process flow cannot be used if we dont have oracle workflow?
    If yes, any of you know the alternative ways to run a schedule mapping sequentially?
    Currently i am running schedule from OWB (Control center manager). And i am not satisfied because the lack of Log / information regarding the job execution. Is there a way to get more detail schedule job logs???
    Thank you in advance guys...

    Oracle is pretty smart when it comes to those things, Im sure the table is not lock because one session is DML:ing it and one is querying it. You can get a deadlock when the sessions are trying to DML for the same row, because update/insert/delete is based on row lock, not table lock.
    What you can do to test these kinds of long operation mappings is to generete the intermediate code from the map - open the map editor and click on genereate - choose intermediate from the dropdown list and click on your target table. Copy that code and run it from i.e toad, do the same thing instantly for the other map and you will found out if your sessions are locking each other.
    Because the owb map is bascly configured in a way that it most of it is unnecessery when your are not running it thrue the repository, like I said. Try to run it they way it is - else create a new one by the code generated in the intermediate mapping.
    Cheers

  • Help with scheduling Process Flows in Workflow in 10gR2

    Dear All
    I am after some help with the scheduling of process flows (PF) in Workflow in OWB 10gR2. I am trying to set up some PF’s to handle the refresh of some staging tables from various source systems. I have created a separate Process Flow Module for each source systems, mainly just to keep them separate and organised. I have a number of mappings which all run fine if execute manually or linked directly to a schedule/job. The problem I am encountering is when I try to run process flows from schedules. I have created process flows ok and have run them manually and they complete fine so I know that the content of the PF is ok. After linking and deploying the jobs I can never get all the process flows to run from the various schedules. What appears to happen is that the first PF works ok and any other PF that are within the same Process Flow Module/packages also runs ok even if its running off as different schedules. However PF’s under the other Process Flow Module fail with the following error
    CC_DAILY_0400
    Description :
    Runtime User : OWBRT_USER
    Started : 2006-08-31 04:00:00.0
    Status      Log
    Error      ORA-20002: 3114: Activity 'PACK_1/CC_DAILY_0400' is not a process.
    ORA-06512: at "OWF_MGR.WF_ENGINE", line 3920
    ORA-06512: at line 1
    Following this error the PF will not even run manually! If I then stop the schedule and either drop or replace this failed PF thus redeploying, the PF then runs fine manually and if the restart the schedule it runs ok the next evening. My problem is that this then appears to impact on the other PF’s which all though have not been touched and ran ok the previous evening they then fail the following evening with the same error
    WS_DAILY_2400
    Description :
    Runtime User : OWBRT_USER
    Started : 2006-09-01 00:00:01.0
    Status      Log
    Error      ORA-20002: 3114: Activity 'PACK_1/WS_DAILY_2400' is not a process.
    ORA-06512: at "OWF_MGR.WF_ENGINE", line 3920
    ORA-06512: at line 1
    ORA-20002: 3114: Activity is not a process.
    I basically can not get both sets to run even though they are on separate modules and separate schedules.Has anyone any idea as to what could be wrong or if I am setting something up in a strange way which would cause these symptoms.
    All help or advice greatly appreciated
    Regards Kevin

    Dear All
    I am after some help with the scheduling of process flows (PF) in Workflow in OWB 10gR2. I am trying to set up some PF’s to handle the refresh of some staging tables from various source systems. I have created a separate Process Flow Module for each source systems, mainly just to keep them separate and organised. I have a number of mappings which all run fine if execute manually or linked directly to a schedule/job. The problem I am encountering is when I try to run process flows from schedules. I have created process flows ok and have run them manually and they complete fine so I know that the content of the PF is ok. After linking and deploying the jobs I can never get all the process flows to run from the various schedules. What appears to happen is that the first PF works ok and any other PF that are within the same Process Flow Module/packages also runs ok even if its running off as different schedules. However PF’s under the other Process Flow Module fail with the following error
    CC_DAILY_0400
    Description :
    Runtime User : OWBRT_USER
    Started : 2006-08-31 04:00:00.0
    Status      Log
    Error      ORA-20002: 3114: Activity 'PACK_1/CC_DAILY_0400' is not a process.
    ORA-06512: at "OWF_MGR.WF_ENGINE", line 3920
    ORA-06512: at line 1
    Following this error the PF will not even run manually! If I then stop the schedule and either drop or replace this failed PF thus redeploying, the PF then runs fine manually and if the restart the schedule it runs ok the next evening. My problem is that this then appears to impact on the other PF’s which all though have not been touched and ran ok the previous evening they then fail the following evening with the same error
    WS_DAILY_2400
    Description :
    Runtime User : OWBRT_USER
    Started : 2006-09-01 00:00:01.0
    Status      Log
    Error      ORA-20002: 3114: Activity 'PACK_1/WS_DAILY_2400' is not a process.
    ORA-06512: at "OWF_MGR.WF_ENGINE", line 3920
    ORA-06512: at line 1
    ORA-20002: 3114: Activity is not a process.
    I basically can not get both sets to run even though they are on separate modules and separate schedules.Has anyone any idea as to what could be wrong or if I am setting something up in a strange way which would cause these symptoms.
    All help or advice greatly appreciated
    Regards Kevin

  • Process Flow hangs

    I'm executing a process flow with several activities.
    After each activity I check for status and error out if the activity fails. If the activity succeeds I go on to the next activity.
    When I execute a map within the process flow, the process flow executes successfully only if the map is the last activity. The map functions to output data to a file.
    The Process Flow hangs if after the execution of the map
    a stored procedure is called. The stored procedure dynamically creates a DBLINK and updates a table.
    When I cancel the execution of the process flow I see the map has successfully outputted all
    records. But why hasn't it gotten to the stored procedure.
    Each activity on its own works fine but in combination with each other they hang the process flow.

    Mahesh,
    I am assuming that RAB stands for Runtime Audit Browser.
    The Runtime Audit Browser tells me how many records were
    rejected but not which ones and why.
    The debugger does not give adequate information in fact it does not issue a "break" where it should.
    Are there any tools OWB provides for "catching", or
    auditting data that cannot for whatever reason be extracted to a file.
    I see that they have a mechanism for capturing problems with data coming from a file into a table. I don't see
    where there is an explanation of error handling going from a table into a file.
    Thanks!
    Marion

  • Process Flow Error

    I ran into a problem with one of my process flows, within my staging area. I have not had this problem before. My process flow looks as follows:
    --- > Map2
    --- > Map3
    --- > Map4
    Map1 --- > Fork --- > Map5 --- > AND --- > Map 9 --- > Map10 --- > Map11 …
    --- > Map6
    --- > Map7
    --- > Map8
    Mappings 1 through 8 ran successfully and populated their staging tables. The process flow showed success of mappings 9 through 16 (the last one) and successful completion of the process flow. However Mapping 9 was still extracting data and had not yet loaded its staging table. I was watching the execution in OEM. Mappings 10 through 16 did not execute after completion of Mapping 9.
    This is a weekly large data load into our data warehouse so I will not be able to run this process flow again until this weekend. Map 9 was running with a degree of parallelism of 16, if it matters, and did not fail.
    Has anyone run into this issue before?
    Thanks.
    Scott

    user10483084 wrote:
    I ran into a problem with one of my process flows, within my staging area. I have not had this problem before. My process flow looks as follows:
    --- > Map2
    --- > Map3
    --- > Map4
    Map1 --- > Fork --- > Map5 --- > AND --- > Map 9 --- > Map10 --- > Map11 …
    --- > Map6
    --- > Map7
    --- > Map8
    Mappings 1 through 8 ran successfully and populated their staging tables. The process flow showed success of mappings 9 through 16 (the last one) and successful completion of the process flow. However Mapping 9 was still extracting data and had not yet loaded its staging table. I was watching the execution in OEM. Mappings 10 through 16 did not execute after completion of Mapping 9.
    This is a weekly large data load into our data warehouse so I will not be able to run this process flow again until this weekend. Map 9 was running with a degree of parallelism of 16, if it matters, and did not fail.
    Has anyone run into this issue before?
    Thanks.
    Scottperhaps there's an error at MAPPING before MAPPING10, but it's not detected
    I suggest you to add 2 transitionto COMPLETE ERROR and COMPLETE WARNING
    from MAP1; AND; MAP9, MAP10, MAP11, ...
    and fill the transition; enumerate ERROR for ERROR, enumerate WARNING for WARNING,
    and the other line; enumerate SUCCESS
    --- > Map2
    --- > Map3
    --- > Map4
    Map1 --- > Fork --- > Map5 --- > AND --- > Map 9 --- > Map10 --- > Map11 …
    --- > Map6
    --- > Map7
    --- > Map8
    so you can know what MAPPING that doesn't success, and you can test it manually with control center, to know the error message
    Thanks

  • Process Flow Validation Warning.

    Hi All,
    OWB Config Details is as follows:
    Oracle 9i Warehouse Builder Client: 9.2.0.2.8
    Oracle 9i Warehouse Builder Repository: 9.2.0.2.0
    1> Following Warning is displayed when Process Flow is validated in Process Editor in OWB.
    VLD-10005: The following path in the process flow is a loop and may cause the process to loop infinitely,
    START->EXT_SHELL_SCRIPT->M_STG_GRADE_OF_SERVICE->EMAIL_ERROR->UPDATE_MAP_PROCESS_ERROR->
    2> A mapping is developed in OWB with some default configuration parameters e.g. In a mapping involves Source Text File as Source and Staging Table as Target.
    Readsize and Bindsize config parameters are set to 5000 respectively. I included the same mapping in process flow as mapping. The above mentioned Config Parameters are modified to 5242880. I committed the changes & synchornized in the process flow editor. I was expecting the modified config parameter will be displayed in the process flow editor. However I found that it still displays old config parameters and not modified config parameters. Is this expected or Did i miss some steps ?
    Can someone please let me know more on these issues?
    Regards,
    Vidyanand

    Please go the OWB forum for these questions:
    Warehouse Builder
    Thanks, Mark

  • Process Flow Warning in OWB.

    Hi All,
    OWB Config Details is as follows:
    Oracle 9i Warehouse Builder Client: 9.2.0.2.8
    Oracle 9i Warehouse Builder Repository: 9.2.0.2.0
    1> Following Warning is displayed when Process Flow is validated in Process Editor in OWB.
    VLD-10005: The following path in the process flow is a loop and may cause the process to loop infinitely,
    START->EXT_SHELL_SCRIPT->M_STG_GRADE_OF_SERVICE->EMAIL_ERROR->UPDATE_MAP_PROCESS_ERROR->
    2> A mapping is developed in OWB with some default configuration parameters e.g. In a mapping involves Source Text File as Source and Staging Table as Target.
    Readsize and Bindsize config parameters are set to 5000 respectively. I included the same mapping in process flow as mapping. The above mentioned Config Parameters are modified to 5242880. I committed the changes & synchornized in the process flow editor. I was expecting the modified config parameter will be displayed in the process flow editor. However I found that it still displays old config parameters and not modified config parameters. Is this expected or Did i miss some steps ?
    Can someone please let me know more on these issues?
    Regards,
    Vidyanand

    Re your first question, you are probably hitting a bug (3035080) that makes OWB generate false warnings. Please ignore the warnings and try to deploy and run the process flow.
    As for your second question - I cannot reproduce this on 9.2.0.2.8. The configuration properties are updated corectly in the SQLLOADER map and I get the same values in the process flow. By the way, why are you not using external tables instead of the SQLLoader maps?
    Regards:
    Igor

  • Regarding Performance Tunning of Mapping & Process Flow

    Hi,
    i have around 60-70 GB data in Target database.
    i need do improve the performance of Mapping and Process flow.
    i have used lookup transformation in mapping.
    Plz give me some tips for improving performance of Process.
    Thanks,

    Please go through a Performance Tuning Book for Oracle 10gR2.
    Most importantly remember that in Oracle 10g, your performace of mappings can be increased manifold by following these steps:
    1. Do not design a mapping wherein you load a table and then select from that table to load another table and so on. This is a bad design.
    2. Keep mappings as simple as possible. In other words if a mapping is complicated in terms of joins or other operators then split the mapping into more than one parts.
    3. Also check that all your source tables should be analyzed using DBMS_STATS. Ensuring this one single step can make your work very easy.
    4. Put indexes where you find your predicate has a very high selectivity. Also keep in mind the column ordering of the index.
    5. Use Set Based operation, since it is always a good idea to achieve the result by running one single query rather than a loop and multiple inserts.
    6. Use APPEND PARALLEL hint while loading the target tables. This will not generate wny redo and save time
    7. Please have a recheck while usng some performance intensive operators like UNION, DISTINCT and AGGREGATION
    8. When using a sequence operator to load a large table, check that the sequence should be cached with some values.
    9. When loading large data HASH JOINS are the most appropriate more often than not. So you can use USE_HASH as the hint for selecting from large tables.
    10. Filter out as much unrequired data from a table as soon as possible in the mapping before doing multiple joins.
    I am sure there are many more ... the above is just a random list that I could remember now. Please go through Oracle Performance Tuning Guide, Tom Kyte's Oracle Expert One on One. Knowledge of performance tuning will grow with experience. I am also learning each day !!!
    Regards
    AP

  • Mapping audit_id in process flow with parameter

    Hello!
    I have a following request.
    I would like to get AUDIT_EXECUTION_ID for certain Mapping in process flow and pass this ID to next mapping (all this with help of parameters I guess??) to get some data from AUDIT tables. How to do this?
    OWB version 10.2.0.1
    Thank you Gorazd
    Edited by: gor on May 9, 2011 1:40 PM

    Hi,
    It is a bit complicated and technical.
    Will give you a hint.
    You need to install run_my_owb_stuff as a function, grant access, and create a public synonym.
    Then within you have to add capturing of the execution_id as an additional custom parameter. (see below).
    Once done. Add a Start parameter with same name to each Process Flow, and Bind it to any activity that you want to use the parameter with.
    -- Override Parameters
    dbms_output.put_line('Stage 3: Overriding Parameters');
    -- override_system_input_params(l_audit_execution_id, p_system_params);
    if p_task_type='PROCESS' then
    -- GRI addition
         l_custom_params := p_custom_params || ', p_execution_id=' || l_audit_execution_id;
    override_custom_input_params(l_audit_execution_id, l_custom_params);
    else
         override_custom_input_params(l_audit_execution_id, p_custom_params);
         end if;
    -- override_custom_input_params(l_audit_execution_id, p_custom_params);
    - Jojo

Maybe you are looking for