Mapping Dependency Execution in OWB.

Hi All,
Oracle 9i Warehouse Builder Client : 9.2.0.4.0
Oracle 9i Warehouse Builder Repository: 9.2.0.2.0
I created 2 mappings / process flows , mapping 1 for a project proj1 pointing to database 1 & mappng 2 for a project proj2 pointing to database 2.
I need to schedule these mappings / process flows as follows in OEM.
Mapping2 should get executed only on successful execution of the mapping 1.
How can i schedule this in OEM?
Many thanks in Advance.
Regards,
Vidyanand

Hi Mahesh,
Thanks for your quick reply.
However I believe since the process flows PF1 & PF2 belongs to different projects & databases. I will not able to create process flow PF3 which consist of 2 process flows from 2 different projects & databases.
To make it clear let us consider following scenario.
In my data ware house application there are 2 databaes, db1 for staging & db2 for target.
I developed two separate mappings M1 in project proj1 which points to db1 & M2 in project proj2 which points to db2. Also i created 2 separate process flows Pf1 & Pf2. I will not able to create process flow Pf3 which consist of Pf1 & Pf2.
Is my understanding is correct? Can you please confirm the same?
Many thanks in Advance.
Regards,
Vidyanand

Similar Messages

  • What are the best approaches for mapping re-start in OWB?

    What are the best approaches for mapping re-start in OWB?
    We are using OWB repository 10.2.0.1.0 and OWB client 10.2.0.1.31. The Oracle version is 10 G (10.2.0.3.0). OWB is installed on Linux.
    We have number of mappings. We built process flows for mappings as well.
    I like to know, what are the best approches to incorportate re-start options in our process. ie a failure of mapping in process flow.
    How do we re-cycle failed rows?
    Are there any builtin features/best approaches in OWB to implement the above?
    Does runtime audit tables help us to build re-start process?
    If not, do we need to maintain our own tables (custom) to maintain such data?
    How did our forum members handled above situations?
    Any idea ?
    Thanks in advance.
    RI

    Hi RI,
    How many mappings (range) do you have in a process flows?Several hundreds (100-300 mappings).
    If we have three mappings (eg m1, m2, m3) in process flow. What will happen if m2 fails?Suppose mappings connected sequentially (m1 -> m2 -> m3). When m2 fails then processflow is suspended (transition to m3 will not be performed). You should obviate cause of error (modify mapping and redeploy, correct data, etc) and then repeat m2 mapping execution from Workflow monitor - open diagram with processflow, select mapping m2 and click button Expedite, choose option Repeat.
    In re-start, will it run m1 again and m2 son on, or will it re-start at row1 of m2?You can specify restart point. "at row1 of m2" - I don't understand what you mean (all mappings run in Set based mode, so in case of error all table updates will rollback,
    but there are several exception - for example multiple target tables in mapping without corelated commit, or error in post-mapping - you must carefully analyze results of error).
    What will happen if m3 fails?Process is suspended and you can restart execution from m3.
    By having without failover and with max.number of errors=0, you achieve re-cycle failed rows to zero (0).This settings guarantee existence only two return result of mapping - SUCCSES or ERROR.
    What is the impact, if we have large volume of data?In my opinion for large volume Set based mode is the prefered processing mode of data processing.
    With this mode you have full range enterprise features of Oracle database - parallel query, parallel DML, nologging, etc.
    Oleg

  • How to make one map dependent on another ?

    Assume you have 2 maps - M1 and M2
    How to execute M2 automatically after successful run of M1 ?
    or if I have process flows PF2 start after success of PF1 without a manual intervention
    I looked at scheduling,.You can only set timing there but don't know how long your map runs.
    Thank You

    To make map dependant of another map you can use process flow. In process flow you can call M2 only if M1 is success. If M1 is failure you may exit the process flow.
    Am not sure how it will work for process flow. May be writing a 3rd process flow and call PF1 & PF2 within it based on success status? or write some PL-Sql wraper procedure for the same???!!! give it a try.

  • Problems with pre-mapping process operator in owb 9i

    Hi,
    I was trying to use the pre-mapping process operator in owb 9i. Problem is that the manual does not specify how the inputs need to be connected to this operator.
    Following is what I went through -
    I created a mapping table operator and a mapping dimension operator and connected these two. Then i created a pre-mapping process operator selecting the LTRIM function. Further I connected one of the table attributes to this pre-mapping operator as input and connected the output of this pre-mapping operator to the appropriate dimension operator attribute.
    On performing Validate, following error message was flashed -
    VLD-2451 : Illegal connection to pre-mapping process operator
    I am trying to learn how to use OWB 9i from the manual. So my interpretation of the use of the pre-mapping process operator may be wrong.
    In any case kindly help,
    Thanks,
    Saju

    Pre-mapping process is use to perform some operations preceding to mapping operation itself.
    For example, if your mapping is designed to incrementally append data to table for the definite time interval (witch is a parameter of the map operation) you might want to perform the table data cleanup for that period. That will allow for reload data number of time.
    In this case you have to define the procedure witch perform cleanup and than include the call to that procedure as a pre-mapping process.
    Other examples of pre- and post mapping process is disabling referential integrity before loading and re-enabling them after loading.
    Anyway, OWB documentation has clear definition for pre- and post-mapping processes.

  • Problem using pre-mapping process operator in owb 9i

    Hi,
    I was trying to use the pre-mapping process operator in owb 9i. Problem is that the manual does not specify how the inputs need to be connected to this operator. Following is what I went through -
    I created a mapping table operator and a mapping dimension operator and connected these two. Then i created a pre-mapping process operator selecting the LTRIM function. Further I connected one of the table attributes to this pre-mapping operator as input and connected the output of this pre-mapping operator to the appropriate dimension operator attribute. On performing Validate, following error message was flashed -
    VLD-2451 : Illegal connection to pre-mapping process operator
    I am trying to learn how to use OWB 9i from the manual. So my interpretation of the use of the pre-mapping process operator may be wrong.
    In any case kindly help,
    Thanks
    Saju

    Pre-mapping process is use to perform some operations preceding to mapping operation itself.
    For example, if your mapping is designed to incrementally append data to table for the definite time interval (witch is a parameter of the map operation) you might want to perform the table data cleanup for that period. That will allow for reload data number of time.
    In this case you have to define the procedure witch perform cleanup and than include the call to that procedure as a pre-mapping process.
    Other examples of pre- and post mapping process is disabling referential integrity before loading and re-enabling them after loading.
    Anyway, OWB documentation has clear definition for pre- and post-mapping processes.

  • Execution of mapping gets hanged in OWB Deployment Manager

    Description:
    Scenario - 1
    1. Source schema (rapid) to staging schema (radar01)
    We have a few mappings where we need to load data from the source to target. These mappings are successfully validated, deployed and executed.
    Both schemas reside in same database.
    Scenario - 2
    2. Staging schema (radar01) to Warehouse schema (radar02)
    We have a few mappings where we need to load from staging to warehouse.
    Please note the staging schema here is a target schema in the scenario-1 and in scenario - 2
    It is the source. The mappings where successfully validated, but a warning message is displayed while deploying.
    Warning:
    VLD-2771: System privileges may not allow extraction from source STG_ULD.
    Skipped the warning and then executed, Till 22nd the mapping was getting executed and data was getting loaded into the target schema from staging.
    From 22nd onwards, while executing these mappings, execution remains hanged forever
    Runtime Audit Browser:
    In Runtime Audit Browser shows the status for the above mappings as BUSY
    Actions Taken:
    1. Purged all previous mappings
    2. Analysed all the runtime repository tables.
    3. Purged the BUSY status mappings also by using the procedure.
    4. Grants given (SELECT ON) on all the staging schema tables to warehouse schema.
    5. Restarted the database.
    Even after the above actions also the problem has not yet been solved. Kindly suggest us how to go about.
    Environment:
    OWB 10g Client - 10.1.0.2.0
    OWB 10g Repository - 10.1.0.1.0
    Workstation OS - Windows XP
    Oracle database - 9.2.0.1.0
    Database server installed on Unix OS (Solaris)

    May be you would like to check at the database of what exactly is getting executed. Just check if there is a query running on either source or target objects. If this is a case then you would like to grab it and optimize the same.
    If this is not a case then it could be that the mapping is waiting for some locks.

  • Owb mapping parts execution order

    Hello.
    Let's imagine that we have a owb mapping with two independent parts (or threads or chains of operators, etc) inside it. Is it possible to investigate which one will be executed first?
    Sorry for my English :)

    There's a Target Load Order property of the mapping

  • The mis-mapping problem when upgrading OWB from 11.1 to 11.2.0.2.

    Hi all,
    Our OWB ETL program was developed initially on OWB 11.1 platform.
    Then after upgrading it from 11.1 to 11.2.0.2, several mappings are mismatched in the deployment PL/SQL package as below:
    CREATE OR REPLACE PACKAGE BODY "MAP_PRODUCT_LIFE_CRT" AS
    CURSOR "T_T_PRODUCT_LIFE_c" IS
    SELECT
    /* T_PRODUCT_LIFE.T_T_PRODUCT_LIFE.INOUTGRP1 */
    "T_T_PRODUCT_LIFE"."INS_TYPE" "ORGAN_ID$1",
    "T_T_PRODUCT_LIFE"."START_DATE" "HOSPITAL_TYPE$1",
    "T_T_PRODUCT_LIFE"."END_DATE" "UNIT_FLAG$1",
    "T_T_PRODUCT_LIFE"."LIFE_RATE" "INS_TYPE",
    You can see that the field mapping is wrong! We have about 100 mappings, but only some of them go wrong.
    We have tried the Metadata Loader and Repository Assistant for the upgrading, but the results are the same.
    We have released all OWB patches on 11.2.0.2, but the results are the same.
    Please help on it!

    Hi David,
    Besides the mega patch 10185523, we also have applied other tens of patches, e.g. 10225308, 11671955, 11705023, 1768535, etc.
    Now we find that the wrong field mapping just appears in the row-based part of the mapping package. So if the set-based mode works well for the mapping execution, no error will happen. That's why we just found the problems in several mappings.
    And one thing is even worse. In the patched OWB 11.2.0.2 env., we tried creating totally new mapping. The result is the same: the wrong field mapping still appears in the row-based part of the mapping package.
    So this problem seems not related to the upgrading.
    Regards.

  • Parallel execution in owb

    Hii all,
    Is it possible to execute mapping in a parallel manner in owb 11.1?If yes how do we do it?Please help
    Thanks
    Edited by: 854918 on Apr 26, 2011 9:15 PM

    Good morning Ajay,
    The checkbox appears on the 4th screen when you use the "New Transformation Wizard" to create a function (directly under checkbox "Function is deterministic").
    IMHO, whether the code is run in parallel or not depends on how the mapping is run; if parallel it would try to run the function in parallel as well (unless - I guess - you've altered the property of the function in the mapping which says it can only be run row-based).
    HTH.
    Regards, Patrick

  • Executing a mapping from sqlplus in OWB 904

    I have used the below syntax to execute OWB 902 mappings as below:
    sqlplus / <<*sql*
    Execute MAP_SOMETHING.MAIN();
    sql
    I deployed the same mapping with new version of OWB 904.. it did not work. What is new syntax to execute the mapping from sqlplus?

    From OWB 9.0.4 on, you must use the execution template to execute your mappings from sql. The template file, that has a brief doc section in its heading is:
    <OWB installation directory>/owb/rtp/sql/sqlplus_exec_template.sql
    Regards:
    Igor

  • Versioning, Deployment, Execution in OWB 9.2

    For the warehouse created using OWB versions 9.0.3, my company used a versioning tool. All the PLS generated using OWB were moved to Source control system and from there deployed to Target Schema.
    Now in 9.0.4 generated code contains location information that gets replaced by physical location only when code is deployed using OWB. So the old system of versioning and deployment doesnt work. We need to deploy using OWB only. In this case we loose all the versioning of mappings.
    Question 1 :
    Will you please suggest a solution how can we do versioning of our code now ? Versioning is very important for us. In case it is not possible then can we go back to old system of deployment and execution by doing some manual workarounds.
    Question 2 :
    Is it necessary to deploy thru OWB only because of Location factor. In that case if we design such that source and target tables are in the same then these mappings will not contain location information and we should be able to deploy them without using OWB ? What are the other disadvantages of using this approach.
    Question 3 :
    When executing mapping using the template provided by OWB, we need to provide location name and execute it from Runtime Repository owner. I guess, When I deploy outside OWB I will not be able to use this method of execution. Default parameters and custome parameters of the mapping are stored in run time repository under the location name at the time of deployment . This information is accessed by the template to execute mapping. ( is it true ??)
    We are not going to use multiple physical configuration for the same mapping, so if we deploy not using owb, can we execute directly in Target schema, without using template. How can we provide the input parameter to Main function in this case ?
    Question 4 :
    Will we be able to get Audit information if we do not use OWB's recommended way of deployment and execution.
    Vibha

    Vibha,
    Question 1:
    What you can do, is store the entire deployment script using the deploy to file option and version control that file. The format will be an ASCII-based XML file that versions exactly the same as you would version the files that OWB generated in previous releases.
    Also, you may be interested in the version control you can do inside of OWB. That is, as of 9.0.4 we enable you to create snapshots of your design, which you could compare and restore at any point in the future. In 9.0.4, we provided a beta UI on top of this feature (scripting would be the other way to go), but in 9.2 we provide a production quality UI. To enable the UI in 9.0.4, close the product, locate preference.properties in <owb home>\owb\bin\admin and change the setting for MCMUIEnabled from false to true. Please do consider migrating to OWB92...
    Question 2:
    The locations were also introduced to support multi deployment (e.g. development, test, production) with potentially different physical setups of the machines, but no design changes necessary. Also, execution uses the location concepts. Did you already read http://otn.oracle.com/products/warehouse/pdf/OWB_RuntimeChanges.pdf, which contains more information about the changes and why these are there?
    Question 3:
    To execute directly from the database, I recommend you download http://otn.oracle.com/sample_code/products/warehouse/files/run_my_owb_stuff.sql and use that as the basis for your execution. The entire architecture changed from 9.0.3 to 9.0.4... and the runtime platform now manages the execution. Via the script above you get a wrapper around the script that OWB provides out of the box (minimum version 9.0.4.10).
    Question 4:
    The generated code contains all the calls to log the audits, so in principle yes. Having said that, if you do not deploy through OWB... it will be difficult to get the correct references that OWB will use during its logging to be resolved. I hope that with the explanation above you will be able to deploy and run inside the OWB context and make the entire environment work for you.
    Thanks,
    Mark.

  • Error when deploying a mapping to Staging in OWB 11gr2

    Hi,
    I recieve the following error when I try to deploy a simple Source to Stage mapping:
    RPE-01012: Cannot deploy PL/SQL maps to this target schema. The target schema must be in the same instance as the Control Center and have the OWB_USER role granted.
    I have spent few hours researching it, but have not yet found a solution.
    My environment:
    DB 11gr2 on Red Hat Linux. (OWB Server is on linux and the same machine as its Repository and the 11gr2 db).
    Source is on an Oracle 10.2 DB server in Texas;
    Target (Staging env) is in California, also on Oracle 10.2.
    I can validate the mapping and the Staging table was deployed sucessfully to the Staging DB.
    Do I need to create a repository and workspace on the Target DB (Oracle 10.2) using Repository Assistant? (If that's the case, then would one also need to create OWBSYS?)
    Please help.

    This is an urgent request. Would appreciate it if someone could point out what I may be doing wrong.
    To summarize, my target DB is on Oracle 10.2 in a remote location, so I ran the 3 scripts that OWB Install documents require: Section: Configuring OWB Installed as a Standalone Instance.
    Cat_web.sql, reset_owbcc_home.sql, and remote_owb_install.sql.
    But, when I run the Repository assistant, I get the error above (asking me to install compatible OWB version.)
    Here are the detailed steps:
    1. Ran clean_owbsys.sql
    2. Ran cat_owb.sql (Against USERS Tablespace)
    3. reset_owbcc_home.sql (?? Ora home for Target???)
    4. Ran remote_owb_install.sql (??Ora Home for Target???) -- I have tried different combinations, but no difference.
    For Repository Assistant, here are the detailed steps:
    1. Provide connect info for the remote db (host:Port:Service)
    2. Selected "Managed Warehouse Build Workspaces"
    3. Selected "Create a new Warehouse Builder Workspace"
    4. Select "Create a workspace with an existing user as workspace owner" (Since we have user already defined for this role. Should I create a new one??)
    5. Entered Workspace Owner's User Name, Password and created WKS1 as the Workspace.
    6. Entered System / and Password to logon
    7. Provided Owbsys user id and Password and clicked on Next
    8. As soon as I hit Next on the OWBSYS Information screen, I get the error:
    "OWB software with the compatible version must be installed locally on the database server machine". If it is installed, acquire the DBA user credentials,
    and run the remote_owb_install.sql script on the Oracle Database server machine..."
    (I have tried with different paths for the reset_owbcc_home and remote_owb_install scripts)

  • Mapping import error in OWB 10.1

    In OWB 10R1, while mapping importing I got the following error.. although the mapping is not existing in the repository.
    PLS-00302:Already committed in repository. Please Synchronize..
    Anyone help in this regard
    Thanks
    Hari
    Edited by: user608775 on May 26, 2009 12:53 PM

    this error is only valid if xp appearance is set to "Windows classic style", under "Windows XP style there is no issue
    Signature: Looking for a footprint, component, model? Might be here > http://ni.kittmaster.com

  • ACE ignoring class map depending on source???

    I have a problem with a the load balancing "not working" properly depending on the source.
    The load balancing decision is done with a secondary cookie (?ld=fe1 or ?ld=fe2). If it appears and the value is fe1 the request should go to serverfarm FE1-app. If the value is fe2 then serverfarm FE2-app should be choosen. If it is not present in the http request then serverfarm FE-app in the class-default is taking over.
    This approach works if "surfing" to the VIP from a certain part of the internal network. It does not work from another part of the network. It seems that cookie is ignored and only the class default triggers.
    The strange thing is that the same approach works for another setup that looks identical (with different rservers and different VIP of course). There the class map for the cookie triggers always.
    My question is now: Why does the ACE seem to ignore the class map for the cookie when coming from a certain part of the network? How can I debug/follow a certain connection or load balancing decision?
    Here is the config:
    rserver host FE1-app
      description frontend app
      ip address 192.168.137.69
      inservice
    rserver host FE2-app
      description frontend app
      ip address 192.168.137.74
      inservice
    serverfarm host FE1-app
      rserver FE1-app 80
        inservice
    serverfarm host FE2-app
      rserver FE2-app 80
        inservice
    serverfarm host FE-app
      rserver FE1-app 80
        inservice
      rserver FE2-app 80
        inservice
    class-map type http loadbalance match-all COOKIE-FE1
      2 match http cookie secondary ld cookie-value "fe1"
    class-map type http loadbalance match-all COOKIE-FE2
      2 match http cookie secondary ld cookie-value "fe2"
    class-map match-all VIP-app
      2 match virtual-address 192.168.138.39 tcp eq www
    policy-map type loadbalance first-match VIP-app-loadbalance
      class COOKIE-FE1
        serverfarm FE1-app
      class COOKIE-FE2
        serverfarm FE2-app
      class class-default
        serverfarm FE-app
    policy-map multi-match INT470
      class VIP-app
        loadbalance vip inservice
        loadbalance policy VIP-app-loadbalance
        loadbalance vip icmp-reply
    interface vlan 470
      description lb_rpfedrift
      ip address 192.168.138.36 255.255.255.240
      alias 192.168.138.35 255.255.255.240
      peer ip address 192.168.138.37 255.255.255.240
      service-policy input remote_mgmt_allow_policy
      service-policy input INT470
      no shutdown

    Hi Federico,
    The source of the request has no relation with the way ACE handles the connections, so, there are probably other differences in the traffic.
    The best way to troubleshoot these kind of connections is taking a traffic capture on the TenGigabit interface connecting the ACE with the switch backplane. Once you have it, you can try to look for differences between the working and failing connections.
    From what you describe, I wouldn't be surprised if the issue comes from the fact that there are several HTTP requests inside the same TCP flow (in which case, by default, the ACE will look only at the first one), so I would suggest you to enable "persistence rebalance" for this VIP. For more details, check the link below:
    http://www.cisco.com/en/US/docs/interfaces_modules/services_modules/ace/vA2_3_0/configuration/slb/guide/classlb.html#wp1062907
    I hope this helps
    Daniel

  • Mapping flat files in OWB

    Hello,
    I am trying to start a job which has flat file as source and table as target.
    When I deploy mapping everything is fine but when I try to execute it this error occurs:
    RPE-01013: SQL Loader reported error condition, number 1.
    SQL*Loader: Release 10.2.0.1.0 - Production on Tue Apr 1 16:42:48 2008
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    SQL*Loader-500: Unable to open file (/home/oracle/u00/oracle/product/owb10.2/owb/bin/unix/C:Documents and SettingsvmuzaDesktopDWHSifarniciForETLM_LOAD_SIFARNIK_PORTFELJA.ctl)
    SQL*Loader-553: file not found
    SQL*Loader-509: System error: No such file or directory
    I assume I have to put the flat file somewhere on the server, not on my local machine. Is that the only way? Or is it possible to define path on local machine?
    If I have to upload flat file on the server where do i put it?
    Thankful for any hint.

    Are you sure that the file has to be on the server? I'm just trying and experiencing the same behavior on 10g, but it makes no sense. 80% of the time I'd run SQL*Loader from the client machine, referencing a csv file somewhere in the network drive, certainly not on the DB server machine.
    Then what's the point of using OWB if I need to copy the files onto the server?

Maybe you are looking for

  • Error Installing Adobe Photoshop CS2

    I tried to install Adobe Photoshop CS2 in section http://www.adobe.com/downloads/cs2_downloads/index.html on a Mac ProBook 2.4 Intel Core i5 processor, 4 GB Memory, Intel HD Graphics 3000 384 MB, install halfway then I see the following information:

  • Change id in gamecenter

    I have this old gay stalker dude and he keeps using my email and id to sign up all over the place sony, disney and dating services i know who it is he is a friend of my parents. anyway he has been using my apple id and has changed the name to goatman

  • SAP Netweaver RFC SDK bindings for Node.js

    Hi all, I have just completed a first version of bindings to the SAP Netweaver RFC SDK for [Node.js|www.nodejs.org]. You are encouraged to inspect and download the source code on [https://github.com/jdorner/node-sapnwrfc|https://github.com/jdorner/no

  • Cant print som mails, specially with pictures

    Some mails I cannont print... the same goes for PictureViewer. The printer (brother hl-1870n) shows that information is received (red light) - but nothing happens... I can however, take a screenshot (cant print it from PictureViewer, same problem) -

  • Problem while querying a XML type

    I have created a view to access the content of a XML type in 10g.This view works on my PC.But when the same tried on server it gives the following error... select * from aw33_view ERROR at line 1: ORA-04031: unable to allocate 1000 bytes of shared me