RAC as the source DB in OWB

Hello there,
I am quite new to OWB and have not encountered any RAC system. My client will migrate to RAC and as such, the source DB in our OWB tables will be affected as well.
How do you set-up in RAC in OWB as the source when there are 2 IP addresses? Can I specify 2 IPs as the hosts? What is the best way to go with it?
Thank you.

Hello, I have trouble changing the source connection.
My source connection name is same as before. Only DB name, schema name is changing now for source connection. Here are the steps i have done.
1. Went to control center manager and unregister the source mapping.
2. Went to Connection explorer and right click on source connection and click the open editor. change the username, password, ip address, DB name.
3. went to control center manager and re-deploy the mapping. It is not working... It says, table or view does not exists.
Is ther any other place i need to change... Somewhere i am missing..
Any help is highly appreciated..
Edited by: Shrinika on Oct 21, 2010 9:52 AM

Similar Messages

  • How do we change the source connection in OWB?

    Hello,
    Currently my source connection has read/write privileges. For security reason, i created read only schema in the source database.
    Now i want to change my username in the source mapping.
    I go to connection Explorer,
    click Location
    click database
    Click oracle
    right click on source name(existing one), click open editor.
    When i click open editor, I see all the user info are grayed out. I could not change anything.
    How do i change the current source connection username?
    Any help is appreciated..

    Hello, I have trouble changing the source connection.
    My source connection name is same as before. Only DB name, schema name is changing now for source connection. Here are the steps i have done.
    1. Went to control center manager and unregister the source mapping.
    2. Went to Connection explorer and right click on source connection and click the open editor. change the username, password, ip address, DB name.
    3. went to control center manager and re-deploy the mapping. It is not working... It says, table or view does not exists.
    Is ther any other place i need to change... Somewhere i am missing..
    Any help is highly appreciated..
    Edited by: Shrinika on Oct 21, 2010 9:52 AM

  • Source Mapping - requires a control centre schema on the source?

    Hi,
    If I create a mapping on a source (that maps source table[s]/view[s] into a source 'staging' table on the same source), when I come to debug the mapping I get an error:
    "DBG1005: The connection information provided does not specify a valid OWB Control Centre target schema. .."
    Is it possible to DEBUG a mapping on the source only within OWB? if so, do I need to install / create a control centre on the source too? - how?
    I can run/debug that same mapping but only if the mapping is created on the TARGET - which seems crazy as then its selecting across database links to push data within the source.
    Chris.

    Hi,
    in OWB a source is that from where you only load data and import object definitions from the data dictionary.
    A target is where you want to deploy to. You can also load data from it like from a normal source. It's both.
    If you create a database modul you're asked if it is a source or a target.
    The deployed objects send and get information to/out of the repository (direct or over a database link). That's why a target schema gets several OWB specific objects and is registered in the repository as a target.
    Change the modul to a target and then you can deploy to it - but then it is not a source.
    Example:
    The staging area of a DWH is a target, the OLTP systems to load from are sources.
    Regards
    Detlef

  • Dynamically passing the source table name to OWB mapping

    I am building a mapping wherein one of the source tables is a view. The view name varies with the time parameter I pass in. I am looking at ways to pass in the time parameter to the mapping procedure such that it first gets the view name from a table and uses that view as the source table to fetch data. Any directions?
    In normal PL/SQL coding, I can first get the view name and use this view name to buld a dynamic query, which can be then executed.

    This is a common question. The best way to do this is to use a synonym.
    Create the synonym in the database and import into OWB. Use the synonym in your mapping. Have your mapping accept a mapping input for the table you want the synonym to point to. Setup a premapping process to re-create the synonym with the table you want to use.
    Here is the procedure that I use. It defaults to a private synonym. Remember, the synonym will be created in the same schema that the mapping is deployed to.
    CREATE OR REPLACE PROCEDURE "CAWDATA"."CREATE_SYNONYM_PRC" ("P_SYNONYM_NAME" IN VARCHAR2,  "P_OBJECT_NAME" IN VARCHAR2,
    "P_IS_PUBLIC_SYNONYM" IN BOOLEAN DEFAULT false)    IS
    BEGIN
    if p_is_public_synonym = true then
    execute immediate 'create or replace public synonym '|| p_synonym_name || ' for '|| p_object_name;
    else
    execute immediate 'create or replace synonym '|| p_synonym_name || ' for '|| p_object_name;
    end if;
    exception
    when others
    then
          raise_application_error(sqlcode,sqlerrm) ;
    END;

  • Best approach to delete records that are not in the source table anymore.

    I have a situation where I need to remove records from dimensions that are not in the source data anymore. Right now we are not maintaing history, i.e. not using SCD but planning for the next release. If we did that it would be easy to figure the latest records. The load is nightly and records are updated and new added.
    The approach that I am considering is to join the dimension tables the the sources on keys and delete what doesn't join. However, is there perhaps some function in OWB that would allow to do this automatically on import so it can be also in place for the future?
    Thanks!

    Bear in mind that deleting dimension records becomes problematic if you have facts attached to them. Just because this record is no longer in the active set doesn't mean that it wasn't used historically, and so have foreign key constraints on it in your database. IF this is the case, a short-term solution would be to add an expiry_date field to the dimension and update the load to set this value when the record disappears rather than to delete it.
    And to do that, use the target dimension as a source table, outer join it to the actual source table on the natural key, and so your update will set expiry_date=nvl(expiry_date,sysdate) to set to sysdate if this record has not already been expired on all records where the outer join fails.
    Further consideration: what do you do if the record is re-inserted into the source table? create a new dimension key? Or remove the expiry date?
    But I will say that I am not a fan of deleting records in most circumstances. What do you do if you discover a calculation error and need to fix that and republish historical cubes? Without the historical data, you lose the ability to do things like that.

  • Problem in executing the mapping from outside OWB

    hi guys,
    i am facing a strange problem in a regular fashion.
    i have developed 4 mappings - the mappings have been validated with warnings, all of column lengths and data length related.
    the mappings have been deployed successfully - which i have also verified by checking the generated package in the backend - it is
    in a valid state.
    but when i am running the package thru the pl/sql block, tho it returns me a retval=0 (and sometimes retval=2) but no data transfer is happening, the same thing is
    occuring when i try to run the mapping thru the deployment manager in the OWB. but through the debugger all the DML operations of
    upsert(insert/update) and delete are working perfect with the correct transformations.
    this is happening for all the mappings that i have developed.
    need urgent help.
    thanks and regards,
    Prasenjit

    i am going crazy now.
    suddenly the mapping is working for the insert/update but for the delete it is giving weird results.
    the primary key column in the source table has a value ",1:EH/L/A,13:2,14:9", which should be transformed to EH/L/A::2::9 - i have used substr and instr to achieve this.
    but in the case of delete operation, thru the debugger i get the expected result - the corresponding row is getting deleted from the target, but theu the deployment manager - another row is getting inserted into the target table.
    on checking the target table, i find the corresponding columns having values - "::::" - don't know the reason.
    thanks and regards.

  • How to extract the cobol files into owb

    Hi
    How to extract the cobol files in owb,I am using OWB9.2.
    Ex:I am having 2 records DCEC4000100 9159000050C those values are 5 bytes 'RAVI and 6 bytes '15000'.
    Using contorl file i am trying to load these into Oracle,But i am not able into insert these records into Oracle.
    Can any tell me how to handle this kind of situation?
    Regards,
    Rupak

    Good morning Ashok,
    Your data is in a SQL*Loader supported datatype, so basically it should not be a problem.
    Have you checked the Database Utilities guide (for Oracle9i for instance: http://www.lc.leidenuniv.nl/awcourse/oracle/server.920/a96652/toc.htm) on how to handle this?
    I'll also repeat my previous question, have you ever been able to load data into the database using SQL*Loader (either using ASCII values or any other datatype)?
    For the OWB part, have you read "Importing Data Definitions" (typically chapter 4)? More specifically, "Specifying Field Properties" in "About Flat File Sources and Targets"?
    In "SQL*Loader Properties" it says the following:
    Type Describes the data type of the field for the SQL*Loader. You can use the Flat
    File Sample Wizard to import the following data types: CHAR, DATE, DECIMAL
    EXTERNAL, FLOAT EXTERNAL, INTEGER EXTERNAL, ZONED , AND ZONED
    EXTERNAL. For complete information on SQL*Loader field and data types, refer to
    Oracle9i Utilities. Currently, only portable datatypes are supported.
    If you check the database utitlities guide (chapter 6 Field List Reference) and look for "SQL*Loader Datatypes", you'll find (packed) decimal under the Nonportable Datatypes. This implies that OWB does not support it.
    This does not mean you can't use SQL*Loader, you'll only have to define everything separate from OWB and call it separately as well.
    Good luck, Patrick

  • The 'Third Party Connections' drop down box on the 'Source Database' step in the Migration Wizard is empty.

    Here are the environmental details:  
    SQLDeveloper 32-bit Version 3.2.20.09 (Build MAIN-09.87)
    Oracle Database:  12C single instance (no RAC)
    Oracle client:  12C 32 bit
    Windows 64 bit desktop
    I am attempting to migrate an MS Access application to Oracle.  I pick the ‘Migrate’ selection from the ‘Tools’ Menu SQL Developer.  This puts me into the Migration Wizard.  When I get to the ‘Source Database’ step of the migration wizard, the ‘Choose the Third Party database from which you are migrating’  drop-down box is not being populated with the current available connections. It is empty.  If I type a valid connection name into the box, SQL Developer takes it on this screen but does not keep it for the next screen/step.  It has no persistence.  I even tried to make sure that each connection was open/connected before beginning the migration process.
    Due to the fact that I am attempting to migrate an MSAccess application, I am mixing 32-bit and 64 bit software.  This may be the issue as to why the drop down is losing addressability.  I have not applied any patches to this version of SQL Developer.

    Hi,
    If you need easy way to migrate Microsoft Access to Oracle probably it is more reasonable to review some tools specially designed to this purpose. For example:
    Access to Oracle converter by Intelligent Converters
    Access to Oracle converter by SpectralCore
    Access to Oracle converter by Interface Computers
    Hope, this will help.
    Sincerely,
    Vlad

  • How to change the source module version

    Hi All,
    In the OWB i created a source module and during those steps i clicked on version 10.1, but during deployment in the deployment managaer when i tried to register the location, its giving me an error as "Source verison 10.1 is not matching with the target version 10.2", can any one please help me or Advice me as to how to change the "source module version"?
    Thanks

    Hi
    what i understood is that the source module is associate with the location and you have provided the 10.1 which you need to change.
    so you can change that under location.
    Cheers
    Nawneet

  • How to implement SQL Over-ride on Source Tables in OWB?

    How can we acheive SQL-Override feature similar to the one available in informatica on OWB.
    We are thinking of using a view to filter the required data.And defiinig the view as the source in the mapping.
    Is this the best approach to go by,any thoughts suggestions?

    Informatica is a pipelined client/server ETL tool which means that if you read from a table and then apply a filter it will read all of the data onto the client (i.e. the application server, not a user machine) and then filter the data so it is not unusual for developers to want to put a lot of the logic into the SQL to reduce the number of records (which sort of defies the purpose of using the tool in the first place).
    OWB is more of a ELT/code generator so if you create a similar mapping with a table and filter it will generate the SQL with the filter included. To be honest, I'm not 100% sure about this bit but I believe if you have a source via dblink (or even ODBC) then that full generated SQL should be sent to the source. In other words, although some people might be more comfortable with custom SQL I dont think it is actually necessary in most if not all cases.

  • Facing problem in implementing the analytical query in OWB

    Hi All,
    I was trying analytical function for my Customer Dimension but am not able to implement the same in my OWB.I have 2 tables who have duplicates customers which cannot be figured out by Customer id as customer id is not unique.So, i have to filter out the information on the basis of First Name, middle name, last name & DOB. Since, each customer has multiple address records( for Permanent, Correspondence, Unknown ).
    I want 1 record for all the three address types.
    Can i do thru the following query???
    If yes, how will i implment the same in OWB.
    SELECT PRENOM, MIDDLE_NAME, NOM, DEBUT, SEXE, OCCUPATION, FAMILY_STATUS,
         EDUC_QUALIFICATION, CUSTOMER_NATIONALITY, SIGLE, ANNUAL_INCOME,
         FATHER_HUSBAND_NAME,TYPE_ADDRESS, ADDRESS1, ADDRESS2, ADDRESS3,
         TOWN, REGION, POSTCODE,ADDRESS_EFFECT_DATE
    FROM
    (SELECT C1.PRENOM, C1.MIDDLE_NAME, C1.NOM, C1.DEBUT, C1.SEXE, C1.OCCUPATION, C1.FAMILY_STATUS,
         C1.EDUC_QUALIFICATION, C1.CUSTOMER_NATIONALITY, C1.SIGLE, C1.ANNUAL_INCOME,
         C1.FATHER_HUSBAND_NAME,C1.TYPE_ADDRESS, C1.ADDRESS1, C1.ADDRESS2, C1.ADDRESS3,
         C1.TOWN, C1.REGION, C1.POSTCODE,C1.ADDRESS_EFFECT_DATE,
    ROW_NUMBER() OVER (PARTITION BY C1.PRENOM, C1.MIDDLE_NAME, C1.NOM, C1.DEBUT,C1.TYPE_ADDRESS
    ORDER BY C1.ADDRESS_EFFECT_DATE DESC)M
    FROM CUST_1 C1
    UNION ALL
    SELECT C1.PR_FIRST_NM, C1.PR_MIDDLE_NM, C1.PR_LAST_NM, C1.PR_DOB, C1.PR_GENDER, C1.PR_OCCUPATION, C1.PR_MARITAL_STATUS,
    C1.PR_EDUCATION_QUAL, C1.PR_NATIONALITY, C1.PR_TITLE, C1.GP_ANNUAL_INC, C1.PR_FATHER_NM,
    C1.AD_TYPE, C1.AD_ADDR1, C1.AD_ADDR2, C1.AD_ADDR3, C1.TOWN, C1.REGION, C1.AD_PIN_CD,C1.REC_UPDT_DT,
    ROW_NUMBER() OVER (PARTITION BY C1.PR_FIRST_NM, C1.PR_MIDDLE_NM, C1.PR_LAST_NM, C1.PR_DOB,C1.AD_TYPE
    ORDER BY C1.REC_UPDT_DT DESC )M
    FROM CUST_2 C1)
    WHERE M = 1
    Please help me out as right now, am using aggregators which are taking lot of time. Loading type on my Customer Dimension table is- Insert/Update
    Thanks in advance
    -Nikita.

    Hello Nikita,
    From your first post I understand that you would like to merge 3 similar (but not complete identical) records into 1 record. If this is the case, then have you thought about using the match-merge operator in OWB?
    With the match-merge operator you can tell the mapping which records are similar by selecting "first_nm", "middle_nm", "last_nm" in the Match-Bin of the operator.
    Then you define how attributes should be merged.
    This way you do not need to aggregate your source data set. Just select all records from both source tables. Put those in the Set-Operator "UNION ALL" and this will be you input for the match-merge operator.
    Regards,
    Ilona

  • What is the source table for repository browser info?

    Hi guys,
    I need to create report for everyday's data refresh, mostly like what repository browser does. So what is the source table for repository browser info? I am in OWB 10gR2. Thanks a lot.

    I believe the runtime audit browser uses the views that start with "RAB" in the control center.
    But it would probably better if you used the Audit Execution views. More information can be found in the "API and Scripting Reference" document...
    http://download.oracle.com/docs/cd/B31080_01/doc/owb.102/b28225/api_2runviews.htm
    In order to get access to the Public views from SQL Plus using a schema other than the control center, I believe you need to grant the ACCESS_PUBLICVIEW_BROWSER sys privilege to each OWB user. Look at Note:434718.1 for more information.

  • IMPORT METADATA IN THE SOURCE MODULE

    Please I have a problem with importing metadata, after I select the correct schema, it show me some tables but, it' doesn't show me my tables.
    Have I to configure a Runtime Repository Connection before importing source metadata.
    Someone know why?
    Thanks
    Message was edited by:
    user455714

    Importing metadata into OWB has noting to do with runtime. It's all about design environment.
    1. Define module (source or target - no matter)
    2. For the module define the source - database link, and may be owner schema. Test connection.
    3. Perform the import.
    If you refer as "your tables" those your've just designed within OWB - you don't have to import them again into different module. Instead you can use them as a source directly.

  • Manage table function's source code in OWB metadata

    Hi all,
    please I would like to know if it possible to code a table function in OWB's metadata.
    I'll try to better explain myself; I know how to use a tf in a mapping but I also would like to manage the table function's source code in OWB metadata repository.
    It appears you cant code a table function as procedures/functions: there's no ref_cursor as input , no pipelined option and so on.
    I'm workin on 10.2.0.4 OWB repos.
    Any information will be appreciated, thanks.
    best regards, David

    ... you are right. You cannot write the code of a table function inside owb. You must write the table function external and create it in the database. In owb you only can call the table function.
    Regards,
    Detlef

  • MERGE Statement - unable to get a stable set of rows in the source tables

    OWB Client: 10.1.0.2.0
    OWB Repository: 10.1.0.1.0
    I am trying to create a MERGE in OWB.
    I get the following error:
    ORA-12801: error signaled in parallel query server P004 ORA-30926: unable to get a stable set of rows in the source tables
    I have read the other posts regarding this and can't seem to get a fix.
    The target table has a unique index on the field that I am matching on.
    The "incoming" data doesn't have a unique index, but I have checked and confirmed that it is unique on the appropriate key.
    The "incoming" data is created by a join and filter in the mapping and I'd rather avoid having to load this data into a new table and add a unique index on this.
    Any help would be great.
    Thanks
    Laura

    Hello Laura,
    The MERGE statement does not require any constraints on its target table or source table. The only requirement is that two input rows cannot update the same target row, meaning that all existing target rows can be matched by at most one input row (otherwise the MERGE would be undeterministic since you don't know which of the input rows you would end up with in the target).
    If a table takes ages to load (and is not really big) I suspect that your mapping is not running in set mode and that it performs a full table scan on source data for each target row it produces.
    If you ARE running in set mode you should run explain plan to get a hint on what is wrong.
    Regarding your original mapping, try to set the target operator property:
    Match by constraint=no constraints
    and then check the Loading properties on each target column.
    Regards, Hans Henrik

Maybe you are looking for