How to 1 logical schema to many physical schemas?

I have a database schema which is instantiated on many different servers. I set up a physical schema pointing to one, and a logical schema pointing to that physical schema. I imported the schema to a model, created interfaces for the tables, and created a package to execute it; and that is all working for that one physical instance.
1) How can I implement that same model, interfaces, and package for each of the physical instances?
1a) Can I change the JDBC parameters at package run time to point to a different database? How?
1b) Can I select a different physical schema for the logical schema at package run time so that I only have to set up a different physical schema for each database? How?
Thank you.

"But if you have a lot of context (for example 1000 stores), you can define a generic physical schema, a logical one. The physical is based on variables (host, port,..). "
Using contexts is working for me, but at least one of my schemas has more than 50 server instances, so this approach would be beneficial. Before I posted this question, I had tried to use variables for the host, port, and SID without success. I used a global variable and gave it default values, but it failed. Then I tried setting the value in a package and creating a scenario, but that too failed. What am I missing?

Similar Messages

  • A separate work schema for a physical schema ?

    Hi,
    The ODI best practices white paper tells us (p.47/48) to create a separate schema and designate that as the work schema for a physical schema. This, however means that the user with which we connect to the database needs to have powerful system privileges like CREATE ANY TABLE, DROP ANY TABLE depending on what you want to do. Many DBA's will object to this. A simpler solution seems to be to set the work schema to be identical to the physical schema.
    Has anybody done this and experienced drawbacks/issues with it ?
    Can anybody give any good arguments in favor of the best practice advising a separate work schema ?
    thanks, Arjan

    Arjan ,
    I would like to suggest my opinion. The main reason for allocating the work Schema separate to the main schema is to have all the $ tables created separated so that while debugging we can choose to have them and drop them later and also mostly when the CKM is enabled the E$ tables are created and they are not dropped. there are also situation where dynamic $ tables are created so each time you run they $ tables are created and in case of failure or de bugging they remain . On the second attempt we have a new dynamic $ table. Think of a scenario where we have both of them in one place and the amount of unnecessary $ tables we have with our DW tables . Also if such a scenario have to happen in production box. Nobody would like to have such a scenario and then you need to either drop them manually or through another pl/sql codes . Thats the main reason its good to have a separate Work Schema.
    Also we use the Work schema for creating stagging table at run time while ODI processing.
    Secondly regarding permission and using the schema i have seen both the kind of environment where user used in topology is either the Work Schema or the Data Schema. Yes you would need to give certain permission or priviledges like Drop , create as $ tables have to have them but some dont give permission like Delete , truncate target tables .
    Now if you make your DBA understand the architecture , he would surely provide the required permission for the Work schema tables to be created and dropped and i dont see any thing to worry, since $ tables gets created and dropped but yes the permission have to be in such a way the user used in ODI data server dont have the permission to drop the Data schema table , if you want have more restricited permission for truncate and delete too, but in development we would surely need truncate and delete too since we tend to load wrong data and logic . Think about it in a different scenarios and requirement.
    Look at this link Cezar has spoken on this topic. -http://odiexperts.com/?p=672
    hope this helps.

  • How to convert logical standby database to Physical Standby Database

    Hi,
    Does anyone know how to convert logical standby database(10.2.0.3) to Physical Standby Database(10.2.0.3)? Is this possible? If so,please give me steps for it.
    Thanks.
    Regards,
    RJ.

    I don't believe this to be possible. One of the steps invloved in the creation of a logical standby is the following:
    ALTER DATABASE OPEN RESETLOGS;
    This makes the logical standby a different incarnation of the db from the primary.
    A physical standby is a binary copy of a primary db, a logical standby is a completely seperate db that just happens to get sql statements applied to it that are generated from a primary db, but is open read/write.
    One avenue, though would be if you had flashback database enabled and you flashed back all the way to when the db was a physical standby and then rolled forward with the archived redo logs.
    It's somewhat unlikely you are in that situation, but that is a possibility.
    Indeed the 11g snapshot standby does this sort of thing:
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28294/manage_ps.htm#sthref740
    jason.
    http://jarneil.wordpress.com

  • Physical Schema and logical schema

    Hi,
    When creating the data server in the topology corresponding to the appropriate technology we are creating a physical schema. But then why do we need to create logical schema. Is it created for execution of the interface? And can multiple physical schemas be mapped to same logical schema?

    Hi
    Physical schema represents the actual connection to the data source or data target. Logical schema represents the logical name associated to that source or target.
    One logical schema can be associated with multiple physical schema along with context, i.e. one logical schema is associated with different physical schema using different context.
    It can be understood with following example:
    You have 3 environments: Dev, QA, Prod, each having different database servers as DB1, DB2, DB3, respectively. Similarly we have 3 context corresponding to Dev, QA and Prod. You create logical schema with name DB_source
    Now you associate physical DB servers to logical schema (DB_source) for each context:
    DEV: DB1
    QA: DB2
    PROD: DB3
    Now when u develop ODI interfaces, you use one context DEV which associates DB_source to DB1. While mentioning context for execution, keep it as "Execution". This means, whatever context you choose during execution, corresponding physical DBs will be used.
    Thus if you change the execution context, corresponding physical schema will be used during execution.
    Let me know if you have further questions !!

  • Context,Physical schema and Logical schema

    Hi,
    How the context,physical schema,logical schema and agent are interrelated.
    Please explain
    Thanks
    Jack

    Hi Jack,
    Context:
    A context is a set of resources allowing the operation or simulation of one or more data processing applications. Contexts allow the same jobs (Reverse, Data Quality Control, Package, etc) to be executed on different databases and/or schemas.
    Its used to run the object(process) in different database.
    Physical Schema:
    The physical schema is a decomposition of the data server, allowing the Datastores (tables, files, etc) to be classified. Objects stored in data servers with this mode of classification can be accessed by specifying the name of the schema attached to the object name.
    Ex
    Oracle classifies its tables by "schema" (or User). Each table is linked to a schema, thus SCOTT.EMP represents the table EMP in the schema SCOTT.
    Logical schema:
    A logical schema is an alias that allows a unique name to be given to all the physical schemas containing the same datastore structures.
    ->The aim of the logical schema is to ensure the portability of the procedures and models on the different physical schemas. In this way, all developments in ODI Designer are carried out exclusively on logical schemas.
    Thanks
    Madha

  • How to Use logical schema in Variable

    Hi,
    I am using ODIFileCopy. I want to use the connection details configured in my Logical or Physical Schema, in ODIFileCopy.
    Please let me know i can use the topology connection in ODIFILECOPY.
    Thanks:)

    Got the answer...
    please find the below link
    http://gerardnico.com/doc/odi/webhelp/en/ref_api/getschemaname.html
    Thanks

  • Can we create logical/ physical schema by running any script (unix/window)?

    I wish to create logical/ physical schema in oracle data integrator
    by running any windows or preferably unix script.
    is there any method or scripts provided in ODI?

    I think you could insert info into Matser Respository tables:
    Connect info: snp_connect
    Logical Schemas : snp_lschema
    Physical Schemas: snp_pschema
    Or you may write some sql to do this.

  • How to list all physical schemas in ODI procedure

    Dear Experts,
    I am trying a requirement which is to execute a set of sqls in all the schemas configured in ODI.
    for example
    1) I have four data servers/physical schemas configured in Physical Architecture under Oracle techonlogy.
    2) Created corresponding 4 logical schemas for Oracle.
    3) Mapped using two different contexts.
    4) one context mapped two schemas and another mapped two other schemas.
    now, I need to exeture the following sql for all the Oracle schemas mapped in context selected.
    just for testing: select dummy from dual
    I would like to execute this sql in ODI procedure for every schema mapped in the selected context in single execution.
    Can any expert help me on this solution?
    - Raja

    Raja,
    I was actually talking about multiple ODI Step command rather than multiple procedure. What you are trying to achieve is difficult unless you specify the logical schema in ODI Procedure , becuase getInfo will throw Null pointer if we dont specify the logical schema ,
    the getSchema needs Logical schema . The other way i can think query in work rep tables and get the schema name and pass it .
    Any way in case your figure out without passing Logical schema . Please share with us , i would be interested in learning the trick .

  • Physical schema vs logical schema in odi

    hi, i am new to odi.I have successfully loaded data metadata to essbase and planning . But still i am not clear why odi uses physical schema when we just uses logical schema while reversing, execution of interfaces etc in designer

    Hi,
    Logical schema will always point to a Physical Schema .
    The aim of the logical schema is to ensure the portability of the procedures and models on the different physical schemas. In this way, all developments in Designer are carried out exclusively on logical schemas.
    A logical schema can have one or more physical implementations on separate physical schemas, but they must be based on data servers of the same technology. A logical schema is always directly linked to a technology.
    To be usable, a logical schema must be declared in a context. Declaring a logical schema in a context consists of indicating which physical schema corresponds to the alias - logical schema - for this context.
    Thanks,
    Sutirtha

  • Logical schema can have one or more physical schemas

    Hi Experts,
    Can a logical schema in ODI will have one or more physical schema, if possible in what case do we have it.
    In the same way can one physical schema have multiple logical schema if possible in what case.
    Please give me your brief explanation on it.
    Thx,
    Sahadeva.

    Hi Sahadeva,
    1) Yes. You can map it through different contexts. The goal is to use the same logical schema to point to a different physical schemas for each environment (Dev, Test, PROD, ...).
    So let's say you have a have the schema SH_DEV on database 1 for Dev and schema SH on database 2 for Dev :
    - Create one SH Logical schema
    - Create two data server : one for database1 and one for database2
    - Create the physical schemas SH_DEV under dataserver1 and SH under dataserver2
    - Create two context : Dev and Prod
    - Map logical SH to dataserver1.SH_DEV through context Dev
    - Map logical SH to dataserver2.SH through context Prod
    2) Technically yes, through different contexts. Though I don't see any use case for this.
    Hope it helps.
    Regards,
    JeromeFr

  • How many physical Network Adapter required for to implement NVGRE Gateway ?

    Hello Guys,
    Would like to know how many physical Network Adapter required to implement network virtualization including the technologies such as Windows Azure Pack| system center 2012 R2 & Hyper-V 2012 R2.
    Thanks 
    NM-BG

    A quick search of TechNet will turn up all sorts of step-by-step and overviews and blogs with lots of information about how to configure and setup.  Minimum of two.  Actual configurations may vary according to business needs and requirements.
    . : | : . : | : . tim

  • How many physical standbys can be created in one data guard environment?

    Hi Experts,
    How many physical standbys can be created from primary in one data guard environment? Why LOG_ARCHIVE_CONFIG limits up to 9 unique database names? Thanks.

    Victor Jin-Oracle wrote:
    Hi Experts,
    How many physical standbys can be created from primary in one data guard environment? Why LOG_ARCHIVE_CONFIG limits up to 9 unique database names? Thanks.
    Post the source of your information abouit the limit of 9 names.
    Even the Oracle doc shows 1..30 names in the same provided although that doesn't necessarily indicate a limit.
    https://docs.oracle.com/database/121/REFRN/refrn10237.htm

  • How many Physical CPUs Arch Linux Supports?

    How many Physical CPUs does Arch Linux Support, the traditional 2?
    Can it support 4 or more Physical CPUs, How?
    By Physical CPUs I take RedHat deffinition :
    Red Hat defines physical CPUs equivalently to sockets, so a multi-core and/or hyperthreading CPU is counted as a single CPU or socket when determining whichedition of Enterprise Linux to deploy.
    Thanks

    Let's put it this way.
    Linux can likely support more cpu's than you can afford.
    That is, unless you are super rich..
    If that is the case..then by all means.....
    ....donate!
    EDIT: After some peeking around.. I found this.
    SGI Altix 4700 incorporates the shared-memory NUMAflex™ architecture, which simplifies software development, workload management and system administration. It supports up to 512 processors under one instance of Linux and as much as 128TB of globally shared memory. Supporting these powerful capabilities is the NUMAlink™ interconnect, which leads the industry in bandwidth and latency for superior performance on cluster applications. The SGI Altix 4700 represents a versatile solution for shared or distributed memory applications of any scale.
    Likely SGI has a few tweaks that haven't percolated into the mainline kernel yet (pending or rejected)..But sgi usually pushes the high end mutiprocessor systems..
    So.. around 512?

  • Logical Database design and physical database implementation

    Hi
    I am an ORACLE DBA basically and we started a proactive server dashboard portal ,which basically reports all aspects of our infrastructure (Dev,QA and Prod,performance,capacity,number of servers,No of CPU,decomissioned date,OS level,Database patch level) etc..
    This has to be done entirely by our DBA team as this is not externally funded project.Now i was asked to do " Logical Database design and physical Database
    implementation"
    Even though i know roughly what's that mean(like designing whole set of tables in star schema format) ,i have never done this before.
    In my mind i have a rough set of tables that can be used but again i think there is lot of engineering involved in this area to make sure that we do it properly.
    I am wondering you guys might be having some recommendations for me in the sense where to start?are there any documents online , are there any book on this topic?Are there any documents which explain this phenomena with examples ?
    Also exactly what is the difference between logical database design vs physical database implementation
    Thanks and Regards

    Logical database design is the process of taking a business or conceptual data model (often described in the form of an Entity-Relationship Diagram) and transforming that into a logical representation of that model using the specific semantics of the database management system. In the case of an RDBMS such as Oracle, this representation would be in the form of definitions of relational tables, primary, unique and foreign key constraints and the appropriate column data types supported by the RDBMS.
    Physical database implementation is the process of taking the logical database design and translating that into the actual DDL statements supported by the target RDBMS that will create the database objects in a target RDBMS database. This will generally include specific physical implementation details such as the specification of tablespaces, use of specialised indexing (bitmap, clustered etc), partitioning, compression and anything else that relates to how data will actually be physically stored inside the database.
    It sounds like you already have a physical implementation? If so, you can reverse engineer this implementation into a design tool such as SQL Developer Data Modeller. This will create a logical design by examining the contents of the Oracle data dictionary. Even if you don't have an existing database, Data Modeller is a good tool to use as a starting point for logical and even conceptual/business models.
    If you want to read anything about logical design, "An Introduction to Database Systems" by Date is always a good starting point. "Database Systems - A Practical Approach to Design, Implementation and Management" by Connolly & Begg is also an excellent reference.

  • Physical schema

    Hi,
    i have an interface and it is running without errors, inserting rows to target table A at schema_A.
    i am creating new interface, and using table A as an source of my new interface.
    after i make drag and drop table A to the flow tab, i recognize that table schema shows different from schema A.
    how can i fix this ? dou you have any idea?

    It should not. Check the logical schema used in both interface for datastore A. Then check the physical schema...Might be you have two different logicalschema pointing the same physical schema.
    Thanks.

Maybe you are looking for