Physical Schema Import

Hi i have a problem
I have 2 schemas a and b ... I am logging in with a main_user that can see both schemas..
I can import them fine using import - Then choose schema(s) and import..
But when i row count its like it cant see the schemas... If i create a synonym for tables within schemas under the main_user its fine.. But i dont know why it cant make out that the tables are within schemas... so a.table_1 b.table_1 etc....
I can see 2 physical schemas in the physical layer...
thanks
Edited by: 791465 on 24-Sep-2010 03:07

2 Database Schemas A, B
I have a Main_user that can see both database Schemas A & B
I can import them Fine into The physical layer with 1 Connection Pool. I can see 2 Schemas.
When i do a Row Count i get a Cant View Table errors for both Schemas..
However if i create a synonym for tables under under the Main_user they row count fine (No Schema Prefix such as A.table1 etc.
Why cant OBIEE in the physical layer recognise the Schema prefix before the table...
Edited by: 791465 on 24-Sep-2010 03:13

Similar Messages

  • How to 1 logical schema to many physical schemas?

    I have a database schema which is instantiated on many different servers. I set up a physical schema pointing to one, and a logical schema pointing to that physical schema. I imported the schema to a model, created interfaces for the tables, and created a package to execute it; and that is all working for that one physical instance.
    1) How can I implement that same model, interfaces, and package for each of the physical instances?
    1a) Can I change the JDBC parameters at package run time to point to a different database? How?
    1b) Can I select a different physical schema for the logical schema at package run time so that I only have to set up a different physical schema for each database? How?
    Thank you.

    "But if you have a lot of context (for example 1000 stores), you can define a generic physical schema, a logical one. The physical is based on variables (host, port,..). "
    Using contexts is working for me, but at least one of my schemas has more than 50 server instances, so this approach would be beneficial. Before I posted this question, I had tried to use variables for the host, port, and SID without success. I used a global variable and gave it default values, but it failed. Then I tried setting the value in a package and creating a scenario, but that too failed. What am I missing?

  • Physical schema conflict exception

    Hi,
    I try to join 2 Customer_ID from different technologies - FILE (file is located on C:\) and ORACLE. Both field are numeric (ODI definition -numeric, Oracle - number), but theI have got message
    Physical schema conflict exception. Should I something to do with KMs or what?
    Thanks a lot
    Gejza

    Yes. it is at staging area and it appears, when I click in property inspector the button "check the expression in DBMS" But perhaps it is normally, because one source is a flat file? I have imported 2 LKMs file to Oracle (SQLLoader) and Fileto SQL..
    Edited by: 841404 on 11.3.2011 2:12

  • Schema Import failed :IMP-00008: unrecognized statement in the export file:

    Hi ,
    I was running a database schema import from a database DMP file.
    I encountered the following error in the log file that was getting generated :
    IMP-00008: unrecognized statement in the export file:
    All the tables got copied except for one. Now what I have thought of is to import that one table again after taking an export from the prod schema.
    But what about the functions,triggers,procedures ?? How do i bring all of them in the schema ?
    If reimporting the schema again the only option ?
    Thanks
    Kk

    Hello,
    In this order:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/exp_imp.htm#i1023988
    Table Objects: Order of Import
    Table objects are imported as they are read from the export file. The export file contains objects in the following order:
    1. Type definitions
    2. Table definitions
    3.Table data
    4.Table indexes
    5. Integrity constraints, views, procedures, and triggers
    6. Bitmap, function-based, and domain indexes
    The order of import is as follows: new tables are created, data is imported and indexes are built, triggers are imported, integrity constraints are enabled on the new tables, and any bitmap, function-based, and/or domain indexes are built. This sequence prevents data from being rejected due to the order in which tables are imported. This sequence also prevents redundant triggers from firing twice on the same data (once when it is originally inserted and again during the import).
    For example, if the emp table has a referential integrity constraint on the dept table and the emp table is imported first, all emp rows that reference departments that have not yet been imported into dept would be rejected if the constraints were enabled.
    When data is imported into existing tables, however, the order of import can still produce referential integrity failures. In the situation just given, if the emp table already existed and referential integrity constraints were in force, many rows could be rejected.
    A similar situation occurs when a referential integrity constraint on a table references itself. For example, if scott's manager in the emp table is drake, and drake's row has not yet been loaded, scott's row will fail, even though it would be valid at the end of the import.

  • Significance of DEFAULT check in Physical Schema

    Hi guys! Can you tell me the significance of Default check in physical schema(in topology manager)?
    When I remove this check from my current physical schema and run any Interface based on the same. I get errors like Delete previous checksome error
    And If i create two or more than two physical schema and everytime when I want to Excecute interfaces based on these different schemas all I need to go to that particular schema and make it my default Schema; Otherwise excecuting Interface gives checksome error if that interface doesnot belong to our default physical schema.
    Can you please comment on this.
    Thankyou.

    Hi Diwakar,
    Adding my comment on this is,
    The below is coded in document,
    "If this box is checked, then the physical schema will be the data server default schema, when no schema has been specified. Only one physical schema can be marked by default."
    The bottom line is for one data server even though it has one/more physical schemas ,atleast one should be default schema.
    So every time if you are using different schema you have to make it as default to took that connection.
    P.S: Experts comments are welcome !!!
    Thanks,
    Guru

  • Issue at create physical schema

    Hello,
    I created a data server using the technology Hyperion Essbase and also put the port because I'm not use the default, then I'am creating a Physical Schema but I can not see any other value than Undefined.
    So I can not continue.
    What is happening?

    You have to manually type in the Application/Database names for example http://2.bp.blogspot.com/__2AaArK5lW8/SNZib4cqZBI/AAAAAAAAAZg/v6SgHh4ux5A/s1600-h/6.png
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Context,Physical schema and Logical schema

    Hi,
    How the context,physical schema,logical schema and agent are interrelated.
    Please explain
    Thanks
    Jack

    Hi Jack,
    Context:
    A context is a set of resources allowing the operation or simulation of one or more data processing applications. Contexts allow the same jobs (Reverse, Data Quality Control, Package, etc) to be executed on different databases and/or schemas.
    Its used to run the object(process) in different database.
    Physical Schema:
    The physical schema is a decomposition of the data server, allowing the Datastores (tables, files, etc) to be classified. Objects stored in data servers with this mode of classification can be accessed by specifying the name of the schema attached to the object name.
    Ex
    Oracle classifies its tables by "schema" (or User). Each table is linked to a schema, thus SCOTT.EMP represents the table EMP in the schema SCOTT.
    Logical schema:
    A logical schema is an alias that allows a unique name to be given to all the physical schemas containing the same datastore structures.
    ->The aim of the logical schema is to ensure the portability of the procedures and models on the different physical schemas. In this way, all developments in ODI Designer are carried out exclusively on logical schemas.
    Thanks
    Madha

  • Physical Schema and logical schema

    Hi,
    When creating the data server in the topology corresponding to the appropriate technology we are creating a physical schema. But then why do we need to create logical schema. Is it created for execution of the interface? And can multiple physical schemas be mapped to same logical schema?

    Hi
    Physical schema represents the actual connection to the data source or data target. Logical schema represents the logical name associated to that source or target.
    One logical schema can be associated with multiple physical schema along with context, i.e. one logical schema is associated with different physical schema using different context.
    It can be understood with following example:
    You have 3 environments: Dev, QA, Prod, each having different database servers as DB1, DB2, DB3, respectively. Similarly we have 3 context corresponding to Dev, QA and Prod. You create logical schema with name DB_source
    Now you associate physical DB servers to logical schema (DB_source) for each context:
    DEV: DB1
    QA: DB2
    PROD: DB3
    Now when u develop ODI interfaces, you use one context DEV which associates DB_source to DB1. While mentioning context for execution, keep it as "Execution". This means, whatever context you choose during execution, corresponding physical DBs will be used.
    Thus if you change the execution context, corresponding physical schema will be used during execution.
    Let me know if you have further questions !!

  • A separate work schema for a physical schema ?

    Hi,
    The ODI best practices white paper tells us (p.47/48) to create a separate schema and designate that as the work schema for a physical schema. This, however means that the user with which we connect to the database needs to have powerful system privileges like CREATE ANY TABLE, DROP ANY TABLE depending on what you want to do. Many DBA's will object to this. A simpler solution seems to be to set the work schema to be identical to the physical schema.
    Has anybody done this and experienced drawbacks/issues with it ?
    Can anybody give any good arguments in favor of the best practice advising a separate work schema ?
    thanks, Arjan

    Arjan ,
    I would like to suggest my opinion. The main reason for allocating the work Schema separate to the main schema is to have all the $ tables created separated so that while debugging we can choose to have them and drop them later and also mostly when the CKM is enabled the E$ tables are created and they are not dropped. there are also situation where dynamic $ tables are created so each time you run they $ tables are created and in case of failure or de bugging they remain . On the second attempt we have a new dynamic $ table. Think of a scenario where we have both of them in one place and the amount of unnecessary $ tables we have with our DW tables . Also if such a scenario have to happen in production box. Nobody would like to have such a scenario and then you need to either drop them manually or through another pl/sql codes . Thats the main reason its good to have a separate Work Schema.
    Also we use the Work schema for creating stagging table at run time while ODI processing.
    Secondly regarding permission and using the schema i have seen both the kind of environment where user used in topology is either the Work Schema or the Data Schema. Yes you would need to give certain permission or priviledges like Drop , create as $ tables have to have them but some dont give permission like Delete , truncate target tables .
    Now if you make your DBA understand the architecture , he would surely provide the required permission for the Work schema tables to be created and dropped and i dont see any thing to worry, since $ tables gets created and dropped but yes the permission have to be in such a way the user used in ODI data server dont have the permission to drop the Data schema table , if you want have more restricited permission for truncate and delete too, but in development we would surely need truncate and delete too since we tend to load wrong data and logic . Think about it in a different scenarios and requirement.
    Look at this link Cezar has spoken on this topic. -http://odiexperts.com/?p=672
    hope this helps.

  • Oracle Physical schema password management

    Hi:
    After setting up an Oracle physical schema and attaching a data store, I did a lot of work reversing tables, views etc and using them in my integrations, procedures etc. I never had to look back at the physical schema until we went into production, and the IT have informed me that there is a password ageing restriction for production systems. I made provisions for other system passwords (Planning, Essbase etc) using variables. But I couldn't get my head around to doing it at the Phsyical schema level.
    Given that this is a classic problem, has anyone run into this sort of issue before, and if yes, what's the best practice to deal with it?
    Any help would be greatly appreciated
    Thanks

    Rajesh,
    I see that you are using variables in Planning, essbase. On the same lines you can uses the variable as a password in the physical schema for database servers as well.
    Not that this is the best practice.
    But if your IT has a password ageing restriction - how tough is it for you to change it in the Topology Manager. That is the same as changing it in a database table and refreshing a variable from it but with less overhead.
    Or you can change your system to use service accounts. Usage of service accounts is the best practice for things like these. Ask them - How do they run automated jobs on Unix ? Do they change the passwords for those accounts every now and then. I think the answer that you will get is "We use service accounts for that"
    Are the database passwords stored in LDAP somewhere. Perhaps you can use JNDI to authenticate your database user/password from LDAP.
    In the Dataserver connection properties, you will see the "JNDI Connection" checkbox. If you select it, it will change the next tab to let you add LDAP details.
    Hope that helps

  • Can we create logical/ physical schema by running any script (unix/window)?

    I wish to create logical/ physical schema in oracle data integrator
    by running any windows or preferably unix script.
    is there any method or scripts provided in ODI?

    I think you could insert info into Matser Respository tables:
    Connect info: snp_connect
    Logical Schemas : snp_lschema
    Physical Schemas: snp_pschema
    Or you may write some sql to do this.

  • Physical schema

    Hi,
    i have an interface and it is running without errors, inserting rows to target table A at schema_A.
    i am creating new interface, and using table A as an source of my new interface.
    after i make drag and drop table A to the flow tab, i recognize that table schema shows different from schema A.
    how can i fix this ? dou you have any idea?

    It should not. Check the logical schema used in both interface for datastore A. Then check the physical schema...Might be you have two different logicalschema pointing the same physical schema.
    Thanks.

  • How to list all physical schemas in ODI procedure

    Dear Experts,
    I am trying a requirement which is to execute a set of sqls in all the schemas configured in ODI.
    for example
    1) I have four data servers/physical schemas configured in Physical Architecture under Oracle techonlogy.
    2) Created corresponding 4 logical schemas for Oracle.
    3) Mapped using two different contexts.
    4) one context mapped two schemas and another mapped two other schemas.
    now, I need to exeture the following sql for all the Oracle schemas mapped in context selected.
    just for testing: select dummy from dual
    I would like to execute this sql in ODI procedure for every schema mapped in the selected context in single execution.
    Can any expert help me on this solution?
    - Raja

    Raja,
    I was actually talking about multiple ODI Step command rather than multiple procedure. What you are trying to achieve is difficult unless you specify the logical schema in ODI Procedure , becuase getInfo will throw Null pointer if we dont specify the logical schema ,
    the getSchema needs Logical schema . The other way i can think query in work rep tables and get the schema name and pass it .
    Any way in case your figure out without passing Logical schema . Please share with us , i would be interested in learning the trick .

  • Physical schema conflict

    Hello,
    I have one requirement where I need to join one table from one schema say schema1 and another table say schema2 and map to 3 rd table of shcema2 only.
    I have reversed the tables with 2 different Physical shcemas, also have Logical schema for each one and when trying to join them in interface getting physical schema conflict error(while validating in interface).
    I created one Logical schema pointing 2 physical shcemas with differnt contexts, still getting the error saying
    java.sql.SQLSyntaxErrorException: ORA-00942: table or view does not exist
    I would appreciate your help
    Preet.

    make sure you have mapped right context with physical schema and you are using the right context to trigger and view data and also the user used in topology have the necessary permission to view edit etc.

  • Physical schema vs logical schema in odi

    hi, i am new to odi.I have successfully loaded data metadata to essbase and planning . But still i am not clear why odi uses physical schema when we just uses logical schema while reversing, execution of interfaces etc in designer

    Hi,
    Logical schema will always point to a Physical Schema .
    The aim of the logical schema is to ensure the portability of the procedures and models on the different physical schemas. In this way, all developments in Designer are carried out exclusively on logical schemas.
    A logical schema can have one or more physical implementations on separate physical schemas, but they must be based on data servers of the same technology. A logical schema is always directly linked to a technology.
    To be usable, a logical schema must be declared in a context. Declaring a logical schema in a context consists of indicating which physical schema corresponds to the alias - logical schema - for this context.
    Thanks,
    Sutirtha

Maybe you are looking for

  • Jabber for Window, phone mode, how to disable "use my computer for calls"

    We have a Citrix environment and only want to control the Cisco deskphone with Jabber for Windows in the phone mode. We use CUCM 8.6.2 and I have installed Jabber for Windows 9.7.0 and all works fine, I can control my deskphone, make and receive call

  • The Creative Zen Touch Movement: Active Creative Us

    Who out there wants a skin for their Creative Zen Touch? Who is tired of the monopoly that Vaja cases has and forces you to buy their horrific case for 70 Big Ones? Who is tired of the Creative Case that is included and is extremely bulky and doesnt

  • Help! Errors in installing 8.0.5 on RedHat Linux 5.2

    Hi, everyone: Oracle free installation support points me here. They said "Unfortunately, the trial release of Oracle8 for Linux is not supportedby this group. You can obtain peer support on the Oracle8 for Linux trialproduct from the discussion group

  • Creating SVG Chart with multiple lines

    Hi All, I want to create a chart with multiple lines. Ex: I have one emp table with empno, sal and deptno columns. I want to print deptno in x axis and no of employees on y axis. I want to print one line for employees with sal > 15000 and one more li

  • JMS - remove persistent messages from memory

    Hi, I have a problem with RAM and JMS. I have to store many JMS-messages (overall size of more than 2GB). By default they are stored in RAM and in the database. How can I configure WL not to store all messages in RAM.