RCU single Schema

Hi
While creating RCU schemas,Is it posible to create BIPLATFORM and MDS both in Single Schema.
e.g
PROD_OBIEE
instead of this
PROD_BIPLATFORM
PROD_MDS
Note: IF create both in single schema it will impact anything or ?
I am trying install OBIEE11g in prodiction machine.
Thanks
Gram.

Hi,
yes it will not affect subject to have sufficient space in production.schema can be from one or multiple database it works
please mark as helpful if answered
thanks

Similar Messages

  • What are the best practices for the RCU's schemas

    Hi,
    I was wondering if there is some best practices about the RCU's schemas created with BIEE.
    I already have discoverer (and application server), so I have a metadata repository for the Application Server. I will upgrade Discoverer 10g to 11, so I will create new schema with RCU in my metada repository (MR) of the Application Server. I'm wondering if I can put the BIEE's RCU schemas in the same database.
    Basically,
    1. is there a standard for the PREFIX ?
    2. If I have multiple components of Fusion in the same Database, I will have multiples PREFIX_MDS schema ? Can they have the same PREFIX ? Or They all need to have a different prefix ?
    For exemple: DISCO_MDS and BIEE_MDS or I can have DEV_MDS and this schema is valid for both Discoverer and BIEE.
    Thank you !

    What are the best practices for exception handling in n-tier applications?
    The application is a fat client based on MVVM pattern with
    .NET framework.
    That would be to catch all exceptions at a single point in the n-tier solution, log it and create user friendly messages displayed to the user. 

  • Error in expdp - single schema

    Hi,
    We are exporting single schema using expdp command and getting following error while export in been taken:
    Command used:
    expdp impexp/********@dbname schemas=schema1 directory=DATA_PUMP_DIR dumpfile=EXPDP_schema1_PRE_231111.dmp LOGFILE=EXPDP_schema1_PRE_231111.log version=compatible content=ALL
    Error
    Processing object type SCHEMA_EXPORT/USER
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    Processing object type SCHEMA_EXPORT/TABLESPACE_QUOTA
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    Processing object type SCHEMA_EXPORT/SYNONYM/SYNONYM
    Processing object type SCHEMA_EXPORT/TYPE/TYPE_SPEC
    Processing object type SCHEMA_EXPORT/DB_LINK
    Processing object type SCHEMA_EXPORT/SEQUENCE/SEQUENCE
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type SCHEMA_EXPORT/TABLE/COMMENT
    Processing object type SCHEMA_EXPORT/TABLE/RLS_POLICY/RLS_POLICY
    Processing object type SCHEMA_EXPORT/PACKAGE/PACKAGE_SPEC
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.UNLOAD_METADATA [PACKAGE:"SCHEMA1".PK_XFACE_IV_LOST_REP_REL_DTLS"]
    ORA-06502: PL/SQL: numeric or value error: NULL index table key value
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPW$WORKER", line 8358
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    0x965e8970 19208 package body SYS.KUPW$WORKER
    0x965e8970 8385 package body SYS.KUPW$WORKER
    0x965e8970 2935 package body SYS.KUPW$WORKER
    0x965e8970 9054 package body SYS.KUPW$WORKER
    0x965e8970 1688 package body SYS.KUPW$WORKER
    0x89194c60 2 anonymous block
    Job "IMPEXP"."SYS_EXPORT_SCHEMA_03" stopped due to fatal error at 09:31:11
    Please suggest what can be done to fix this...

    Thanks for the reply..
    but it is really strange that how compile time errors effect the overall export and import operation... ideailly it should deploy object as it is...
    any other suggestion except to fix code? jst wondering what to do in situation where DBA's are not allowed to touch source codes..

  • Single schema or multiple schemas

    Hello
    A few years ago I worked on a greenfield project where we were building a system to serve 20 or so different departments with 1200+ users. The approach the existing DBA had taken was to use a single schema for all objects and for all apps. The big advantage I found with this approach was for development, we could have one database and each developer could develop whatever they wanted in their own schema. Name resolution meant that they could override the main copy of whichever object they were working on just by having it in their schema. Unfortunately the project went nowhere and I left after 3 months so I never got to see what issues were raised with the system in production. So I'm wondering, has anyone else taken this approach and if so what would you say are the main things to be wary of? Especially things that aren't a problem when the objects are distributed between multiple schemas.
    Thank you in advance
    David

    user12065404 wrote:
    Hi Ed
    Thank you for your reply. I think I need to clarify what I meant a little.
    On a number of sites I've been to there have been multiple applications spread out over multiple schemas for the same business unit. To draw on the emp example it would be like having the EMP schema containing the emp table and the DEPT schema containing the dept table, with a single HR application (for a single buisness unit/department) selecting from both. If you scale this up to 000s of tables and 30 odd schemas, it's a bit of a nightmare from the perspective of code location and permissions because each department needs to look across schemas to get all of the data to do their job. Effectively tables and code had become segregated into separate schemas by business "function" and it means that the applicaitons are so intertwined with the different schemas that to have more than one business unit means you either need a completely separate database or you have to look at VPD.
    App         SCHEMA      Table
    HR          emp         emp
    dept        dept
    Invoicing   emp         emp
    invoicing   inv
    contracts   contract
    orders      ORDER
    Ordering    emp         emp
    contracts   contract
    orders      ORDER
    orders     orderitem
    ...The advantage I saw with a single schema for all the data for a single business unit is the one you have mentioned - i.e. you can very easily set up a new business unit that is totally separate from the other by having a new schema and point the same version of the same application at it. Or failing that, using VPD. And on top of that, security is simplified - no needing direct grants from one schema to the other and especially not requiring WITH GRANT OPTION in the case of a view that pulls data from various schemas and rights to select from the view have to be grated to a role.
    App         SCHEMA      Table
    HR          abc_ltd     emp
    abc_ltd     dept
    Invoicing   abc_ltd     emp
    abc_ltd     inv
    abc_ltd     contract
    abc_ltd     ORDER
    Ordering    abc_ltd     emp
    abc_ltd     contract
    abc_ltd     ORDER
    abc_ltd     orderitem
    ...The other big advantage I saw was from a development and testing perspective. You can have one testing database supporting lots of developers working on lots of different projects against the same core data set and code. The multiple schema setup I described above means that effectively you need to have one testing database per project - which becomes unmanageable when databases start growing to the TB range.
    David
    Edited by: user12065404 on 26-Mar-2010 08:30
    Saw a typo in my first exampleOK, now it is a lot more clear.
    What you have here is a problem of some data (employee) being enterprise data, and other data being application data.
    I take it your "invoiceing" app is used by the sales side of the business, and "ordering" is used by the purchasing side (my last assignment when I was an apps analyst was the purchasing/inventory control for an auto manufacturer).
    With what I know at this point, I'd have an HR schema, an INVOICING schema, and an ORDERING schema. (Well actually, I'd have SALES and PURCHASING, because INVOICEING and ORDERING are going to have to integrate with other stuff in those areas). Even though invoicing and ordering both have a CONTRACT table, I'd think the properties of a contract for one would be different than the other, so the table design itself would be different. How about tables SALES_CONTRACT and PURCHASING_CONTRACT. Same concept for ORDERS, and both sales and purchasing would need header and details .... SALES_ORDER_HEADER, SALES_ORDER_DETAIL, PURCH_ORDER_HEADER, PURCH_ORDER_DETAIL.
    These two apps need for employee data should be pulled directly from the HR emp table, with access limited to stored procedures that control what data is exposed. Or from materialized views. Never try to maintain the same data in two places at once. It will never maintain consistency.

  • Multi Schema to single schema.

    Hi,
    I am new to streams.
    Can it be possible to push the changes from multi schemas to a single schema? Structure for the consolidated schema will be same as other source schemas. Juast an additional column in target tables. I would like populate a unique id for each schema.
    Any example is much appreciated.
    Thanks.

    It is possible. But you have to change the schema name in the changes by using a dml-handler or a transformation rule. For a dml handler I have an example:
    CREATE OR REPLACE PROCEDURE emp_dml_handler(in_any IN SYS.AnyData) IS
    lcr SYS.LCR$_ROW_RECORD;
    rc PLS_INTEGER;
    command VARCHAR2(10);
    old_values SYS.LCR$_ROW_LIST;
    BEGIN
    -- Access the LCR
    rc := in_any.GETOBJECT(lcr);
    -- Get the object command type
    command := lcr.GET_COMMAND_TYPE();
    -- Check for DELETE command on the employees table
    IF command = 'DELETE33' THEN
    -- Set the command_type in the row LCR to INSERT
    lcr.SET_COMMAND_TYPE('INSERT');
    -- Set the object_name in the row LCR to EMP_DEL
    lcr.SET_OBJECT_NAME('DESTINATION.EMPLOYEES');
    -- Get the old values in the row LCR
    old_values := lcr.GET_VALUES('old');
    -- Set the old values in the row LCR to the new values in the row LCR
    lcr.SET_VALUES('new', old_values);
    -- Set the old values in the row LCR to NULL
    lcr.SET_VALUES('old', NULL);
    -- Add a SYSDATE value for the timestamp column
    -- lcr.ADD_COLUMN('new', 'TIMESTAMP', SYS.AnyData.ConvertDate(SYSDATE));
    -- Apply the row LCR as an INSERT into the emp_del table
    lcr.EXECUTE(true);
    END IF;
    END;
    BEGIN
    DBMS_APPLY_ADM.SET_DML_HANDLER(
    object_name => 'source.employees',
    object_type => 'TABLE',
    operation_name => 'INSERT',
    error_handler => false,
    user_procedure => 'strm_emp_admin.emp_dml_handler',
    apply_database_link => NULL,
    apply_name => NULL);
    END;
    Regards,
    Martien

  • Using both 837 P multiple and 837 P single schema

    I have one application which is using 837P Muliple achema........Now i have another application which needs to use 837P single schema....
    but how to use 837P Single schema......As both have same namespace and root node..........so i cant use schema from that application also and i cant deploy my schema also....Is there any solution for this situation.....

    Can you elaborate on what you mean by "Multiple schema" vs "Single schema"... it seems like the single schema could be satisfied by the multiple schema (assuming a difference between repeating vs non-repeating structure(s)).
    Assuming you need to deploy both schemas several options are available.
    If the documents are received on different transport URIs, you can dedicate a different BizTalk application for each schema type.
    If the documents are received on different transport URIs, you can change the namespace of one of the schemas.
    If the documents are received on the same transport URIs, you may run into issues disassembling the messages, as the disassembler (EDI or flat file) may match to the wrong schema during the Probe() method call.  This option will require more discussion
    to determine what you may need to do.
    David Downing... If this answers your question, please Mark as the Answer. If this post is helpful, please vote as helpful.

  • Full Expdp & Try to import single schema

    Hi All,
    I have full expdp backup & i want to do single schema import.
    When i m trying to import single schema it started full import.
    Export :-
    expdp praful/praful directory=dump dumpfile=full.dmp logfile=full.log full=y
    Import :-
    impdp praful/praful directory=dump dumpfile=full.dmp logfile=imp_praful.log remap_schema=PRAFUL:TEST
    Kindly give a suggetion for single schema import from full export dump.
    Regards,
    Praful

    Hi,
    Don't forget to include the parameter SCHEMAS - which provides the details, wher as remap_schema - suffice the details for remapping instead of importing to current schema which you are connected with respect to current session.
    - Pavan Kumar N

  • Multi schema versus single schema set up

    Hi,
    I have a question regarding the environment set up of an Oracle datawarehouse regarding multischema versus single schema set up. My database has several schemas on it representing the functional areas of the warehouse. I have installed the Runtime Repository on the Database and registered each target schema with it. Now that I have started developing, I have noticed the way in which OWB builds database links for mappings which have source and target objects residing in different schemas. I therefore have a major concern regarding the performance of such a set up as opposed to one with a single schema. If anyone has had any experience of working with both environments please could they advise me.
    Your comments will be most appropriate.
    Take care
    Mitesh

    The requirement of sigle or multi schema is driven by the business requirement and it is a policy descion than a techical one.
    Depending upon requirement you can have not only multiple schemas but also multiple instances or databses or server for OLTP , Staging and Star
    Normally for good performance staging and star (which means one for each layer) a schema is created as it offers good performance over single schema set up.

  • RCU creation schema

    Hi
    I am using OBIEE11g in linux box in UAT.Now we have plan to move production.In production without DBA privileges user credential ,my DB team will give 2 schemas one for PROD_BIPLATFORM and PROD_MDS.Now can i possible to install based this schema in OBIEE11g istallation.(Without create RCU)
    Thanks
    Gram.

    Why would you want to do that? Just have someone from the DBA team come and put the needed connection information and SYSDBA password in when you're installing. If that's not possible, are they just creating empty schemas or are they moving the tables within from some other environment?

  • How to import a single schema into multiple target schema?

    Hi,
    I want to import a schema named sd from my source database to my target database in  schema sd1 and sd2.
    I have tried using the following impdp parameter which is throwing error :-
    impdp system/system schemas=sd remap_schema=sd:sd1,sd:sd2 directory=backup dumpfile=8.12.13.dmp logfile=8.12.13.log
    Please help me.

    If you look up the syntax in the docs, you will see that while you can remap many schemas into one, you cannot remap one schema into many.

  • Create materialized view throws insufficient privileges in single schema

    I'm trying to create a complex materialized view in my own schema, using only tables from my own schema:
    CREATE MATERIALIZED VIEW MYSCHEMA.MYMVIEW AS
        SELECT
            A.ONE_THING,
            B.ANOTHER_THING,
            C.A_THIRD_THING
        FROM MYSCHEMA.TABLEA A
        JOIN MYSCHEMA.TABLEB B
            ON A.COL1 = B.COL1
        JOIN MYSCHEMA.TABLEC C
            ON A.COL2 = C.COL2The line JOIN MYSCHEMA.TABLEB B throws an ORA-01031: insufficient privileges error, highlighting TABLEB.
    I understand that grants need to be explicit on tables when creating objects across schemas, but this code is operating within my own schema only; i created and own all the tables in this schema.
    What's going on?
    Thanks!

    Perhaps it is the tool that i am using that highlights the wrong item, because as it turns out, i don't have the CREATE MATERIALIZED VIEW permission after all (the permission was mistakenly granted to a different user instead of to me).
    SELECT * FROM SESSION_PRIVS;returns CREATE VIEW, but does not return CREATE MATERIALIZED VIEW.
    Sorry to have wasted your time.

  • RCU or Manual repository creation ?

    hi everyone,
    The master/work repositories could be created either by Repository Creation Utility(RCU) or by Master Repository creation wizard of ODI Studio
    RCU creates a single schema for both work and master repositories. But if we choose to create repositories using ODI Studio, it is be possible to create Master and Work repositories in two separate schema.
    Which would be the best practice to create repositories ? Using single schema for Master and Work or having separate schemas?

    Hi,
    As per my experience, creating Master and work in "separate" schema is always recommended.
    Thanks,
    Guru

  • (11.1.1.5) RCU or Manual repository creation ?

    hi everyone,
    The master/work repositories could be created either by Repository Creation Utility(RCU) or by Master Repository creation wizard of ODI Studio
    RCU creates a single schema for both work and master repositories. But if we choose to create repositories using ODI Studio, it is be possible to create Master and Work repositories in two separate schema.
    Which would be the best practice to create repositories ? Using single schema for Master and Work or having separate schemas?

    Hi,
    As per my experience, creating Master and work in "separate" schema is always recommended.
    Thanks,
    Guru

  • XML Schema file in data definition

    Hi,
    I am creating a custom RTF template for a seeded data definition in the eBusiness Suite. The data definition has an XML Schema file loaded however this does not include all the data required for the template. Running the report to just generate the XML (i.e. no template) I can see that the additional data is available at this level just not included in the XSD.
    I have built the template including the additional elements from the XML and expected to have to create a custom data definition as well to add the additional elements to the XSD. However the template is working against the seeded definition and successfully picking up the elements that are not included in the XSD.
    Has anybody else come across this situation? And are there any potential issues? I could create a custom data definition as well to load an amended XSD but this would be a more invasive change as it would involve end dating the seeded data definition which I would prefer not to do. The report is not being generated from a concurrent request so I cannot just create a custom request and leave the seeded report untouched.
    Thanks,
    Sarah

    The XML Schema is only required if you are using a PDF template (the template file itself is PDF) with field-mapping. For other template types the schema is optional.
    The XML Schema is used to define a "class" of XML documents, so to speak. The define the semantics and structure of the XML documents more thoroughly.
    If you have closely related reports, you could use a single schema to standardize (in a sense) the XML to be output from the class of related reports. In that case you could have a single schema file uploaded to multiple reports/data-definitions.
    The W3C has good documentation on XML Schema.
    If you want to see samples, you can use tools such as dtd2xs to create sample schemas from your DTD.

  • Biztalk schema exposing as WCF service with multiple operations

    Hi,
    I am working on a design pattern to expose the BizTalk schema as a WCF service and thought of taking expert’s advice. I have a single schema with 30 different nodes and each one is a different message type. For this 30 message types we need to expose WCF
    endpoints to customer for consuming from their application.
    When I exposed this schema as WCF service, a single receive location got created in BizTalk application and in IIS there is one .svc file. I defined 30 operations for this while exposing the endpoint with wizard.
    Currently I came up with the below 2 approaches but I did not like them as there is more redundancy
    Create 30 different orchestrations and each orchestration subscribes to Message Box direct binding with message type as filter.
    Create a single orchestration that binds to single receive location( this location got created by wizard..) and access the operation/message type in the expression shape. Based on this value create 30 IF conditions for each message type. For this the
    single orchestration receive shape should have a message type of xml document and then access the type. Is it possible?
    I don’t like the above approaches as it is very difficult to maintain and also difficult to add for any new message type.
    Does anyone has good ideas for doing this?
    Thanks
    JB

    "Canonical Schemas" are used where multiple source messages need to be addressed by a SINGLE Operation !!!
    Every operation anyway translates to a separate process. Forget that we're dealing with BizTalk for a minute, when you're writing plain we services, you introduce new service contracts (operations) when you deal with different data contracts. To come back,
    in BizTalk your data contracts are your schemas. So when you have multiple schemas, each requiring different processing then you need different orchestrations.
    Let us for a minute assume you've defined a single service schema (very possible with a service header and a service details where the service data contract could differ). Now you use this schema to expose ONE service endpoint (good till now). In that service
    you're suggesting to ADD decisions (something like a switch statement where you'll do 'x' set of activities is the message is of 'x' type and so on... How is this sustainable? Every time you introduce an new operation, you'll have to tear down and redeploy
    the new orchestration which will affect all other operations. During the introduction of ONE op you may touch other parts of the orchestration (multiple branches in a parallel shape anyways make the orchestration very unwieldy from a code management/support
    standpoint) and introduce errors in existing ops.
    Regards. 

Maybe you are looking for

  • Deleting Old Outlook E-mail Address and Adding a New One

    I am using an SMT5800 and have recently moved. In my previous location, my phone was synced to my workplace Outlook E-mail address. Since moving, I have tried repeatedly to update my Outlook E-mail settings to match my new workplace. Unfortunately, w

  • Can't figure out what is wrong with my form

    Aloha all! Need some help from my DW friends.  For some reason my form isn't emailing all the fields.  I have sent several tests and only 2 of the fields are showing, the name and the comments.  I can't see why the other wouldn't be working, it looks

  • Alignment of check boxes in LIveCycle

    Hi, This is my first try using LiveCycle. I have a form that is almost filled with check boxes. I cannot figure out how to algin the boxes-horizontally, vertically, left, right and to each other. I have used the align tab but it only aligns one direc

  • How can i make a family account

    ok so heres the story my brother has an ipod 5 and i have an ipad mini 2 and we wanna use our seperate apple id because when i go to purchases i see all of his history and i like to keep mine clean and tidy for some reason... i want my ipad mini 2 to

  • Bind a huge object to jndi tree in sun one 7

    Hi we are loading all the master tables in to objects at the app server startup (SUN ONE 7) and then binding the whole object to the jndi tree. All the tables data in text files comes to about 5 MB and the serilized object with data to about 8 MB. Bu