Transfer whole schema between databases

Hi,
I have a question on how to move an entire schema between two databases which are of the same version and patchset.
I identified the possible methods as follows:
1) Using exp/imp or expdp/impdp, but that's quite slow on larger schemas and the transported data might not be physically the same as on the source database (indexes are rebuilt, tables get loaded with data again....).
2) Using transportable tablespaces. This is fast, the physical data layout is the same as on the source database, but unfortunately only the objects in the transported tablespace get transferred. Schema objects like PL/SQL packages, private synonyms, functions.... and grants aren't transferred since they are stored in the SYSTEM tablespace.
Is there any other way to move a schema between databases? Maybe some combination of expdp/impdp (transfer of user object definitions stored in the SYSTEM tablespace like PL/SQL code, view definitions, etc.) and transportable tablespaces (transfer data segments like tables, indexes....)
From what I know RMAN can't do that since the smallest "logical object" it can restore/recover is the whole tablespace with tablespace point in time recovery.
Thanks in advance for any answer.
Regards,
Jure

user633661 wrote:
Hi,
I have a question on how to move an entire schema between two databases which are of the same version and patchset.
I identified the possible methods as follows:
1) Using exp/imp or expdp/impdp, but that's quite slow on larger schemas and the transported data might not be physically the same as on the source database (indexes are rebuilt, tables get loaded with data again....).
2) Using transportable tablespaces. This is fast, the physical data layout is the same as on the source database, but unfortunately only the objects in the transported tablespace get transferred. Schema objects like PL/SQL packages, private synonyms, functions.... and grants aren't transferred since they are stored in the SYSTEM tablespace.
Is there any other way to move a schema between databases? Maybe some combination of expdp/impdp (transfer of user object definitions stored in the SYSTEM tablespace like PL/SQL code, view definitions, etc.) and transportable tablespaces (transfer data segments like tables, indexes....)
From what I know RMAN can't do that since the smallest "logical object" it can restore/recover is the whole tablespace with tablespace point in time recovery.
Thanks in advance for any answer.
Regards,
JureIf you have only one business schema, then you may go with duplication of database.
Otherwise, you can go with Schema Based Replication with Streams or as you said combination of expdp/impdp and transport.

Similar Messages

  • Difference Between Database Schema and Database User Account ??

    First i would like to know what an Oracle Database schema exactly is ?
    And what is it for and also What is it's use ?
    Later the exact "Difference Between Database Schema and Database User Account".
    I googled about it but i'm unable to find out it's The difference,
    Please try to explain in a simple manner........
    Thank you very Much in Advance..........

    user13655582 wrote:
    Greate example. but i would like add one more point..
    a user is a schema and schema is a user. but when this applies to the user that he become a schema, only if he has got some objects. so we can say while creating the user which dosent contain any objects called simple user account but afterwords when he has got an objects then we can say its a schema as the above user has shown you through the exampleIt is just word-play. There is nothing in like a 'status' indicator to say "USER_A is just a user but USER_B has become a schema". Many people use the terms "user" and "schema" interchangeably, and some very good DBAs will insist this is correct. I don't disagree with them even though I usually try to make the distinction.
    In a typical application, all of the objects (tables, procedures, etc) that implement the app will be owned by a "application schema owner" - a user which exists for the sole purpose of owning the schema. Said user will have no privileges at all .. especially not CREATE SESSION. Except for possible maintenance activity no one ever actually connects as this user. User accounts are then given to actual humans who connect with their own accounts. The accounts have the necessary system and object privileges granted to them - via a role. One could also create a special account (let's call it APP_ADMIN) that is used for maintenance and batch operations within the application.

  • Difference between Database, instance and schema

    Hi
    I don't know the clear difference between database, instance and schema.. In the broad sense, everything seems to be the same. If anyone can differentiate between them with some example, that will be of great help.
    Thank you.

    Hi,
    In Oracle, people often do use "database" and "instance" interchangeably, or together as one term: "database instance".
    A "schema" is quite different. A schema is a collection of objects (such as tables, indexes or procedures). Every object belongs to exactly one schema, and every schema is a part of exactly one database. (Names may not be unique, however. For example, your database may have one schema named SCOTT, and another schema named USER1. Each of those schemas may have a table named EMP. The EMP table in the SCOTT schema, or SCOTT.EMP, is not the same as the EMP table in the USER1 schema, or USER1.EMP. A different database, my database for example, may also have schemas called SCOTT and USER1, which are completely independent of the SCOTT and USER1 in your database.)

  • Copy whole schema to new database

    Hi...what is ther best way to copy whole schema (25GB) to new database.
    Thanks

    Hello,
    Best way is to use either conventional export/import or datapump.
    exp system/password owner=userA file=userA.dmp log=userA.log
    Note: Make sure tablespace, any special roles or db-link exists before importing or cloning schema
    imp system/password fromUser=userA toUser=userA file=userA.dmp log=userAimport.logSee following link for more infomation and examples to use datapump
    http://www.oracle-base.com/articles/10g/OracleDataPump10g.php
    Regards
    OrionNet

  • Move Data Between Databases

    Dear All,
    I have 2 databases one of them in Egypt and the other in USA.
    I want to transfer data of one schema between the 2 databases.
    I currently use export and import utility to do this action but the size of data is very large so I want an automatic way to do this action.
    what can I use to solve this problem?
    can I use advanced replication or dataguard or what.
    if I use advanced replication data will be transferred through a database link over the internet which will be a very slow operation.
    what can I do,please help.

    Hello Airtifa,
    The main reasons to use a VPN would be
    a) dedicated connection, so more reliable
    b) no or little need to encrypt data, as data doesn't travel across the public Internet
    It should have a positive effect on performance, though you may be suffering from high latency in the network.
    What I'm trying to say is:
    there is not much to tune in a sqlnet connection of whatever type (database links are just ordinary connections), apart from the Session Data Unit, which should be a multiple of the MTU of the network card (so the card doesn't have to fragment packets).
    You are really at the mercy of the (bandwith of the) connection.
    If the connection is slow, the link will be slow.
    If the connection is fast, the link shouldn't pose extra problems.
    Database links aren't slow. Networks can be.
    Also not using the array interface in an efficient way can be a killer.
    Sybrand Bakker
    Senior Oracle DBA

  • Transfer data from one database to another

    Hello,
    We have an application which uses an Oracle 9i server database. On this server we have two databases ( one for test and one for production ).
    From time to time it's required to transfer data from production database to test database.
    The steps followed by the ones which made the transfer are:
    a) drop the user schema ( the one which contains all data)
    b) recreate the user
    c) import the data ( using import utility )
    d) compile invalid object ( what this will mean I don't know)
    It's not possible to delete only the content of the tables and then import data from a dump file? In this way the steps c) and d) will no longer be required. If this is possible can you tell me if is there a command which will delete all data from all table ( almost 2000 tables)?
    Thank you fro your help!

    I think ..following a,b,c,d is the best way to maintain consistency. you are just thinking about the data in the tables, but what about the changes in views,procedures,functions etc....
    compile invalid object ( what this will mean I don't know)create a file xxx.sql with following code and execute it from sql*plus
    set echo off
    set head off
    set feedback off
    set timing off
    spool compile_invalid_obj.sql
    select 'ALTER '||DECODE(OBJECT_TYPE,'PACKAGE BODY','PACKAGE',OBJECT_TYPE)||' '||OBJECT_NAME||' '||DECODE(OBJECT_TYPE,'PACKAGE','COMPILE PACKAGE;','PACKAGE BODY','COMPILE PACKAGE;','COMPILE;')
    from user_objects
    where status = 'INVALID';
    spool off
    set feedback on
    set echo on
    set timing on
    @compile_invalid_obj.sql
    HTH
    --Chaitanya                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Using TOAD and SQL Developer to compare db objects in schemas in databases

    Hi All,
    My primary objective was to compare objects in schemas in two different databases and find out the differences,
    Execute DDL's in the database where objects are missing and syn schemas in two different databases.
    So I need to compare schemas in databases. Which tool would be helpful and will be user friendly to make a comparison of database objects existing in schemas in two different databases.
    I'd like to see if I can get a list of pro and cons between Toad and SQL Developer for comparing schemas pros and cons.
    Could you please also help me on navigation in SQL Developer to compare schemas.
    Connect to Source
    Connect to Target
    Compare schemas with different object types
    Find out differences
    Generate DDL's for the missing objects or for the objects in difference report
    Run them in missing instace(Source/Target)
    Make sure both are in sync.
    Thanks All

    Hi,
    Most dba type tools have this kind of functionality - personally i use plsqldeveloper but this is very similar to most other tools on the market. SQL Developer seems to rely on you needing the change management pack licence which you likely dont have and would be quite a large cost to roll out.
    I'd compare plsqldeveloper/toad/sqlnavigator etc and maybe the tools from redgate which look pretty good though i haven't used them.
    Regards,
    Harry

  • Best way to move LOB between database

    I am using Oracle 10.2.0.4. Please share your experience what is the best to move LOBs between databases. Lob size is 40 gb. I have tried using data pump, and parallel insert with nologging but it is taking almost 1 hour to transfer 1 gb.
    Thanks for any suggestion.

    N Gasparotto wrote:
    For data pump I am using impdp over network_link .Would be much more efficient to copy the file accross the network and then run impdp locally. Did you also tried also parallel impdp ?
    Nicolas.Thank I will be trying your suggestion to move the copy. Currently do to space issue I cannot try it but I have already requested for the designated mount point.
    I tried using parallel hint. Our sever had 8 cpu so I tried parallel 16. Although current parallelism was 16 ,
    worker parallelism was 1. They were only 2 workers and other worker was always waiting. In other words, the two workers did not execute simultaneously. So, I guess data pump did not run parallely.
    Should not more than one worker be executing simultaneously for it to run parallel? I did not understand what worker parallelism meant?

  • Alert log:  failed to establish dependency between database and diskgroup

    Hi, I have a database 11.2.0 with ASM. When I startup an instance I see this error in alert log:
    ERROR: failed to establish dependency between database SPBUFOR and diskgroup resource ora.SPBUFOR_FLASH.dg
    SPBUFOR_FLASH is a diskgroup where a flash recovery area is located.
    Database is opened cleanly but I need to know that everything works fine.
    How can I resolve this issue ?

    ASM ALERT LOG:
    Wed Oct 05 15:35:12 2011
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 3
    Using LOG_ARCHIVE_DEST_1 parameter default value as /u01/app/oracle/product/11.2.0/grid/dbs/arch
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =0
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    NOTE: Volume support enabled
    Starting up:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Automatic Storage Management option.
    Using parameter settings in server-side spfile +CSA_DATA/asm/asmparameterfile/registry.253.758208241
    System parameters with non-default values:
    large_pool_size = 12M
    instance_type = "asm"
    remote_login_passwordfile= "EXCLUSIVE"
    asm_diskgroups = "CSA_FLASH"
    asm_diskgroups = "SPBUFOR_DATA"
    asm_diskgroups = "SPBUFOR_FLASH"
    asm_power_limit = 1
    diagnostic_dest = "/u01/app/oracle"
    Wed Oct 05 15:35:13 2011
    PMON started with pid=2, OS id=24108
    Wed Oct 05 15:35:14 2011
    VKTM started with pid=3, OS id=24112 at elevated priority
    VKTM running at (10)millisec precision with DBRM quantum (100)ms
    Wed Oct 05 15:35:14 2011
    GEN0 started with pid=4, OS id=24118
    Wed Oct 05 15:35:14 2011
    DIAG started with pid=5, OS id=24122
    Wed Oct 05 15:35:14 2011
    PSP0 started with pid=6, OS id=24127
    Wed Oct 05 15:35:14 2011
    DIA0 started with pid=7, OS id=24131
    Wed Oct 05 15:35:14 2011
    MMAN started with pid=8, OS id=24135
    Wed Oct 05 15:35:14 2011
    DBW0 started with pid=9, OS id=24139
    Wed Oct 05 15:35:14 2011
    LGWR started with pid=10, OS id=24143
    Wed Oct 05 15:35:14 2011
    CKPT started with pid=11, OS id=24148
    Wed Oct 05 15:35:14 2011
    SMON started with pid=12, OS id=24157
    Wed Oct 05 15:35:14 2011
    RBAL started with pid=13, OS id=24163
    Wed Oct 05 15:35:14 2011
    GMON started with pid=14, OS id=24173
    Wed Oct 05 15:35:14 2011
    MMON started with pid=15, OS id=24177
    Wed Oct 05 15:35:14 2011
    MMNL started with pid=16, OS id=24181
    ORACLE_BASE from environment = /u01/app/oracle
    Wed Oct 05 15:35:15 2011
    SQL> ALTER DISKGROUP ALL MOUNT
    NOTE: Diskgroups listed in ASM_DISKGROUPS are
    CSA_FLASH
    SPBUFOR_DATA
    SPBUFOR_FLASH
    Diskgroup with spfile:CSA_DATA
    NOTE: cache registered group CSA_DATA number=1 incarn=0x75fe76df
    NOTE: cache began mount (first) of group CSA_DATA number=1 incarn=0x75fe76df
    NOTE: cache registered group CSA_FLASH number=2 incarn=0x760e76e0
    NOTE: cache began mount (first) of group CSA_FLASH number=2 incarn=0x760e76e0
    NOTE: cache registered group SPBUFOR_DATA number=3 incarn=0x760e76e1
    NOTE: cache began mount (first) of group SPBUFOR_DATA number=3 incarn=0x760e76e1
    NOTE: cache registered group SPBUFOR_FLASH number=4 incarn=0x761e76e2
    NOTE: cache began mount (first) of group SPBUFOR_FLASH number=4 incarn=0x761e76e2
    NOTE: Loaded library: /opt/oracle/extapi/64/asm/orcl/1/libasm.so
    NOTE: Assigning number (1,0) to disk (ORCL:CSA_D1)
    NOTE: Assigning number (1,1) to disk (ORCL:CSA_D2)
    NOTE: Assigning number (2,0) to disk (ORCL:CSA_F1)
    NOTE: Assigning number (2,1) to disk (ORCL:CSA_F2)
    NOTE: Assigning number (3,0) to disk (ORCL:SPBUFOR_D1)
    NOTE: Assigning number (3,1) to disk (ORCL:SPBUFOR_D2)
    NOTE: Assigning number (3,2) to disk (ORCL:SPBUFOR_D3)
    NOTE: Assigning number (4,0) to disk (ORCL:SPBUFOR_F1)
    NOTE: Assigning number (4,1) to disk (ORCL:SPBUFOR_F2)
    NOTE: start heartbeating (grp 1)
    kfdp_query(CSA_DATA): 6
    kfdp_queryBg(): 6
    NOTE: cache opening disk 0 of grp 1: CSA_D1 label:CSA_D1
    NOTE: F1X0 found on disk 0 au 2 fcn 0.0
    NOTE: cache opening disk 1 of grp 1: CSA_D2 label:CSA_D2
    NOTE: cache mounting (first) external redundancy group 1/0x75FE76DF (CSA_DATA)
    NOTE: cache recovered group 1 to fcn 0.25573
    NOTE: LGWR attempting to mount thread 1 for diskgroup 1 (CSA_DATA)
    NOTE: LGWR found thread 1 closed at ABA 4.2486
    NOTE: LGWR mounted thread 1 for diskgroup 1 (CSA_DATA)
    NOTE: LGWR opening thread 1 at fcn 0.25573 ABA 5.2487
    NOTE: cache mounting group 1/0x75FE76DF (CSA_DATA) succeeded
    NOTE: cache ending mount (success) of group CSA_DATA number=1 incarn=0x75fe76df
    NOTE: start heartbeating (grp 2)
    kfdp_query(CSA_FLASH): 8
    kfdp_queryBg(): 8
    NOTE: cache opening disk 0 of grp 2: CSA_F1 label:CSA_F1
    NOTE: F1X0 found on disk 0 au 2 fcn 0.0
    NOTE: cache opening disk 1 of grp 2: CSA_F2 label:CSA_F2
    NOTE: cache mounting (first) external redundancy group 2/0x760E76E0 (CSA_FLASH)
    NOTE: cache recovered group 2 to fcn 0.49881
    NOTE: LGWR attempting to mount thread 1 for diskgroup 2 (CSA_FLASH)
    NOTE: LGWR found thread 1 closed at ABA 3.5793
    NOTE: LGWR mounted thread 1 for diskgroup 2 (CSA_FLASH)
    NOTE: LGWR opening thread 1 at fcn 0.49881 ABA 4.5794
    NOTE: cache mounting group 2/0x760E76E0 (CSA_FLASH) succeeded
    NOTE: cache ending mount (success) of group CSA_FLASH number=2 incarn=0x760e76e0
    NOTE: start heartbeating (grp 3)
    kfdp_query(SPBUFOR_DATA): 10
    kfdp_queryBg(): 10
    NOTE: cache opening disk 0 of grp 3: SPBUFOR_D1 label:SPBUFOR_D1
    NOTE: F1X0 found on disk 0 au 2 fcn 0.0
    NOTE: cache opening disk 1 of grp 3: SPBUFOR_D2 label:SPBUFOR_D2
    NOTE: cache opening disk 2 of grp 3: SPBUFOR_D3 label:SPBUFOR_D3
    NOTE: cache mounting (first) external redundancy group 3/0x760E76E1 (SPBUFOR_DATA)
    NOTE: cache recovered group 3 to fcn 0.317867
    NOTE: LGWR attempting to mount thread 1 for diskgroup 3 (SPBUFOR_DATA)
    NOTE: LGWR found thread 1 closed at ABA 3.8570
    NOTE: LGWR mounted thread 1 for diskgroup 3 (SPBUFOR_DATA)
    NOTE: LGWR opening thread 1 at fcn 0.317867 ABA 4.8571
    NOTE: cache mounting group 3/0x760E76E1 (SPBUFOR_DATA) succeeded
    NOTE: cache ending mount (success) of group SPBUFOR_DATA number=3 incarn=0x760e76e1
    NOTE: start heartbeating (grp 4)
    kfdp_query(SPBUFOR_FLASH): 12
    kfdp_queryBg(): 12
    NOTE: cache opening disk 0 of grp 4: SPBUFOR_F1 label:SPBUFOR_F1
    NOTE: F1X0 found on disk 0 au 2 fcn 0.0
    NOTE: cache opening disk 1 of grp 4: SPBUFOR_F2 label:SPBUFOR_F2
    NOTE: cache mounting (first) external redundancy group 4/0x761E76E2 (SPBUFOR_FLASH)
    NOTE: cache recovered group 4 to fcn 0.16114
    NOTE: LGWR attempting to mount thread 1 for diskgroup 4 (SPBUFOR_FLASH)
    NOTE: LGWR found thread 1 closed at ABA 2.1922
    NOTE: LGWR mounted thread 1 for diskgroup 4 (SPBUFOR_FLASH)
    NOTE: LGWR opening thread 1 at fcn 0.16114 ABA 3.1923
    NOTE: cache mounting group 4/0x761E76E2 (SPBUFOR_FLASH) succeeded
    NOTE: cache ending mount (success) of group SPBUFOR_FLASH number=4 incarn=0x761e76e2
    kfdp_query(CSA_DATA): 13
    kfdp_queryBg(): 13
    NOTE: Instance updated compatible.asm to 11.2.0.0.0 for grp 1
    SUCCESS: diskgroup CSA_DATA was mounted
    kfdp_query(CSA_FLASH): 14
    kfdp_queryBg(): 14
    NOTE: Instance updated compatible.asm to 11.2.0.0.0 for grp 2
    SUCCESS: diskgroup CSA_FLASH was mounted
    kfdp_query(SPBUFOR_DATA): 15
    kfdp_queryBg(): 15
    NOTE: Instance updated compatible.asm to 11.2.0.0.0 for grp 3
    SUCCESS: diskgroup SPBUFOR_DATA was mounted
    kfdp_query(SPBUFOR_FLASH): 16
    kfdp_queryBg(): 16
    NOTE: Instance updated compatible.asm to 11.2.0.0.0 for grp 4
    SUCCESS: diskgroup SPBUFOR_FLASH was mounted
    SUCCESS: ALTER DISKGROUP ALL MOUNT
    SQL> ALTER DISKGROUP ALL ENABLE VOLUME ALL
    SUCCESS: ALTER DISKGROUP ALL ENABLE VOLUME ALL
    NOTE: diskgroup resource ora.CSA_DATA.dg is online
    NOTE: diskgroup resource ora.CSA_FLASH.dg is online
    NOTE: diskgroup resource ora.SPBUFOR_DATA.dg is online
    NOTE: diskgroup resource ora.SPBUFOR_FLASH.dg is online
    Wed Oct 05 15:35:44 2011
    Starting background process ASMB
    Wed Oct 05 15:35:44 2011
    ASMB started with pid=18, OS id=24330
    Thu Oct 06 11:48:19 2011
    SQL> alter diskgroup SPBUFOR_DATA check all
    NOTE: starting check of diskgroup SPBUFOR_DATA
    kfdp_checkDsk(): 17
    kfdp_checkDsk(): 18
    Thu Oct 06 11:48:30 2011
    kfdp_checkDsk(): 19
    SUCCESS: check of diskgroup SPBUFOR_DATA found no errors
    SUCCESS: alter diskgroup SPBUFOR_DATA check all
    Thu Oct 06 11:48:56 2011
    SQL> alter diskgroup SPBUFOR_FLASH check all
    NOTE: starting check of diskgroup SPBUFOR_FLASH
    kfdp_checkDsk(): 20
    kfdp_checkDsk(): 21
    SUCCESS: check of diskgroup SPBUFOR_FLASH found no errors
    SUCCESS: alter diskgroup SPBUFOR_FLASH check all

  • How to export and import from AIX to solaris a whole schema with different

    how to export and import from AIX to solaris a whole schema with different character set
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P1 character set (possible charset conversion)
    Import terminated successfully with warnings.

    The conversion between the character sets on the two servers are handled automatically. Export from one server will export the data given the character set defined on that server. Import into another server will import the data given the character set defined on that server, performing a conversion if necessary. This can be a problem when going from single-byte to a multi-byte character set. In your case, the warning is simply for your information, shoule be no real problem given the character sets you're using.
    For more info...check out (pertains to 9i):
    Export
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96652/ch01.htm#1006791
    Character Ste Migration
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96529/ch10.htm#1656
    HTH

  • Transforming the Raster Data between Databases

    Hi
    I have a Requirement to Import and Export The Raster Data Between Data bases .. Could Any Body Give Me Hint...
    Rajanikanthc

    NedimAT pointed you to the 10g manual. If your database is 11g, please check it out from the 11g manual, Section 3.21 Transferring GeoRaster Data Between Databases:
    http://download.oracle.com/docs/cd/E11882_01/appdev.112/e11827/geor_operations.htm#CIHECFGC
    Depending on your requirements and the source and target database environments, you may also consider using Transportable Tablespace to transfer the raster data, particularly when you want to transfer only part of the database based on queries or cropping areas. Check section 3.22:
    http://download.oracle.com/docs/cd/E11882_01/appdev.112/e11827/geor_operations.htm#CHDGEGHB
    hope this helps
    Jeffrey

  • Possible to replicate cn=schema between masters that don't share a LDBM?

    I have multiple directory servers in my environment. One set has suffix A, another set has suffix B and a 3rd set has suffix C. I would like to replicate just the cn=schema between all of the masters for consistency reasons. Is it possible to replicate cn=schema between masters that don't share or replicate a user created database with each other?
    These servers are DSEE 6.3.1 and are all controlled by the same DSCC.
    Thank you,
    Matt

    I already know how to manually copy the 99user.ldif file to all of the servers. I am trying to get away from having to manually copy schema changes between servers every time a new attribute or objectclass is added to one set of servers.
    Is there a way to replicate the cn=schema DIT alone?

  • Replication based on condition for whole Schema

    Hi There
    Oracle 11.1.0.7
    We are using oracle spatial and we have got huge tables of size more than 50 GB . I just started configuring GG on test and i would like to know if it suits our requirement.
    In this regards i have certain queries :
    1) Can i configure GG for whole schema based on conditions , for example in the schema there are 300 tables , out of those i want all 250 to replicate at the target as is and the rest 50 should replicate on where approved = 'YES' . Is this possible ?
    2) If its possible to filter data during replication , i would like to know if we can filter using a query which has a 'JOIN'
    3) Should the target database (Oracle 11.1.0.7) have the tables with the data as on before starting GG so that when we start replicating both source and target should be exactly the same.
    I appreciate if some one can post a response.
    Thanks

    You can filter data using a WHERE clause, or even using SQLEXEC, mapping OUT values to target columns. A join could be handled in the output from a stored procedure.
    You can do all tables or a portion thereof. If names are similar, they can be wildcarded. Otherwise, list them.
    Not sure what you mean by APPROVED.
    For initial load, the stock answer is this:
    GoldgenGate recommends using native RDBMS tools for the initial load. Restore a backup, that kind of drill. You can use one of the four GG initial load options, but given the size you mentioned, it would probably be much faster to create the target via Oracle Data Pump or RMAN.

  • Transfer of Goods between Center (2 Steps MM-SD)

    Hi Friends;
    I have a little problem to which I have done several tests and did not have much success to complete the activities related to the transfer of goods between centers using the procedure I describe below:
    - ME21N -> Creation of Purchase for transfer using the category UB (Stock Transport Order);
    - VL10B -> Process documents for Shipping (using Purchase Order);
    - VL02N ->  To Process Delivery  (Picking and deposit);
    - J1B3N -> Issue of the tax document (invoice);
    - MB0A -> Entry of Goods to Purchase Order;
    The problem is that when I am doing the PO (First part of the process) I notify the PO type, in the field I meet the supplier center provider, I meet the item and the center of will receipt the goods. After all information completed on the screens of Purchase Order (ME21N) the system displays the following message.
    u201CNot possible to determine shipping data for material 9999999u201D
    u201CMessage no. 06280u201D
    As this is the beginning all of process, I am not able to continue. I already made several checks which I will describe below, and still am not successful in completing the process.
    u2022     The material is already expanded to the Center Supplier and Receiver;
    u2022     The material is already expanded to SD View, with the sales organization, distribution channel and sector of activity properly;
    u2022     Customer (center) already expanded to the Organization of Sales and Channel Distribution properly;
    u2022     On the table is expanded KNVV you both for the organization's sales center provider for the center as recepitor;
    u2022     On IMG:
        o     It was the assign of the sales organization of the center (Materials Management -> Purchasing -> Purchase Order -> Set up     
            Transport Order -> Define Shipping Data for Plants);
        o     It  was assigned to the Shipping Points (Logistics Execution -> Shipping -> Basic Shipping Functions-> Shipping points and
            Good Receiving Point Determination -> Assign shipping Point). Therefore the u201CShipping Conditionsu201D is properly linking to
            Custumer Data Master and  Loading Group is properly to Goods Master Data.
    Even with all these settings made, I am not having success in the process of creating of PO  and start the procedure for transfer of material between centers and so I ask the help of colleagues to help me find what is still missing or set up in process that can make it flow.
    Beforehand, Thanks a lot.
    Vinicius

    Hi ,
    GOTO IMG -> MM->PURCHASING ->Purchase Order ->Set up Stock Transport Order->Define Shipping Data for Plants-
    here for both receiving and supplying plant give the same sales org , distridution channel and division .
    And in receiving plant maintain the customer no .
    And one more thing do the PO as document type NB (standard order ) , then maintain delivery type NLCC against NB in customizing (GOTO IMG -> MM->PURCHASING ->Purchase Order ->Set up Stock Transport Order ->Assign Delivery Type and Checking Rule )
    Regards,
    Anupam
    Edited by: Anupam Halder on Jul 28, 2009 7:07 PM

  • How do you get your old scores on your games when you transfer the games between devices ive signed into game center and my scores are there but when I open the games Im at 0 again

    how do you get your old scores on your games when you transfer the games between devices ive signed into game center and my scores are there but when I open the games Im at 0 again

    Did you do this?
    iOS: Transferring information from your current iPhone, iPad, or iPod touch to a new device

Maybe you are looking for

  • HOW TO PASS THE DATA FROM SELECTION SCREEN TO STANDARD TRANSACTION?

    HI, HOW TO PASS THE DATA FROM SELECTION SCREEN TO STANDARD TRANSACTION? thanks, samba.

  • Flash chart with dynamic sql query does not display

    Hi, In my schema SIVOA I have a lot of tables declared like this :   CREATE TABLE "SIVOA"."EVV_xxxx"    (     "CLEF_VAR" NUMBER(4,0),      "DATE1" DATE,      "VALEUR" NUMBER(16,8) Only the last part of the name changes "xxxx". For example E009, E019,

  • Formula variable for date in report

    Hi All, I have to display the last date value of a particular date in the report. For a better understanding, I have Customer, Equipment, Service date, Amount In the report, I have to display Customer, Equipment, Last Service date, Last Service Amoun

  • Multi-page images?

    Post Author: martorell CA Forum: General Hi!I try to insert a multipage image (tiff) and CR only shows the first image of file. Is possible show all tiff images? Thanks!!

  • Kernel panic "kernel NULL pointer dereference in pacpi_set_dmamode"

    Hi. After upgrade to linux-3.6.2-1 my PC fails to boot with this panic: http://imageshack.us/photo/my-images/5/imag0085.jpg/ How do I fix it? I can provide more specific info if needed. Last edited by thesame (2012-10-31 20:29:41)