Impdp DB link

Hi,
I'm going to test using impdp through db_link(to source db)
1. Can I do a full import but exclude schemas as normal impdp/expdp?
2. Will TRANSFORM=SEGMENT_ATTRIBUTES:n create objects in the default schema for a user?
2b. If I'd like all Indexes in a index tablespace and other in data tablespace what would be the command?
Would this command work?
impdp system@database full=y network_link=REMOTE_DB directory=data_pump logfile=impdpfull.log exclude=schema:IN ('SYS','SYSTEM') flashback_time=\"to_timestamp\(\'$DATE\',\'yyyy-mm-dd hh24:mi:ss\'\)\" TRANSFORM=SEGMENT_ATTRIBUTES:n
Regards
933746

933746 wrote:
hitgon wrote:
impdp -help
Edited by: hitgon on Jun 26, 2012 3:19 PMHi, Thanks for your feedback.
Ok to be more concrete will TRANSFORM=SEGMENT_ATTRIBUTES:n just remove tablespace clause from create statment in impdp and therefor all objects will be created as default for their respective users?
Regards
933746Just to add an update to this thread:
TRANSFORM=SEGMENT_ATTRIBUTES:n will remove storage and tablespace clause and create objects in default tablespace.

Similar Messages

  • Importing a few Oracle 10g schemas into oracle 11g database

    Hi,
    I have the task to import a few schemas that are on oracle 10g database (10.2.0.4) into oracle 11gR2 database (11.2.0.2). Now I have only oracle 10g export and import utility (the 11g server is in not where I am given the import rights).
    Can I do the export of the 10g schemas using the 10g export utility and then import into 11g database using the 10g import utility? I ran a test to do that and found it works but would also like to have expert opinion on it.
    Also another question is: if the source schema is 11g - suppose- in that case can the 10g export/import utility still work? I mean export the schema from say db1 (which is 11g) using 10g export utility and import them into 11g database db2 (which is again 11g) using 10g import? I am asking because I don't have access to 11g server and trying to find what I can do with the 10g utilities.
    Thanks,

    set up a database link to the 10gR2 version database in the 11gR2 database.
    Now using the 11gR2 client impdp utility use
    impdp network_link=<link to 10gR2 database>
    Done.
    Your 'objection' you don't have access to 11gR2 doesn't count as you can install a 11gR2 client.
    Also generally speaking you should not use a lower version imp to imp to a higher version database
    Also exp and imp are deprecated in 10g and higher.
    Sybrand Bakker
    Senior Oracle DBA

  • Expdp showing ORA-31655

    Hi,
    whenever i tried to export using expdp i got ORA-31655.
    Please find detail snapshot below:
    $ expdp TABLESPACES=DP_TEST file=data_expdp.dmp directory=DMP_DIR;
    Export: Release 11.2.0.1.0 - Production on Tue Jan 19 16:30:26 2010
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Username: sys@MYDB as sysdba
    Password:
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Legacy Mode Active due to the following parameters:
    Legacy Mode Parameter: "file=data_expdp.dmp" Location: Command Line, Replaced with: "dumpfile=data_expdp.dmp"
    Legacy Mode has set reuse_dumpfiles=true parameter.
    Starting "SYS"."SYS_EXPORT_TABLESPACE_01": sys/********@MYDB AS SYSDBA TABLESPACES=DP_TEST dumpfile=data_expdp.dmp directory=DMP_DIR reuse_dumpfiles=true
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    ORA-31655: no data or metadata objects selected for job
    Job "SYS"."SYS_EXPORT_TABLESPACE_01" completed with 1 error(s) at 16:30:42
    The same command is working file in another(fresh 11.2.0.1) database
    We have just upgraded our 10.2.01 database to 11.2.0.1 and compatible parameter has been set to11.2.0.1.0.
    I also tried with include parameter but still showing ORA-31655.
    Please help :(

    Hi Anurag,
    Thanks a lot.
    Problem has been solved.
    I did all the tests using SYS (test object creation, expdp ....) which was wrong.
    During my tests All the objects were created unde SYS and SYS is never exported.
    expdp/impdp: database links  not imported
    I created some objects in other schema and finally expdp has been done successfully.
    Thanks
    Obaid

  • Expdp / impdp - Private synonyms using DB link not imported

    Hi
    I'd appreciate any help with this.
    We are taking an export (expdp) of a database on Oracle EE 10.2.0.3 in Prod and then importing it into an identical DB on 10.2.0.5 in Dev.
    I doubt the minor version difference is an issues, but I mention it for completeness.
    Our expdp in prod used to look like this:
    expdp agdba/x directory=DATA_PUMP_DIR dumpfile=X.DMP logfile=X.LOG schemas=A,B,C,D
    We have changed it to:
    expdp agdba/x directory=DATA_PUMP_DIR dumpfile=X.DMP logfile=X.LOG full=y EXCLUDE=SCHEMA:"IN ('E', 'F')"
    ( so basically do a full export, but exclude schemas E & F )
    The impdp in dev has not changed (we import one schema at a time):
    impdp agdba/x DUMPFILE=X.DMP LOGFILE=imp_X.LOG SCHEMAS=A DIRECTORY=DATA_PUMP_DIR TABLE_EXISTS_ACTION=SKIP
    DB user AGDBA has been granted the DBA role...
    The Issue:_
    Private synonyms that use a DB link are NOT imported, but private synonyms that do NOT use a DB link ARE imported.
    I can fix the issue by simply recreating the private synonyms using the DB link - e.g.:
    CREATE SYNONYM EVENTLOGTBL FOR EVENTLOGTBL@FCISTOSMS;
    Things were working fine until we change the expdp to use FULL. I see a few posts about synonyms and FULL=Y issues, but nothing quite like our problem.
    Any ideas?
    Thanks,
    Andreas

    Andreas Hess wrote:
    So the problem is expdp FULL=Y for some reason does not export synonyms that refer to objects via DB links. I doubt that ... More than few times a week I refresh non-prod databases from a prod db, (it is schema level export not FULL=Y), and two of the schema own db links, and others have synonyms pointing to them. I have never encountered synonym problem (except that if db_links are invalid then impdp takes extremely long time to timeout while compiling pl/sql code.). We normally change db_links after the impdp is done, or (while it is importing table via another session).
    I just re-ran sqlfile option with include=db_link,synonym option and i can see statements that create synonyms and create db links.
    However i think the order in which impdp runs might be the source of your problem ... see this
    Starting "ME"."SYS_SQL_FILE_SCHEMA_01":  ME/******** dumpfile=xxxx.%u.dmp directory=xxxx_exp schemas=schema1,schema2,schema3 sqlfile=schema.sql include=db_link,synonym
    Processing object type SCHEMA_EXPORT/SYNONYM/SYNONYM       <<<< Synonyms are created first
    Processing object type SCHEMA_EXPORT/DB_LINK               <<<< db_links come later
    Job "ME"."SYS_SQL_FILE_SCHEMA_01" successfully completed at 08:32:02So, it is conceivable that if you drop synonyms/db_links from your schemas before impdp, some synonyms will fail to create properly. This could be a version specific issue, mine is 11.2 and I don't see failures for synonym creation, i.e. no error messages while impdp. Next time you could try to drop objects except db_links and see if you still have the same issue.
    Raj

  • Moving Large amount of data using IMPDP with network link

    Hi Guru,
    Here we are having a requirement to move 2TB of data from production to non-prod using Network_link parameter. What is the process to make it fast.
    Previously we did it but it took 7 days for importing data and index .
    Here i am having an idea can you please guide me is it good to make import faster .
    Step 1) import only metadata .
    Step 2) import only table data using table_exists_action=append or truncate.( Here indexes are allready created in step 1 and import will be fast as per my plan.)
    Please help me the better way if we can.
    Thanks & Regards,
    Venkata Poorna Prasad.S

    You might want to check these as well:
    DataPump Import (IMPDP) Over NETWORK_LINK Is Sometimes Very Slow (Doc ID 1439691.1)
    DataPump Import Via NETWORK_LINK Is Slow With CURSOR_SHARING=FORCE (Doc ID 421441.1)
    Performance Problems When Transferring LOBs Using IMPDP With NETWORK_LINK (Doc ID 1488229.1)

  • ORA-00001: unique constraint @ impdp with table_exists_action=truncate

    Hi everybody
    I can't understand why my data pump import execution with parameter TABLE_EXISTS_ACTION=TRUNCATE returned ORA-00001 (unique constraint violation), while importing schema tables from a remote database where the source tables have the same unique constraints as the corresponding ones on the target database.
    Now my question is "If the table would be truncated, why I get unique constraint violation when inserting records from a table where the same unique constraint is validated?
    Here are the used parameter file content and the impdp logfile.
    parfile
    {code}
    DIRECTORY=IMPEXP_LOG_COLL2
    CONTENT=DATA_ONLY
    NETWORK_LINK=PRODUCTION
    PARALLEL=1
    TABLE_EXISTS_ACTION=TRUNCATE
    EXCLUDE=STATISTICS
    {code}
    logfile
    {code}
    Import: Release 10.2.0.1.0 - Production on Gioved� 22 Ottobre, 2009 15:33:44
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Connesso a: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    FLASHBACK automatically enabled to preserve database integrity.
    Starting "IMPEXP"."PROVA_REFRESH_DBCOLL": impexp/********@dbcoll SCHEMAS=test_pump LOGFILE=test_pump.log parfile=refresh_dbcoll.par JOB_NAME=prova_refresh_dbcoll
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 523 MB
    ORA-31693: Table data object "TEST_PUMP"."X10000000_TRIGGER" failed to load/unload and is being skipped due to error:
    ORA-00001: unique constraint (TEST_PUMP.SYS_C00726627) violated
    ORA-31693: Table data object "TEST_PUMP"."X10000000_BASIC" failed to load/unload and is being skipped due to error:
    ORA-00001: unique constraint (TEST_PUMP.SYS_C00726625) violated
    Job "IMPEXP"."PROVA_REFRESH_DBCOLL" completed with 2 error(s) at 15:34:04
    {code}
    Thank you
    Bye Alessandro

    I forgot to read the last two lines of the documentation about TABLE_EXISTS_ACTION where it says:
    "TRUNCATE cannot be used on clustered tables or over network links."
    So it seems that it ignored the clause for the use of NETWORK_LINK and unreasonably opted for an APPEND action instead of throwing an error to highlight the conflicting parameters in the used configuration.
    Bye Alessandro

  • How to filter parent and child tables data in impdp?

    Hi Friends,
    i hope everyone doing good..
    I have two tables in my schema like below.
    STUDENT_DTLS:-
    SID Sname CID
    1     venkat     001
    2     uppy     002
    3     deepu     003
    4     trinath     003
    5     naveen     004
    6     ramesh     005
    COURSE_DTLS:-
    CID CNAME Fee
    001     Oracle     2000
    002     java     6000
    003     sieble     4000
    004     informatica     3000
    005     obiee 4000
    if I load only few Student details say sid=1,2,3 into another environment using impdp,corresponding to those students course details from COURSE_DTLS table only need to be loaded.Not all course details.
    Iam Expecting Result in new schema as below:-
    SID Sname CID
    1     venkat     001
    2     uppy     002
    3     deepu     003
    CID CNAME Fee
    001     Oracle     2000
    002     java     6000
    003     sieble     4000
    But when i am using the below import scrit,The COURSE_DTLS is loading first and its query option(select cid from STUDENT_DTLS) is returning no rows.Because at that time no data is insertred into STUDENT_DTLS yet.
    secondly it is importing data into STUDENT_DTLS with query option where sid=in (1,2,3) which is loaded propelry.
    Thirdly it is trying to enable primary key and Foreign key constraints which is also failing because of no data in COURSE_DTLS table.
    i have changed the order of tables in TABLES clause like tables=COURSE_DTLS ,STUDENT_DTLS.But still COURSE_DTLS table is loading first.
    impdp pa_data_sub/******* NETWORK_LINK=qdssih30 tables=STUDENT_DTLS,COURSE_DTLS query='STUDENT_DTLS:"where SID IN(1,2,3)"','COURSE_DTLS:"where cid in(select cid from STUDENT_DTLS)"', CONTENT=all table_exists_action=replace directory=DATA_PUMP_DIR logfile=exp_join.log
    output:
    Total estimation using BLOCKS method: 128 KB
    Processing object type TABLE_EXPORT/TABLE/TABLE
    . . imported "PA_DATA_SUB"."COURSE_DTLS" 0 rows
    . . imported "PA_DATA_SUB"."STUDENT_DTLS" 3 rows
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type TABLE_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Processing object type TABLE_EXPORT/TABLE/CONSTRAINT/REF_CONSTRAINT
    ORA-39083: Object type REF_CONSTRAINT failed to create with error:
    ORA-02298: cannot validate (PA_DATA_SUB.SYS_C0040905) - parent keys not found
    Failing sql is:
    ALTER TABLE "PA_DATA_SUB"."STUDENT_DTLS" ADD FOREIGN KEY ("CID") REFERENCES "PA_DATA_SUB"."COURSE_DTLS" ("CID") ENABLE
    Processing object type TABLE_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "PA_DATA_SUB"."SYS_IMPORT_TABLE_01" completed with 1 error(s) at 17:28:13
    Please suggest me how to achieve this task in a single import command?
    Is there any option in impdp to prioritize the table to be loaded first?
    Thanks,
    Venkat Vadlamudi

    You can't be guaranteed which order the data will be loaded in, so using a query that examines a table that is being imported is not going to always work the way you want it. The failing constraints makes sense because there is no data.
    So, in your case, course_dtls gets imported first and the table student-dtls exists, but it not populated yet, so the select cid from student_dtls comes back with 0 rows. The second query works fine.
    You might be able to change the query to be:
    'COURSE_DTLS:"where cid in(select cid from STUDENT_DTLS@qdssih30 where SID IN(1,2,3)"'
    This way, you query the source table pointed to by the network link and you use the same subquery from the first table.
    This should work for you... I think anyway.
    Dean

  • Changing database to noarchivelog mode during impdp

    Hi Guys,
    I am using Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bit Production on RHEL5.
    I am in the middle on the impdp and I want my database to run on noarchivelog as it creates a lot of archivelogs.
    Will it be proper if I stop the import:
    e.g Import> STOP_JOB=IMMEDIATE
    and shutdown the database put the database on noarchivelog mode.
    then opne database and
    impdp hr/hr@mydb ATTACH=myfulljob Will these work fine?
    Or should I execute:
    alter system archive log stop;
    Please Help !!!!!!!!!!!

    Vinod Dhandapani wrote:
    Running the database in Noarchivelog mode means you should be ready to lose data incase there is any crash. I would recommend you to change the size of the redologs so that it will not create frequent archivelogs. Which does nothing about the space required for the archivelogs. If you are generating redo at the rate of 1gb/hour, you will consume archivelog dest space at the rate of 1gb/hour. Doesn't matter if those files are 10mb each or 100mb each.
    If space is the problem then use a shell script to move the archivelog to another location (Tape)....And hopefully that shell script is invoking rman to do the work, so that in the event of recovery, rman will know that backups were taken and where they are located.
    >
    Incase you want to convert the database from archivelog to noarchivelog, follow the steps provided in the following link.
    http://racadmin.blogspot.com/2011/06/convert-archivelog-mode-to-noarchivelog.html
    Thanks

  • How can I replace existing oracle-sequences using impdp?

    I use impdp to transport table data from one database-instance to another (on a regular base).
    I do this by using a database link.
    Here is the content of the parameterfile for the impdp:
    NETWORK_LINK=dbl_backoffice
    INCLUDE=TABLE:" in ('SSU_CUSTOMERS','SSU_ORDERS','SSU_ORDERLINES')"
    INCLUDE=SEQUENCE:" in ('SEQ_ORDERNR','SEQ_CUSTNR')"
    TABLE_EXISTS_ACTION=REPLACE
    Now, the mentioned tabels are neatly replaced however, the sequence-objects aren't because impdp reports :
    ORA-31684: Object type SEQUENCE:"SSU"."SEQ_ORDERNR'" already exists
    ORA-31684: Object type SEQUENCE:"SSU"."SEQ_CUSTNR'" already exists
    It is understandable, because the impdp will issue create sequence statements and fails because they're already there (and there is no such thing as a import-parameter as SEQUENCE_EXISTS_ACTION=REPLACE or whatsoever).
    I don't want to do a complete schema import, since I only want to refresh front-office data with a part of the back-office data.

    You need a plugin called Flip4Mac to create a wmv on a Mac
    It plugs in to Quicktime.
    WMV is generally unavailable on MAcs cos its a Windows thing ( Windows Media Video)
      just like PorRes is not available generally on a Windows system.

  • Netwrok link

    Hai
    i have a doubt regarding the import behaviour.please clear it.i have pasted it below.
    =========See behaviour of IMPDP while using NETWORK_LINK parameter.
    Say you have DB1 on Host1 and DB2 on Host2. Now, you want to import Table1 from DB1 to DB2.
    Table1 is owned by User1 at DB1 and you want it to be owned by USER2 at DB2.
    1. Create database link at DB2 which connects to DB1 and uses USER1 for connecting. TNSNAME for connection from Host2 to Host1 is DB1.WORLD.COM:
    create database link mylink connect tu user1 identified by 'user1_password' using 'DB1.WORLD.COM';
    Here DB1.WORLD.COM is tns entry in tnsnames.ora at HOST2.
    2. Execute IMPDP:
    impdp user2/user2_password TABLES = TABLE1 DIRECTORY=MYDIR NETWORK_LINK=MYLINK;
    NETWORK_LINK is the name of database link in the database where IMPDP connects to.
    So, impdp connect to DB2, search for Database link named 'MYLINK' and connects to DB1 and selects Table1 from DB1(Oracle connects to DB1 using connection properties of DBLINK - in our case, username is USER1, password is USER1_PASSWORD, TNSNAME is DB1.WORLD.COM).
    ====================================================
    i read the following in the forum.here i have a doubt that here in import what does directory denotes.does it denote the location where the export dumpfile will bw placed and will the user who import the data will be having access on the directory.please say me.is it necessary to mention this directory while importing.
    actully my doubt is that the source directory will be in one server where the dumpfile will be placed and the server from where we runthe import will be different.so how does the import utility in some other server recognize the dumpfile existing in the directory in some other server for performing the import operation.please clear my doubt

    i have 2 servers A and B.A is the source server having db1 and B is the server having DB2.the schema existing in server A is RAJ.the user performs the export of the employees table on the source server A like this
    create directory a1 as '/u01/app';
    grant read,write on directory a1 to raj
    expdp raj/raj schemas=raj tables=employees dumpfile=exp1.dmp directory=a1
    this command peroforms the export and creates the dumpfile and places it in the a1 directory.
    the user created a link as sid in tnsnames.ora of source instance is db1
    create database link mks connect to raj identified by raj using db1;
    so this "create database link" command we have to implement in the source where i have performed the export
    now i have to import the data in the db2 database in the target server.i have to give in the target database like this
    impdp raj/raj tables=employees directory=a1 network_link=mks
    so when i implement this in target server the data will be imprted in the target server through the link
    here my doubt is that whether we have to implement the database link creation in the source server?
    and is it neccessary that the import query directory name should match the directory existing in the source server that is 'a1'

  • Expdp full and impdp full fails with error....pls help..!(Oracle 11g)

    Hi
    I wanted to refresh my dev environment quickly..
    As we also wanted to migrate from exp/imp to datapump, I tried these..
    Only thing worked for me is impdp on network mode using remap schema.
    1)First i wanted to implement expdp and impdp standard procedures(Full database export/import)
    My expdp failed with the following error on a full export..
    Pls see below:
    C:\Documents and Settings\oracle2>expdp system/<p.w> full=Y directory=DATA_PUMP_DIR dumpfile=OAPFULL.dmp logfile=OAPFULL.log
    Export: Release 11.1.0.7.0 - Production on Thursday, 26 April, 2012 16:12:49
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYSTEM"."SYS_EXPORT_FULL_02":  system/******** full=Y directory=DATA_PUMP_DIR dumpfile=OAPFULL.dmp logfile=OAPFULL.log
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 4.351 GB
    Processing object type DATABASE_EXPORT/TABLESPACE
    Processing object type DATABASE_EXPORT/PASSWORD_VERIFY_FUNCTION
    Processing object type DATABASE_EXPORT/PROFILE
    Processing object type DATABASE_EXPORT/SYS_USER/USER
    Processing object type DATABASE_EXPORT/SCHEMA/USER
    Processing object type DATABASE_EXPORT/ROLE
    Processing object type DATABASE_EXPORT/GRANT/SYSTEM_GRANT/PROC_SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/GRANT/SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/ROLE_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/DEFAULT_ROLE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLESPACE_QUOTA
    Processing object type DATABASE_EXPORT/RESOURCE_COST
    Processing object type DATABASE_EXPORT/SCHEMA/DB_LINK
    Processing object type DATABASE_EXPORT/TRUSTED_DB_LINK
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/SEQUENCE
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/DIRECTORY/DIRECTORY
    Processing object type DATABASE_EXPORT/DIRECTORY/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/CONTEXT
    Processing object type DATABASE_EXPORT/SCHEMA/PUBLIC_SYNONYM/SYNONYM
    Processing object type DATABASE_EXPORT/SCHEMA/SYNONYM
    ORA-39014: One or more workers have prematurely exited.
    ORA-39029: worker 1 with process name "DW01" prematurely terminated
    ORA-31672: Worker process DW01 died unexpectedly.
    Job "SYSTEM"."SYS_EXPORT_FULL_02" stopped due to fatal error at 16:14:562)As the above failed, I tried impdp with network_link import and full=y and its a different issue!
    impdp system/<p.w> NETWORK_LINK=OAPLIVE.WORLD full=y logfile=OAPDRfull2.log table_exists_
    action=REPLACE
    Import: Release 11.1.0.7.0 - Production on Thursday, 26 April, 2012 13:36:01
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYSTEM"."SYS_IMPORT_FULL_01":  system/******** NETWORK_LINK=OAPLIVE.WORLD full=y logfile=OAPDRfull2.log table_exists_action=REPLACE
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 4.670 GB
    Processing object type DATABASE_EXPORT/TABLESPACE
    ORA-31684: Object type TABLESPACE:"SYSAUX" already exists
    *************lots more of object already exists errors here............(space limit so cant paste)***************
    ORA-31684: Object type SYNONYM:"GEOPROD"."EXT_POSTAIM_PRESORT_61" already exists
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.DISPATCH_WORK_ITEMS [SYNONYM:"GEOPROD"."EXT_POSTAIM_PRESORT_61"]
    ORA-31600: invalid input value 200001 for parameter HANDLE in function CLOSE
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.DBMS_METADATA", line 569
    ORA-06512: at "SYS.DBMS_METADATA", line 4731
    ORA-06512: at "SYS.DBMS_METADATA", line 792
    ORA-06512: at "SYS.DBMS_METADATA", line 4732
    ORA-06512: at "SYS.KUPW$WORKER", line 2718
    ORA-03113: end-of-file on communication channel
    ORA-02055: distributed update operation failed; rollback required
    ORA-02063: preceding lines from OAPLIVE.WORLD
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPW$WORKER", line 7858
    ----- PL/SQL Call Stack -----
      object      line  object
      handle    number  name
    242F2F2C     18256  package body SYS.KUPW$WORKER
    242F2F2C      7885  package body SYS.KUPW$WORKER
    242F2F2C      8657  package body SYS.KUPW$WORKER
    242F2F2C      1545  package body SYS.KUPW$WORKER
    241DDF3C         2  anonymous block
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.UNLOAD_METADATA []
    ORA-31642: the following SQL statement fails:
    SELECT unique ku$.seq# from sys.metanametrans$ ku$ WHERE ku$.htype='DATABASE_EXPORT' AND ku$.model='ORACLE' AND NOT ( ku$.seq#>=(select a.seq# from sys.metanametrans$ a where
       a.model='ORACLE' and a.htype='DATABASE_EXPORT' and a.name ='DATABASE_EXPORT/SCHEMA/SYNONYM')) order by ku$.seq#
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.DBMS_METADATA_INT", line 5002
    ORA-01427: single-row subquery returns more than one row
    ORA-06512: at "SYS.DBMS_
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.KUPW$WORKER", line 7853
    ----- PL/SQL Call Stack -----
      object      line  object
      handle    number  name
    242F2F2C     18256  package body SYS.KUPW$WORKER
    242F2F2C      7885  package body SYS.KUPW$WORKER
    242F2F2C      2744  package body SYS.KUPW$WORKER
    242F2F2C      8523  package body SYS.KUPW$WORKER
    242F2F2C      1545  package body SYS.KUPW$WORKER
    241DDF3C         2  anonymous block
    Job "SYSTEM"."SYS_IMPORT_FULL_01" stopped due to fatal error at 13:39:48So its basically these 2 issues I want to know how to fix.
    1-expdp error cause and fix
    2-impdp error cause and fix -Also how to avoid object alreay exists error?
    Also for example the package etc. is the same name..but i want new package from LIVE so it means if the same name package, view, etc is there, it wouldnt get updated?
    Any way to overcome this?
    I need it exactly same as LIVE...(with a few exceptions which is small enough i can do after impdp finishes fine)
    Pleaseeeeeeee help!!
    Thanks&Regards.......

    Hi..
    Thanks for the links..I applied the tips on each of them but it didnt work.
    Also my database is 11g so it is not true that this happens on 10g only.
    Things tried:
    1)I tried with different values on parallel parameter but same error
    2)I applied the following:
    alter system set open_cursors=1024 scope=spfile;
    alter system set "_optimizer_cost_based_transformation"=off;
    commit;
    The 3rd link was bit better
    I tried to find out where exactly the error was causesusing
    expdp attach =SYS_EXPORT_FULL_03
    But I cant figure out what the object : PUBLIC
    oracle/context/isearch/GetPage
    is..?
    Does this needs to be excluded from the export?if so, how?
    Can someone help how to fix the error now?
    Processing object type DATABASE_EXPORT/SCHEMA/SYNONYM
    ORA-39014: One or more workers have prematurely exited.
    ORA-39029: worker 1 with process name "DW01" prematurely terminated
    ORA-31672: Worker process DW01 died unexpectedly.
    Job "SYSTEM"."SYS_EXPORT_FULL_03" stopped due to fatal error at 11:29:32
    C:\Documents and Settings\ora2>expdp attach=SYS_EXPORT_FULL_03
    Export: Release 11.1.0.7.0 - Production on Tuesday, 01 May, 2012 11:35:38
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Username: system
    Password:
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Job: SYS_EXPORT_FULL_03
      Owner: SYSTEM
      Operation: EXPORT
      Creator Privs: TRUE
      GUID: 8499C802F52A414A8BCACE552DDF6F11
      Start Time: Tuesday, 01 May, 2012 11:37:56
      Mode: FULL
      Instance: geooap
      Max Parallelism: 1
      EXPORT Job Parameters:
      Parameter Name      Parameter Value:
         CLIENT_COMMAND        system/******** parfile=h:\datapump\oapfull.par
      State: IDLING
      Bytes Processed: 0
      Current Parallelism: 1
      Job Error Count: 0
      Dump File: H:\datapump\oapfull.dmp
        bytes written: 4,096
    Worker 1 Status:
      Process Name: DW01
      State: UNDEFINED
      Object Schema: PUBLIC
      Object Name: oracle/context/isearch/GetPage
      Object Type: DATABASE_EXPORT/SCHEMA/PUBLIC_SYNONYM/SYNONYM
      Completed Objects: 766
      Total Objects: 766
      Worker Parallelism: 1

  • Exclude schema grants - impdp

    Hi Experts,
    I am getting the following error during import using impdp
    Failing sql is:
    GRANT SELECT ON "CM_DBA"."PARTY" TO "CM_USR"
    ORA-39083: Object type OBJECT_GRANT failed to create with error:
    ORA-01917: user or role 'CM_USR' does not exist
    Failing sql is:
    GRANT SELECT ON "CM_DBA"."PARTY_RLN" TO "CM_USR"
    How to exclude CM_USR during import in windows?
    Thanks for the help.
    Regards,
    Richard

    This link will be helpful
    http://psoug.org/reference/datapump.html
    Here just find out word EXCLUDE a good explanation is given

  • Impdp import schema to different tablespace

    Hi,
    I have PRTDB1 schema in PRTDTS tablespace, I need encrypt this tablespace.
    and I have done this steps:
    1. I created new encrypted PRTDTS_ENC tablespace.
    2. I take export PRTDB1 schema. (expdp system/system schema=PRTDB1 directory=datapump dumpfile=data.dmp logfile=log1.log)
    and I want import this dump to new PRTDTS_ENC tablespace.
    can you give me impdp command, which I need?

    Hello,
    ORA-14223: Deferred segment creation is not supported for this table
    Failing sql is:
    CREATE TABLE "PRTDB1"."DATRE_CONT" ("BRON" NUMBER NOT NULL ENABLE, "PCDID" NUMBER(9,0) NOT NULL ENABLE, "OWNERSID" NUMBER NOT NULL ENABLE, "VALUESTR" VARCHAR2(4000 CHAR), "VALNUM" NUMBER, "VALDATE" DATE, "VALXML" "XMLTYPE") SEGMENT CREATION DEFERRED PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING So it means that these Tables "need" a Segment but, they are created with the option SEGMENT CREATION DEFERRED.
    A workaround is to create a Segment for these Tables on the Source Database before exporting them. Then, at the Import you won't have the option SEGMENT CREATION DEFERRED. The following Note of MOS detail this Solution:
    - *Error Ora-14223 Deferred Segment Creation Is Not Supported For This Table During Datapump Import [ID 1293326.1]*
    The link below gives more informations about the DEFERRED SEGMENT CREATION:
    http://docs.oracle.com/cd/E14072_01/server.112/e10595/tables002.htm#CHDGJAGB
    With the DEFERRED SEGMENT CREATION feature, a Segment is created automatically at the first insert on the Table. So the Tables for which the error ORA-14223 occurs are empty.
    You may also create them separately by using a script generated by the SQLFILE parameter in which you change the option SEGMENT CREATION DEFERRED by SEGMENT CREATION IMMEDIATE.
    Hope this help.
    Best regards,
    Jean-Valentin

  • Trigger-Body isnt remaped during IMPDP and REMAP_SCHEMA in Oracle 10g

    Hello,
    we are facing the following problem:
    During an IMPDP we are remaping the schema. The schema includes triggers. The trigger bodys are of normal syntax without a relation to the schema/user.
    If we start the import with datapump using the remap_schema option the triggers are remap to the new schema. But the trigger bodys point to the old schema.
    For example:
    CREATE OR REPLACE TRIGGER "OLD_SCHEMA".NAME_OF_TRIGGER
    BEFORE INSERT ON "OLD_SCHEMA"."TABLE_NAME" REFERENCING OLD AS OLD NEW AS NEW FOR EACH ROW BEGIN
    END;
    After using the REMAP_SCHEMA option:
    CREATE OR REPLACE TRIGGER "NEW_SCHEMA".NAME_OF_TRIGGER
    BEFORE INSERT ON "OLD_SCHEMA"."TABLE_NAME" REFERENCING OLD AS OLD NEW AS NEW FOR EACH ROW BEGIN
    END;
    So the Trigger Body in the NEW_SCHEMA uses a link to the OLD_SCHEMA and the trigger doesn't work. We wrote the trigger without referencing to a schema. So the OLD_SCHEMA resp. NEW_SCHEMA syntax must be added by datapump.
    The question: Is there a way to remap the triggers correctly in the new schema or is this behavior "normal" to Oracle 10g Datapump?
    We are using Oracle 10g Database 10.2.0.4 with the latest Patches on Windows Server 2003 SP2.
    Thank you for you attention.

    Hello,
    I had the same problem and you can't fix it by changing a parameter of impdp. You really need to change the source of those triggers so they don't use hard-code schema references.
    Solution would be to change the source code of those triggers. Or to use the "script" parameter of impdp and recreate the triggers yourself after the impdp has finished.
    For more info check Metalink note: 750783.1
    I hope this helps.
    Regards,
    Michiel.

  • Impdp taking long time for only few MBs data...

    Hi All,
    I have one query related to impdp. I have one expdp file and size is 47M. When I restore this dmp using impdp it will take more time. Also initially table_data loaded finsih very fast but then later on alter function/procedure/view taking a lot time almost 4 to 5 hrs.
    I have no idea why its taking long time... Earlier I can see one DB link is failed and it given error for "TNS name could not resolved" so I create DB link also before run impdp but the same result. Can any one suggest what could be the cause to taking long time for only 47MB data?
    Note - Both expdp and impdp database version is 11.2.0.3.0. If I import the same dmp file into 11.2.0.1.0 then its done in few mins.
    Thanks...

    Also Read
    Checklist For Slow Performance Of DataPump Export (expdp) And Import (impdp) [ID 453895.1]
    DataPump Import (IMPDP) is Very Slow at Object/System/Role Grants, Default Roles [ID 1267951.1]

Maybe you are looking for

  • What does the "remove option : press C" do in IMG 1.9?

    I've noticed that you can press the "c" button to remove the option of a program in IMG 1.9. What does it do? I don't see anything happening. Is there a IMG 1.9 guide online somewhere? IMG 1.9 seems to have new features, and I'd like to have a better

  • Multiple rows in a single but with different column names

    Hello All, I have a table in which data is stored in the below displayed fashion. TRANS_ID Ty_1     NU_1 38960     BP     5215153159 38960    EM     [email protected]           38960    FX     5115538033 38960    HP     4419965055 38960 For some weir

  • Mp3 and importing

    Hi Whenever i import an Mp3 file from a CD or my hard drive i tunes does not recognise it as an imported file via I Tunes. Therefore i can't get any Track Names etc and am fed up of having to put them in myself. Is there any way to aviod this??

  • GT784WNV not working in Bridge Mode! Fed up with lack of Customer Service

    I don't even know where to start but I am this close to just cancelling Verizon and switching to cable internet.  If you want to skip the rant, jump down a paragraph. I recently moved from a house a few miles away in which I had a solid DSL connectio

  • In oracle 10g is the same Oracle Net Listener than...?

    Hi forum members again I'm vey new in oracle server10g, this time I have a doubt about oracle components , that is the same the Oracle Net Listner than the Oracle SQL*Net listener, if is not the same, please tell me the diference Thanks in advance