Expdp & impdp of R12 schema into 11g DB

Hi,
i need to take a backup of GL schema from R12 instance using expdp &
import it into 11g DB( standalone DB) using impdp.
i have used the following for expdp in R12(10g DB)
$expdp system/<pass> schemas=gl directory=dump_dir dumpfile=gl.dmp logfile=gl.log
export was successfull.
How to import into an 11g DB( This is on a different machine).
regards,
Charan

Refer to "Oracle® Database Utilities 11g Release 1 (11.1)" Manual.
Data Pump Import
http://download.oracle.com/docs/cd/B28359_01/server.111/b28319/dp_import.htm#i1007653
Oracle® Database Utilities 11g Release 1 (11.1)
http://download.oracle.com/docs/cd/B28359_01/server.111/b28319/toc.htm

Similar Messages

  • Using expdp/impdp to backup schemas to new tablespace

    Hello,
    I have tablespace A for schemas A1 and A2, and I wish back up these schemes to tablespace B using schema names B1 and B2 (so the contents of schemas A1 and A2 are copied into schemas B1 and B2, respectively, to use as backups in case something happens to schemas A1 or A2 or tablespace A).
    I began by creating tablespace B, and schemas B1 and B2. Then I attempted to populate schemas B1 and B2 by doing the following:
    EXPORT SCHEMAS:
    expdp a1/a1password@myIdentifier ESTIMATE=BLOCKS DUMPFILE=myDpumpDirectory:a1.dmp LOGFILE=myDpumpDirectory:a1_export.log SCHEMAS=a1 COMPRESSION=METADATA_ONLY
    expdp a2/a2password@myIdentifier ESTIMATE=BLOCKS DUMPFILE=myDpumpDirectory:a2.dmp LOGFILE=myDpumpDirectory:a2_export.log SCHEMAS=a2 COMPRESSION=METADATA_ONLY
    IMPORT SCHEMAS:
    impdp b1/b1password@myIdentifier DUMPFILE=myDpumpDirectory:a1.dmp LOGFILE=myDpumpDirectory:b1_import.log REMAP_SCHEMA=a1:b1
    impdp b2/b2password@myIdentifier DUMPFILE=myDpumpDirectory:a2.dmp LOGFILE=myDpumpDirectory:b2_import.log REMAP_SCHEMA=a2:b2
    This resulted in backing up schema A1 into schema B1, and schema A2 into B2, but the tablespaces for schemas B1 and B2 remained tablespace A (when I wanted them to be tablespace B).
    I will drop schemas B1 and B2, create new schemas, and try again. What command should I use to get the tablespace correct this time?
    Reviewing the documentation for data pump import
    http://docs.oracle.com/cd/E11882_01/server.112/e22490/dp_import.htm#SUTIL300
    specifically the section titled REMAP_TABLESPACE, I'm thinking that I could just add a switch to the above import commands to remap tablespace, such as:
    impdp b1/b1password@myIdentifier DUMPFILE=myDpumpDirectory:a1.dmp LOGFILE=myDpumpDirectory:b1_import.log REMAP_SCHEMA=a1:b1 REMAP_TABLESPACE=a:b
    impdp b2/b2password@myIdentifier DUMPFILE=myDpumpDirectory:a2.dmp LOGFILE=myDpumpDirectory:b2_import.log REMAP_SCHEMA=a2:b2 REMAP_TABLESPACE=a:b
    Is that correct?
    Also, is it OK to use the same export commands above, or should they change to support the REMAP_TABLESPACE?

    Hi,
    if i understand correctly, you want to import  A1:B1 and  A2:B2 with the Respective Tablespace. You are using the expdp with ESTIMATE it can not help you
    You can use something like that with one dump file
    expdp system/password directory=<myDpumpDirectory> dumpfile=A1_A2_Export.dmp logfile=A1_A2_Export.log schemas=A1,A2
    impdp system/password directory=<myDpumpDirectory> dumpfile=A1_A2_Export.dmp logfile=A1_A2_Import.log remap_schemas=<A1:B1,A2:B2> REMAP_TABLESPACE=<TAB1:TAB2>
    HTH

  • Expdp impdp fails from 10g to 11g db version

    Hello folks,
    Export DB Version : 10.2.0.4
    Import DB Version : 11.2.0.1
    Export Log File
    Export: Release 10.2.0.4.0 - Production on Wednesday, 03 November, 2010 2:19:20
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - Production
    With the Partitioning, Data Mining and Real Application Testing options
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 45 GB
    . . exported "DYM"."CYCLE_COUNT_MASTER" 39.14 GB 309618922 rows
    Master table "DYM"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    Dump file set for DYM.SYS_EXPORT_SCHEMA_01 is:
    Job "DYM"."SYS_EXPORT_SCHEMA_01" successfully completed at 02:56:49
    Import Log File
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Master table "SYSTEM"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    ORA-31693: Table data object "DYM_PRJ4"."CYCLE_COUNT_MASTER" failed to load/unload and is being skipped due to error:
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    Job "SYSTEM"."SYS_IMPORT_FULL_01" completed with 1 error(s) at 10:54:38
    from 10g expdp to 11g impdp is not allowed ? any thoughts appreciated ??

    Nope , I do not see any error file.
    Current log# 2 seq# 908 mem# 0:
    Thu Nov 04 11:58:20 2010
    DM00 started with pid=530, OS id=1659, job SYSTEM.SYS_IMPORT_FULL_02
    Thu Nov 04 11:58:20 2010
    DW00 started with pid=531, OS id=1661, wid=1, job SYSTEM.SYS_IMPORT_FULL_02
    Thu Nov 04 11:58:55 2010
    DM00 started with pid=513, OS id=1700, job SYSTEM.SYS_IMPORT_FULL_02
    Thu Nov 04 11:58:55 2010
    DW00 started with pid=520, OS id=1713, wid=1, job SYSTEM.SYS_IMPORT_FULL_02
    Thu Nov 04 12:00:54 2010
    Thread 1 cannot allocate new log, sequence 909
    Private strand flush not complete
    Current log# 2 seq# 908 mem# 0: ####################redo02.log
    Thread 1 advanced to log sequence 909 (LGWR switch)
    Current log# 3 seq# 909 mem# 0: ###################redo03.log
    Thu Nov 04 12:01:51 2010
    Thread 1 cannot allocate new log, sequence 910
    Checkpoint not complete
    Current log# 3 seq# 909 mem# 0:###################redo03.log

  • Errors when importing 10.2.0.4  Solaris xml schema into 11g Linux

    We are migrating our database from 10g on Solaris to 11g on Linux. Used expdt for full DB but getting this error - (ORA-39139: Data Pump does not support XMLSchema objects. TABLE_DATA:"SPRUBATCH"."ACCNT_XML_STG_50" will be skipped.) for the schema containing XML. So used exp for this particular schema. But Getting the following errors during import:
    IMP-00058: ORACLE error 604 encountered
    ORA-00604: error occurred at recursive SQL level 1
    ORA-22814: attribute or element value is larger than specified in type
    ORA-06512: at "XDB.DBMS_XMLSCHEMA_INT", line 72
    ORA-06512: at "XDB.DBMS_XMLSCHEMA", line 33
    ORA-06512: at line 1
    . . importing table             "ACCNT_XML_ALL_52"ORA-24550: signal received: [si_signo=11] [si_errno=0] [si_code=1] [si_int=0] [si_ptr=0x7000000000]
    [si_addr=0x1043f94e8]
    Memory fault
    Should we deinstall 11g XDB and bring in 10g XDB ? It seems like the "XDB.DBMS_XMLSCHEMA" has changed from 10 to 11?
    Or there is upgrade script needs to be run?
    Who has done this upgrade before and appreciate if you can share your experience with us.
    Thanks!

    Hi Srini,
    here is the export log:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, Oracle Label Security, Data Mining and Real Application Testing options
    Export done in US7ASCII character set and AL16UTF16 NCHAR character set
    server uses WE8MSWIN1252 character set (possible charset conversion)
    . exporting pre-schema procedural objects and actions
    . exporting foreign function library names for user SPRUBATCH
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions for user SPRUBATCH
    About to export SPRUBATCH's objects ...
    . exporting database links
    . exporting sequence numbers
    . exporting cluster definitions
    . about to export SPRUBATCH's tables via Conventional Path ...
    . . exporting table ACCNT_XML_ALL_50 0 rows exported
    . . exporting table ACCNT_XML_ALL_51 0 rows exported
    . . exporting table ACCNT_XML_ALL_52 1712 rows exported
    . . exporting table ACCNT_XML_STG_50 0 rows exported
    . . exporting table ACCNT_XML_STG_51 0 rows exported
    . . exporting table ACCNT_XML_STG_52 1 rows exported
    . . exporting table AsOfDate1714_TAB 0 rows exported
    . . exporting table AsOfDate1736_TAB 0 rows exported
    . . exporting table BRNCH_XML_ALL_50 0 rows exported
    . . exporting table BRNCH_XML_ALL_51 0 rows exported
    . . exporting table BRNCH_XML_ALL_52 99 rows exported
    . . exporting table BRNCH_XML_STG_50 0 rows exported
    . . exporting table BRNCH_XML_STG_51 0 rows exported
    . . exporting table BRNCH_XML_STG_52 1 rows exported
    . . exporting table CNSTR_XCPTN_TBL 2 rows exported
    EXP-00091: Exporting questionable statistics.
    . . exporting table ChangeDatestamp1730_TAB 0 rows exported
    . . exporting table ChangeDatestamp1756_TAB 0 rows exported
    . . exporting table ChangedBy1731_TAB 0 rows exported
    . . exporting table ChangedBy1757_TAB 0 rows exported
    . . exporting table Comments1715_TAB 0 rows exported
    . . exporting table ComplaintsFromDate1716_TAB 0 rows exported
    . . exporting table ComplaintsToDate1717_TAB 0 rows exported
    . . exporting table DueDate1718_TAB 0 rows exported
    . . exporting table File1742_TAB 0 rows exported
    . . exporting table FileName1738_TAB 0 rows exported
    . . exporting table FileNumber1741_TAB 0 rows exported
    . . exporting table FileReceivedDate1740_TAB 0 rows exported
    . . exporting table FileSize1739_TAB 0 rows exported
    . . exporting table FirmCrdNumber1720_TAB 0 rows exported
    . . exporting table FirmCrdNumber1743_TAB 0 rows exported
    . . exporting table FirmDeletedFiles1746_TAB 0 rows exported
    . . exporting table FirmName1721_TAB 0 rows exported
    . . exporting table FirmName1747_TAB 0 rows exported
    . . exporting table MatterID1719_TAB 0 rows exported
    . . exporting table MsgType1722_TAB 0 rows exported
    . . exporting table MsgType1748_TAB 0 rows exported
    . . exporting table NewStatus1732_TAB 0 rows exported
    . . exporting table NewStatus1758_TAB 0 rows exported
    . . exporting table PublishDate1723_TAB 0 rows exported
    . . exporting table RejectComment1733_TAB 0 rows exported
    . . exporting table RejectComment1759_TAB 0 rows exported
    . . exporting table RequestGroupId1724_TAB 0 rows exported
    . . exporting table RequestGroupId1749_TAB 0 rows exported
    . . exporting table RequestId1725_TAB 0 rows exported
    . . exporting table RequestId1734_TAB 0 rows exported
    . . exporting table RequestId1750_TAB 0 rows exported
    . . exporting table RequestId1760_TAB 0 rows exported
    . . exporting table RequestStatus1726_TAB 0 rows exported
    . . exporting table RequestType1727_TAB 0 rows exported
    . . exporting table RequestType1751_TAB 0 rows exported
    . . exporting table SRM_RQ_XML_ALL 50 rows exported
    . . exporting table SRM_RQ_XML_STG 1 rows exported
    . . exporting table SRM_SST_SC_XML_ALL 43 rows exported
    . . exporting table SRM_SST_SC_XML_STG 1 rows exported
    . . exporting table SRM_XF_XML_ALL 83 rows exported
    . . exporting table SRM_XF_XML_STG 1 rows exported
    . . exporting table SST_SRM_SC_XML_STG 0 rows exported
    . . exporting table SchemaVersion1728_TAB 0 rows exported
    . . exporting table Status1755_TAB 0 rows exported
    . . exporting table ValidFiles1754_TAB 0 rows exported
    . exporting synonyms
    . exporting views
    . exporting stored procedures
    . exporting operators
    . exporting referential integrity constraints
    . exporting triggers
    . exporting indextypes
    . exporting bitmap, functional and extensible indexes
    . exporting posttables actions
    . exporting materialized views
    . exporting snapshot logs
    . exporting job queues
    . exporting refresh groups and children
    . exporting dimensions
    . exporting post-schema procedural objects and actions
    . exporting statistics
    Export terminated successfully with warnings.
    And import log:
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Export file created by EXPORT:V10.02.01 via conventional path
    Warning: the objects were exported by SPRUBATCH, not by you
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8MSWIN1252 character set (possible charset conversion)
    . importing SPRUBATCH's objects into SPRUBATCH
    IMP-00058: ORACLE error 604 encountered
    ORA-00604: error occurred at recursive SQL level 1
    ORA-22814: attribute or element value is larger than specified in type
    ORA-06512: at "XDB.DBMS_XMLSCHEMA_INT", line 72
    ORA-06512: at "XDB.DBMS_XMLSCHEMA", line 33
    ORA-06512: at line 1
    IMP-00058: ORACLE error 604 encountered
    ORA-00604: error occurred at recursive SQL level 1
    ORA-22814: attribute or element value is larger than specified in type
    ORA-06512: at "XDB.DBMS_XMLSCHEMA_INT", line 72
    ORA-06512: at "XDB.DBMS_XMLSCHEMA", line 33
    ORA-06512: at line 1
    IMP-00058: ORACLE error 604 encountered
    ORA-00604: error occurred at recursive SQL level 1
    ORA-22814: attribute or element value is larger than specified in type
    ORA-06512: at "XDB.DBMS_XMLSCHEMA_INT", line 72
    ORA-06512: at "XDB.DBMS_XMLSCHEMA", line 33

  • Can I use expdp/impdp for APEX schemas to setup a new server?

    If setting up a new development server that will replace an old server, can I simply use datapump tp export the APEX_040100 and FLOWS_FILES schemas on the old server, and just import the schemas on the new server? That would seem much simpler and faster than exporting all the applications in the APEX console, and importing the applications, much less steps involved.

    Pl post database and OS versions. Is there a reason you cannot clone the entire database from the old server to the new ?
    HTH
    Srini

  • Expdp+Impdp: Does the user have to have DBA privilege?

    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) his own schema?
    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) other (=not his own) schemas ?
    If he is not allowed: Which GRANT is necessary to be able to perform such expdp/impdp operations?
    Peter
    Edited by: user559463 on Feb 28, 2010 7:49 AM

    Hello,
    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) his own schema?Yes, a User can always export its own objects.
    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) other (=not his own) schemas ?Yes, if this User has EXP_FULL_DATABASE and IMP_FUL_DATABASE Roles.
    So, you can create a User and GRANT it EXP_FULL_DATABASE and IMP_FULL_DATABASE Roles and, being connected
    to this User, you could export/import any Object from / to any Schemas.
    On databases, on which there're a lot of export/import operations, I always create a special User with these Roles.
    NB: In DataPump you should GRANT also READ, WRITE Privileges on the DIRECTORY (if you use "dump") to the User.
    Else, be accurate on the choice of your words, as previously posted, DBA is a Role not a Privilege which has another meaning.
    Hope this help.
    Best regards,
    Jean-Valentin

  • Importing a few Oracle 10g schemas into oracle 11g database

    Hi,
    I have the task to import a few schemas that are on oracle 10g database (10.2.0.4) into oracle 11gR2 database (11.2.0.2). Now I have only oracle 10g export and import utility (the 11g server is in not where I am given the import rights).
    Can I do the export of the 10g schemas using the 10g export utility and then import into 11g database using the 10g import utility? I ran a test to do that and found it works but would also like to have expert opinion on it.
    Also another question is: if the source schema is 11g - suppose- in that case can the 10g export/import utility still work? I mean export the schema from say db1 (which is 11g) using 10g export utility and import them into 11g database db2 (which is again 11g) using 10g import? I am asking because I don't have access to 11g server and trying to find what I can do with the 10g utilities.
    Thanks,

    set up a database link to the 10gR2 version database in the 11gR2 database.
    Now using the 11gR2 client impdp utility use
    impdp network_link=<link to 10gR2 database>
    Done.
    Your 'objection' you don't have access to 11gR2 doesn't count as you can install a 11gR2 client.
    Also generally speaking you should not use a lower version imp to imp to a higher version database
    Also exp and imp are deprecated in 10g and higher.
    Sybrand Bakker
    Senior Oracle DBA

  • EXPDP/IMPDP : Clone schema with datapump

    Hello,
    I have a full export with the following parameters:
    DIRECTORY=dir$export
    DUMPFILE=datapump_exp_FULL.dmp
    LOGFILE=datapump_exp_FULL.log
    FULL=Y
    And I want to import only one schema into a new schema with the following parameters:
    DIRECTORY=dir$export
    DUMPFILE=datapump_exp_FULL.dmp
    LOGFILE=datapump_imp_SCHEMA2.log
    SCHEMAS=SCHEMA1
    REMAP_SCHEMA=SCHEMA1:SCHEMA2
    TABLE_EXISTS_ACTION=REPLACE
    TRANSFORM=OID:N
    But for what reason ever IMPDP tries to import other schemas like SYSMAN, APEX... also. This throws a load of ORA-31684 (object already exists), what means TABLE_EXISTS_ACTION=REPLACE doesn't work also.
    I also tried to import without SCHEMAS=SCHEMA1, but didn't succeed anyway...
    So, how can I force IMPDP to import only SCHEMA1, without touching the others ?
    Regards,
    Mynz

    either use INCLUDE PARAMETER as per documentation
    http://docs.oracle.com/cd/E11882_01/server.112/e10701/dp_import.htm#i1007761OR
    EXCLUDE=SCHEMA:"IN ('OUTLN','SYSTEM','SYSMAN','FLOWS_FILES','APEX_030200','APEX_PUBLIC_USER','ANONYMOUS')"OR
    Using NOT-LIKE with EXCLUDETake care of those quotation marks..

  • Schema Expdp-IMPdp , i have the below errors in my imp log

    Hello All,
    I am doing some exp/imp of schemas in a OBI environment and size around 300GB of each shema,  after quiet a long time.., and not sure of the imp log errors whether to ignore or rerun my import with any changes or drop schemas and reimport ......
    My IMP statement is
    impdp system/********@OBIDEV schemas=ODS,DWH,STG,GEOGRAPHY directory=dumpdir dumpfile=BI2011_SCHEMAS.dmp logfile=impBI2011_SCHEMAS.log
    Errors from the imp log :
    Processing object type SCHEMA_EXPORT/USER
    ORA-31684: Object type USER:"GEOGRAPHY" already exists
    ORA-31684: Object type USER:"DWH" already exists
    ORA-31684: Object type USER:"STG" already exists
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    ORA-39083: Object type ROLE_GRANT failed to create with error:
    ORA-01919: role 'ROLE_SRC' does not exist
    Failing sql is:
    GRANT "ROLE_SRC" TO "STG" WITH ADMIN OPTION
    ORA-39083: Object type ROLE_GRANT failed to create with error:
    ORA-01919: role 'ROLE_STG' does not exist
    Failing sql is:
    GRANT "ROLE_STG" TO "ODS" WITH ADMIN OPTION
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    a
    --- and after some lines again the below errors, many failing sql's to created the tables ..
    ORA-39083: Object type TABLE:"STG"."STG_ART_STOCK_OIG_XT" failed to create with error:
    ORA-06564: object BI_SRC_FILES does not exist
    Failing sql is:
    CREATE TABLE "STG"."STG_ART_STOCK_OIG_XT" ("YEARWEEK" NUMBER, "ARTICLE_NUMBER" NUMBER, "BU_CODE" NUMBER, "ART_QTY_STOCK" NUMBER, "PM_CODE" VARCHAR2(10 BYTE)) ORGANIZATION EXTERNAL ( TYPE ORACLE_LOADER DEFAULT DIRECTORY "BI_SRC_FILES" ACCESS PARAMETERS ( RECORDS DELIMITED BY NEWLINE CHARACTERSET WE8MSWIN1252 STRING SIZES ARE IN BYTES BADFILE 'ART_STOCK_OIG.bad'
    ORA-39083: Object type TABLE:"STG"."STG_ART_STOCK_INGKA_XT" failed to create with error:
    ORA-06564: object BI_SRC_FILES does not exist
    ORA-39083: Object type TABLE:"STG"."STG_ART_PMCODE_INGKA_XT" failed to create with error:
    ORA-06564: object BI_SRC_FILES does not exist
    Failing sql is:
    CREATE TABLE "STG"."STG_IFPM_XT" ("YEAR" VARCHAR2(255 BYTE), "TERTIAL" VARCHAR2(255 BYTE), "YW" VARCHAR2(255 BYTE), "STORE_NAME" VARCHAR2(255 BYTE), "BUDGET" VARCHAR2(255 BYTE), "ACTUAL_SALES" VARCHAR2(255 BYTE), "CUSTOMERS" VARCHAR2(255 BYTE), "VISITORS" VARCHAR2(255 BYTE), "REST_SALES" VARCHAR2(255 BYTE), "EXIT_CAFE_SALES" VARCHAR2(255 BYTE), "SWE_FOOD_SALES" VARCHAR2
    ORA-39083: Object type TABLE:"STG"."HFB_PA_EXT" failed to create with error:
    ORA-06564: object TEMP does not exist
    Failing sql is:
    CREATE TABLE "STG"."HFB_PA_EXT" ("HFB_ID" NUMBER, "HFB_DESC" VARCHAR2(100 BYTE), "RA_ID" NUMBER, "RA_DESCR" VARCHAR2(100 BYTE), "PA_ID" NUMBER, "PA_DESCR" VARCHAR2(100 BYTE)) ORGANIZATION EXTERNAL ( TYPE ORACLE_LOADER DEFAULT DIRECTORY "TEMP" ACCESS PARAMETERS ( records delimited  by newline
        skip 1
        fields  terminated by ','
        missing field values are null
          ) LOCAT
    ORA-39151: Table "GEOGRAPHY"."CAHC" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    . . imported "ODS"."ZZ_ODS_ART_STOCK_T"                  26.81 GB 321434766 rows
    imported "ODS"."ODS_PM_CODE_T":"SYS_P15548"          259.4 MB 3418214 rows  (are the sys tables effected here)
    Then the rows started importing ....
    Share all the possible ways to avoid the above all errors in the imp log as it has to be shared with the client ,    Thanks for the swif response and help in advance.
    Regards,
    RanG

    HI,
    RanG wrote:
    Hello All,
    I am doing some exp/imp of schemas in a OBI environment and size around 300GB of each shema,  after quiet a long time.., and not sure of the imp log errors whether to ignore or rerun my import with any changes or drop schemas and reimport ......
    My IMP statement is
    impdp system/********@OBIDEV schemas=ODS,DWH,STG,GEOGRAPHY directory=dumpdir dumpfile=BI2011_SCHEMAS.dmp logfile=impBI2011_SCHEMAS.log
    Errors from the imp log :
    Processing object type SCHEMA_EXPORT/USER
    ORA-31684: Object type USER:"GEOGRAPHY" already exists
    ORA-31684: Object type USER:"DWH" already exists
    ORA-31684: Object type USER:"STG" already exists
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    ORA-39083: Object type ROLE_GRANT failed to create with error:
    ORA-01919: role 'ROLE_SRC' does not exist
    Failing sql is:
    GRANT "ROLE_SRC" TO "STG" WITH ADMIN OPTION
    ORA-39083: Object type ROLE_GRANT failed to create with error:
    ORA-01919: role 'ROLE_STG' does not exist
    Failing sql is:
    GRANT "ROLE_STG" TO "ODS" WITH ADMIN OPTION
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    a
    --- and after some lines again the below errors, many failing sql's to created the tables ..
    ORA-39083: Object type TABLE:"STG"."STG_ART_STOCK_OIG_XT" failed to create with error:
    ORA-06564: object BI_SRC_FILES does not exist
    Failing sql is:
    CREATE TABLE "STG"."STG_ART_STOCK_OIG_XT" ("YEARWEEK" NUMBER, "ARTICLE_NUMBER" NUMBER, "BU_CODE" NUMBER, "ART_QTY_STOCK" NUMBER, "PM_CODE" VARCHAR2(10 BYTE)) ORGANIZATION EXTERNAL ( TYPE ORACLE_LOADER DEFAULT DIRECTORY "BI_SRC_FILES" ACCESS PARAMETERS ( RECORDS DELIMITED BY NEWLINE CHARACTERSET WE8MSWIN1252 STRING SIZES ARE IN BYTES BADFILE 'ART_STOCK_OIG.bad'
    ORA-39083: Object type TABLE:"STG"."STG_ART_STOCK_INGKA_XT" failed to create with error:
    ORA-06564: object BI_SRC_FILES does not exist
    ORA-39083: Object type TABLE:"STG"."STG_ART_PMCODE_INGKA_XT" failed to create with error:
    ORA-06564: object BI_SRC_FILES does not exist
    Failing sql is:
    CREATE TABLE "STG"."STG_IFPM_XT" ("YEAR" VARCHAR2(255 BYTE), "TERTIAL" VARCHAR2(255 BYTE), "YW" VARCHAR2(255 BYTE), "STORE_NAME" VARCHAR2(255 BYTE), "BUDGET" VARCHAR2(255 BYTE), "ACTUAL_SALES" VARCHAR2(255 BYTE), "CUSTOMERS" VARCHAR2(255 BYTE), "VISITORS" VARCHAR2(255 BYTE), "REST_SALES" VARCHAR2(255 BYTE), "EXIT_CAFE_SALES" VARCHAR2(255 BYTE), "SWE_FOOD_SALES" VARCHAR2
    ORA-39083: Object type TABLE:"STG"."HFB_PA_EXT" failed to create with error:
    ORA-06564: object TEMP does not exist
    Failing sql is:
    CREATE TABLE "STG"."HFB_PA_EXT" ("HFB_ID" NUMBER, "HFB_DESC" VARCHAR2(100 BYTE), "RA_ID" NUMBER, "RA_DESCR" VARCHAR2(100 BYTE), "PA_ID" NUMBER, "PA_DESCR" VARCHAR2(100 BYTE)) ORGANIZATION EXTERNAL ( TYPE ORACLE_LOADER DEFAULT DIRECTORY "TEMP" ACCESS PARAMETERS ( records delimited  by newline
        skip 1
        fields  terminated by ','
        missing field values are null
          ) LOCAT
    ORA-39151: Table "GEOGRAPHY"."CAHC" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    . . imported "ODS"."ZZ_ODS_ART_STOCK_T"                  26.81 GB 321434766 rows
    imported "ODS"."ODS_PM_CODE_T":"SYS_P15548"          259.4 MB 3418214 rows  (are the sys tables effected here)
    Then the rows started importing ....
    Share all the possible ways to avoid the above all errors in the imp log as it has to be shared with the client ,    Thanks for the swif response and help in advance.
    Regards,
    RanG
    Because the table already exists. You can use the TABLE_EXISTS_ACTION option, so you can tell impdp what to do if the table it is trying to create already exists.
    TABLE_EXISTS_ACTION={SKIP | APPEND | TRUNCATE | REPLACE}
    impdp system/********@OBIDEV schemas=ODS,DWH,STG,GEOGRAPHY directory=dumpdir dumpfile=BI2011_SCHEMAS.dmp logfile=impBI2011_SCHEMAS.log table_exists_action=replace
    Thank you

  • How to copy one schema into another

    I need to copy one schema into another. That means, all the tables, views, constraints, etc, and all the data as well.
    I have a dmp file of source-schema (created daily with expdp by our backup schedule).
    Can I use impdp to import this dmp file into dest-schema? I'm thinking about something like this:
    The dmp file was created with:
    expdp source-schema/pwd DIRECTORY=dmpdir DUMPFILE=myproject.dmp
    Create the copy on dest-schema with:
    impdp dest-schema/pwd DIRECTORY=dmpdir DUMPFILE=myproject.dmp
    (dest-schema is an existing schema on the same machine. I'll drop all the tables before executing the command. Is there anything else I should do first?)
    If this approach is not feasible, kindly suggest me an alternative.
    Thank you.
    Edited by: mgro on Apr 16, 2010 3:22 AM (changed 2nd command: expdp to impdp)

    Thank you both. Dean, I've made a note of your proposed command. May use it next time.
    In the end I used the following command and it worked for me. I added TABLE_EXISTS_ACTION=REPLACE since the destination was not empty.
    First I ran these command logged in as SYS
    GRANT Import Full Database TO dest-schema;
    GRANT Create Any Directory TO dest-schema;
    GRANT Unlimited Tablespace TO dest-schema;
    impdp
    dest-schema/pwd
    DIRECTORY=dmpdir
    DUMPFILE=myproject.dmp
    REMAP_SCHEMA=source-schema:dest-schema
    TABLE_EXISTS_ACTION=REPLACE
    LOGFILE=implog.log
    There were some errors though. The import process complained about constraints referencing non-existent primary keys.
    It turns out that the source schema has some constraints that are marked with status DISABLED. These constraints either:
    - are not part of the DUMP file
    or
    - are in the DUMP file but skipped during import
    Don't know if this is a bug or by design behavior.

  • Expdp/impdp :: Constraints in Parent child relationship

    Hi ,
    I have one table parent1 and tables child1, child2 and chld3 have foreign key created on this parent1.
    Now I want to do some deletion on parent1. But since number of records are very high on parent1 , we are going with expdp / impdp with querry option.
    I have taken query level expdp on parent1. Now I dropped parent1 with cascade constraints option and all the foreign keys created by child1 ,2 and 3 which references parent1 are automatically dropped.
    Now If i fire the impdp for the query level dump file , will these foreign key constraints will get created automatically on child1 ,2 and 3 or I need to manually re-create it ??
    Regards,
    Anu

    Hi,
    The FK's will not be in the dumpfile - see the example code below where i generate a sqlfile following pretty much the process you would have done. THis is because the FK is part of the DDL for the child table not the parent.
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    OPS$ORACLE@EMZA3>create table a (col1 number);
    Table created.
    OPS$ORACLE@EMZA3>alter table a add primary key (col1);
    Table altered.
    OPS$ORACLE@EMZA3>create table b (col1 number);
    Table created.
    OPS$ORACLE@EMZA3>alter table b add constraint x foreign key (col1) references a(col1);
    Table altered.
    OPS$ORACLE@EMZA3>
    EMZA3:[/oracle/11.2.0.1.2.DB/bin]# expdp / include=TABLE:\"=\'A\'\"
    Export: Release 11.2.0.3.0 - Production on Fri May 17 15:45:50 2013
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    Starting "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04":  /******** include=TABLE:"='A'"
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    . . exported "OPS$ORACLE"."A"                                0 KB       0 rows
    Master table "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04" successfully loaded/unloaded
    Dump file set for OPS$ORACLE.SYS_EXPORT_SCHEMA_04 is:
      /oracle/11.2.0.3.0.DB/rdbms/log/expdat.dmp
    Job "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04" successfully completed at 15:45:58
    Import: Release 11.2.0.3.0 - Production on Fri May 17 15:46:16 2013
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    Master table "OPS$ORACLE"."SYS_SQL_FILE_FULL_01" successfully loaded/unloaded
    Starting "OPS$ORACLE"."SYS_SQL_FILE_FULL_01":  /******** sqlfile=a.sql
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Job "OPS$ORACLE"."SYS_SQL_FILE_FULL_01" successfully completed at 15:46:17
    -- CONNECT OPS$ORACLE
    ALTER SESSION SET EVENTS '10150 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10904 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '25475 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10407 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10851 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '22830 TRACE NAME CONTEXT FOREVER, LEVEL 192 ';
    -- new object type path: SCHEMA_EXPORT/TABLE/TABLE
    CREATE TABLE "OPS$ORACLE"."A"
       (    "COL1" NUMBER
       ) SEGMENT CREATION IMMEDIATE
      PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255
    NOCOMPRESS LOGGING
      STORAGE(INITIAL 16384 NEXT 16384 MINEXTENTS 1 MAXEXTENTS 505
      PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1
      BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
      TABLESPACE "SYSTEM" ;
    -- new object type path: SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    ALTER TABLE "OPS$ORACLE"."A" ADD PRIMARY KEY ("COL1")
      USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255
      STORAGE(INITIAL 16384 NEXT 16384 MINEXTENTS 1 MAXEXTENTS 505
      PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1
      BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
      TABLESPACE "SYSTEM"  ENABLE;
    -- new object type path: SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    DECLARE I_N VARCHAR2(60);
      I_O VARCHAR2(60);
      NV VARCHAR2(1);
      c DBMS_METADATA.T_VAR_COLL;
      df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS';
    stmt varchar2(300) := ' INSERT INTO "SYS"."IMPDP_STATS" (type,version,flags,c1,c2,c3,c5,n1,n2,n3,n4,n5,n6,n7,n8,n9,n10,n11,n12,d1,cl1) VALUES (''I'',6,:1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13,NULL,:14,:15,NULL,:16,:17)';
    BEGIN
      DELETE FROM "SYS"."IMPDP_STATS";
      c(1) := 'COL1';
      DBMS_METADATA.GET_STAT_INDNAME('OPS$ORACLE','A',c,1,i_o,i_n);
      EXECUTE IMMEDIATE stmt USING 0,I_N,NV,NV,I_O,0,0,0,0,0,0,0,0,NV,NV,TO_DATE('2013-05-17 15:43:24',df),NV;
      DBMS_STATS.IMPORT_INDEX_STATS('"' || i_o || '"','"' || i_n || '"',NULL,'"IMPDP_STATS"',NULL,'"SYS"');
      DELETE FROM "SYS"."IMPDP_STATS";
    END;
    /Regards,
    Harry
    http://dbaharrison.blogspot.com/

  • Migrate schemas to 11g

    We have 10g RAC environment, and we are going to create new 11g non-rac in Linux environment and need to import schemas(approx 200GB) from 10g RAC, before importing to 11g database, what are things need to consider database level,after importing what are things need to check performence prospective
    Thanks
    Edited by: Autoconfig on Jul 20, 2011 10:41 PM

    We have 10g RAC environment, and we are going to create new 11g non-rac in Linux environment and need to import schemas(approx 200GB) from 10g RAC, before importing to 11g database, what are things need to consider database level,after importing what are things need to check performence prospectiveI suggest to check below notes:
    Minimizing Downtime During Production Upgrade [ID 478308.1]
    Master Note For Oracle Database Upgrades and Migrations [ID 1152016.1]
    Different Upgrade Methods For Upgrading Your Database [ID 419550.1]
    Please check my blog also
    http://heliosguneserol.wordpress.com/2010/06/17/move-to-oracle-database-11g-release-2-wiht-mike-dietrich/
    In this pdf you can see patch of to upgrade db from x to n wiht many senerios wiht all related metalinks notes which is created by Oracle worker Mike Dietrich
    @Helios,
    Here OP is asking to migrate schema using EXPDP/IMPDP and intention is after and before migrating what are the considerations need to take care?
    for example: Instance level SGA parameters? is it source is AMM or not?
    after importing what should SGA we need to set ? is it should be AMM or not?
    any issues in optimization?
    Any chance of decreasing Performance?

  • Expdp/impdp quick question

    I just did a export of around 20 schemas from my production database.
    expdp userid=system/pwd schemas=('a','b',...'x') dumpfile=xxx logfile=yyyy directory=....
    Is is possible for me to import just one particular schema into my test database by specifying the schema of interest??? (in this case schema a)
    impdp userid=system/pwd schemas=('a') dumpfile=.... logfile=... directory=...

    Sorry, you'll need to close and repost. I am not authorized to move the thread.
    Regards,
    Brandye Barrington
    Certification Forum Moderator
    Certification Program Manager

  • Data Pump - expdp / impdp utility question

    HiAll,
    As a basic exercise to learn Data pump, I am trying to export schema scott and then want to load the dump file into a test_t schema in the same database.
    --1. created dir object from sysdba
    CREATE DIRECTORY dpump_dir1 AS 'C:\output_dir';
    --2. created dir on the OS as c:/output_dit
    --3. run expdp
    expdp system/*****@orcl schemas=scott DIRECTORY=dpump_dir1 JOB_NAME=hr DUMPFILE=scott_orcl_nov5.dmp PARALLEL=4
    --4. create test_t schema and grant dba to it.
    --5. run impdp
    impdp system/*****@orcl schemas=test_t DIRECTORY=dpump_dir1 JOB_NAME=hr DUMPFILE=scott_orcl_nov5.dmp PARALLEL=8
    it fails here as ORA39165 : schema test_t not found. However the schema test_t does exist.
    So, I am not sure why it should give this error. It seems that the schema of Scott from the expdp dump file can not be loaded to any other schema but only to a schema named scott...Is it right? If yes then how can I load all the objects of schema say scott to another schema say test_t? It would be helpful if you can please show the respective expdp and impdp command.
    Thanks a lot
    KS

    The test_t schema does not exist in the export dump file, you should remap from input scott schema to the target test_t schema.
    REMAP_SCHEMA : http://download.oracle.com/docs/cd/E11882_01/server.112/e22490/dp_import.htm#SUTIL927
    Nicolas.

  • EXP/IMP..of table having LOB column to export and import using expdp/impdp

    we have one table in that having colum LOB now this table LOB size is approx 550GB.
    as per our knowldge LOB space can not be resused.so we have alrady rasied SR on that
    we are come to clusion that we need to take backup of this table then truncate this table and then start import
    we need help on bekow ponts.
    1)we are taking backup with expdp using parellal parameter=4 this backup will complete sussessfully? any other parameter need to keep in expdp while takig backup.
    2)once truncate done,does import will complete successfully..?
    any SGA PGA or undo tablespace size or undo retention we need to increase.. to completer susecfully import? because its production Critical database?
    current SGA 2GB
    PGA 398MB
    undo retention 1800
    undo tbs 6GB
    please any one give suggestion to perform activity without error...also suggest parameter need to keep during expdp/impdp
    thanks an advance.

    Hi,
    From my experience be prepared for a long outage to do this - expdp is pretty quick at getting lobs out but very slow at getting them back in again - a lot of the speed optimizations that may datapump so excellent for normal objects are not available for lobs. You really need to test this somewhere first - can you not expdp from live and load into some test area - you don't need the whole database just the table/lob in question. You don;t want to find out after you truncate the table that its going to take 3 days to import back in....
    You might want to consider DBMS_REDEFINITION instead?
    Here you precreate a temporary table (with same definitiion as the existing one), load the data into it from the existing table and then do a dictionary switch to swap them over - giving you minimal downtime. I think this should work fine with LOBS at 10g but you should do some research and see if it works fine. You'll need a lot of extra tablespace (temporarily) for this approach though.
    Regards,
    Harry

Maybe you are looking for

  • ESYU: Blanket PO approve시 ASL과 Sourcing Rules이 생성되기 위해 요구되는 steps

    Purpose Oracle Purchasing - Version: 11.5.8 to 11.5.10 Blanket Purchase Order approval 동안 ASL과 Sourcing Rules이 생성되게 하기 위해 요구되는 steps에는 무엇이 있는지 알아본다. Solution Profile Options: - PO: Allow Auto-generate Sourcing Rules => CREATE AND UPDATE - PO: Allow A

  • Why doesn't the apple store give the same 3 year guarantee that John Lewis gives

    Hello! It would appear that John Lewis are also selling the new macbook pros for the same price as the apple store, but with a 3 year guarntee (presumably you would also get the 1 year warranty provided by Apple too), why does the Apple Store not als

  • Trying to print on dvd with discus software?

    i am using the discus software recommended by Old Toad here. i feel like i have the settings right - 80 mm, etc. but the disk missed the middle of the dvd completely and the whole picture is off center. what am i doing wrong? deb

  • How jsp can call applet functions

    good morning to all java experts i have one problem that is in my application i have one applet in the server side which contains functions like encryption, decryption, hash functions. whenever any client wants to store some information on server sid

  • Help with Signatures

    Hello, I am trying to sign a PDF with Adobe Reader XI on a MacBook Air.  I had an older version of Adobe that worked fine on another computer.  I've created a signature in the preferences and clicked it to be allowed to be used for signatures.  The s