Data Pump -Importing one table index

Is it possible to import one table index alone(any table ex emp ) .If it can be how should the param should look like ..
Thanks and Regards
harris

I can't think of anything that would prevent this from working. You just need to make sure that the large table does not have any ref constraints, or other associations with the other tables that may get screwed up while the other users are using the database.
Dean

Similar Messages

  • Data pump import a table in to a different schema in 11g

    Hi All
    I have 11.2 Oracle database and i have a requirement to import few tables in to a new schema using my previous month export. I can not import whole schema as it is very large. I check REMAP_TABLE option but it just create the table in same schema and only rename the table.
    Eg i have TABLE GAS.EMPLOYE_DATA i want to import this to GASTEST.EMPLOYEE_DATA
    IS there way to do this using datapump?
    Appriciate any advice.

    Hello,
    You may use the parameter INCLUDE so as to select only one Table:
    REMAP_SCHEMA=SCOTT:SCOTT_TEST
    INCLUDE=TABLE:"IN ('EMP')"Hope this help.
    Best regards,
    Jean-Valentin

  • Data Pump Import

    I am having trouble using the data pump import tool. I am working with Oracle 10.2.4.0 on Windows Server 2008. I am trying to import a lot of data (several schemas, thousands of tables, hundreds of GB of data) from the same version of Oracle on Linux. Since it's so much data and will take so long to import, I am trying to make sure I can get it to work for one table or one schema before I attempt everything. So I'm trying to import the TEST_USER.TABLE1 table using the following command:
    impdp.exe dumpfile=big_file.dmp logfile=big_file_import.log tables=test_user.table1 remap_datafile=\'/oradata1/data1.dbf\':\'D:\oracle\product\10.2.0\oradata\orcl\data1.dbf\'
    I supply sys as sysdba for the login when it asks (not sure how you get that info into the initial command). However, I am getting the following error:
    user 'TEST_USER' does not exist
    My understanding was that the data pump utility would create all the requisite schemas for me. Is that not the case? The target database is a new install so the source schemas don't exist there.
    Even if I create the test_user schema by hand, then I get an error stating that the tablespace doesn't exist:
    ORA-00959: tablespace 'TS_1' does not exist
    Then even doing that, which I don't want to have to do to begin with, doesn't work. Then it gives me something about how the user doesn't have the proper privileges on the tablespace.
    Isn't the data pump utility supposed to make this stuff automatic? Do I really have to create all the schemas and tablespaces by hand? That's going to take a long time. Am I doing something wrong here?
    Thanks,
    Dave

    user5482172 wrote:
    Hemant K Chitale wrote:
    tables=test_user.table1 The "TABLES" mode does NOT create the database accounts.
    The FULL mode creates tablespaces and database accounts before importing data.
    The SCHEMAS mode creates database accounts before importing data-- but expects Tablespaces to pre-exist so that the tables and indexes can be created in the correct tablespaces.
    Hemant K Chitale
    http://hemantoracledba.blogspot.com
    That is consistent with what I'm seeing. So basically there is no way to import just a few tables or schemas and have it create all the necessary users and tablespaces for you. That seems a little arbitrary, but whatever. Why make things easy on your users when you can be all Oracley about it. Anyway, I'll give it a try and let you know what happens.
    Can I ask where you got this information? I'm sure you're right, but I can't find anything stating this.
    Thanks,
    Dave
    Edited by: user5482172 on Feb 18, 2010 10:51 PMAfter some testing, this appears to be correct. The system will not create tablespaces under the SCHEMAS mode and it will not create schemas under the TABLES mode; only under the FULL mode will everything be created for you. I find that unintuitive and extremely unfriendly, but that's Oracle for you. Have you ever tried doing this in SQL Server? You open the GUI, tell it where the dump file is, the name of the database, where to put the data files and you're done. Oracle has a real knack for making things way more complicated than they need to be. End Oracle rant.

  • Error Data Pump Import

    Hi All,
    I am having errors below when trying to use data pump to import one table from the dump file (from a pre-prod db) into another database (prod_db) same version. I used the expdp to generate the dump file.
    Gettings errors -
    ORA-39083
    ORA-00959
    ORA-39112
    Any suggestions and/or advice would be appreciated.
    Thank you
    Import: Release 10.2.0.1.0 - 64bit Production
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP and Data Mining options
    Master table "SYSTEM"."SYS_IMPORT_TABLE_01" successfully loaded/unloaded
    Starting "SYSTEM"."SYS_IMPORT_TABLE_01": system/******** DIRECTORY=DATA_PUMP_DIR DUMPFILE=EXP_Mxxxx_1203.DMP LOGFILE=Mxxxx_CLAIMS.LOG TABLES=CLAIMS
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    ORA-39083: Object type TABLE failed to create with error:
    ORA-00959: tablespace 'OXFORD_DATA_01' does not exist
    Failing sql is:
    CREATE TABLE "Mxxxx"."CLAIMS" ("CLAIM_NUM" VARCHAR2(25) NOT NULL ENABLE, "LINE_NUM" NUMBER(3,0) NOT NULL ENABLE, "PAID_FLAG" CHAR(1), "ADJ_FLAG" CHAR(1), "CLAIM_TYPE" VARCHAR2(20), "MEM_ID" VARCHAR2(20), "MEM_SUF" VARCHAR2(20), "MEM_BEG_DATE" DATE, "MEM_REL" CHAR(2), "MEM_NAME" VARCHAR2(40), "MEM_DOB" DATE, "MEM_SEX" CHAR(1), "REC_DATE" DATE, "PAID_DATE" DATE, "FROM_DATE" DATE, "
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39112: Dependent object type INDEX:"Mxxxx"."CLAIMS_IDX1" skipped, base object type TABLE:"Mxxxx"."CLAIMS" creation failed
    ORA-39112: Dependent object type INDEX:"Mxxxx"."CLAIMS_PK" skipped, base object type TABLE:"Mxxxx"."CLAIMS" creation failed
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"Mxxxx"."CLAIMS_IDX1" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"Mxxxx"."CLAIMS_PK" creation failed
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    ORA-39112: Dependent object type TABLE_STATISTICS skipped, base object type TABLE:"Mxxxx"."CLAIMS" creation failed
    Job "SYSTEM"."SYS_IMPORT_TABLE_01" completed with 6 error(s) at 13:49:33
    impdp username/password DIRECTORY=DATA_PUMP_DIR DUMPFILE=EXP_Mxxxx_1203.DMP LOGFILE=Mxxxx_CLAIMS.LOG TABLES=CLAIMS

    Ok,
    Thank you for the advice-
    So I did use REMAP_TABLESPACE option and it did import and load the file but should I be concerned with 4 errors? Looks like have to rebuilt indexes which is fine.
    Master table "Mxxxx"."SYS_IMPORT_TABLE_01" successfully loaded/unloaded
    Starting "Mxxxx"."SYS_IMPORT_TABLE_01": Mxxxx/******** DIRECTORY=DATA_PUMP_DIR DUMPFILE=EXP_Mxxxx_1203.DMP LOGFILE=Mxxxx_CLAIMS.LOG TABLES=CLAIMS REMAP_TABLESPACE=OXFORD_DATA_01:Mxxxx_DATA_01
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    . . imported "Mxxxx"."CLAIMS" 605.3 MB 1715554 rows
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    ORA-39083: Object type INDEX failed to create with error:
    ORA-00959: tablespace 'USER_INDEX_01' does not exist
    Failing sql is:
    CREATE INDEX "Mxxxx"."CLAIMS_IDX1" ON "Mxxxx"."CLAIMS" ("MEM_ID") PCTFREE 10 INITRANS 2 MAXTRANS 255 NOLOGGING STORAGE(INITIAL 8388608 NEXT 8388608 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "USER_INDEX_01" PARALLEL 1
    ORA-39083: Object type INDEX failed to create with error:
    ORA-00959: tablespace 'USER_INDEX_01' does not exist
    Failing sql is:
    CREATE UNIQUE INDEX "Mxxxx"."CLAIMS_PK" ON "Mxxxx"."CLAIMS" ("CLAIM_NUM", "LINE_NUM") PCTFREE 10 INITRANS 2 MAXTRANS 255 NOLOGGING STORAGE(INITIAL 8388608 NEXT 8388608 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "USER_INDEX_01" PARALLEL 1
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"Mxxxx"."CLAIMS_IDX1" creation failed
    ORA-39112: Dependent object type INDEX_STATISTICS skipped, base object type INDEX:"Mxxxx"."CLAIMS_PK" creation failed
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "Mxxxx"."SYS_IMPORT_TABLE_01" completed with 4 error(s) at 16:57:12

  • Import one table taking toom uch time

    Hi All,
    i have database 10g running with ASM under HPUX 04 processor with 08Gb of RAM.
    I have a problem of importing one table with 6M lines, it takes too much time even 02 days the data is not yet importing, i don't know where is the problem.
    I tried also to use data pump but when exporting the table from the other server in RAC it gives error bellow:
    ORA-39014: One or more workers have prematurely exited.
    ORA-39029: worker 2 with process name "DW04" prematurely terminated
    ORA-31671: Worker process DW04 had an unhandled exception.
    ORA-12801: error signaled in parallel query server P029, instance ab-db2:abdb2 (2)
    ORA-29913: error in executing ODCIEXTTABLEOPEN callout
    ORA-31626: job does not exist
    ORA-06512: at "SYS.ORACLE_DATAPUMP", line 19
    ORA-06512: at "SYS.KUPW$WORKER", line 1342
    ORA-06512: at line 2
    Thanks fro your help for this problem.
    regards
    raitsarevo

    This the command i used:
    time expdp abillity/4dd1ct3d dumpfile=cb_coupons.dmp logfile=cb_coupons.log directory=exp_dir parallel=4 tables=cb_coupons
    and this is a portion of my alert log file, no error signaled:
    Fri Apr 18 15:57:35 2008
    ALTER SYSTEM SET service_names='abdb','SYS$SYS.KUPC$C_1_20080418155733.ABDB' SCOPE=MEMORY SID='abdb1';
    Fri Apr 18 15:57:35 2008
    ALTER SYSTEM SET service_names='SYS$SYS.KUPC$C_1_20080418155733.ABDB','abdb','SYS$SYS.KUPC$S_1_20080418155733.ABDB' SCOPE=MEMORY SID='abdb1';
    kupprdp: master process DM00 started with pid=212, OS id=10470
    to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_TABLE_09', 'ABILLITY', 'KUPC$C_1_20080418155733', 'KUPC$S_1_20080418155733', 0);
    kupprdp: worker process DW01 started with worker id=1, pid=215, OS id=10621
    to execute - SYS.KUPW$WORKER.MAIN('SYS_EXPORT_TABLE_09', 'ABILLITY');
    kupprdp: worker process DW02 started with worker id=2, pid=219, OS id=11777
    to execute - SYS.KUPW$WORKER.MAIN('SYS_EXPORT_TABLE_09', 'ABILLITY');
    Fri Apr 18 15:59:41 2008
    ALTER SYSTEM SET service_names='SYS$SYS.KUPC$S_1_20080418155733.ABDB','abdb' SCOPE=MEMORY SID='abdb1';
    Fri Apr 18 15:59:41 2008
    ALTER SYSTEM SET service_names='abdb' SCOPE=MEMORY SID='abdb1';
    Fri Apr 18 16:01:16 2008
    Thread 1 advanced to log sequence 62141
    Current log# 5 seq# 62141 mem# 0: +ASM_DG2/abdb/onlinelog/group_5.286.618685253
    Current log# 5 seq# 62141 mem# 1: +ASM_DG1/abdb/onlinelog/group_5.10591.618685257
    Thanks for your help

  • How do u save datas more than one table using net beans ide using JSF

    Hi,
    I am new to JSF.
    I save / delete / update / New master table using POJO (Plain Old Java Objects), database - oracle and Toplink Persistence Unit.
    How do u save data more than one table using net beans ide using JSF (I am using POJO) ?
    and also Tell me the reference book for JSF.
    Thanks in advance.
    regards,
    N.P.Siva

    SivaNellai wrote:
    I am new to JSF.
    So, I am using net beans IDE 6.1 from sun microsystem. It is a free software.No, you don't drag'n'drop if you're new to JSF. Switch to source code mode. Write code manually, with the help of IDE for the speed up.
    So, please guide me the reference books, articles. I need the basic understanding of JSF, net beans IDE.[JSF: The Complete Reference|http://www.amazon.com/JavaServer-Faces-Complete-Reference/dp/0072262400] is a good book. The [JSF specification document|http://jcp.org/aboutJava/communityprocess/final/jsr252/index.html] is also a good reading to understand what JSF is and how it works. There are also javadocs and tlddocs of Sun JSF Mojarra.

  • Exclude DBMS_SCHEDULER jobs in DATA PUMP import

    Hello,
    I need to exclude all DBMS_SCHEDULER jobs during DATA PUMP import.
    I tried to do this with: EXCLUDE=JOB but this only works on DBMS_JOB.
    Is there a way to exclude all DBMS_SCHEDULER jobs during import?
    Kind Regards

    There are PROCOBJ that can be excluded (Procedural objects in the selected schemas) but I'm affraid it exclude not only DBMS_SCHEDULER jobs.
    Any ideas?

  • Data pump import error with nested tables

    So the problem is somewhat long :)
    Actually the problems are two - why and how oracle are treating OO concept and why data pump doesn't work?
    So scenario for the 1st one:
    1) there is object type h1 and table of h1
    2) there is unrelated object type row_text and table of row_text
    3) there is object type h2 under h1 with attribute as table of row_text
    4) there is table tab1 with column b with data type as table of h1. Of course column b is stored as nested table.
    So how do you think - how many nested tables Oracle will create? The correct answer is 2. One explicitly defined and one hidden with system
    generated name for the type h2 which is under type h1. So the question is WHY? Why if I create an instance of supertype Oracle tries to adapt
    it for the subtype as well? Even more if I create another subtype h3 under h1 another hidden nested table appears.
    This was the first part.
    The second part is - if I do schema export and try to import it in another schema I got error saying that oracle failed to create storage table for
    nested table column b. So the second question is - if Oracle has created such a mess with hidden nested tables how to import/export to another
    schema?
    Ok and here is test case to demonstrate problems above:
    -- creating type h1 and table of it
    SQL> create or replace type h1 as object (a number)
      2  not final;
      3  /
    Type created.
    SQL> create or replace type tbl_h1 as table of h1;
      2  /
    Type created.
    -- creating type row_text and table of it
    SQL> create or replace type row_text as object (
      2    txt varchar2(100))
      3  not final;
      4  /
    Type created.
    SQL> create or replace type tbl_row_text as table of row_text;
      2  /
    Type created.
    -- creating type h2 as subtype of h1
    SQL> create or replace type h2 under h1 (some_texts tbl_row_text);
      2  /
    Type created.
    SQL> create table tab1 (a number, b tbl_h1)
      2  nested table b
      3  store as tab1_nested;
    Table created.
    -- so we have 2 nested tables now
    SQL> select table_name, parent_table_name, parent_table_column
      2  from user_nested_tables;
    TABLE_NAME                     PARENT_TABLE_NAME
    PARENT_TABLE_COLUMN
    SYSNTfsl/+pzu3+jgQAB/AQB27g==  TAB1_NESTED
    TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H2")."SOME_TEXTS"
    TAB1_NESTED                    TAB1
    B
    -- another subtype of t1
    SQL> create or replace type h3 under h1 (some_texts tbl_row_text);
      2  /
    Type created.
    -- plus another nested table
    SQL> select table_name, parent_table_name, parent_table_column
      2  from user_nested_tables;
    TABLE_NAME                     PARENT_TABLE_NAME
    PARENT_TABLE_COLUMN
    SYSNTfsl/+pzu3+jgQAB/AQB27g==  TAB1_NESTED
    TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H2")."SOME_TEXTS"
    SYSNTfsl/+pz03+jgQAB/AQB27g==  TAB1_NESTED
    TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H3")."SOME_TEXTS"
    TAB1_NESTED                    TAB1
    B
    SQL> desc "SYSNTfsl/+pzu3+jgQAB/AQB27g=="
    Name                                      Null?    Type
    TXT                                                VARCHAR2(100)OK let it be and now I'm trying to export and import in another schema:
    [oracle@xxx]$ expdp gints/xxx@xxx directory=xxx dumpfile=gints.dmp logfile=gints.log
    Export: Release 11.2.0.1.0 - Production on Thu Feb 4 22:32:48 2010
    <irrelevant rows skipped>
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    . . exported "GINTS"."TAB1"                                  0 KB       0 rows
    . . exported "GINTS"."SYSNTfsl/+pz03+jgQAB/AQB27g=="         0 KB       0 rows
    . . exported "GINTS"."TAB1_NESTED"                           0 KB       0 rows
    . . exported "GINTS"."SYSNTfsl/+pzu3+jgQAB/AQB27g=="         0 KB       0 rows
    Master table "GINTS"."SYS_EXPORT_SCHEMA_01" successfully loaded/unloaded
    ******************************************************************************And now import. In order to create types transformation of OIDs is applied and also remap_schema
    Although it fails to create the table.
    [oracle@xxx]$ impdp gints1/xxx@xxx directory=xxx dumpfile=gints.dmp logfile=gints_imp.log remap_schema=gints:gints1 transform=OID:n
    Import: Release 11.2.0.1.0 - Production on Thu Feb 4 22:41:48 2010
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Release 11.2.0.1.0 - Production
    Master table "GINTS1"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
    Starting "GINTS1"."SYS_IMPORT_FULL_01":  gints1/********@xxx directory=xxx dumpfile=gints.dmp logfile=gints_imp.log remap_schema=gints:gints1 transform=OID:n
    Processing object type SCHEMA_EXPORT/USER
    ORA-31684: Object type USER:"GINTS1" already exists
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
    Processing object type SCHEMA_EXPORT/TYPE/TYPE_SPEC
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    ORA-39083: Object type TABLE:"GINTS1"."TAB1" failed to create with error:
    ORA-02320: failure in creating storage table for nested table column B
    ORA-00904: : invalid identifier
    Failing sql is:
    CREATE TABLE "GINTS1"."TAB1" ("A" NUMBER, "B" "GINTS1"."TBL_H1" ) SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    ORA-39083: Object type INDEX_STATISTICS failed to create with error:
    ORA-01403: no data found
    ORA-01403: no data found
    Failing sql is:
    DECLARE I_N VARCHAR2(60);   I_O VARCHAR2(60);   c DBMS_METADATA.T_VAR_COLL;   df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS'; BEGIN  DELETE FROM "SYS"."IMPDP_STATS";   c(1) :=   DBMS_METADATA.GET_STAT_COLNAME('GINTS1','TAB1_NESTED',NULL,'TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H2")."SOME_TEXTS"',1);  DBMS_METADATA.GET_STAT_INDNAME('GINTS1','TAB1_NESTED',c,1,i_o,i_n);   INSERT INTO "
    ORA-39083: Object type INDEX_STATISTICS failed to create with error:
    ORA-01403: no data found
    ORA-01403: no data found
    Failing sql is:
    DECLARE I_N VARCHAR2(60);   I_O VARCHAR2(60);   c DBMS_METADATA.T_VAR_COLL;   df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS'; BEGIN  DELETE FROM "SYS"."IMPDP_STATS";   c(1) :=   DBMS_METADATA.GET_STAT_COLNAME('GINTS1','TAB1_NESTED',NULL,'TREAT(SYS_NC_ROWINFO$ AS "GINTS"."H3")."SOME_TEXTS"',1);  DBMS_METADATA.GET_STAT_INDNAME('GINTS1','TAB1_NESTED',c,1,i_o,i_n);   INSERT INTO "
    Processing object type SCHEMA_EXPORT/TABLE/STATISTICS/TABLE_STATISTICS
    Job "GINTS1"."SYS_IMPORT_FULL_01" completed with 4 error(s) at 22:41:52So any idea how to make export/import of such tables?
    TIA
    Gints

    Tom Kyte has said it repeatedly ... I will repeat it here for you.
    The fact that Oracle allows you to build object tables is not an indication that you should.
    Store your data relationally and build object_views on top of them.
    http://www.morganslibrary.org/reference/object_views.html
    If you model properly, and store properly, you don' have any issues.

  • 10g to 11gR2 Upgrade using Data Pump Import

    Hi,
    I am intending to move a 10g database from one windows server to another. However there is also a requirement to upgrade this database to 11gR2. Therefore I was going to combine the 2 in one movement by -
    1. take a full data pump export of the source 10g database
    2. create a new empty 11g database on the target environment
    3. import the dump file into the target database
    However I have a couple of queries running over in my mind about this approach -
    Q1. What happens with the SYSTEM and SYS objects from SYSTEM and SYSAUX during the import ? Given that I will have in effect already created a new dictionary on the empty target database - will any import of SYSTEM or SYS simply produce errror messages which should be ignored ?
    Q2. should I use EXCLUDE on SYS and SYSTEM ( is this EXCLUDE better on the export or import side ) ?
    Q3. what happens if there are things such as scheduled jobs etc on the source system - since these would be stored in SYSTEM owned tables, how would I bring these across to the new target 11g database ?
    thanks,
    Jim

    Jim Thompson wrote:
    Hi,
    I am intending to move a 10g database from one windows server to another. However there is also a requirement to upgrade this database to 11gR2. Therefore I was going to combine the 2 in one movement by -
    1. take a full data pump export of the source 10g database
    2. create a new empty 11g database on the target environment
    3. import the dump file into the target database
    However I have a couple of queries running over in my mind about this approach -
    Q1. What happens with the SYSTEM and SYS objects from SYSTEM and SYSAUX during the import ? Given that I will have in effect already created a new dictionary on the empty target database - will any import of SYSTEM or SYS simply produce errror messages which should be ignored ?
    You wont get error related to system , sysaux tablespace bacsei these wont be exported at all. Schemas like "SYS, CTXSYS, MDSYS and ORDSYS" are never exported using datapump. Thats why oracle reommends not to create any objects under sys,system schema.
    Q2. should I use EXCLUDE on SYS and SYSTEM ( is this EXCLUDE better on the export or import side ) ?
    Not required, as datadictionary schemas wont be exported, so spcefying EXCLUDE wont do anything
    Q3. what happens if there are things such as scheduled jobs etc on the source system - since these would be stored in SYSTEM owned tables, how would I bring these across to the new target 11g database ?
    DDL will get export and imported into new database. For example if you have schema A and you had defined job whos owner is A, then while importing this information also gets imported. For user defined Jobs there are view like USER_SCHEDULER_JOBS etc, So dont worry about the user jobs, these will be created.
    Also see
    MOS note - Schema's CTXSYS, MDSYS and ORDSYS are Not Exported [ID 228482.1]
    Export system or sys schema
    I also run following test in my db which shows i cannt export the objects under sys schema;
    SQL> show user
    USER is "SYS"
    SQL> create table pump (id number);
    Table created.
    SQL> insert into pump values (1);
    1 row created.
    SQL> insert into pump values (2);
    1 row created.
    SQL> exit
    Disconnected from Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Pr
    oduction
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    C:\Documents and Settings\rnagi\My Documents\Ranjit Doc\Performance\SQL>expdp tables=sys.pump logfile=test.log
    Export: Release 11.2.0.1.0 - Production on Mon Feb 27 18:11:29 2012
    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
    Username: / as sysdba
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Produc
    tion
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYS"."SYS_EXPORT_TABLE_01":  /******** AS SYSDBA tables=sys.pump logfi
    le=test.log
    Estimate in progress using BLOCKS method...
    Processing object type TABLE_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    *ORA-39166: Object SYS.PUMP was not found.*
    *ORA-31655: no data or metadata objects selected for job*
    *Job "SYS"."SYS_EXPORT_TABLE_01" completed with 2 error(s) at 18:11:57*

  • Data Pump import to a sql file error :ORA-31655 no data or metadata objects

    Hello,
    I'm using Data Pump to export/import data, one requirement is to import data to a sql file. The OS is window.
    I made the follow export :
    expdp system/password directory=dpump_dir dumpfile=tablesdump.dmp content=DATA_ONLY tables=user.tablename
    and it works, I can see the file TABLESDUMP.DMP in the directory path.
    then when I tried to import it to a sql file:
    impdp system/password directory=dpump_dir dumpfile=tablesdump.dmp sqlfile=tables_export.sql
    the log show :
    ORA-31655 no data or metadata objects selected for job
    and the sql file is created empty in the directory path.
    I'm not DBA, I'm a Java developer , Can you help me?
    Thks

    Hi, I added the command line :
    expdp system/system directory=dpump_dir dumpfile=tablesdump.dmp content=DATA_ONLY schemas=ko1 tables=KO1QT01 logfile=capture.log
    the log in the console screen is (is in Spanish), no log file was cerated in the directory path.
    Export: Release 10.2.0.1.0 - Production on Martes, 26 Enero, 2010 12:59:14
    Copyright (c) 2003, 2005, Oracle. All rights reserved.
    Conectado a: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    UDE-00010: se han solicitado varios modos de trabajo, schema y tables.
    (English error)
    UDE-00010: multiple job modes requested,schema y tables.
    This is why I used tables=user.tablename instead, is this right ?
    Thks

  • Error during data pump import with SQL developer

    Hello,
    I try to trasnfer data from one database to another one through data pump via SQL Developper (data amount is quite important) exporting several tables.
    Tables export is doing fine, but I encounter the following error when I import the file (I try data only and data + DDL).
    "Exception: ORA-39001: argument value invalid dbms_datapump.get_status(64...=
    ORA-39001: argument value invalid
    ORA-39000: ....
    ORA-31619: ...
    The file is in the right place, data pump folder of the new database. User is the same on both base, database version are similar.
    Do you have any idea of the problem ?
    Thanks

    With query SELECT * FROM v$version;
    Environment source
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    PL/SQL Release 11.2.0.1.0 - Production
    "CORE     11.2.0.1.0     Production"
    TNS for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production
    Environment target
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    PL/SQL Release 11.2.0.1.0 - Production
    "CORE     11.2.0.1.0     Production"
    TNS for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production

  • Fatal error in data pump import

    Hi
    Please help me in solving this error while importing data from dmp file
    Import: Release 10.1.0.2.0 - Production on Thursday, 15 February, 2007 3:54
    Copyright (c) 2003, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Master table "SYSTEM"."SYS_SQL_FILE_FULL_02" successfully loaded/unloaded
    Starting "SYSTEM"."SYS_SQL_FILE_FULL_02": system/******** directory=data_pump dumpfile=prashant_dp.
    dmp SQLFILE=prashant_imp.sql logfile=prashant_imp.log
    Processing object type TABLE_EXPORT/TABLE/TABLE
    ORA-39125: Worker unexpected fatal error in KUPW$WORKER.PUT_DDLS while calling DBMS_METADATA.CONVERT
    ORA-06502: PL/SQL: numeric or value error
    LPX-00210: expected '<' instead of 'n'
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    0x540a0b7c 13460 package body SYS.KUPW$WORKER
    0x540a0b7c 5810 package body SYS.KUPW$WORKER
    0x540a0b7c 11153 package body SYS.KUPW$WORKER
    0x540a0b7c 2891 package body SYS.KUPW$WORKER
    0x540a0b7c 3530 package body SYS.KUPW$WORKER
    0x540a0b7c 6395 package body SYS.KUPW$WORKER
    0x540a0b7c 1208 package body SYS.KUPW$WORKER
    0x534beb10 2 anonymous block
    Job "SYSTEM"."SYS_SQL_FILE_FULL_02" stopped due to fatal error at 03:54

    here are trace files [http://files.mail.ru/B05EBC4402A342D0AE1E88BFB04191DB|http://files.mail.ru/B05EBC4402A342D0AE1E88BFB04191DB]
    error here:
    KUPM:12:47:19.571: Error detected by MCP
    KUPM:12:47:19.572: ORA-39014: One or more workers have prematurely exited.
    KUPM:12:47:19.572: ORA-39029: worker 1 with process name "DW00" prematurely terminated
    KUPM:12:47:19.572: ORA-31671: Worker process DW00 had an unhandled exception.
    ORA-39079: unable to enqueue message RQ
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.KUPC$QUE_INT", line 965
    ORA-00931: missing identifier
    ORA-06512: at "SYS.KUPW$WORKER", line 16746
    ORA-06512: at "SYS.KUPW$WORKER", line 19035
    ORA-06512: at "SYS.KUPW$WORKER", line 8191
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.MAIN []
    ORA-06512: at "SYS.KUPW$WORKER", line 1705
    ORA-44002: invalid object name
    ORA-06512: at line 2
    KUPM:12:47:19.577: In restart_worker for worker 1...
    KUPM:12:47:19.577: Error being processed is:  -26457
    KUPM:12:47:19.577: worker id is:
    KUPM:12:47:19.577: Worker error is: 0
    KUPM:12:47:19.577: Exited main loop...
    KUPM:12:47:19.577: Returned to MAIN
    KUPV:12:47:19.577: Update request for job: SYSTEM.SYS_IMPORT_FULL_03, func: 1
    KUPM:12:47:19.578: Entered state: UNDEFINED
    KUPM:12:47:19.578: In RESPOND_TO_START
    KUPC:12:47:19.578: Before ENQ: Sending Type: 2041 ID:
    KUPC:12:47:19.578:  RP,KUPC$C_1_20130603124648,MCP,KUPC$A_1_20130603124649,10,Y
    kwqberlst !retval block
    kwqberlst rqan->lagno_kwqiia  5
    kwqberlst rqan->lascn_kwqiia > 0 block
    kwqberlst rqan->lascn_kwqiia  5
    kwqberlst ascn 2375213 lascn 22
    KUPM:12:47:19.579: In check_workers...
    KUPM:12:47:19.579: Live worker count is:  0
    KUPM:12:47:19.579: worker id is:
    KUPM:12:47:19.579: Worker error is: 0
    KUPM:12:47:19.579: Job is completing

  • Data pump import from 11g to 10g

    I have 2 database: first is 11.2.0.2.0 and second is 10.2.0.1.0
    In 10g i created database link on 11g
    CREATE DATABASE LINK "TEST.LINK"
    CONNECT TO "monservice" IDENTIFIED BY "monservice"
    USING '(DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = host)(PORT = port))) (CONNECT_DATA = (SID = sid)))';
    And execute this query for test dbLink which work fine:
    select * from v$[email protected];
    After it i try to call open function:
    declare
    h number;
    begin
    h := dbms_datapump.open('IMPORT', 'TABLE', 'TEST.LINK', null, '10.2');
    end;
    and get exception: 39001. 00000 - "invalid argument value"
    if i remove 'TEST.LINK' from the arguments it works fine
    Edited by: 990594 on 26.02.2013 23:41

    Hi Richard Harrison,
    result for import from 11g to 10g:
    impdp user/pass@dburl schemas=SCHEMANAME network_link=TEST_LINK version=10.2
    Connected to: Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    ORA-39001: invalid argument value
    ORA-39169: Local version of 10.2.0.1.0 cannot work with remote version of 11.2.0.2.0
    result for export from 11g to 10g:
    expdp user/pass@dburl schemas=SCHEMANAME network_link=TEST_LINK version=10.2
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.2.0 - 64 bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing option
    ORA-39006: internal error
    ORA-39065: unexpected master process exception in DISPATCH
    ORA-04052: error occurred when looking up remote object SYS.KUPM$MCP@TEST_LINK
    ORA-00604: error occurred at recursive SQL level 3
    ORA-06544: PL/SQL: internal, error, arguments: [55916], [], [], [], [], [], [], []
    ORA-06553: PLS-801: internal error [55916]
    ORA-02063: preceding 2 lines from TEST_LINK
    ORA_39097: Data Pump job encountered unexpected error -4052

  • Importing one table question

    I have a DMP file and I want to import all tables except one (it's a large table). I know I can do that with a parameter file, but my question is, can I.........
    import all but the 1 table (the large one), start the import of that 1 table and have people start using the DB while that one table is importing. This large table does not have critical data they need to access right away so my thought was that I could import everything else first (small amount of data), start the import of the large table and they users could access the DB while that 1 table is importing.
    Due to special circumstances IMP/EXP is their only backup solution (please no lectures on that, I KNOW, I know....., but it is what it is)

    I can't think of anything that would prevent this from working. You just need to make sure that the large table does not have any ref constraints, or other associations with the other tables that may get screwed up while the other users are using the database.
    Dean

  • Data pump import problem

    i have exported Hr schema from one macine using following code it export successfully but i tried to import this exported file in another machine it gave error. one thing more how i use package to import file as we exported by using data pump using package plz help
    C:\Documents and Settings\remote>impdp hr/hr DIRECTORY=DATA_PUMP_DIR DUMPFILE=YAHOO.DMP full=y
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-31626: job does not exist
    ORA-31637: cannot create job SYS_IMPORT_FULL_01 for user HR
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPV$FT_INT", line 663
    ORA-39080: failed to create queues "KUPC$C_1_20090320121353" and "KUPC$S_1_20090
    320121353" for Data Pump job
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPC$QUE_INT", line 1665
    ORA-01658: unable to create INITIAL extent for segment in tablespace SYSTEM
    DECLARE
    handle NUMBER;
    BEGIN
    handle := dbms_datapump.open (
    operation => 'EXPORT' ,job_mode => 'SCHEMA');
    dbms_datapump.add_file(
    handle => handle
    *,filename => 'YAHOO.dmp'*
    *,directory => 'DATA_PUMP_DIR'*
    *,filetype => 1);*
    dbms_datapump.metadata_filter(
    handle => handle
    *,name => 'SCHEMA_EXPR'*
    *,value => 'IN(''HR'')');*
    dbms_datapump.start_job(
    handle => handle);
    dbms_datapump.detach(handle => handle);
    end;

    +*<Moderator edit - deleted contents of MOS Doc 752374.1 -  pl do not post such contents - it is a violation of your Support agreement - locking this thread>*+                                                                                                                                                                                                                                                                                                                               

Maybe you are looking for

  • How can an external table handle data with line feed between delimiters?

    I have defined an external table as below. My data is pipe delimited and comes from a DOS system. I already remove any carriage returns before putting the file into the DATA_DIR for reading. But I have found that some of my VARCHAR fields have embede

  • Problems turning off required fields in contact forms

    Hello I have created a form using the "simple contact forms' widget, but I am struggling to turn off required fields. I can see the tick box, but it greyed out, any ideas? Thanks

  • Is bound ?

    Hi As a newbie to ABAP, I would like to know the meaning of the phrase 'is bound' as in: 'IF htmlb_event IS BOUND AND htmlb_event->name = 'button'. ' Can you also refer me to a tutorial on OO ABAP? Thank you yuval

  • Iphoto,imovie,garage band quit unexpectedly

    Hello there! Is anyone here can help me regarding my problem.I have Mac OS X v 10.7.5. Now, my problem is  my ilife '11, abode reader and skype cant open. It says everytime I click it...for example iphoto quit unexpectedly. Pls anyone can help me her

  • Help With Ora-19046:Out-of-line table cannot be shared by two top-level tab

    Hi! My tutor asked me to store and query the xml files in a "native" way in xml db. First I tried to register the schemas on which those xml files are based. There are about 30 of schemas asscociated with each other, and I have to use "FORCE" to regi