Impdp full database require user??

Hi All,
I have taken a expdp backup of full database using "indiamart" user which have the sys privilege.
Now on the other machine i hv created a database using dbca.
Now i want to import that backup here, do i need to create "indiamart" user here also???
Plz guide me what steps i need to follow..
Thanks
Nimish Garg

You need a user with the DATAPUMP_IMP_FULL_DATABASE role granted. Neither the same name nor DBA (or even SYSDBA) privilege is required.
http://download.oracle.com/docs/cd/E11882_01/server.112/e10701/dp_import.htm#i1011935
Werner

Similar Messages

  • Impdp full database advice

    I need to impdp from a full expdp dump. There will be 5 11g exp dmp.
    I want to be sure before I do the import the following steps are right:
    1). drop users cascaed; (to drop all the users except sys/system)
    2). using impdp to import the exp dump files.
    Anything else I need to do ?
    Thank you .

    Pl post exact OS and database versions, along with the complete expdp command used. You will need to ensure that the target database has the same characterset (or a superset characterset like AL32UTF8). Pre-create the tablespaces if the target database has a different file layout than the source.
    http://docs.oracle.com/cd/E11882_01/server.112/e22490/dp_import.htm#i1007653
    HTH
    Srini

  • Import all users and their objects without doing full database import

    Hi Guys,
    I have a task to that involves Importing all existing users and their objects in production to the test database without doing a full database import - Please how do i do this?
    Pls i need your help urgently.

    SQL> select * from v$version;
    BANNER
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Product
    PL/SQL Release 10.2.0.1.0 - Production
    CORE 10.2.0.1.0 Production
    TNS for Linux: Version 10.2.0.1.0 - Production
    NLSRTL Version 10.2.0.1.0 - Production
    I tried to import objects and data from a user from a FULL dump file. File was created with the following command:
    server is: SQL*Plus: Release 10.2.0.1.0 - Production on Wed May 26 15:34:05 2010
    exp full=y FILE="full.dmp" log=full.log
    Now I imported:
    imp file=full.dmp log=full.log INCTYPE=SYSTEM
    imp fromuser=user1 file=full.dmp
    Results: not all the user procedures have been imported:
    SQL> select count(*) from user_procedures;
    the Original
    COUNT(*)
    134
    the current:
    select count(*) from user_procedures;
    COUNT(*)
    18
    I also tried these alternatives:
    exp tablespaces="user1_data" FILE="user1.dmp" log=user1.log
    exp LOG=user1.log TABLESPACES=user1_data FILE=user1_data.dmp
    exp LOG=user1owner.log owner=user1 FILE=user1owner.dmp
    expdp DIRECTORY=dpump_dir1 dumpfile=servdata_pump version=compatible SCHEMAS=user1
    impdp directory=data_pump_dir dumpfile=servdata_pump.dmp :
    ORA-39213: Metadata processing is not available
    SQL> execute dbms_metadata_util.load_stylesheets
    BEGIN dbms_metadata_util.load_stylesheets; END;
    ERROR at line 1:
    ORA-31609: error loading file "kualter.xsl" from file system directory
    "/usr/lib/oracle/xe/app/oracle/product/10.2.0/server/rdbms/xml/xsl"
    ORA-06512: at "SYS.DBMS_METADATA_UTIL", line 1793
    ORA-06512: at line 1
    file kualter.xsl does not exist in XE !!
    imp owner=user1 rows=n ignore=y
    imp full=y file=user1_data.dmp
    imp full=y file=full.dmp tablespaces=user1_data,user1_index rows=n ignore=y
    So, I do not understand why user1 objects are not imported:
    see this part of the first import log:
    Export file created by EXPORT:V10.02.01 via conventional path
    import done in US7ASCII character set and AL16UTF16 NCHAR character set
    import server uses WE8ISO8859P1 character set (possible charset conversion)
    . importing SYS's objects into SYS
    . importing USER1's objects into USER1
    . . importing table .........................
    why only 18 rows?
    if you have an suggestion, you are welcome, as I do not have any other idea...
    ren
    Edited by: ronpetitpatapon on May 26, 2010 12:38 PM
    Edited by: ronpetitpatapon on May 26, 2010 12:41 PM
    Edited by: ronpetitpatapon on May 26, 2010 1:03 PM

  • EXPORT Problem. when i choose USER mode, it is exporting full database

    Dear OTN Members,
    we have a system ORacle 8i, 8.1.3 database, we created a user with DBA role, when we want to export data for a perticular user, after giving USER MODE Option, it is doing full database export including sys, system etc.
    what is the reason?
    Thanks in advance

    Any ideas please,
    Still im looking for this solution. I have a " user" with DBS role, we are trying to export only this user, it is coming up to type of mode like
    1)E ntire database 2)U 3) T ables >
    we are choosing user mode, but all the time it si taking Entire databse, including, Systtem and other users eetc.
    im struggling since 20 days. no success. Please

  • Impdp full Db export to remote database

    Hi techies,
    I have taken a full database export using expdp on my local server/database DEV1.
    We have Linux RHEL 5.6 and Oracle 11.2.0.2 versions.
    The DBA from other group has created a empty database DEV2 on other server.
    I can not connect to that server and start the impdp process there.
    Is there any way to start the impdp process from my local server and do the import to the remote database.
    Thanks,
    Rane

    No.
    Import needs to be done on the target database. If import is done from a file, the file needs to be local. If you dont have room forth local dump file, and you have a network link from the target db pointing to the remote db, then you can use that instead of the dump file.
    Hope this helps.
    Dean

  • Full database import(impdp) error

    HI ppl,
    Plz help..
    My full database import is failing due to
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.LOAD_METADATA [TABLE_DATA:"SYS"."SYS_IMPORT_FULL_04"]
    SELECT process_order, flags, xml_clob, NVL(dump_fileid, :1), NVL(dump_position, :2), dump_length, dump_allocation, grantor, object_row, object_schema, object_long_name, processing_status, processing_state, base_object_type, base_object_schema, base_object_name, property, size_estimate, in_progress FROM "SYS"."SYS_IMPORT_FULL_04" WHERE process_order between :3 AND :4 AND processing_state <> :5 AND duplicate = 0 ORDER BY process_order
    ORA-25153: Temporary Tablespace is Empty
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.KUPW$WORKER", line 6272
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    c000000267348288 14916 package body SYS.KUPW$WORKER
    c000000267348288 6293 package body SYS.KUPW$WORKER
    c000000267348288 3511 package body SYS.KUPW$WORKER
    c000000267348288 6882 package body SYS.KUPW$WORKER
    c000000267348288 1259 package body SYS.KUPW$WORKER
    c00000026fb2de68 2 anonymous block
    Job "SYS"."SYS_IMPORT_FULL_04" stopped due to fatal error at 14:56:10
    regards

    Hi..
    ya temporary tablespace is present,and also contains 2gb freespace,with 2 datafiles.Are you sure its a datafile.It must be a tempfile. :)
    Secondly, have you checked the default temporary tablespace of the database.Is it the one that you have mentioned.
    SELECT PROPERTY_NAME,PROPERTY_VALUE FROM DATABASE_PROPERTIES where PROPERTY_NAME='DEFAULT_TEMP_TABLESPACE';
    Anand

  • Impdp does not create user/schema

    I'm an Oracle noob. I'm trying to copy (expdp/impdp) schema (no data) from one machine to another. I do not want any remapping. I just want empty table structure created on targetHOST.
    Error I get is:
        ORA-39083: Object type PROCACT_SCHEMA failed to create with error:
        ORA-31625: Schema ZABBIX is needed to import this object, but is unaccessible
        ORA-01435: user does not exist
    My understanding is that if the user performing import has 'IMPORT FULL DATABASE' privilege, it'll create the users/schemas in the targetHOST. I granted it to 'scott' and tried with that too, same error.
    - DB version: SQL*Plus: Release 11.2.0.1.0 Production on Thu Oct 24 14:27:41 2013
    - If it matters the DB version on sourceHOST is 11.2.0.***3***
    - OS: Linux targetHOST 2.6.18-308.el5 #1 SMP Fri Jan 27 17:17:51 EST 2012 x86_64 x86_64 x86_64 GNU/Linux
    Import errors:
        (0)oracle@targetHOST$  sqlplus system/manager
        SQL*Plus: Release 11.2.0.1.0 Production on Thu Oct 24 14:27:41 2013
        Copyright (c) 1982, 2009, Oracle.  All rights reserved.
        SQL> select * from session_privs;
        PRIVILEGE
        IMPORT FULL DATABASE
        CREATE SESSION
        .......200 other privileges.......
        202 rows selected.
        SQL> ^D
        (0)oracle@targetHOST$
        (0)oracle@targetHOST$ impdp system/oracle123 directory=TEST_DIR dumpfile=ZABBIX.dmp logfile=impdpZabbix.log
        Import: Release 11.2.0.1.0 - Production on Thu Oct 24 14:14:51 2013
        Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.
        Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
        With the Partitioning, OLAP, Data Mining and Real Application Testing options
        Master table "SYSTEM"."SYS_IMPORT_FULL_01" successfully loaded/unloaded
        Starting "SYSTEM"."SYS_IMPORT_FULL_01":  system/******** directory=TEST_DIR dumpfile=ZABBIX.dmp logfile=impdpZabbix.log
        Processing object type SCHEMA_EXPORT/PRE_SCHEMA/PROCACT_SCHEMA
        ORA-39083: Object type PROCACT_SCHEMA failed to create with error:
        ORA-31625: Schema ZABBIX is needed to import this object, but is unaccessible
        ORA-01435: user does not exist
        Failing sql is:
        BEGIN
        sys.dbms_logrep_imp.instantiate_schema(schema_name=>SYS_CONTEXT('USERENV','CURRENT_SCHEMA'), export_db_name=>'XXX.YYY.COM', inst_scn=>'7788478540892');COMMIT; END;
        Processing object type SCHEMA_EXPORT/TABLE/TABLE
        ORA-39083: Object type TABLE:"ZABBIX"."TABLE1" failed to create with error:
        ORA-01918: user 'ZABBIX' does not exist
        Failing sql is:
        CREATE TABLE "ZABBIX"."TABLE1" ("COLUMN1" VARCHAR2(20 BYTE) NOT NULL ENABLE, "COLUMN2" VARCHAR2(20 BYTE), "COLUMN3" VARCHAR2(20 BYTE)) SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAU
        Processing object type SCHEMA_EXPORT/TABLE/INDEX/INDEX
        Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
        Job "SYSTEM"."SYS_IMPORT_FULL_01" completed with 2 error(s) at 14:14:53
        (5)oracle@targetHOST$
    Export was done using following command (on a different machine)
    (0)oracle@sourceHOST$ expdp zabbix/zabbix schemas=ZABBIX content=METADATA_ONLY directory=TEST_DIR dumpfile=ZABBIX.dmp logfile=expdpZABBIX.log
    Complete export log available here.:

    I figured it out on my own. Required privilege is 'CREATE USER'. I don't have system's password so I granted CREATE USER to zabbix, made it DBA and it seems to be exporting more than it did last time:
    Processing object type SCHEMA_EXPORT/USER
    Processing object type SCHEMA_EXPORT/SYSTEM_GRANT
    Processing object type SCHEMA_EXPORT/ROLE_GRANT
    Processing object type SCHEMA_EXPORT/DEFAULT_ROLE
    Yet to import but looks good so far..
    Appreciate you taking the time..
    Thanks...

  • Expdp full and impdp full fails with error....pls help..!(Oracle 11g)

    Hi
    I wanted to refresh my dev environment quickly..
    As we also wanted to migrate from exp/imp to datapump, I tried these..
    Only thing worked for me is impdp on network mode using remap schema.
    1)First i wanted to implement expdp and impdp standard procedures(Full database export/import)
    My expdp failed with the following error on a full export..
    Pls see below:
    C:\Documents and Settings\oracle2>expdp system/<p.w> full=Y directory=DATA_PUMP_DIR dumpfile=OAPFULL.dmp logfile=OAPFULL.log
    Export: Release 11.1.0.7.0 - Production on Thursday, 26 April, 2012 16:12:49
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYSTEM"."SYS_EXPORT_FULL_02":  system/******** full=Y directory=DATA_PUMP_DIR dumpfile=OAPFULL.dmp logfile=OAPFULL.log
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 4.351 GB
    Processing object type DATABASE_EXPORT/TABLESPACE
    Processing object type DATABASE_EXPORT/PASSWORD_VERIFY_FUNCTION
    Processing object type DATABASE_EXPORT/PROFILE
    Processing object type DATABASE_EXPORT/SYS_USER/USER
    Processing object type DATABASE_EXPORT/SCHEMA/USER
    Processing object type DATABASE_EXPORT/ROLE
    Processing object type DATABASE_EXPORT/GRANT/SYSTEM_GRANT/PROC_SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/GRANT/SYSTEM_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/ROLE_GRANT
    Processing object type DATABASE_EXPORT/SCHEMA/DEFAULT_ROLE
    Processing object type DATABASE_EXPORT/SCHEMA/TABLESPACE_QUOTA
    Processing object type DATABASE_EXPORT/RESOURCE_COST
    Processing object type DATABASE_EXPORT/SCHEMA/DB_LINK
    Processing object type DATABASE_EXPORT/TRUSTED_DB_LINK
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/SEQUENCE
    Processing object type DATABASE_EXPORT/SCHEMA/SEQUENCE/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/DIRECTORY/DIRECTORY
    Processing object type DATABASE_EXPORT/DIRECTORY/GRANT/OWNER_GRANT/OBJECT_GRANT
    Processing object type DATABASE_EXPORT/CONTEXT
    Processing object type DATABASE_EXPORT/SCHEMA/PUBLIC_SYNONYM/SYNONYM
    Processing object type DATABASE_EXPORT/SCHEMA/SYNONYM
    ORA-39014: One or more workers have prematurely exited.
    ORA-39029: worker 1 with process name "DW01" prematurely terminated
    ORA-31672: Worker process DW01 died unexpectedly.
    Job "SYSTEM"."SYS_EXPORT_FULL_02" stopped due to fatal error at 16:14:562)As the above failed, I tried impdp with network_link import and full=y and its a different issue!
    impdp system/<p.w> NETWORK_LINK=OAPLIVE.WORLD full=y logfile=OAPDRfull2.log table_exists_
    action=REPLACE
    Import: Release 11.1.0.7.0 - Production on Thursday, 26 April, 2012 13:36:01
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Starting "SYSTEM"."SYS_IMPORT_FULL_01":  system/******** NETWORK_LINK=OAPLIVE.WORLD full=y logfile=OAPDRfull2.log table_exists_action=REPLACE
    Estimate in progress using BLOCKS method...
    Processing object type DATABASE_EXPORT/SCHEMA/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 4.670 GB
    Processing object type DATABASE_EXPORT/TABLESPACE
    ORA-31684: Object type TABLESPACE:"SYSAUX" already exists
    *************lots more of object already exists errors here............(space limit so cant paste)***************
    ORA-31684: Object type SYNONYM:"GEOPROD"."EXT_POSTAIM_PRESORT_61" already exists
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.DISPATCH_WORK_ITEMS [SYNONYM:"GEOPROD"."EXT_POSTAIM_PRESORT_61"]
    ORA-31600: invalid input value 200001 for parameter HANDLE in function CLOSE
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 105
    ORA-06512: at "SYS.DBMS_METADATA", line 569
    ORA-06512: at "SYS.DBMS_METADATA", line 4731
    ORA-06512: at "SYS.DBMS_METADATA", line 792
    ORA-06512: at "SYS.DBMS_METADATA", line 4732
    ORA-06512: at "SYS.KUPW$WORKER", line 2718
    ORA-03113: end-of-file on communication channel
    ORA-02055: distributed update operation failed; rollback required
    ORA-02063: preceding lines from OAPLIVE.WORLD
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 95
    ORA-06512: at "SYS.KUPW$WORKER", line 7858
    ----- PL/SQL Call Stack -----
      object      line  object
      handle    number  name
    242F2F2C     18256  package body SYS.KUPW$WORKER
    242F2F2C      7885  package body SYS.KUPW$WORKER
    242F2F2C      8657  package body SYS.KUPW$WORKER
    242F2F2C      1545  package body SYS.KUPW$WORKER
    241DDF3C         2  anonymous block
    ORA-39126: Worker unexpected fatal error in KUPW$WORKER.UNLOAD_METADATA []
    ORA-31642: the following SQL statement fails:
    SELECT unique ku$.seq# from sys.metanametrans$ ku$ WHERE ku$.htype='DATABASE_EXPORT' AND ku$.model='ORACLE' AND NOT ( ku$.seq#>=(select a.seq# from sys.metanametrans$ a where
       a.model='ORACLE' and a.htype='DATABASE_EXPORT' and a.name ='DATABASE_EXPORT/SCHEMA/SYNONYM')) order by ku$.seq#
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.DBMS_METADATA_INT", line 5002
    ORA-01427: single-row subquery returns more than one row
    ORA-06512: at "SYS.DBMS_
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.KUPW$WORKER", line 7853
    ----- PL/SQL Call Stack -----
      object      line  object
      handle    number  name
    242F2F2C     18256  package body SYS.KUPW$WORKER
    242F2F2C      7885  package body SYS.KUPW$WORKER
    242F2F2C      2744  package body SYS.KUPW$WORKER
    242F2F2C      8523  package body SYS.KUPW$WORKER
    242F2F2C      1545  package body SYS.KUPW$WORKER
    241DDF3C         2  anonymous block
    Job "SYSTEM"."SYS_IMPORT_FULL_01" stopped due to fatal error at 13:39:48So its basically these 2 issues I want to know how to fix.
    1-expdp error cause and fix
    2-impdp error cause and fix -Also how to avoid object alreay exists error?
    Also for example the package etc. is the same name..but i want new package from LIVE so it means if the same name package, view, etc is there, it wouldnt get updated?
    Any way to overcome this?
    I need it exactly same as LIVE...(with a few exceptions which is small enough i can do after impdp finishes fine)
    Pleaseeeeeeee help!!
    Thanks&Regards.......

    Hi..
    Thanks for the links..I applied the tips on each of them but it didnt work.
    Also my database is 11g so it is not true that this happens on 10g only.
    Things tried:
    1)I tried with different values on parallel parameter but same error
    2)I applied the following:
    alter system set open_cursors=1024 scope=spfile;
    alter system set "_optimizer_cost_based_transformation"=off;
    commit;
    The 3rd link was bit better
    I tried to find out where exactly the error was causesusing
    expdp attach =SYS_EXPORT_FULL_03
    But I cant figure out what the object : PUBLIC
    oracle/context/isearch/GetPage
    is..?
    Does this needs to be excluded from the export?if so, how?
    Can someone help how to fix the error now?
    Processing object type DATABASE_EXPORT/SCHEMA/SYNONYM
    ORA-39014: One or more workers have prematurely exited.
    ORA-39029: worker 1 with process name "DW01" prematurely terminated
    ORA-31672: Worker process DW01 died unexpectedly.
    Job "SYSTEM"."SYS_EXPORT_FULL_03" stopped due to fatal error at 11:29:32
    C:\Documents and Settings\ora2>expdp attach=SYS_EXPORT_FULL_03
    Export: Release 11.1.0.7.0 - Production on Tuesday, 01 May, 2012 11:35:38
    Copyright (c) 2003, 2007, Oracle.  All rights reserved.
    Username: system
    Password:
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Job: SYS_EXPORT_FULL_03
      Owner: SYSTEM
      Operation: EXPORT
      Creator Privs: TRUE
      GUID: 8499C802F52A414A8BCACE552DDF6F11
      Start Time: Tuesday, 01 May, 2012 11:37:56
      Mode: FULL
      Instance: geooap
      Max Parallelism: 1
      EXPORT Job Parameters:
      Parameter Name      Parameter Value:
         CLIENT_COMMAND        system/******** parfile=h:\datapump\oapfull.par
      State: IDLING
      Bytes Processed: 0
      Current Parallelism: 1
      Job Error Count: 0
      Dump File: H:\datapump\oapfull.dmp
        bytes written: 4,096
    Worker 1 Status:
      Process Name: DW01
      State: UNDEFINED
      Object Schema: PUBLIC
      Object Name: oracle/context/isearch/GetPage
      Object Type: DATABASE_EXPORT/SCHEMA/PUBLIC_SYNONYM/SYNONYM
      Completed Objects: 766
      Total Objects: 766
      Worker Parallelism: 1

  • Full database refresh

    Hello Gurus,
    I am need to take make full database refresh from one database to other, Can someone let me know step how to do it?
    select * from V$VERSIONBANNER                                                          
    Oracle Database 10g Enterprise Edition Release 10.2.0.5.0 - 64bi
    PL/SQL Release 10.2.0.5.0 - Production                          
    CORE     10.2.0.5.0     Production                                        
    TNS for HPUX: Version 10.2.0.5.0 - Production                   
    NLSRTL Version 10.2.0.5.0 - Production      

    CKPT wrote:
    jgarry wrote:
    I'm with Ed and Brendan on this. 3 main reasons:
    1. Legacy has some special requirements.
    2. Definitely want different schema names, this has saved my bacon sooo many times. Even so, people still get it wrong (including me, and I'm totally paranoid about it to begin with).
    3. There are many more copies of production than actual production, for various reasons (test, several dev, special projects, future app versions...) . We would rather have one db with many schemata than many db's, right?
    Now if we could just rename user...Agreed some extent, I can say to refresh entire database the preferred method is RMAN restore, where it can manage very easily
    how?
    1) To import again we need to create a new database
    2) Need to create all the tablespaces with same names or we can use REMAP option
    3) need to monitor entire import process,
    4) need to take care of other things like , temp utilization and all.
    5) need to perform export, Need to copy those dumpfiles, We should have enough space on mount points of both source & destinations. ;-)
    I would just caution there that you shouldn't think that because you are taking rman backups you have no need to also regularly export the db. A good backup policy includes both physical backup (rman) and logical backup of the data (export).
    If you talk with couple of schemas the preferred method is EXPDP/IMPDP, totally agreed.
    If in case of database size is huge, then better to go with RMAN to refresh database. Either DUPLICATE or manual way restore.
    as ED said, still we could have chance to test RMAN backups too. Multiple advantages.
    Let me know if I'm wrong. :)

  • Full database exp/imp  between RAC  and single database

    Hi Experts,
    we have a RAC database oracle 10GR2 with 4 node in linux. i try to duplicate rac database into single instance window database.
    there are same version both database. during importing, I need to create 4 undo tablespace to keep imp processing.
    How to keep one undo tablespace in single instance database?
    any experience of exp/imp RAC database into single instance database to share with me?
    Thanks
    Jim
    Edited by: user589812 on Nov 13, 2009 10:35 AM

    JIm,
    I also want to know can we add the exclude=tablespace on the impdp command for full database exp/imp?You can't use exclude=tablespace on exp/imp. It is for datapump expdp/impdp only.
    I am very insteresting in your recommadition.
    But for a full database impdp, how to exclude a table during full database imp? May I have a example for this case?
    I used a expdp for full database exp. but I got a exp error in expdp log as ORA-31679: Table data object "SALE"."TOAD_PLAN_TABLE" has long columns, and longs can not >be loaded/unloaded using a network linkHaving long columns in a table means that it can't be exported/imported over a network link. To exclude this, you can use the exclude expression:
    expdp user/password exclude=TABLE:"= 'SALES'" ...
    This will exclude all tables named sales. If you have that table in schema scott and then in schema blake, it will exclude both of them. The error that you are getting is not a fatal error, but that table will not be exported/imported.
    the final message as
    Master table "SYSTEM"."SYS_EXPORT_FULL_01" successfully loaded/unloaded
    Dump file set for SYSTEM.SYS_EXPORT_FULL_01 is:
    F:\ORACLEBACKUP\SALEFULL091113.DMP
    Job "SYSTEM"."SYS_EXPORT_FULL_01" completed with 1 error(s) at 16:50:26Yes, the fact that it did not export one table does not make the job fail, it will continue on exporting all other objects.
    . I drop database that gerenated a expdp dump file.
    and recreate blank database and then impdp again.
    But I got lots of error as
    ORA-39151: Table "SYSMAN"."MGMT_ARU_OUI_COMPONENTS" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    ORA-39151: Table "SYSMAN"."MGMT_BUG_ADVISORY" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    ......ORA-31684: Object type TYPE_BODY:"SYSMAN"."MGMT_THRESHOLD" already exists
    ORA-39111: Dependent object type TRIGGER:"SYSMAN"."SEV_ANNOTATION_INSERT_TR" skipped, base object type VIEW:"SYSMAN"."MGMT_SEVERITY_ANNOTATION" >already exists
    and last line as
    Job "SYSTEM"."SYS_IMPORT_FULL_01" completed with 2581 error(s) at 11:54:57Yes, even though you think you have an empty database, if you have installed any apps or anything, it may create tables that could exist in your dumpfile. If you know that you want the tables from the dumpfile and not the existing ones in the database, then you can use this on the impdp command:
    impdp user/password table_exists_action=replace ...
    If a table that is being imported exists, DataPump will detect this, drop the table, then create the table. Then all of the dependent objects will be created. If you don't then the table and all of it's dependent objects will be skipped, (which is the default).
    There are 4 options with table_exists_action
    replace - I described above
    skip - default, means skip the table and dependent objects like indexes, index statistics, table statistics, etc
    append - keep the existing table and append the data to it, but skip dependent objects
    truncate - truncate the existing table and add the data from the dumpfile, but skip dependent objects.
    Hope this helps.
    Dean

  • [Q] ORA-30510 error while doing full database import

    We have ORACLE 9.2.0.6 on Redhat LINUX AS 3.4. while we doing "full database import" there has ORA-30510 come out. I searched Metalink and Google, but not found much information. The error messages on import are:
    . importing OUTLN's objects into OUTLN
    . importing SYSTEM's objects into SYSTEM
    . importing TKOWNER's objects into TKOWNER
    IMP-00017: following statement failed with ORACLE error 30510:
    "CREATE TRIGGER "TKOWNER".TRIG_DONOTANALYZE before analyze on schema decla"
    "re begin if sys.dictionary_obj_name = 'MYWTKEMPLOYEE' then raise_ap"
    "plication_error(-20101, 'Do Not ANALYZE MYWTKEMPLOYEE'); end if; end; "
    IMP-00003: ORACLE error 30510 encountered
    ORA-30510: system triggers cannot be defined on the schema of SYS user
    . importing SYS's objects into SYS
    . importing TKOWNER's objects into TKOWNER
    About to enable constraints...
    Import terminated successfully with warnings.

    877410 wrote:
    Hi I get the following ierrors while importing a schema from prod to dev database... can anyone help ,thanks!
    impdp system DIRECTORY=DATA_PUMP_DIR DUMPFILE=abcdprod.DMP LOGFILE=abcdprod.log REMAP_SCHEMA=abcdprod:abcddev
    ORA-39002: invalid operation
    ORA-31694: master table "SYSTEM"."SYS_IMPORT_FULL_01" failed to load/unload
    ORA-31644: unable to position to block number 170452 in dump file "/ots/oracle/echo/wxyz/datapump/abcdprod.DMPpost complete command line for the expdp that made the abcdprod.DMP file

  • How to determine who performed the last full database dump before cumulative dump

    Hello,
    ASE 15.7 SP100 allows cumulative backups, and if cumulative dump is tried before a full dump, ASE shows this error:
    You cannot execute an incremental dump before a full database dump.
    From my experiments, it seems that it does not matter how the full dump is performed (by native isql, or by a 3rd party API). As long as a full dump is done on a database, ASE seems to be keeping track of all the changed pages since the last full dump. So this means that you can perform a full dump using a 3rd party API, and then on isql, if you run a cumulative dump command, the cumulative dump succeeds, which is based on the last full dump by another library!
    So, my question is: is there any way to programmatically determine how (by native isql or 3rd party) the last full backup was performed? I believe $SYBASE_HOME/ASE-15_0/dumphist contains this info, but it requires 'enable dump history' to be set first, and I am looking for a solution which does not involve checking a disk file.
    Thanks,
    Ali

    Dear Mr Razib,
    I have not explored the feature but ASE autiding might provide you the possibility to access information on past database dumps via SQL.
    Apart from that - I am not aware of an SQL interface to the dumphist file (would be nice to have, I agree)
    Enabling dump history is definitley highly recommended  (in my opinion) .
    There is yet another feature which might help you to prevent an DBA from dumping databases and transactions to various locations.
    When you create a DUMP CONFIGURATION and additionally set parameter
            enforce dump configuration
    ASE will prevent normal (free style) DUMP commands but enforce the use of an existing dump configuration. The mechanism is not fool proof (nothing prevents from creating yet another dump configuration on the fly) - but at least something.
    With kind regards
    Tilman Model-Bosch

  • Getting error in full database export

    hi,
    I am exporting full database i am getting this error shown below:
    operating system:windows 2003 server
    oracle:oracle 10g
    C:\Documents and Settings\Administrator.SRCSD>exp system/systemsms file=d:\21041
    0.dmp log=d:\210410.log full=y
    Export: Release 10.1.0.2.0 - Production on ╟ط╟╤╚┌╟┴ ╟╚╤وط 21 11:15:42 2010
    Copyright (c) 1982, 2004, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Produc
    tion
    With the Partitioning, OLAP and Data Mining options
    Export done in AR8MSWIN1256 character set and AL16UTF16 NCHAR character set
    server uses AR8ISO8859P6 character set (possible charset conversion)
    About to export the entire database ...
    . exporting tablespace definitions
    . exporting profiles
    . exporting user definitions
    . exporting roles
    . exporting resource costs
    . exporting rollback segment definitions
    . exporting database links
    . exporting sequence numbers
    . exporting directory aliases
    . exporting context namespaces
    . exporting foreign function library names
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions
    . exporting system procedural objects and actions
    . exporting pre-schema procedural objects and actions
    . exporting cluster definitions
    EXP-00056: ORACLE error 4021 encountered
    ORA-04021: timeout occurred while waiting to lock object
    ORA-06512: at "SYS.DBMS_METADATA", line 1511
    ORA-06512: at "SYS.DBMS_METADATA", line 1548
    ORA-06512: at "SYS.DBMS_METADATA", line 1864
    ORA-06512: at "SYS.DBMS_METADATA", line 3707
    ORA-06512: at "SYS.DBMS_METADATA", line 3689
    ORA-06512: at line 1
    EXP-00056: ORACLE error 4021 encountered
    ORA-04021: timeout occurred while waiting to lock object
    ORA-06512: at "SYS.DBMS_METADATA", line 1511
    ORA-06512: at "SYS.DBMS_METADATA", line 1548
    ORA-06512: at "SYS.DBMS_METADATA", line 1864
    ORA-06512: at "SYS.DBMS_METADATA", line 3707
    ORA-06512: at "SYS.DBMS_METADATA", line 3689
    ORA-06512: at line 1
    EXP-00000: Export terminated unsuccessfully
    C:\Documents and Settings\Administrator.SRCSD>
    plz reply me appropriate solution.
    thanks.

    Possibly bug 3386590 (4021: TIMEOUT OCCURRED WHILE WAITING TO LOCK OBJECT), upgrade to 10gR2 or even better 11gR2.
    Werner

  • Full database export hangs (Oracle9i rel2)

    I recently migrated successfully from oracle8i rel 8.0. to oracle9i release 2,But now am not able to do a full database export.has anyone experienced this and whats the remedy?.
    the text i get is listed below.
    Microsoft Windows [Version 5.2.3790]
    (C) Copyright 1985-2003 Microsoft Corp.
    C:\Documents and Settings\Administrator>exp system/manager@ORCL file=C:\orclbk\p
    oll.dmp full=y
    Export: Release 9.2.0.1.0 - Production on Tue Aug 9 18:36:13 2005
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Connected to: Oracle9i Enterprise Edition Release 9.2.0.1.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.1.0 - Production
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    About to export the entire database ...
    . exporting tablespace definitions
    . exporting profiles
    . exporting user definitions
    . exporting roles
    . exporting resource costs
    . exporting rollback segment definitions
    . exporting database links
    . exporting sequence numbers
    . exporting directory aliases
    . exporting context namespaces
    . exporting foreign function library names
    . exporting PUBLIC type synonyms
    . exporting private type synonyms
    . exporting object type definitions
    EXP-00090: cannot pin type "OE"."CUSTOMER_TYP"
    EXP-00056: ORACLE error 22303 encountered
    OCI-22303: type "OE"."CUSTOMER_TYP" not found
    EXP-00056: ORACLE error 24323 encountered
    ORA-24323: value not allowed
    EXP-00000: Export terminated unsuccessfully
    C:\Documents and Settings\Administrator>

    You have to run the export with the SYS user. This error may arise on any plattform ranging form 9.0.1 to 10.2.0.
    C:\>exp "'sys/oracle as sysdba'" full=y

  • Full database backup and restore

    I have a doubt on the syntax for a full database backup and restore. For a single schema, the syntax for backup & restore is :-
    exp JISPREPCORBILLINGPRD501/JISPREPCORBILLINGPRD501 owner=JISPREPCORBILLINGPRD501 STATISTICS=none
    log=JISPREPCORBILLINGPRD501_exp.log grants=y constraints=Y rows=y compress=N
    file=JISPREPCORBILLINGPRD501
    imp <userid/password> file=<export dump file location> fromuser=<export user name> touser=<import user name>
    grants=y constraints=Y rows=y log=<import log file location>
    I hope, my question is clear.
    Please, help in solving the doubt.
    regards

    hi,
    thanks, for your answer. I took a export of one schema of the database. But, when I tried to import the schema to the destination database, it showed
    imported with errors with the existence of functions showing. The commands used by me for export and import are as follows:-
    exp JISPREPCORBILLINGPRD501/JISPREPCORBILLINGPRD501 owner=JISPREPCORBILLINGPRD501 STATISTICS=none log=JISPREPCORBILLINGPRD501_exp.log grants=y constraints=Y rows=y compress=N file=JISPREPCORBILLINGPRD501
    imp JISPREPCORBILLINGPRD501/JISPREPCORBILLINGPRD501 file=JISPREPCORBILLINGPRD501 fromuser=JISPREPCORBILLINGPRD501 touser=JISPREPCORBILLINGPRD501 grants=y constraints=Y rows=y log=JISPREPCORBILLINGPRD501_imp.log
    Do, we need to drop the existing schema before importing or somethinig else. Please, help.
    regards

Maybe you are looking for