Full export dump to import schema

Hi,
I have an exp dump(datapump) of database, the command use for this:
+/******* directory=expdump DUMPFILE=HYRT.dmp logfile=HYRT_exp.log FULL=y+*
Now, I have imported the dump with the below command:
T24OWNER/******** directory=dor dumpfile=HYRT.dmp logfile=DGFRdmp.impdp.log REMAP_SCHEMA=USER:OWNER REMAP_TABLESPACE=DATAL:BDATA,INDEL:INDEX*
I have got several errors.
Is this errors occured because of the import command where I used imported a schema alone?
If I use full =y parfile in the import command, will these errors can be avoided.
Thanks in advance,
nonuday
Edited by: Nonuday on Apr 29, 2012 10:56 PM
Edited by: Nonuday on Apr 29, 2012 10:57 PM

Hi Vishwanath,
The import got completed.
I have got sum of 16164 errors and I have posted each of its kind below.
1
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
abcd
9
bf d6
0cfc6e4Sm6mfaMYwsbW2JygBepcwg/BKmJ4VZy/pO06UXsVUMejsissTcGWYR4qeK4TPqfDK
q7UPH+SmKP6nW9zmxMZnuK1VDzM0Iv9O4E4SvvsnHWn+EPF9hR+oBFe3fEro6RQ5R5Ejd1nr
e+fAK010dExf76gg/c6ZB3YxGPHDOqkGI4lV6HNsd57gKLwTd0bxan5UwJriIpt7Vjc=
Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/FUNCTION
ORA-31685: Object type FUNCTION:"SYSTEM"."WM$_CHECK_INSTALL" failed due to insufficient privileges. Failing sql is:
CREATE FUNCTION "SYSTEM"."WM$_CHECK_INSTALL" return boolean is begin return false ; end;
Processing object type DATABASE_EXPORT/SCHEMA/PROCEDURE/PROCEDURE
ORA-31685: Object type PROCEDURE:"OUTLN"."ORA$GRANT_SYS_SELECT" failed due to insufficient privileges. Failing sql is:
CREATE PROCEDURE "OUTLN"."ORA$GRANT_SYS_SELECT" as
begin
EXECUTE IMMEDIATE 'GRANT SELECT ON OUTLN.OL$ TO SELECT_CATALOG_ROLE';
EXECUTE IMMEDIATE 'GRANT SELECT ON OUTLN.OL$HINTS TO SELECT_CATALOG_ROLE';
EXECUTE IMMEDIATE 'GRANT SELECT ON OUTLN.OL$NODES TO SELECT_CATALOG_ROLE';
end;
ORA-31685: Object type PROCEDURE:"SYSTEM"."ORA$_SYS_REP_AUTH" failed due to insufficient privileges. Failing sql is:
CREATE PROCEDURE "SYSTEM"."ORA$_SYS_REP_AUTH" as
begin
EXECUTE IMMEDIATE 'GRANT SELECT ON SYSTEM.repcat$_repschema TO SYS ' ||
'WITH GRANT OPTION';
EXECUTE IMMEDIATE 'GRANT SELECT ON SYSTEM.repcat$_repprop TO SYS ' ||
'WITH GRANT OPTION';
EXECUTE IMMEDIATE 'GRANT SELECT ON SYSTEM.def$_aqcall TO SYS ' ||
'WITH GRANT OPTION';
EXECUTE
Processing object type DATABASE_EXPORT/SCHEMA/FUNCTION/ALTER_FUNCTION
ORA-31685: Object type ALTER_FUNCTION:"SYSTEM"."WM$_CHECK_INSTALL" failed due to insufficient privileges. Failing sql is:
ALTER FUNCTION "SYSTEM"."WM$_CHECK_INSTALL" COMPILE PLSQL_OPTIMIZE_LEVEL= 2 PLSQL_CODE_TYPE= INTERPRETED PLSQL_DEBUG= FALSE REUSE SETTINGS TIMESTAMP '2010-11-01 16:05:09'
Processing object type DATABASE_EXPORT/SCHEMA/PROCEDURE/ALTER_PROCEDURE
ORA-31685: Object type ALTER_PROCEDURE:"OUTLN"."ORA$GRANT_SYS_SELECT" failed due to insufficient privileges. Failing sql is:
ALTER PROCEDURE "OUTLN"."ORA$GRANT_SYS_SELECT" COMPILE PLSQL_OPTIMIZE_LEVEL= 2 PLSQL_CODE_TYPE= INTERPRETED PLSQL_DEBUG= FALSE REUSE SETTINGS TIMESTAMP '2010-11-01 16:05:09'
ORA-31685: Object type ALTER_PROCEDURE:"SYSTEM"."ORA$_SYS_REP_AUTH" failed due to insufficient privileges. Failing sql is:
ALTER PROCEDURE "SYSTEM"."ORA$_SYS_REP_AUTH" COMPILE PLSQL_OPTIMIZE_LEVEL= 2 PLSQL_CODE_TYPE= INTERPRETED PLSQL_DEBUG= FALSE REUSE SETTINGS TIMESTAMP '2010-11-01 16:05:09'
Processing object type DATABASE_EXPORT/SCHEMA/VIEW/VIEW
ORA-31685: Object type VIEW:"SYSTEM"."MVIEW_WORKLOAD" failed due to insufficient privileges. Failing sql is:
CREATE FORCE VIEW "SYSTEM"."MVIEW_WORKLOAD" ("WORKLOADID", "IMPORT_TIME", "QUERYID", "APPLICATION", "CARDINALITY", "RESULTSIZE", "LASTUSE", "FREQUENCY", "OWNER", "PRIORITY", "QUERY", "RESPONSETIME") AS select
a.collectionid# as workloadid,
a.collecttime as import_time,
a.queryid# as queryid,
a.application,
a.cardinality,
a.resultsize,
a.qdate as lastuse,
a.frequency,
a.uname as o
Processing object type DATABASE_EXPORT/SCHEMA/VIEW/GRANT/OWNER_GRANT/OBJECT_GRANT
ORA-39112: Dependent object type OBJECT_GRANT:"SYSTEM" skipped, base object type VIEW:"SYSTEM"."MVIEW_WORKLOAD" creation failed
Processing object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
Processing object type DATABASE_EXPORT/SCHEMA/PASSWORD_HISTORY
ORA-39083: Object type PASSWORD_HISTORY failed to create with error:
ORA-06550: line 1, column 33:
PLS-00201: identifier 'SYS.DBMS_PSWMG_IMPORT' must be declared
ORA-06550: line 1, column 10:
PL/SQL: Item ignored
These are the errors in the log..
I have done import from many client export dump before but they are schema level export only.
This one is full database export where I have imported only the schema from the full exp dump. Is this the reason why am getting so many errors or anything else can you throw some light.
Thanks in advance.
nonuday
Edited by: Nonuday on Apr 29, 2012 11:03 PM

Similar Messages

  • Import of Table data only from Full export Dump

    I want to import data inot specified tables of a databse only. Is it possible to do this from a dump file which is the full export of the another database. The database i want to import into is the mirror of the database whose export dump i want to use. Let me know if this is possible and if yes then how i could do this.
    Thanking in advance.

    Please check the following link for full instructions:
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96652/part1.htm#435787
    If on another release, the same documentation is available for it on OTN as well.
    Thanks!

  • Import few schemas out of a full DB dump file into a new DB

    Hi,
    I have a 11g database, ORCL.
    I have created a new database, TEST, where i want only 3 schemas of ORCL database to be imported.
    I have taken an export dump file of the 3 schemas (A,B and C), with which the below import command works:
    impdp system/xxxx directory=EXPDP dumpfile=schema3.dmp logfile=schema3.log REMAP_TABLESPACE=A_TBS:USERS,B_TBS:USERS,C_TBS:USERS
    where A_TBS, B_TBS and C_TBS are the tablespaces of the respective schemas A,B and C in ORCL database
    USERS is the default tablespace in the TEST database
    But If I take a full export dump of the entire database and want to import only the 3 required schemas, then is the below command correct???
    impdp system/xxxx directory=EXPDP dumpfile=full_dump.dmp logfile=full_dump.log SCHEMAS=A,B,C REMAP_TABLESPACE=A_TBS:USERS,B_TBS:USERS,C_TBS:USERS
    Please guide.
    Thanks in Advance.

    If you do a FULL export and FULL import, the import also attempts to create Tablespaces that do not exist. However, it will attempt to create the tablespaces with the same file names (including directory paths) as present on the source database. That is why it is always preferable to precreate Tablespaces at the new database before running the import -- so that you can control the sizes and names of datafiles, even change tablespace parameters if necessary. REMAP would be used if you want to import to different tablespace names. When precreating tablespaces, the size required really depends on the size of the data.
    At the source you might have a tablespace with datafiles adding up to 5GB but used segments of 3GB only (Again, within the 3GB, 80% of the data might have been deleted before the export !). So the size required at the new database might be 3GB or only 600MB !! It is best to precreate a few datafiles with some reasonable sizes and have them with AUTOEXTEND ON so that the tablespaces grow to the actual size that is really required.
    Hemant K Chitale

  • Full Expdp & Try to import single schema

    Hi All,
    I have full expdp backup & i want to do single schema import.
    When i m trying to import single schema it started full import.
    Export :-
    expdp praful/praful directory=dump dumpfile=full.dmp logfile=full.log full=y
    Import :-
    impdp praful/praful directory=dump dumpfile=full.dmp logfile=imp_praful.log remap_schema=PRAFUL:TEST
    Kindly give a suggetion for single schema import from full export dump.
    Regards,
    Praful

    Hi,
    Don't forget to include the parameter SCHEMAS - which provides the details, wher as remap_schema - suffice the details for remapping instead of importing to current schema which you are connected with respect to current session.
    - Pavan Kumar N

  • I try the first time to export or import schemas

    Hi
    I have never used Data moving utilities. I read some from guides and blogs, but I confused.
    Can you help me with simple examples export and import data* and directory_
    Thanks

    I think you want this.
    -- on windows os
    create or replace directory exportfiles on 'C:\exportfiles';
    --- exportfiles folder created on C: directory and we are created Oracle directory on this folder.
    grant read, write on directory exportfiles to mahir;
    -- we are give read, write priviligies to  mahir user on this  directory
    C:\> expdp  mahir/mahir@testdb DIRECTORY=exportfiles FILE=mahirshemadump.dmp SCHEMAS=mahir;
    -- exporting DUMP mahir users schema to exportfiles directory
    C:\> impdp  mahir/mahir@testdb1 DIRECTORY=exportfiles FILE=mahirshemadump.dmp SCHEMAS=mahir;
    -- and importing this dump to other database.
    -- in testdb1 create exportfiles directory too.
    and read same link on Sunny kichloo's post:
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/exp_imp.htm
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_export.htm
    Edited by: Mahir M. Quluzade on May 10, 2011 11:31 AM

  • Finding create user statement using the export dump

    I have an Oracle export dump and it is a schema level export only.
    Is there any way that I can get the create user statement from that dump using import full=y show=y ?
    How do I identify the create user statement once I did the show=y

    Yes with full export this is possible because full export will create all schemas but this is not possible with schema level export because schema level import does not recreate schema.

  • Getting Datapump Export Dump file to the local machine

    I apologize to everyone as this is a duplicate post.
    Re: Getting Datapump Export Dump file to the local machine
    My initial thread(started yesterday)was in 'Database General' and didn't get much response today. Where do i post questions on EXPORT/IMPORT utilities?
    Anyway, here is my problem:
    I want to take the export dump of itemrep schema in orcl database (in a remote machine). I have an Oracle server (10G Rel2) running in my local Windows machine. I have created a user john with necessary EXPORT/IMPORT privileges in my local db. Then i created a Directory object,ie a folder named datapump in my local hard drive and granted READ WRITE privileges to john.
    So john, who is a user in my local machine's oracle db is going to run the expdp utility.
    expdp john/jhendrix@my_local_db_alias SCHEMAS=itemrep directory=datapump logfile=itemrepexp.log
    The above command will fail because it will look for itemrep schema inside my local db, not the remote db where the itemprep is actually located. And you can't qualify the schemaname with its db in the SCHEMAS parameter (like SCHEMAS=itemrep@orcl).
    Can anyone provide me a solution for this?

    I think you can initiate the datapump exp utility from your client machine to export a schema in a remote database.But, Upon execution,oracle looks for the directory in the remote database and not on your local machine.
    You're inovoking expdp from a client (local DB) to export data from a remote DB.
    So, With this method, you can create the dumpfiles only on the Remote server and not on the local Machine.
    You can perform a direct import instead of export using the NETWORK_LINK option.
    Create a DBlink from your local and Remote DB and verify the connection.
    Then,Initiate the Impdp from Your local machine DB using the parameter network_link=<db_link of the Remote DB> to import the schema.
    The advantage of this option eliminates the Dumpfile creation at the Server side.
    There are no dumpfiles during the import process. the Data is imported directly to the target schema.

  • Error while import schema in HANA DB

    Hi...
    I am trying to perform an export-import of a HANA DB user instance at schema level including all objects (in my case tables and synonyms) and data.
    Export and Import works very well except importing a schema which contains a synonym referred to other existing schema .
    Here is a test case
    - Schema A with a Table and Schema B to contain a synonym for schema A's table.
    - Exported both schemas to CSV which is done successfully and everything looked ok in export location.
    - Imported Schema A and then Schema B with the following command.
    Command:
    IMPORT "<bkup_schema>"."*" FROM '<bkp_location>' WITH RENAME SCHEMA <bkup_schema> TO <brand_new_schema>;
    The schema A (with table) imported fine.  Schema B (with synonym) gives the below error :
    "SAP DBTech JDBC: [2048]: column store error: table import failed:  [2003] An index already exists with the same name "
    When trying to use "REPLACE" clause in the command the error changes as below:
    SAP DBTech JDBC: [2048]: column store error: table import failed:  [30117] Binary import failed (cannot execute drop statement)
    Has anybody faced this type of error before??

    Folks, I am running into the same issue. What I figure out was the following:
    1) I turn off the autocommit
    2) I run the import for the first time
    import "<SCHEMA_NAME>"."*" as binary from '<location>' with replace;
    rollback;
    3) I get the error as described
    SAP DBTech JDBC: [2048]: column store error: table import failed:  [30117] Binary import failed (cannot execute drop statement)
    4) I run this import once again and commit and surprisingly the error is gone and the import is successful.
    import "<SCHEMA_NAME>"."*" as binary from '<location>' with replace;
    commit;

  • Export and re-import on same database

    For one reason or another we want to do a full export and re-import all the data back into the same tablespaces... is there a command to imp to tell it to truncate and over-write all the data before importing? or do I have to manually drop all the tables? (by manual I mean write a script to do it) -- and/or drop all the schemas

    Not sure, why you want to export and import back the same thing into the same database, but anyways to answer your question, yes it is possible with Oracle10g by using TABLE_EXISTS_ACTION={SKIP | APPEND | TRUNCATE | REPLACE}. With Oracle9i, you may have to generate a dynamic script to truncate tables....make sure you exclude SYS/SYSTEM tables from the truncate script.........did I say except SYS/SYSTEM and so on :-)
    HTH
    Thanks
    Chandra Pabba

  • Export dump(logical backup) runing since yesterday

    Hi to All,
    We are taking full export dump(logical backup) of database.till last week its fine on sunday we change the password of administrator and we forgot to change to password in this scheduled scripts.and yesterday i run this backup manullay in morning and i saw its affecting our performance I cancel this task in middle only I haven't deleted that files from backup location and I had given the password in that scirpts and that scripts runing and that scripts started runing at there aproprite time but the incomplete backup is existed at backup location and the backup which started at scheduled time is runing since yesterday before its fininsh in 2 or 3 hours but know its runing since yesterday.Please give me some solution.
    Thanks and Regards
    Mohd Khaja

    On windows it gets a bit hairy as there really is no clean and nice way of doing it.There are a couple of options.
    1. If you can rely on dateformat not changing, you can use a static substring expression. For example, the following might work w/ finnish locale echo %date:~3,2%%date:~6,2%%date:~9,4%Similarly, when you know the dateformat you can tokenize the output of 'date /t' and discard the tokens you don't want.
    2. You can set dateformat to your liking and then just use %date% in your script
    3. You can run a "SELECT to_char(sysdate,..." into a file and then read that file in your script.
    4. Simon Sheppard also has a solution you could use as a basis. I have a slight issue with the approach, but that could just be me.
    5. Use gnuwin32 or similar ;)
    Also note that %date% env var is set automatically from w2k onwards, so some of the solutions might not work w/ older versions.

  • Export/Import full database dump fails to recreate the workspaces

    Hi,
    We are studying the possibility of using Oracle Workspace to maintain multiple versions of our business data. I recently did some tests with the import/export of a dump that contains multiple workspaces. I'm not able to import the dump successfully because it doesn't recreate the workspaces on the target database instance, which is an empty Oracle database freshly created. Before to launch the import with the Oracle Data Pump utility, I have the "LIVE" workspace that exists in the WMSYS.WM$WORKSPACES_TABLE table. After the import is done, there is nothing in that table...
    The versions of the Oracle database and Oracle Workspace Manager are the same on source and target database:
    Database version (from the V$VERSION view):
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    PL/SQL Release 10.2.0.4.0 - Production
    CORE     10.2.0.4.0     Production
    TNS for 64-bit Windows: Version 10.2.0.4.0 - Production
    NLSRTL Version 10.2.0.4.0 - Production
    Workspace Manager version (from the WM_INSTALLATION view):
    OWM_VERSION: 10.2.0.4.3
    In order to recreate the tablespaces successfully during the full import of the dump, the directory structure for the tablespaces is the same on the target database's computer. I used the instructions given in this document, section "1.6 Import and Export Considerations" at page 1-19:
    http://www.wyswert.com/documentation/oracle/database/11.2/appdev.112/e11826.pdf
    Considering that the release of Oracle database used is version 10.2.0.4, I use the following command to import the dump since it doesn't contain the WMSYS schema:
    impdp system/<password>@<database> DIRECTORY=data_pump_dir DUMPFILE=expfull.dmp FULL=y TABLE_EXISTS_ACTION=append LOGFILE=impfull.log
    The only hints I have in the import log file are the following block of lines:
    1st one:
    ==============================
    Traitement du type d'objet DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    ORA-39083: Echec de la création du type d'objet PROCACT_SYSTEM avec erreur :
    ORA-01843: ce n'est pas un mois valide
    SQL en échec :
    BEGIN
    if (system.wm$_check_install) then
    return ;
    end if ;
    begin execute immediate 'insert into wmsys.wm$workspaces_table
    values(''LIVE'',
    ''194'',
    ''-1'',
    2nd one at the end of the import process:
    ==============================
    Traitement du type d'objet DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    ORA-39083: Echec de la création du type d'objet PROCACT_SCHEMA avec erreur :
    ORA-20000: Workspace Manager Not Properly Installed
    SQL en échec :
    BEGIN
    declare
    compile_exception EXCEPTION;
    PRAGMA EXCEPTION_INIT(compile_exception, -24344);
    begin
    if (system.wm$_check_install) then
    return ;
    end if ;
    execute immediate 'alter package wmsys.ltadm compile';
    execute immediate 'alter packag
    The target operating system is in french... here is a raw translation:
    1st one:
    ==============================
    Treatment of object type DATABASE_EXPORT/SYSTEM_PROCOBJACT/PRE_SYSTEM_ACTIONS/PROCACT_SYSTEM
    ORA-39083: Object type creation failed PROCACT_SYSTEM with error :
    ORA-01843: invalid month
    failed SQL :
    BEGIN
    if (system.wm$_check_install) then
    return ;
    end if ;
    begin execute immediate 'insert into wmsys.wm$workspaces_table
    values(''LIVE'',
    ''194'',
    ''-1'',
    2nd one at the end of the import process:
    ==============================
    Treatment of object type DATABASE_EXPORT/SCHEMA/POST_SCHEMA/PROCACT_SCHEMA
    ORA-39083: Object type creation failed PROCACT_SCHEMA with error :
    ORA-20000: Workspace Manager Not Properly Installed
    failed SQL :
    BEGIN
    declare
    compile_exception EXCEPTION;
    PRAGMA EXCEPTION_INIT(compile_exception, -24344);
    begin
    if (system.wm$_check_install) then
    return ;
    end if ;
    execute immediate 'alter package wmsys.ltadm compile';
    execute immediate 'alter packag
    By the way, the computer of the source database is Vista 64 in english and the destination computer is Vista 64 in french, do you think this has anything to do with these error messages? The parameters of the NLS_SESSION_PARAMETERS view are the same in the source and target database...
    Thanks in advance for your help!
    Joel Autotte
    Lead Developer
    Edited by: 871007 on Jul 13, 2011 7:31 AM

    I tried to import the full database dump with the "imp" command and I had the same error message with more detail:
    . Import d'objets SYSTEM dans SYSTEM
    IMP-00017: Echec de l'instruction suivante avec erreur ORACLE 1843 :
    "BEGIN "
    "if (system.wm$_check_install) then"
    " return ;"
    " end if ;"
    " begin execute immediate 'insert into wmsys.wm$workspaces_ta"
    "ble"
    " values(''LIVE'',"
    " ''194'',"
    " ''-1'',"
    " ''SYS'',"
    " *to_date(''09-DEC-2009 00:00:00'', ''DD-MON-YYYY HH"*
    *"24:MI:SS''),"*
    " ''0'',"
    " ''UNLOCKED'',"
    " ''0'',"
    " ''0'',"
    " ''37'',"
    " ''CRS_ALLNONCR'',"
    " to_date(''06-APR-2011 14:24:57'', ''DD-MON-YYYY HH"
    "24:MI:SS''),"
    " ''0'',"
    " '''')'; end;"
    "COMMIT; END;"
    IMP-00003: Erreur ORACLE 1843 rencontrée
    ORA-01843: ce n'est pas un mois valide
    ORA-06512: à ligne 5
    The call to the TO_DATE function with the string of the date format is incompatible with the format and language set in the NLS parameters of my database. Since I know that the "imp" and "exp" commands work client side (unlike "impdp" and "expdp"), I did a full export of the source database from the computer on which I try to import the dump to have the same language exported in the dump and I was able to import the dump successfully.
    But, I would like to use the Data Pump tool instead of the old import and export tools. How would it be possible to set the NLS parameters that "impdp" must use before to launch the import of the dump?
    The NLS parameters are set to the "AMERICAN" settings in both source and destination database when I validate that with the "NLS_DATABASE_PARAMETERS" view. The dump I did with "expdp" on the source database exported the data with the "AMERICAN" settings as well...
    On the destination database, the computer language is in french and I guess it is considered by the Data Pump import tool since it outputs the log in the language of the computer and it tries to do a TO_DATE with the wrong date format because it works with the "CANADIAN FRENCH" settings...
    Any suggestions? I'll continue to read on my side and re-post if I find the solution.
    Thanks!
    Edited by: 871007 on Jul 19, 2011 8:42 AM
    Edited by: 871007 on Jul 19, 2011 8:43 AM
    Edited by: 871007 on Jul 19, 2011 8:43 AM

  • HOW TO IMPORT DATA MORE THAN ONCE FROM THE SAME EXPORT DUMP FILE?

    before asking my question i'd like to mention that i'm a french spoke...
    so my english is a little bit bad. sorry for that.
    my problem is : IMPORT
    how to import data a SECOND TIME from an export dump file within oracle?
    My Export dump file was made successfully (Full Export) and then i
    tried to import datas for the first time.
    I got this following message in my logfile: I ADDED SOME COMMENTS
    Warning: the objects were exported by L1, not by you
    . importing SYSTEM's objects into SYSTEM
    REM ************** CREATING TABLESPACES *****
    REM *********************************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "USER_DATA" DATAFILE 'E:\ORANT\DATABASE\USR1ORCL.ORA' SI"
    "ZE 3145728 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAX"
    "EXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "ROLLBACK_DATA" DATAFILE 'E:\ORANT\DATABASE\RBS1ORCL.ORA"
    "' SIZE 10485760 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS "
    "1 MAXEXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    etc........
    IMP-00017: following statement failed with ORACLE error 1119:
    "CREATE TABLESPACE "L1" DATAFILE 'E:\ORADATA\L1.DBF' SIZE 1048576000 "
    "DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAXEXTENTS 121 PCTIN"
    "CREASE 50) ONLINE PERMANENT"
    IMP-00003: ORACLE error 1119 encountered
    ORA-01119: error in creating database file 'E:\ORADATA\L1.DBF'
    ORA-09200: sfccf: error creating file
    OSD-04002: unable to open file
    O/S-Error: (OS 3) The system cannot find the path specified
    --->etc..........
    the drive E: with the folder E:\ORADATA didn't exist, but after
    all that i created it.
    see below, before my IMPORT statement
    REM ********************* CREATING USER *********
    REM ********************************************
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "L1" IDENTIFIED BY VALUES 'A6E0DAA6865E7627' DEFAULT TABLESPACE"
    " "L1" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'L1' does not exist
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "MLCO" IDENTIFIED BY VALUES '56AC6447B7D50467' DEFAULT TABLESPA"
    "CE "MLCO" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'MLCO' does not exist
    ETC.......
    REM ********************* GRANTING ROLES ***********
    REM ************************************************
    IMP-00017: following statement failed with ORACLE error 1917:
    "GRANT ALTER ANY TABLE to "L1" "
    IMP-00003: ORACLE error 1917 encountered
    ORA-01917: user or role 'L1' does not exist
    ETC.........
    IMP-00017: following statement failed with ORACLE error 1918:
    "ALTER USER "L1" DEFAULT ROLE ALL"
    IMP-00003: ORACLE error 1918 encountered
    ORA-01918: user 'L1' does not exist
    -- that is normal, since the creation of the
    tablespace failed !!
    REM******************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLLBACK SEGMENT RB_TEMP STORAGE (INITIAL 10240 NEXT 10240 MINEXTENT"
    "S 2 MAXEXTENTS 121) TABLESPACE "SYSTEM""
    IMP-00015: following statement failed because
    . importing SCOTT's objects into SCOTT
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "EVT_PROFILE_SEQ" MINVALUE 1 MAXVALUE 999999999999999999999"
    "999999 INCREMENT BY 1 START WITH 21 CACHE 20 NOORDER NOCYCLE"
    ETC............
    importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00003: ORACLE error 1435 encountered
    ORA-01435: user does not exist
    REM *************** IMPORTING TABLES *******************
    REM ****************************************************
    . importing SYSTEM's objects into SYSTEM
    . . importing table "AN1999_BDAT" 243 rows imported
    . . importing table "BOPD" 112 rows imported
    . . importing table "BOINFO_AP" 49
    ETC................
    . . importing table "BO_WHF" 2 rows imported
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLE "DEF$_CALL" ("DEFERRED_TRAN_DB" VARCHAR2(128),
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ROLES" FOR "SYS"."DBA_ROLES""
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ERRORS" FOR "SYS"."DBA_ERRORS""
    IMP-00008: unrecognized statement in the export file:
    . importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00008: unrecognized statement in the export file:
    J
    Import terminated successfully with warnings.
    -------------------------------------b]
    So after analysing this log file, i created
    the appropriate drives and folders... as the
    import statement doesn't see them.
    E:\ORADATA G:\ORDATA etc...
    And i started to [b]IMPORT ONE MORE TIME. with:
    $ IMP73 sys/pssw Full=Y FILE=c:\temp\FOLD_1\data_1.dmp BUFFER=64000
    COMMIT=Y INDEXFILE=c:\temp\FOLD_1\BOO_idx.sql
    LOG=c:\temp\FOLD_1\BOO_log.LOG DESTROY=Y IGNORE=Y;
    after that i could not see the users nor the
    tables created.
    and the following message appeared in the log file:
    Warning: the objects were exported by L1, not by you
    . . skipping table "AN1999_BDAT"
    . . skipping table "ANPK"
    . . skipping table "BOAP"
    . . skipping table "BOO_D"
    ETC.....skipping all the tables
    . . skipping table "THIN_PER0"
    . . skipping table "UPDATE_TEMP"
    Import terminated successfully without warnings.
    and only 2 new tablespaces (originally 3) were
    created without any data in ( i check that in
    the Oracle Storage manager : the tablespaces exit
    with 0.002 used space; originally 60 M for each !!)
    so,
    How to import data (with full import option) succefully
    MORE THAN ONE TIME from an exported dump file ?
    Even if we have to overwrite tablespaces , tables and users.
    thank you very much

    The Member Feedback forum is for suggestions and feedback for OTN Developer Services. This forum is not monitored by Oracle support or product teams and so Oracle product and technology related questions will not be answered. We recommend that you post this thread to the appropriate Database forum.
    The main URL is:
    http://forums.oracle.com/forums/index.jsp?cat=18

  • Full Export/Import Errors with Queue tables and ApEx

    I'm trying to take a full export of an existing database in order to build an identical copy in another database instance, but the import is failing each time and causing problems with queue tables and Apex tables.
    I have used both the export utility and Data Pump (both with partial and full exports) and the same problems are occurring.
    After import, queue tables in my schema are unstable. They cannot be dropped using the queue admin packages as they throw ORA-24002: QUEUE_TABLE <table> does not exist exceptions. Trying to drop the tables causes the ORA-24005: must use DBMS_AQADM.DROP_QUEUE_TABLE to drop queue tables error
    As a result, the schema cannot be dropped at all unless manual data dictionary clean up steps (as per metalink) are done.
    The Apex import fails when creating foreign keys to WWV_FLOW_FILE_OBJECTS$PART. It creates the table ok, but for some reason the characters after the $ are missing so the referencing tables try to refer to WWV_FLOW_FILE_OBJECTS$ only.
    I am exporting from Enterprise Edition 10.2.0.1 and importing into Standard edition 10.2.0.1, but we are not using any of the features not available in standard, and I doubt this would cause the issues I'm getting.
    Can anyone offer any advice on how I can resolve these problems so a full import will work reliably?

    Thanks for the lead!
    After digging around MetaLink some more, it sounds like I'm running into Bug 5875568 (MetaLink Note:5875568.8) which is in fact related to the multibyte character set. The bug is fixed in the server patch set 10.2.0.4 or release 11.1.0.6.

  • Missing grant with full export/full import.

    We have the following problem which is unexplained:
    Using Oracle 9.0.1.4 we full export a database and we import it in full mode in another database which is also running Oracle 9.0.1.4. The database has 1 application schema which owns a lot of PL/SQL object packages and triggers.
    In the source database, all objects are valid and schema owner has SELECT privileges on V$SESSION and execute permission on DBMS_ALERT and DBMS_PIPE.
    I cannot tell you if these privileges have been granted directly or with a role.
    In the target database, package bodys and triggers using V$SESSION and DBMS_ALERT/DBMS_PIPE are invalid because schema owner has not the corresponding privileges.
    Schema owner already existed in the target database. Is it normal than the corresponding privileges have not been imported ?
    I cannot give you the exp/imp command lines but I'm sure full export/full import have been run (I've checked the import log).
    Thanks.
    Message was edited by:
    Pierre Forstmann

    See MetaLink Note:97902.1
    When performing a full export of a database, both objects belonging to SYS and object privileges on SYS's objects (e.g. admin views such as DBA_SEGMENTS, etc.) are not exported.
    To generate the missing privileges you can use the follwing script
    Connect to the original database (the one full export was done from) as SYSTEM using sqlplus and run this script:
    set hea off
    set pagesize 0
    set feedback off
    spool objs_upg.sql
    select 'grant '||privilege||' on '||owner||'.'||
    table_name||' to '||grantee||' '||
    decode(grantable,'YES','WITH Grant option')||';'
    from dba_tab_privs
    where owner = 'SYS'
    select 'grant '||privilege||' ('||column_name||') '||
    ' on '||owner||'.'||table_name||' to '||grantee||' '||
    decode(grantable,'YES','WITH Grant option')||';'
    from dba_col_privs
    where owner = 'SYS'
    spool off

  • User & Schema name of Export dump file

    I have an exported dump file. I dont have any log file of that export. I want to Import it in another system. How do I know concerned user or schema name of that dump file.
    Is there any way to know it???

    dbanirav wrote:
    I have an exported dump file. I dont have any log file of that export. I want to Import it in another system. How do I know concerned user or schema name of that dump file.
    Is there any way to know it???made with exp or expdp?
    what OS name & version?
    How do I ask a question on the forums?
    SQL and PL/SQL FAQ

Maybe you are looking for

  • Best Practice Problems

    Trying to be a good developer :) and so i read many articles on the best way to do things and try to adopt these in my developments. For my recent JSF developments this has included dropping JSTL and trying to implement everything in 'pure JSF' way.

  • Cracked screen... Cold weather?

    So last night I was using my phone and took it out of my sweatshirt pocket and noticed it was cracked after being outside in the cold. I hadn't dropped it or smashed it into anything. It looks as though I pressed the home button and it cracked up fro

  • Syncing issues with iPhone 4 & iTunes

    Anyone else having problems with syncing between iPhone 4 and iTunes 9.2? Keep getting syncing error messages when I'm trying to add music - either required file not found or duplicate file name is specified. iTunes then hangs. Leaves me having to in

  • Database Read only

    Hi All, Can SAP application connect to read only SQL Server database ? We are trying to use secondary database of log shipping, for our operational adhoc reporting of production data. Please let me know. Thanks Mushtaq

  • Has anyone experienced a activity actual start date moving when updating progress

    When updating an activity progress in a small schedule i notice that one of my activities which had already started  on march 19th and was about 10% complete from a prior update jumped to March 23rd as start date. What is going on???