EXPDP/IMPDP : Clone schema with datapump

Hello,
I have a full export with the following parameters:
DIRECTORY=dir$export
DUMPFILE=datapump_exp_FULL.dmp
LOGFILE=datapump_exp_FULL.log
FULL=Y
And I want to import only one schema into a new schema with the following parameters:
DIRECTORY=dir$export
DUMPFILE=datapump_exp_FULL.dmp
LOGFILE=datapump_imp_SCHEMA2.log
SCHEMAS=SCHEMA1
REMAP_SCHEMA=SCHEMA1:SCHEMA2
TABLE_EXISTS_ACTION=REPLACE
TRANSFORM=OID:N
But for what reason ever IMPDP tries to import other schemas like SYSMAN, APEX... also. This throws a load of ORA-31684 (object already exists), what means TABLE_EXISTS_ACTION=REPLACE doesn't work also.
I also tried to import without SCHEMAS=SCHEMA1, but didn't succeed anyway...
So, how can I force IMPDP to import only SCHEMA1, without touching the others ?
Regards,
Mynz

either use INCLUDE PARAMETER as per documentation
http://docs.oracle.com/cd/E11882_01/server.112/e10701/dp_import.htm#i1007761OR
EXCLUDE=SCHEMA:"IN ('OUTLN','SYSTEM','SYSMAN','FLOWS_FILES','APEX_030200','APEX_PUBLIC_USER','ANONYMOUS')"OR
Using NOT-LIKE with EXCLUDETake care of those quotation marks..

Similar Messages

  • Clone schema with current db objects and data.

    Hi Cloud expert,
    I need feature to clone existing schema and create new schema with schema as a service. I created profile that export schema with database objects and data. I created template that use the profile created. While provisioning the template I found all the database objects with template. Now I need a solution to clone the current schema with all the objects at current time. The way I tried contains all the objects that were exists while exporting the schema. I need the solution to clone schema so that it will contains all the objects and data right now.
    While reading document I found snapclone but that is not the exactly the solution I need. I simply need to clone schema with latest db object and data with Self Service portal.
    Thank You in advance.
    Dilli R. Maharjan

    One option is to run BACKUP/RESTORE, once you have restored you run:
    SELECT 'ALTER TABLE ' + quotename(s.name) + '.' + quotename(o.name) +
           ' DISABLE TRIGGER ALL '
    FROM   sys.schemas s
    JOIN   sys.objects o ON o.schema_id = s.schema_id
    WHERE  o.type = 'U'
      AND  EXISTS (SELECT *
                   FROM   sys.triggers tr
                   WHERE  tr.parent_id = o.object_id)
    Copy result and paste into a query window do to run. Next:
    SELECT 'DELETE ' + quotename(s.name) + '.' + quotename(o.name)
    FROM   sys.schemas s
    JOIN   sys.objects o ON o.schema_id = s.schema_id
    WHERE  o.type = 'U'
    Run this result until you don't get any more errors.
    Run the first query again, but change DISABLE to ENABLE.
    This is more long-winding than scripting, particularly if you have lots of data. But you know that you will not make any changes whatsoever to the schema.
    The scripting in SSMS generally does it right, I believe, but some items are not scripted by default. If you run BACKUP/RESTORE, you know that you get a faithful copy.
    Of course, the best way is to keep your code under version control and take the version control system as your ultimate truth.
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Expdp/impdp oracle schema

    Hello All,
    We are running 11gr2 on HPUX.
    I am support to take a dump of a schema (say XYZ) and import it to an other server (server2) running same database with same OS.
    Just wanted to confirm the steps what I am proceeding.
    1> expdp xyz/xyz schemas=xyz directory=dump_xyz dumpfile=xyz.dmp logfile=xyz.log
    (here one question, after expdp do I need to give system/pass or xyz/syz???????)
    2> after exp done, then import it to other side,
    3> Once import done then create a sql script on server 1 and move it to server2 and then run it here to have same grant and privs for the user XYZ on server 2 also.
    or is there any way to exp all its grants, privs along with the objects when I was taking dump on server1 ?????
    Thanks in ADV!!!!!

    Thanks budy for this valuable information.
    Will keep in  mind for future reference.
    Now as per my scenario, if I created user xyz on target first and then imported into it.
    This user heaving all privs (as I mentioned above) and my basic need is, this user must have only select to all objects.
    select privilege from dba_sys_privs where grantee='XYZ';
    PRIVILEGES
    CREATE DATABASE LINK
    UNLIMITED TABLESPACE
    CREATE SESSION
    which priv do I need to revoke to prevent this user to create or midify the object. and what priv do I need to give this user to select any object ?????
    Please suggest.
    and also while executing suggested sql on target :
    SQL> select DBMS_METADATA.GET ('USER','XYZ') from dual;
    select DBMS_METADATA.GET ('USER','APPS_QUERY') from dual
    ERROR at line 1:
    ORA-00904: "DBMS_METADATA"."GET": invalid identifier

  • IMPDP into schema with truncated tables, best way to do this?

    Hello all,
    I'm a bit of a noob witih datapump.
    I've got a schema that I've truncated all tables. I have a full schema export I took awhile back, and I'm wanting to import this into the schema to basically 'reset' it.
    First time run, I got the :
    ORA-39151: Table "xyz.tablename" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    I've been reading through, and see suggestions to add to the par file:
    CONTENT=DATA_ONLY TABLE_EXISTS_ACTION=APPEND
    And I've seen others use the option for:
    table_exists_action=replace
    I'm wondering if I could get suggestions best approach to use...I basically want to put the data back into the tables, and have the indexes rebuilt....what is the best course of action here?
    Thank you in advance,
    cayenne

    If all you want is to reload the data, then
    impdp user/password ... content=data_only table_exists_action=append
    This will only load the data into the existing tables and it will append the data to those tables. If the table definitions may have changed, then you would remove the content=data_only and use table_exists_action=replace.
    If you need to load dependent objects like indexes, constraints, etc, then don't do content=data_only and use table_exists_action=replace.
    Hope this helps.
    Dean

  • Datapump expdp impdp and sequences

    Hi all. I have a 10g XE database on my dev machine witha whole load of test data in it. I want to copy the schema to another machine without the data which I can do with expdp usr/pwd CONTENT=metadata_only and I can import it with impdp and everyhing works tickety boo. Except that all the sequences are populated where the test data left off. Can someone please tell me how to copy the schema with the sequences reset please ? I'm guessing either I can export the schema resetting the sequences (somehow) or export EXCLUDING the sequeces and create them seperately (somehow). Thanks for reading.

    I don't think you can reset the sequences directly. You can run the import to an sql file and then use search/replace on it:
    $ impdp user/pass dumpfile=test.dmp directory=MY_DIR include=SEQUENCE sqlfile=sequences.sql
    You will have several lines like this inside "sequences.sql":
    CREATE SEQUENCE "USER"."SEQ_NAME" MINVALUE 1 MAXVALUE 99999999 INCREMENT BY 1 START WITH 1857 CACHE 20 ORDER CYCLE ;
    Then just use regular expressions to replace "START WITH NNNN" by "START WITH 1"

  • Expdp & impdp of R12 schema into 11g DB

    Hi,
    i need to take a backup of GL schema from R12 instance using expdp &
    import it into 11g DB( standalone DB) using impdp.
    i have used the following for expdp in R12(10g DB)
    $expdp system/<pass> schemas=gl directory=dump_dir dumpfile=gl.dmp logfile=gl.log
    export was successfull.
    How to import into an 11g DB( This is on a different machine).
    regards,
    Charan

    Refer to "Oracle® Database Utilities 11g Release 1 (11.1)" Manual.
    Data Pump Import
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28319/dp_import.htm#i1007653
    Oracle® Database Utilities 11g Release 1 (11.1)
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28319/toc.htm

  • Impdp import schema to different tablespace

    Hi,
    I have PRTDB1 schema in PRTDTS tablespace, I need encrypt this tablespace.
    and I have done this steps:
    1. I created new encrypted PRTDTS_ENC tablespace.
    2. I take export PRTDB1 schema. (expdp system/system schema=PRTDB1 directory=datapump dumpfile=data.dmp logfile=log1.log)
    and I want import this dump to new PRTDTS_ENC tablespace.
    can you give me impdp command, which I need?

    Hello,
    ORA-14223: Deferred segment creation is not supported for this table
    Failing sql is:
    CREATE TABLE "PRTDB1"."DATRE_CONT" ("BRON" NUMBER NOT NULL ENABLE, "PCDID" NUMBER(9,0) NOT NULL ENABLE, "OWNERSID" NUMBER NOT NULL ENABLE, "VALUESTR" VARCHAR2(4000 CHAR), "VALNUM" NUMBER, "VALDATE" DATE, "VALXML" "XMLTYPE") SEGMENT CREATION DEFERRED PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING So it means that these Tables "need" a Segment but, they are created with the option SEGMENT CREATION DEFERRED.
    A workaround is to create a Segment for these Tables on the Source Database before exporting them. Then, at the Import you won't have the option SEGMENT CREATION DEFERRED. The following Note of MOS detail this Solution:
    - *Error Ora-14223 Deferred Segment Creation Is Not Supported For This Table During Datapump Import [ID 1293326.1]*
    The link below gives more informations about the DEFERRED SEGMENT CREATION:
    http://docs.oracle.com/cd/E14072_01/server.112/e10595/tables002.htm#CHDGJAGB
    With the DEFERRED SEGMENT CREATION feature, a Segment is created automatically at the first insert on the Table. So the Tables for which the error ORA-14223 occurs are empty.
    You may also create them separately by using a script generated by the SQLFILE parameter in which you change the option SEGMENT CREATION DEFERRED by SEGMENT CREATION IMMEDIATE.
    Hope this help.
    Best regards,
    Jean-Valentin

  • Expdp+Impdp: Does the user have to have DBA privilege?

    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) his own schema?
    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) other (=not his own) schemas ?
    If he is not allowed: Which GRANT is necessary to be able to perform such expdp/impdp operations?
    Peter
    Edited by: user559463 on Feb 28, 2010 7:49 AM

    Hello,
    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) his own schema?Yes, a User can always export its own objects.
    Is a "normal" user (=Without DBA privilege) allowed to export and import (with new expdp/impdp) other (=not his own) schemas ?Yes, if this User has EXP_FULL_DATABASE and IMP_FUL_DATABASE Roles.
    So, you can create a User and GRANT it EXP_FULL_DATABASE and IMP_FULL_DATABASE Roles and, being connected
    to this User, you could export/import any Object from / to any Schemas.
    On databases, on which there're a lot of export/import operations, I always create a special User with these Roles.
    NB: In DataPump you should GRANT also READ, WRITE Privileges on the DIRECTORY (if you use "dump") to the User.
    Else, be accurate on the choice of your words, as previously posted, DBA is a Role not a Privilege which has another meaning.
    Hope this help.
    Best regards,
    Jean-Valentin

  • EXP/IMP..of table having LOB column to export and import using expdp/impdp

    we have one table in that having colum LOB now this table LOB size is approx 550GB.
    as per our knowldge LOB space can not be resused.so we have alrady rasied SR on that
    we are come to clusion that we need to take backup of this table then truncate this table and then start import
    we need help on bekow ponts.
    1)we are taking backup with expdp using parellal parameter=4 this backup will complete sussessfully? any other parameter need to keep in expdp while takig backup.
    2)once truncate done,does import will complete successfully..?
    any SGA PGA or undo tablespace size or undo retention we need to increase.. to completer susecfully import? because its production Critical database?
    current SGA 2GB
    PGA 398MB
    undo retention 1800
    undo tbs 6GB
    please any one give suggestion to perform activity without error...also suggest parameter need to keep during expdp/impdp
    thanks an advance.

    Hi,
    From my experience be prepared for a long outage to do this - expdp is pretty quick at getting lobs out but very slow at getting them back in again - a lot of the speed optimizations that may datapump so excellent for normal objects are not available for lobs. You really need to test this somewhere first - can you not expdp from live and load into some test area - you don't need the whole database just the table/lob in question. You don;t want to find out after you truncate the table that its going to take 3 days to import back in....
    You might want to consider DBMS_REDEFINITION instead?
    Here you precreate a temporary table (with same definitiion as the existing one), load the data into it from the existing table and then do a dictionary switch to swap them over - giving you minimal downtime. I think this should work fine with LOBS at 10g but you should do some research and see if it works fine. You'll need a lot of extra tablespace (temporarily) for this approach though.
    Regards,
    Harry

  • XE11: expdp/impdp

    Hello,
    i would like to use expdp and impdp.
    As i installed XE11 on Linux, i unlocked the HR account:
    ALTER USER hr ACCOUNT UNLOCK IDENTIFIED BY hr;
    and use the expdp:
    expdp hr/hr DUMPFILE=hrdump.dmp DIRECTORY=DATA_PUMP_DIR SCHEMAS=HR
    LOGFILE=hrdump.log
    This quits with:
    ORA-39006: internal error
    ORA-39213: Metadata processing is not available
    The alert_XE.log reported:
    ORA-12012: error on auto execute of job "SYS"."BSLN_MAINTAIN_STATS_JOB"
    ORA-06550: line 1, column 807:
    PLS-00201: identifier 'DBSNMP.BSLN_INTERNAL' must be declared
    I read some entries here and did:
    sqlplus sys/******* as sysdba @?/rdbms/admin/catnsnmp.sql
    sqlplus sys/******* as sysdba @?/rdbms/admin/catsnmp.sql
    sqlplus sys/******* as sysdba @?/rdbms/admin/catdpb.sql
    sqlplus sys/******* as sysdba @?/rdbms/admin/utlrp.sql
    I restarted the database, but the result of expdp was the same:
    ORA-39006: internal error
    ORA-39213: Metadata processing is not available
    What's wrong with that? What can i do?
    Do i need "BSLN_MAINTAIN_STATS_JOB" or can this set ro FALSE?
    I created the database today on 24.07. and the next run for "BSLN_MAINTAIN_STATS_JOB"
    is on 29.07. ?
    In the Windows-Version it is working correct, but not in the Linux-Version.
    Best regards

    Hello gentlemen,
    back to the origin:
    'Is expdp/impdp working on XE11'
    The answer is simply yes.
    After a view days i found out that:
    - no stylesheets installed are required for this operation
    - a simple installation is enough
    And i did:
    SHELL:
    mkdir /u01/app > /dev/null 2>&1
    mkdir /u01/app/oracle > /dev/null 2>&1
    groupadd dba
    useradd -g dba -d /u01/app/oracle oracle > /dev/null 2>&1
    chown -R oracle:dba /u01/app/oracle
    rpm -ivh oracle-xe-11.2.0-1.0.x86_64.rpm
    /etc/init.d/./oracle-xe configure responseFile=xe.rsp
    ./sqlplus sys/********* as sysdba @/u01/app/oracle/product/11.2.0/xe/rdbms/admin/utlfile.sql
    SQLPLUS:
    ALTER USER hr IDENTIFIED BY hr ACCOUNT UNLOCK;
    GRANT CONNECT, RESOURCE to hr;
    GRANT read, write on DIRECTORY DATA_PUMP_DIR TO hr;
    expdp hr/hr dumpfile=hr.dmp directory=DATA_PUMP_DIR schemas=hr logfile=hr_exp.log
    impdp hr/hr dumpfile=hr.dmp directory=DATA_PUMP_DIR schemas=hr logfile=hr_imp.log
    This was carried out on:
    OEL5.8, OEL6.3, openSUSE 11.4
    For explanation:
    We did the style-sheet-installation for XE10 to have the expdp/impd functionality.
    Thanks for your assistance
    Best regards
    Achim
    Edited by: oelk on 16.08.2012 10:20

  • Easyest way to "copy-paste" a schema with data?

    Hi,
    what is the easiest way to make a copy of a schema with data to another (new) schema?
    Thanks

    Hi Jernej Kase
    Q:what is the easiest way to make a copy of a schema with data to another (new) schema?
    A: Oracle 10g datapump’s Network_link option allows us to copy data between schemas without even creating an export dump file.
    I have tried this option and it reduced my copy procedure timings from 2 hours ( with regular exp/imp) to 15 mins.
    all you need to do is
    1. create a DB link pointing to target schema.
    2. run the impdb command
    example:
    impdp uname/pwd@dbname directory=<dir name> logfile=<logfile name> network_link=<dblink name> remap_schema=source_user:target_user parallel=2
    Krishna.B

  • Dataexport with Datapump

    Hi,
    everytime I try to export data with datapump
    I get the following error:
    expdp xxxx/xxxx schemas=SCOTT directory=datapump_dir ...
    ORA-13206: internal error
    ORA-39213: Metadata processing is not available
    The directory is created and the grants given.
    Somebody who knows why this error happens?
    Regards
    Schoeib

    it doesn't exists there!??Yes, I did search on my XE installation, and it isn't there, while it is on 9i, 10.1 and 10.2 Enterprise Edition.
    The Pl/Sql block I posted (which didn't work to you) works to me, albeit I don't have that file. This means that the default XE database already has STYLESHEETS LOADED. Hence my above question : are you trying to use your own database, instead of the default one ?
    Did the installation complete without problems ?

  • Export schema with data

    Good Morning All,
    I would like to export schema with data to another.
    Any help?
    Thanks in advance,
    EB NY

    If you have a few tables...quick and dirty way is to
    create new_table as select * from old_schema.tableOtherwise depending on what version you are on and you have access, use exp/data pump.
    Alot of good pages on using expdp but the general syntax is
    1st, make the directories
    connect / as sysdba
    grant create any directory to org_schema_user;
    create or replace directory dump_directory as '/my_directory_on_the_file_system/';
    grant read, write on directory dump_directory to org_schema_user;
    grant read, write on directory dump_directory to new_schema_user;export
    expdp user/password schemas=start_schema directory=dump_directory dumpfile=my_dmp_file.dmp logfile=my_dmp_file_export.logimport
    impdp new_user/password remap_schema=old_schema:new_schema directory=dump_directory dumpfile=my_dmp_file.dmp logfile=my_dmp_file_import.log

  • Expdp/impdp :: Constraints in Parent child relationship

    Hi ,
    I have one table parent1 and tables child1, child2 and chld3 have foreign key created on this parent1.
    Now I want to do some deletion on parent1. But since number of records are very high on parent1 , we are going with expdp / impdp with querry option.
    I have taken query level expdp on parent1. Now I dropped parent1 with cascade constraints option and all the foreign keys created by child1 ,2 and 3 which references parent1 are automatically dropped.
    Now If i fire the impdp for the query level dump file , will these foreign key constraints will get created automatically on child1 ,2 and 3 or I need to manually re-create it ??
    Regards,
    Anu

    Hi,
    The FK's will not be in the dumpfile - see the example code below where i generate a sqlfile following pretty much the process you would have done. THis is because the FK is part of the DDL for the child table not the parent.
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    OPS$ORACLE@EMZA3>create table a (col1 number);
    Table created.
    OPS$ORACLE@EMZA3>alter table a add primary key (col1);
    Table altered.
    OPS$ORACLE@EMZA3>create table b (col1 number);
    Table created.
    OPS$ORACLE@EMZA3>alter table b add constraint x foreign key (col1) references a(col1);
    Table altered.
    OPS$ORACLE@EMZA3>
    EMZA3:[/oracle/11.2.0.1.2.DB/bin]# expdp / include=TABLE:\"=\'A\'\"
    Export: Release 11.2.0.3.0 - Production on Fri May 17 15:45:50 2013
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    Starting "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04":  /******** include=TABLE:"='A'"
    Estimate in progress using BLOCKS method...
    Processing object type SCHEMA_EXPORT/TABLE/TABLE_DATA
    Total estimation using BLOCKS method: 0 KB
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    . . exported "OPS$ORACLE"."A"                                0 KB       0 rows
    Master table "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04" successfully loaded/unloaded
    Dump file set for OPS$ORACLE.SYS_EXPORT_SCHEMA_04 is:
      /oracle/11.2.0.3.0.DB/rdbms/log/expdat.dmp
    Job "OPS$ORACLE"."SYS_EXPORT_SCHEMA_04" successfully completed at 15:45:58
    Import: Release 11.2.0.3.0 - Production on Fri May 17 15:46:16 2013
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning option
    Master table "OPS$ORACLE"."SYS_SQL_FILE_FULL_01" successfully loaded/unloaded
    Starting "OPS$ORACLE"."SYS_SQL_FILE_FULL_01":  /******** sqlfile=a.sql
    Processing object type SCHEMA_EXPORT/TABLE/TABLE
    Processing object type SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    Processing object type SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    Job "OPS$ORACLE"."SYS_SQL_FILE_FULL_01" successfully completed at 15:46:17
    -- CONNECT OPS$ORACLE
    ALTER SESSION SET EVENTS '10150 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10904 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '25475 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10407 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '10851 TRACE NAME CONTEXT FOREVER, LEVEL 1';
    ALTER SESSION SET EVENTS '22830 TRACE NAME CONTEXT FOREVER, LEVEL 192 ';
    -- new object type path: SCHEMA_EXPORT/TABLE/TABLE
    CREATE TABLE "OPS$ORACLE"."A"
       (    "COL1" NUMBER
       ) SEGMENT CREATION IMMEDIATE
      PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255
    NOCOMPRESS LOGGING
      STORAGE(INITIAL 16384 NEXT 16384 MINEXTENTS 1 MAXEXTENTS 505
      PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1
      BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
      TABLESPACE "SYSTEM" ;
    -- new object type path: SCHEMA_EXPORT/TABLE/CONSTRAINT/CONSTRAINT
    ALTER TABLE "OPS$ORACLE"."A" ADD PRIMARY KEY ("COL1")
      USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255
      STORAGE(INITIAL 16384 NEXT 16384 MINEXTENTS 1 MAXEXTENTS 505
      PCTINCREASE 50 FREELISTS 1 FREELIST GROUPS 1
      BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT)
      TABLESPACE "SYSTEM"  ENABLE;
    -- new object type path: SCHEMA_EXPORT/TABLE/INDEX/STATISTICS/INDEX_STATISTICS
    DECLARE I_N VARCHAR2(60);
      I_O VARCHAR2(60);
      NV VARCHAR2(1);
      c DBMS_METADATA.T_VAR_COLL;
      df varchar2(21) := 'YYYY-MM-DD:HH24:MI:SS';
    stmt varchar2(300) := ' INSERT INTO "SYS"."IMPDP_STATS" (type,version,flags,c1,c2,c3,c5,n1,n2,n3,n4,n5,n6,n7,n8,n9,n10,n11,n12,d1,cl1) VALUES (''I'',6,:1,:2,:3,:4,:5,:6,:7,:8,:9,:10,:11,:12,:13,NULL,:14,:15,NULL,:16,:17)';
    BEGIN
      DELETE FROM "SYS"."IMPDP_STATS";
      c(1) := 'COL1';
      DBMS_METADATA.GET_STAT_INDNAME('OPS$ORACLE','A',c,1,i_o,i_n);
      EXECUTE IMMEDIATE stmt USING 0,I_N,NV,NV,I_O,0,0,0,0,0,0,0,0,NV,NV,TO_DATE('2013-05-17 15:43:24',df),NV;
      DBMS_STATS.IMPORT_INDEX_STATS('"' || i_o || '"','"' || i_n || '"',NULL,'"IMPDP_STATS"',NULL,'"SYS"');
      DELETE FROM "SYS"."IMPDP_STATS";
    END;
    /Regards,
    Harry
    http://dbaharrison.blogspot.com/

  • Create dump file with datapump, with read right for everybody

    Hello,
    I have a problem under Linux : I am creating dump files with datapump.
    Those dump files are owned by a dba group, with no read access for users not in the dba group.
    Is there a way that the datapump utility creates dump files with a read access given to any user ?
    Franck

    Unlike "exp", when using "expdp", the dumpfile is created by the server process. The server process is forked from the database instance. It inherits the umask settings that are present when the database instance is started.
    (Therefore, the only way to change the permissions would be to change the umask for the oracle database server id and restart the database instance --- which is NOT what I would recommend).
    umask is set so that all database files created (e.g. with CREATE TABLESPACE or ALTER TABLESPACE ADD DATAFILE) are created with "secure" permissions preventing others from overwriting them -- of course, this is relevant if your database files are on FileSystem.
    Hemant K Chitale

Maybe you are looking for

  • Not Burning DVDDL+R

    I recently went all out for this girl and ended up having to make a Dual Layer dvd. After it went through all the stages of burning up to muxing I was met with this prompt. and the dvd ejected. I slid the dvd back in and it stopped making noises afte

  • BMC Remedy Integration with SCORCH and AD

    Hi All, We are running BMC Remedy for ticketing. The goal is to automatically add a computer to a security group once the ticket is approved. So, user raises a ticket with his computer details. The workflow for approval in BMC is already set, once ap

  • Combo of 2 problems due to new server at ISP which they deny.

    My co-op ISP is playing havoc with a new server, and we've even been without phone service several days for many hours at a time. They deny anything to do with the current death of my Firefox browser even though there were no problems until they star

  • Flash MX2004 not compatable with Flash CS5?

    I have Flash MX2004 but find when trying to open a flash file created using Flash CS5, the file wiill not open and give me an error ' Unsupported file fotrmat'. Please explain

  • How do i specify scan lines for graphics/text?

    hi there, a program is asking that i place fonts and graphics within scan lines 410-486...how do i do this? i have 1920 x 1080 footage... thanks, steven