Export dump size precheck

Hi
Before do export of oracle users or database is there any option to check what will be the export dumpfile size.

check by using commands what options do you have..
$exp help=y
$expdp help=yexp help=y
Export: Release 11.1.0.7.0 - Production on Wed May 4 02:57:46 2011
Copyright (c) 1982, 2007, Oracle. All rights reserved.
You can let Export prompt you for parameters by entering the EXP
command followed by your username/password:
Example: EXP SCOTT/TIGER
Or, you can control how Export runs by entering the EXP command followed
by various arguments. To specify parameters, you use keywords:
Format: EXP KEYWORD=value or KEYWORD=(value1,value2,...,valueN)
Example: EXP SCOTT/TIGER GRANTS=Y TABLES=(EMP,DEPT,MGR)
or TABLES=(T1:P1,T1:P2), if T1 is partitioned table
USERID must be the first parameter on the command line.
Keyword Description (Default) Keyword Description (Default)
USERID username/password FULL export entire file (N)
BUFFER size of data buffer OWNER list of owner usernames
FILE output files (EXPDAT.DMP) TABLES list of table names
COMPRESS import into one extent (Y) RECORDLENGTH length of IO record
GRANTS export grants (Y) INCTYPE incremental export type
INDEXES export indexes (Y) RECORD track incr. export (Y)
DIRECT direct path (N) TRIGGERS export triggers (Y)
LOG log file of screen output STATISTICS analyze objects (ESTIMATE)
ROWS export data rows (Y) PARFILE parameter filename
CONSISTENT cross-table consistency(N) CONSTRAINTS export constraints (Y)
OBJECT_CONSISTENT transaction set to read only during object export (N)
FEEDBACK display progress every x rows (0)
FILESIZE maximum size of each dump file
FLASHBACK_SCN SCN used to set session snapshot back to
FLASHBACK_TIME time used to get the SCN closest to the specified time
QUERY select clause used to export a subset of a table
RESUMABLE suspend when a space related error is encountered(N)
RESUMABLE_NAME text string used to identify resumable statement
RESUMABLE_TIMEOUT wait time for RESUMABLE
TTS_FULL_CHECK perform full or partial dependency check for TTS
VOLSIZE number of bytes to write to each tape volume
TABLESPACES list of tablespaces to export
TRANSPORT_TABLESPACE export transportable tablespace metadata (N)
TEMPLATE template name which invokes iAS mode export
Export terminated successfully without warnings.

Similar Messages

  • My Databse size is 10 gb what will be my export dump size?

    My Databse size is 10 gb what will be my export dump size?
    Why is there is difference ?

    Hi,
    an Oracle Block constists not only of the row-data itself,
    but also has a block header.
    This is made up ofdifferent structures which will use about
    - 60 bytes of fixed administrative information for the block
    - 24 bytes per transaction slot (depending of the value inittrans there will be a number of preallocated slots for transaction pointers, default 1 for tables 2 for indexes, in every block header) this socalled ITL consists of three parts:
    - Undo segment number the transaction is using for the before images
    - Undo slot number in the undo segment
    - sequence number (this will be incremented at commit time)
    2 bytes for every row which is inserted into the block.
    Also the rows have a header each (congtaining the number of columns belonging to the row for example) and every column in a row has header also(containing the informatin of the datatype and the length of the value for the column.
    You see that there is a lot more info in a block than just the rowdata.

  • Export Dump Size

    Oracle Version 10.2.0.4
    My database total size(allocated) is 120 GB
    Used is 79 GB
    When I take export, can i assume database export dump file size would be around 70 GB ?

    Dean Gagne wrote:
    If you use data pump to export data, you can estimate the size of the dump file using following parameter:
    ESTIMATE_ONLY=YThis is not really true. All that is being estimated is the size of data, it does not include the size of the metadata.
    I'm not sure if there is a good way to estimate how large the dumpfile from Data Pump will be.
    DeanAgreed!
    But the estimated size may at least gives some idea.

  • Export dump size,How to predict

    We had a client database of 45GB but the dump was only 1.5 gb.THere were no views,SP,synonyms,primary key etc except table and indexes but that was performing wonderfully well .
    Now we got db of 12 gb with more 6gb of dump.But here we got 20,000 views,100's of synonyms,
    How does oracle generates ?

    Maran,<br>
    <br>
    I have created pipes and compressed in Unix.How to do it in Windows<br>A solution could be to place the dump file under a compressed directory.<br>
    <br>
    Or, make some scripts, examples here<br>
    <br>
    Nicolas.

  • Export minimize dump size

    hi all ;
    i have 10.2.0.2 run on Redhat. i have to take backup a schema and import it to a test db.
    this schema contains 20 tables and some of them have miilions records. i try to export of this schema and use ;
    exp system/password fille=exp.dmp log=exp.lod owner=schemaowner
    but dump size is nearly 1GB and this is too big for me. as i mentioned that this dump will use for test and i dont have to take all the records. for example last 7 days of records is enough for me.
    i cant take this export without row at least some of records is needed.
    how can i do this export and import ???
    thanks...

    Hello,
    I prefer to use Parameter Files like the following ones:
    For the Original Export:
    TABLES=(
    <schema>.<table_1>,
    <schema>.<table_2>,
    <schema>.<table_n>
    FILE=<path>\<dump_file_name>.dmp
    LOG=<path>\<log_file_name>.dmp
    COMPRESS=N
    DIRECT=Y
    For the Original Import:
    FROMUSER=<schema>
    TOUSER=<schema>
    TABLES=(<table_1>,<table_2>,...<table_n>)
    FILE=<path>\<dump_file_name>.dmp
    LOG=<path>\<log_file_name>.dmp
    IGNORE=Y
    FEEDBACK=10000Then you have just to execute the statements below:
    exp <user>/<password> PARFILE=<export_parameter_file>
    imp <user>/<password> PARFILE=<import_parameter_file>Hope this help.
    Best regards,
    Jean-Valentin

  • Why export dump shows less than the actual size?

    why export dump shows less than the actual size?

    i tried to find to find out the logical size using the following command
    select sum(bytes/1024/1024) from dba_segments;
    SUM(BYTES/1024/1024)
    739when i tried the following query removing index segement:
    select sum(bytes/1024/1024) from dba_segments where SEGMENT_TYPE \<\> 'INDEX' ;
    SUM(BYTES/1024/1024)
    535What other type of information is not exported in a full database export?

  • HOW TO IMPORT DATA MORE THAN ONCE FROM THE SAME EXPORT DUMP FILE?

    before asking my question i'd like to mention that i'm a french spoke...
    so my english is a little bit bad. sorry for that.
    my problem is : IMPORT
    how to import data a SECOND TIME from an export dump file within oracle?
    My Export dump file was made successfully (Full Export) and then i
    tried to import datas for the first time.
    I got this following message in my logfile: I ADDED SOME COMMENTS
    Warning: the objects were exported by L1, not by you
    . importing SYSTEM's objects into SYSTEM
    REM ************** CREATING TABLESPACES *****
    REM *********************************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "USER_DATA" DATAFILE 'E:\ORANT\DATABASE\USR1ORCL.ORA' SI"
    "ZE 3145728 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAX"
    "EXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "ROLLBACK_DATA" DATAFILE 'E:\ORANT\DATABASE\RBS1ORCL.ORA"
    "' SIZE 10485760 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS "
    "1 MAXEXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    etc........
    IMP-00017: following statement failed with ORACLE error 1119:
    "CREATE TABLESPACE "L1" DATAFILE 'E:\ORADATA\L1.DBF' SIZE 1048576000 "
    "DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAXEXTENTS 121 PCTIN"
    "CREASE 50) ONLINE PERMANENT"
    IMP-00003: ORACLE error 1119 encountered
    ORA-01119: error in creating database file 'E:\ORADATA\L1.DBF'
    ORA-09200: sfccf: error creating file
    OSD-04002: unable to open file
    O/S-Error: (OS 3) The system cannot find the path specified
    --->etc..........
    the drive E: with the folder E:\ORADATA didn't exist, but after
    all that i created it.
    see below, before my IMPORT statement
    REM ********************* CREATING USER *********
    REM ********************************************
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "L1" IDENTIFIED BY VALUES 'A6E0DAA6865E7627' DEFAULT TABLESPACE"
    " "L1" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'L1' does not exist
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "MLCO" IDENTIFIED BY VALUES '56AC6447B7D50467' DEFAULT TABLESPA"
    "CE "MLCO" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'MLCO' does not exist
    ETC.......
    REM ********************* GRANTING ROLES ***********
    REM ************************************************
    IMP-00017: following statement failed with ORACLE error 1917:
    "GRANT ALTER ANY TABLE to "L1" "
    IMP-00003: ORACLE error 1917 encountered
    ORA-01917: user or role 'L1' does not exist
    ETC.........
    IMP-00017: following statement failed with ORACLE error 1918:
    "ALTER USER "L1" DEFAULT ROLE ALL"
    IMP-00003: ORACLE error 1918 encountered
    ORA-01918: user 'L1' does not exist
    -- that is normal, since the creation of the
    tablespace failed !!
    REM******************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLLBACK SEGMENT RB_TEMP STORAGE (INITIAL 10240 NEXT 10240 MINEXTENT"
    "S 2 MAXEXTENTS 121) TABLESPACE "SYSTEM""
    IMP-00015: following statement failed because
    . importing SCOTT's objects into SCOTT
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "EVT_PROFILE_SEQ" MINVALUE 1 MAXVALUE 999999999999999999999"
    "999999 INCREMENT BY 1 START WITH 21 CACHE 20 NOORDER NOCYCLE"
    ETC............
    importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00003: ORACLE error 1435 encountered
    ORA-01435: user does not exist
    REM *************** IMPORTING TABLES *******************
    REM ****************************************************
    . importing SYSTEM's objects into SYSTEM
    . . importing table "AN1999_BDAT" 243 rows imported
    . . importing table "BOPD" 112 rows imported
    . . importing table "BOINFO_AP" 49
    ETC................
    . . importing table "BO_WHF" 2 rows imported
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLE "DEF$_CALL" ("DEFERRED_TRAN_DB" VARCHAR2(128),
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ROLES" FOR "SYS"."DBA_ROLES""
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ERRORS" FOR "SYS"."DBA_ERRORS""
    IMP-00008: unrecognized statement in the export file:
    . importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00008: unrecognized statement in the export file:
    J
    Import terminated successfully with warnings.
    -------------------------------------b]
    So after analysing this log file, i created
    the appropriate drives and folders... as the
    import statement doesn't see them.
    E:\ORADATA G:\ORDATA etc...
    And i started to [b]IMPORT ONE MORE TIME. with:
    $ IMP73 sys/pssw Full=Y FILE=c:\temp\FOLD_1\data_1.dmp BUFFER=64000
    COMMIT=Y INDEXFILE=c:\temp\FOLD_1\BOO_idx.sql
    LOG=c:\temp\FOLD_1\BOO_log.LOG DESTROY=Y IGNORE=Y;
    after that i could not see the users nor the
    tables created.
    and the following message appeared in the log file:
    Warning: the objects were exported by L1, not by you
    . . skipping table "AN1999_BDAT"
    . . skipping table "ANPK"
    . . skipping table "BOAP"
    . . skipping table "BOO_D"
    ETC.....skipping all the tables
    . . skipping table "THIN_PER0"
    . . skipping table "UPDATE_TEMP"
    Import terminated successfully without warnings.
    and only 2 new tablespaces (originally 3) were
    created without any data in ( i check that in
    the Oracle Storage manager : the tablespaces exit
    with 0.002 used space; originally 60 M for each !!)
    so,
    How to import data (with full import option) succefully
    MORE THAN ONE TIME from an exported dump file ?
    Even if we have to overwrite tablespaces , tables and users.
    thank you very much

    The Member Feedback forum is for suggestions and feedback for OTN Developer Services. This forum is not monitored by Oracle support or product teams and so Oracle product and technology related questions will not be answered. We recommend that you post this thread to the appropriate Database forum.
    The main URL is:
    http://forums.oracle.com/forums/index.jsp?cat=18

  • Export dump and the Oracle performance

    My prod system is 10g r2 on an AIX box with 4 CPU ( oracle only uses two) and 4 G memory. It also installs with a BAAN 5c system. I noticed that when I did the export dump, the Oracle system was slow to other query processes by the BAAN and the message "Memory Notification: Library Cache Object loaded into SGA
    Heap size 8356K exceeds notification threshold (2048K)" were shown up in alert log. Although there is a solution in the metalink to avoid the notification, I think it does not solve the problem underneath. Some system maintenance processes also caused that. So How can I track the real cause(s)? Maybe my original configuration need to reconfigure.

    Oracle was slow because running export is rather resource intensive (lots and lots of I/O). You generally want to avoid generating an export when the system is being actively used. As an aside, I hope that you're not using the export as your primary backup...
    The message you're getting (presumably in Enterprise Manager) would appear to be both reasonable and probably not the source of the slowdown. I'm not shocked to learn that BAAN has some objects which are relatively large, running an export brings them into the library cache, which generates the alert. I would expect that you'd have to get that alert in order to do an export, so it may be reasonable to turn it off if this is a regular thing.
    Justin

  • Avoiding BLOB in export dump

    Hi All,
    I am taking export of entire database. Since this database contains tables with BLOB datatype, my export dump's size is 4GB where as usually it would be around 100-200MB. I want to exclude these tables in my export. Is there a way to do so. My database contains around 2000 tables.
    Any help on this would be deeply appreciated.

    this is one way i can think.
    user user_tab_columns to identify tables with BLOB datatype.
    add these list in the EXCLUDE parameter in expdp.

  • Sharing some info about Export dumps

    Just to share some info about exports happenig in our system
    I have a 20GB Database but the dump size is 10GB it keeps growing..the data too grows but the dump has also getting increased significantly...
    Has anyone seen encountered soemthing like this...
    Strange but to live with it
    Oracle 8i on Win2k

    Is this your Production Database? Did you schedule this export? During peak/off peak business hours? Production
    Yes
    Off Peak hours
    Apart from this export, what database backup procedure you have in place? Full Backups/Incremental all in place
    You said, you take export regularly as part of procedure. Is this a strange behavior today or did you observe this previously as well??I have been observing since i joined this company.It was 14GB and Dump was around 7 ..now the DB is 20GB and dump is around 10GB
    I was just thinkking for a TB database how much the export would be...:)
    Adding on i had a 100Gb db generating 8Gb dump...Looked strange..)BUt that 100GB did not have so amny objects like what i have now...
    Thanks

  • How to import external table, which exist in export dump file.

    My export dump file has one external table. While i stated importing into my developement instance , I am getting the error "ORA-00911: invalid character".
    The original definition of the extenal table is as given below
    CREATE TABLE EXT_TABLE_EV02_PRICEMARTDATA
    EGORDERNUMBER VARCHAR2(255 BYTE),
    EGINVOICENUMBER VARCHAR2(255 BYTE),
    EGLINEITEMNUMBER VARCHAR2(255 BYTE),
    EGUID VARCHAR2(255 BYTE),
    EGBRAND VARCHAR2(255 BYTE),
    EGPRODUCTLINE VARCHAR2(255 BYTE),
    EGPRODUCTGROUP VARCHAR2(255 BYTE),
    EGPRODUCTSUBGROUP VARCHAR2(255 BYTE),
    EGMARKETCLASS VARCHAR2(255 BYTE),
    EGSKU VARCHAR2(255 BYTE),
    EGDISCOUNTGROUP VARCHAR2(255 BYTE),
    EGREGION VARCHAR2(255 BYTE),
    EGAREA VARCHAR2(255 BYTE),
    EGSALESREP VARCHAR2(255 BYTE),
    EGDISTRIBUTORCODE VARCHAR2(255 BYTE),
    EGDISTRIBUTOR VARCHAR2(255 BYTE),
    EGECMTIER VARCHAR2(255 BYTE),
    EGECM VARCHAR2(255 BYTE),
    EGSOLATIER VARCHAR2(255 BYTE),
    EGSOLA VARCHAR2(255 BYTE),
    EGTRANSACTIONTYPE VARCHAR2(255 BYTE),
    EGQUOTENUMBER VARCHAR2(255 BYTE),
    EGACCOUNTTYPE VARCHAR2(255 BYTE),
    EGFINANCIALENTITY VARCHAR2(255 BYTE),
    C25 VARCHAR2(255 BYTE),
    EGFINANCIALENTITYCODE VARCHAR2(255 BYTE),
    C27 VARCHAR2(255 BYTE),
    EGBUYINGGROUP VARCHAR2(255 BYTE),
    QTY NUMBER,
    EGTRXDATE DATE,
    EGLISTPRICE NUMBER,
    EGUOM NUMBER,
    EGUNITLISTPRICE NUMBER,
    EGMULTIPLIER NUMBER,
    EGUNITDISCOUNT NUMBER,
    EGCUSTOMERNETPRICE NUMBER,
    EGFREIGHTOUTBOUNDCHARGES NUMBER,
    EGMINIMUMORDERCHARGES NUMBER,
    EGRESTOCKINGCHARGES NUMBER,
    EGINVOICEPRICE NUMBER,
    EGCOMMISSIONS NUMBER,
    EGCASHDISCOUNTS NUMBER,
    EGBUYINGGROUPREBATES NUMBER,
    EGINCENTIVEREBATES NUMBER,
    EGRETURNS NUMBER,
    EGOTHERCREDITS NUMBER,
    EGCOOP NUMBER,
    EGPOCKETPRICE NUMBER,
    EGFREIGHTCOSTS NUMBER,
    EGJOURNALBILLINGCOSTS NUMBER,
    EGMINIMUMORDERCOSTS NUMBER,
    EGORDERENTRYCOSTS NUMBER,
    EGRESTOCKINGCOSTSWAREHOUSE NUMBER,
    EGRETURNSCOSTADMIN NUMBER,
    EGMATERIALCOSTS NUMBER,
    EGLABORCOSTS NUMBER,
    EGOVERHEADCOSTS NUMBER,
    EGPRICEADMINISTRATIONCOSTS NUMBER,
    EGSHORTPAYMENTCOSTS NUMBER,
    EGTERMCOSTS NUMBER,
    EGPOCKETMARGIN NUMBER,
    EGPOCKETMARGINGP NUMBER,
    EGWEIGHTEDAVEMULTIPLIER NUMBER
    ORGANIZATION EXTERNAL
    ( TYPE ORACLE_LOADER
    DEFAULT DIRECTORY EV02_PRICEMARTDATA_CSV_CON
    ACCESS PARAMETERS
    LOCATION (EV02_PRICEMARTDATA_CSV_CON:'VPA.csv')
    REJECT LIMIT UNLIMITED
    NOPARALLEL
    NOMONITORING;
    While importing , when i seen the log file , it is failing the create the external table. Getting the error "ORA-00911: invalid character".
    Can some one suggest how to import external tables
    Addressing this issue will be highly appriciated.
    Naveen

    Hi Srinath,
    When i observed the create table syntax of external table from import dump log file, it show few lines as below. I could not understand these special characters. And create table definationis failing with special character viz ORA-00911: invalid character
    ACCESS PARAMETERS
    LOCATION (EV02_PRICEMARTDATA_CSV_CON:'VPA.csv').
    I even observed the create table DDL from TOAD. It is same as i mentioned earlier
    Naveen

  • IMPORT a VIEW from an export dump file

    I'm trying to import a view from an export dump file. In using the Import Utility, do I use the TABLE parameter and give the View name ie: TABLES=view1
    null

    Only I want to do is import ONE view from an export dmp file that contains hundereds of table, many views and procedures. I do not want to reimport everything from the export file.

  • Logical standby error with export dump

    oracal 10g have a setup logical standby and when i am running export dump from logical i got this error.
    EXP-00008: ORACLE error 16224 encountered
    ORA-16224: Database Guard is enabled
    EXP-00000: Export terminated unsuccessfully
    can someone help me out how can it posibal to take export dump and datapump from logical standby.
    thanks

    16224, 00000, "Database Guard is enabled"
    // *Cause: Operation could not be performed because database guard is enabled
    // *Action: Verify operation is correct and disable database guard                                                                                                                                                                                                                                                                                                                                                                                       

  • Is it possible to import and export dump between oracle 9 and oracle 11?

    Is it possible to import and export dump between oracle 9 and oracle 11?
    Source DB : Oracle 11g(unix0
    Oracle client : Oracle 9(Windows)
    Export import utility : Oracle 9's
    Destination DB : oracle 9

    I am getting the Following Error and export utility is not responding after this.
    Export: Release 9.2.0.1.0 - Production on Thu Jul 15 14:37:01 2010
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Export done in WE8MSWIN1252 character set and AL16UTF16 NCHAR character set
    server uses AL32UTF8 character set (possible charset conversion)

  • Do I need to export multiple sizes for all standard print sizes? i.e.Say I crop the image in LR to 6'5" x 9' to fit within an 11" x 14" (overall) mat of a framed print.. Now let's say I want to reprint that same image in a 12" x 16" size ... to fit within

    Do I need to export multiple sizes for all standard print sizes? i.e.Say I crop the image in LR to 6'5" x 9' to fit within an 11" x 14" (overall) mat of a framed print.. Now let's say I want to reprint that same image in a 12" x 16" size ... to fit within a mat of with overall size of 20" x 24"? Or will the 1st export be "locked" at 6.5" X 9" (proportionately)?

    So, for the first one Export it cropped to 6.5 x 9 - and in the Export Dialog box there is a capability to rename the file for exported output - which would allow you to add something like the crop or intention to the base name of the file.  e.g. for _IMG0846.nef - setup the output filename to _IMG0846-65x9 for 11x14.jpg and output it at, say 300 or 360 dpi.  So, it has a unique name and purpose in your output folder.
    For best results, now, re-crop the picture in LR - just open the crop [R] dialog again, reframe the image to the new ratio, 12x16 and Export again - with an adjustment to the filename. e.g. _IMG0846-12x16 for 20x24.jpg
    So, now you have two clearly identified image files in your output area and you can provide those to whomever is printing for you, etc.  Discard them when you're done since you always have the original to return to to re-frame again as necessary.
    This way you get output sharpening and resolution (in pixels) optimized for your output print size.  Maybe not super important for smallish prints-but more important for larger ones.   Output at 100 quality.

Maybe you are looking for