Size of export dump file

I read an article on EXPORT of dump file cannot over the limit of 2Gig. This is because prior to 8.1.3 there is no large file support for Oracle Import, Export, or SQL*Loader utilties.
How bout the current version? 9i or 10g? Can the exp dump file be larger than 2G?

Under AIX Unix, there is a fsize parameter for every user in /etc/security/limits file. if fsize = -1 then user will have authority to create any file size. But to do this change, you should have root access under Unix enviroment.

Similar Messages

  • HOW TO IMPORT DATA MORE THAN ONCE FROM THE SAME EXPORT DUMP FILE?

    before asking my question i'd like to mention that i'm a french spoke...
    so my english is a little bit bad. sorry for that.
    my problem is : IMPORT
    how to import data a SECOND TIME from an export dump file within oracle?
    My Export dump file was made successfully (Full Export) and then i
    tried to import datas for the first time.
    I got this following message in my logfile: I ADDED SOME COMMENTS
    Warning: the objects were exported by L1, not by you
    . importing SYSTEM's objects into SYSTEM
    REM ************** CREATING TABLESPACES *****
    REM *********************************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "USER_DATA" DATAFILE 'E:\ORANT\DATABASE\USR1ORCL.ORA' SI"
    "ZE 3145728 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAX"
    "EXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "ROLLBACK_DATA" DATAFILE 'E:\ORANT\DATABASE\RBS1ORCL.ORA"
    "' SIZE 10485760 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS "
    "1 MAXEXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    etc........
    IMP-00017: following statement failed with ORACLE error 1119:
    "CREATE TABLESPACE "L1" DATAFILE 'E:\ORADATA\L1.DBF' SIZE 1048576000 "
    "DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAXEXTENTS 121 PCTIN"
    "CREASE 50) ONLINE PERMANENT"
    IMP-00003: ORACLE error 1119 encountered
    ORA-01119: error in creating database file 'E:\ORADATA\L1.DBF'
    ORA-09200: sfccf: error creating file
    OSD-04002: unable to open file
    O/S-Error: (OS 3) The system cannot find the path specified
    --->etc..........
    the drive E: with the folder E:\ORADATA didn't exist, but after
    all that i created it.
    see below, before my IMPORT statement
    REM ********************* CREATING USER *********
    REM ********************************************
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "L1" IDENTIFIED BY VALUES 'A6E0DAA6865E7627' DEFAULT TABLESPACE"
    " "L1" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'L1' does not exist
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "MLCO" IDENTIFIED BY VALUES '56AC6447B7D50467' DEFAULT TABLESPA"
    "CE "MLCO" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'MLCO' does not exist
    ETC.......
    REM ********************* GRANTING ROLES ***********
    REM ************************************************
    IMP-00017: following statement failed with ORACLE error 1917:
    "GRANT ALTER ANY TABLE to "L1" "
    IMP-00003: ORACLE error 1917 encountered
    ORA-01917: user or role 'L1' does not exist
    ETC.........
    IMP-00017: following statement failed with ORACLE error 1918:
    "ALTER USER "L1" DEFAULT ROLE ALL"
    IMP-00003: ORACLE error 1918 encountered
    ORA-01918: user 'L1' does not exist
    -- that is normal, since the creation of the
    tablespace failed !!
    REM******************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLLBACK SEGMENT RB_TEMP STORAGE (INITIAL 10240 NEXT 10240 MINEXTENT"
    "S 2 MAXEXTENTS 121) TABLESPACE "SYSTEM""
    IMP-00015: following statement failed because
    . importing SCOTT's objects into SCOTT
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "EVT_PROFILE_SEQ" MINVALUE 1 MAXVALUE 999999999999999999999"
    "999999 INCREMENT BY 1 START WITH 21 CACHE 20 NOORDER NOCYCLE"
    ETC............
    importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00003: ORACLE error 1435 encountered
    ORA-01435: user does not exist
    REM *************** IMPORTING TABLES *******************
    REM ****************************************************
    . importing SYSTEM's objects into SYSTEM
    . . importing table "AN1999_BDAT" 243 rows imported
    . . importing table "BOPD" 112 rows imported
    . . importing table "BOINFO_AP" 49
    ETC................
    . . importing table "BO_WHF" 2 rows imported
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLE "DEF$_CALL" ("DEFERRED_TRAN_DB" VARCHAR2(128),
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ROLES" FOR "SYS"."DBA_ROLES""
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ERRORS" FOR "SYS"."DBA_ERRORS""
    IMP-00008: unrecognized statement in the export file:
    . importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00008: unrecognized statement in the export file:
    J
    Import terminated successfully with warnings.
    -------------------------------------b]
    So after analysing this log file, i created
    the appropriate drives and folders... as the
    import statement doesn't see them.
    E:\ORADATA G:\ORDATA etc...
    And i started to [b]IMPORT ONE MORE TIME. with:
    $ IMP73 sys/pssw Full=Y FILE=c:\temp\FOLD_1\data_1.dmp BUFFER=64000
    COMMIT=Y INDEXFILE=c:\temp\FOLD_1\BOO_idx.sql
    LOG=c:\temp\FOLD_1\BOO_log.LOG DESTROY=Y IGNORE=Y;
    after that i could not see the users nor the
    tables created.
    and the following message appeared in the log file:
    Warning: the objects were exported by L1, not by you
    . . skipping table "AN1999_BDAT"
    . . skipping table "ANPK"
    . . skipping table "BOAP"
    . . skipping table "BOO_D"
    ETC.....skipping all the tables
    . . skipping table "THIN_PER0"
    . . skipping table "UPDATE_TEMP"
    Import terminated successfully without warnings.
    and only 2 new tablespaces (originally 3) were
    created without any data in ( i check that in
    the Oracle Storage manager : the tablespaces exit
    with 0.002 used space; originally 60 M for each !!)
    so,
    How to import data (with full import option) succefully
    MORE THAN ONE TIME from an exported dump file ?
    Even if we have to overwrite tablespaces , tables and users.
    thank you very much

    The Member Feedback forum is for suggestions and feedback for OTN Developer Services. This forum is not monitored by Oracle support or product teams and so Oracle product and technology related questions will not be answered. We recommend that you post this thread to the appropriate Database forum.
    The main URL is:
    http://forums.oracle.com/forums/index.jsp?cat=18

  • How to import external table, which exist in export dump file.

    My export dump file has one external table. While i stated importing into my developement instance , I am getting the error "ORA-00911: invalid character".
    The original definition of the extenal table is as given below
    CREATE TABLE EXT_TABLE_EV02_PRICEMARTDATA
    EGORDERNUMBER VARCHAR2(255 BYTE),
    EGINVOICENUMBER VARCHAR2(255 BYTE),
    EGLINEITEMNUMBER VARCHAR2(255 BYTE),
    EGUID VARCHAR2(255 BYTE),
    EGBRAND VARCHAR2(255 BYTE),
    EGPRODUCTLINE VARCHAR2(255 BYTE),
    EGPRODUCTGROUP VARCHAR2(255 BYTE),
    EGPRODUCTSUBGROUP VARCHAR2(255 BYTE),
    EGMARKETCLASS VARCHAR2(255 BYTE),
    EGSKU VARCHAR2(255 BYTE),
    EGDISCOUNTGROUP VARCHAR2(255 BYTE),
    EGREGION VARCHAR2(255 BYTE),
    EGAREA VARCHAR2(255 BYTE),
    EGSALESREP VARCHAR2(255 BYTE),
    EGDISTRIBUTORCODE VARCHAR2(255 BYTE),
    EGDISTRIBUTOR VARCHAR2(255 BYTE),
    EGECMTIER VARCHAR2(255 BYTE),
    EGECM VARCHAR2(255 BYTE),
    EGSOLATIER VARCHAR2(255 BYTE),
    EGSOLA VARCHAR2(255 BYTE),
    EGTRANSACTIONTYPE VARCHAR2(255 BYTE),
    EGQUOTENUMBER VARCHAR2(255 BYTE),
    EGACCOUNTTYPE VARCHAR2(255 BYTE),
    EGFINANCIALENTITY VARCHAR2(255 BYTE),
    C25 VARCHAR2(255 BYTE),
    EGFINANCIALENTITYCODE VARCHAR2(255 BYTE),
    C27 VARCHAR2(255 BYTE),
    EGBUYINGGROUP VARCHAR2(255 BYTE),
    QTY NUMBER,
    EGTRXDATE DATE,
    EGLISTPRICE NUMBER,
    EGUOM NUMBER,
    EGUNITLISTPRICE NUMBER,
    EGMULTIPLIER NUMBER,
    EGUNITDISCOUNT NUMBER,
    EGCUSTOMERNETPRICE NUMBER,
    EGFREIGHTOUTBOUNDCHARGES NUMBER,
    EGMINIMUMORDERCHARGES NUMBER,
    EGRESTOCKINGCHARGES NUMBER,
    EGINVOICEPRICE NUMBER,
    EGCOMMISSIONS NUMBER,
    EGCASHDISCOUNTS NUMBER,
    EGBUYINGGROUPREBATES NUMBER,
    EGINCENTIVEREBATES NUMBER,
    EGRETURNS NUMBER,
    EGOTHERCREDITS NUMBER,
    EGCOOP NUMBER,
    EGPOCKETPRICE NUMBER,
    EGFREIGHTCOSTS NUMBER,
    EGJOURNALBILLINGCOSTS NUMBER,
    EGMINIMUMORDERCOSTS NUMBER,
    EGORDERENTRYCOSTS NUMBER,
    EGRESTOCKINGCOSTSWAREHOUSE NUMBER,
    EGRETURNSCOSTADMIN NUMBER,
    EGMATERIALCOSTS NUMBER,
    EGLABORCOSTS NUMBER,
    EGOVERHEADCOSTS NUMBER,
    EGPRICEADMINISTRATIONCOSTS NUMBER,
    EGSHORTPAYMENTCOSTS NUMBER,
    EGTERMCOSTS NUMBER,
    EGPOCKETMARGIN NUMBER,
    EGPOCKETMARGINGP NUMBER,
    EGWEIGHTEDAVEMULTIPLIER NUMBER
    ORGANIZATION EXTERNAL
    ( TYPE ORACLE_LOADER
    DEFAULT DIRECTORY EV02_PRICEMARTDATA_CSV_CON
    ACCESS PARAMETERS
    LOCATION (EV02_PRICEMARTDATA_CSV_CON:'VPA.csv')
    REJECT LIMIT UNLIMITED
    NOPARALLEL
    NOMONITORING;
    While importing , when i seen the log file , it is failing the create the external table. Getting the error "ORA-00911: invalid character".
    Can some one suggest how to import external tables
    Addressing this issue will be highly appriciated.
    Naveen

    Hi Srinath,
    When i observed the create table syntax of external table from import dump log file, it show few lines as below. I could not understand these special characters. And create table definationis failing with special character viz ORA-00911: invalid character
    ACCESS PARAMETERS
    LOCATION (EV02_PRICEMARTDATA_CSV_CON:'VPA.csv').
    I even observed the create table DDL from TOAD. It is same as i mentioned earlier
    Naveen

  • IMPORT a VIEW from an export dump file

    I'm trying to import a view from an export dump file. In using the Import Utility, do I use the TABLE parameter and give the View name ie: TABLES=view1
    null

    Only I want to do is import ONE view from an export dmp file that contains hundereds of table, many views and procedures. I do not want to reimport everything from the export file.

  • Export dump file and log file  name as sysdate in script

    Hi to All,
    Can anybody help me to give logical backup export dump file namd log file name is as contain sysdate in its name so that we can Uniquelly identified it.
    Regards
    DXB_DBA

    On windows it gets a bit hairy as there really is no clean and nice way of doing it.There are a couple of options.
    1. If you can rely on dateformat not changing, you can use a static substring expression. For example, the following might work w/ finnish locale echo %date:~3,2%%date:~6,2%%date:~9,4%Similarly, when you know the dateformat you can tokenize the output of 'date /t' and discard the tokens you don't want.
    2. You can set dateformat to your liking and then just use %date% in your script
    3. You can run a "SELECT to_char(sysdate,..." into a file and then read that file in your script.
    4. Simon Sheppard also has a solution you could use as a basis. I have a slight issue with the approach, but that could just be me.
    5. Use gnuwin32 or similar ;)
    Also note that %date% env var is set automatically from w2k onwards, so some of the solutions might not work w/ older versions.

  • Export dump files.

    hi all.
    i am using Enterprise Linux server version 5.1.19.6 /Red Hat and Oracle database 11G R2.I want
    to export dump files automatically like we are doing in windows xp with scheduled Task.
    Is it Possible in Linux?plz help me out thanks in advance.

    Mehwish wrote:
    hi all.
    i am using Enterprise Linux server version 5.1.19.6 /Red Hat and Oracle database 11G R2.I want
    to export dump files automatically like we are doing in windows xp with scheduled Task.
    Is it Possible in Linux?plz help me out thanks in advance.Yes,possible,you can use cron job refer
    http://adminschoice.com/crontab-quick-reference
    But for backup recovery operation you should be use rman(recover manager),and through enterprise manager you can schedule backup job.
    http://download.oracle.com/docs/cd/E10317_01/doc/backup.102/e05407/osb_rman_backup.htm
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10734/rcmrecov.htm

  • Where is my EXPORT dump file ? How to do BRTools off line table reorgan ?

    Hello,
    Here was my BRTools Export output message :
    About to export specified tables via Direct Path ...
    Current user changed to SAPPRD
    . . exporting table                          CDHDR   60357362 rows exported
    Export terminated successfully without warnings.
    BR0280I BRSPACE time stamp: 2011-08-09 16.53.04
    BR1160I 1 table exported by EXP utility
    BR0280I BRSPACE time stamp: 2011-08-09 16.53.04
    BR0670I Enter 'c[ont]' to continue, 'b[ack]' to go back, 's[top]' to abort:
    But where is my Export dump file ?
    Besides, could you tell me the procedures of using BRTools Export & Import to do off-line table reorgan ?
    Due to some reasons, I can't do online table reorgan.

    Hi Ming,
    I noted the OSS note 646681 - Reorganizing tables with BRSPACE in my previous message, already. You don't need any additional effort such as creating script, if you use brspace as I noted in my last message.
    Stop the SAP system, but not database then execute the command, below;
    1) brspace -u / -f tbreorg -t "CDHDR"
    2) At the incoming menu steps, select --> (8 - Table reorganization mode (mode)) .. offline
    3) Then continue with "c"
    At the end of this steps, table and related indexes will be reorganized.
    You can update statistics by using the command, below after you complete the steps, above respectfully;
    brconnect -u / -c -f stats -t all -f collect -p 4
    Best regards,
    Orkun Gedik

  • Getting Datapump Export Dump file to the local machine

    I apologize to everyone as this is a duplicate post.
    Re: Getting Datapump Export Dump file to the local machine
    My initial thread(started yesterday)was in 'Database General' and didn't get much response today. Where do i post questions on EXPORT/IMPORT utilities?
    Anyway, here is my problem:
    I want to take the export dump of itemrep schema in orcl database (in a remote machine). I have an Oracle server (10G Rel2) running in my local Windows machine. I have created a user john with necessary EXPORT/IMPORT privileges in my local db. Then i created a Directory object,ie a folder named datapump in my local hard drive and granted READ WRITE privileges to john.
    So john, who is a user in my local machine's oracle db is going to run the expdp utility.
    expdp john/jhendrix@my_local_db_alias SCHEMAS=itemrep directory=datapump logfile=itemrepexp.log
    The above command will fail because it will look for itemrep schema inside my local db, not the remote db where the itemprep is actually located. And you can't qualify the schemaname with its db in the SCHEMAS parameter (like SCHEMAS=itemrep@orcl).
    Can anyone provide me a solution for this?

    I think you can initiate the datapump exp utility from your client machine to export a schema in a remote database.But, Upon execution,oracle looks for the directory in the remote database and not on your local machine.
    You're inovoking expdp from a client (local DB) to export data from a remote DB.
    So, With this method, you can create the dumpfiles only on the Remote server and not on the local Machine.
    You can perform a direct import instead of export using the NETWORK_LINK option.
    Create a DBlink from your local and Remote DB and verify the connection.
    Then,Initiate the Impdp from Your local machine DB using the parameter network_link=<db_link of the Remote DB> to import the schema.
    The advantage of this option eliminates the Dumpfile creation at the Server side.
    There are no dumpfiles during the import process. the Data is imported directly to the target schema.

  • User & Schema name of Export dump file

    I have an exported dump file. I dont have any log file of that export. I want to Import it in another system. How do I know concerned user or schema name of that dump file.
    Is there any way to know it???

    dbanirav wrote:
    I have an exported dump file. I dont have any log file of that export. I want to Import it in another system. How do I know concerned user or schema name of that dump file.
    Is there any way to know it???made with exp or expdp?
    what OS name & version?
    How do I ask a question on the forums?
    SQL and PL/SQL FAQ

  • Database Restore Using Export Dump File

    I have searched a lot of books and web sites, but cannot find the exact procedure as to how to restore full oracle database using the export dump file and the import utility.
    Does anyone have any ideas.
    Paltform: W2K AD Svr, Oracle 9.2.0.6.
    Please advice

    Hi, I was able to create the db using dbca, and then import.
    When I first imported, I got messages all over that user or objects did not exist. So I commented the line which would invoke the CreateDBCatalog.sql script.
    And now I get this,
    C:\>imp '/ as sysdba' file=z:\export\tuethusat\dbname_full_export.dmp full=y
    Import: Release 9.2.0.1.0 - Production on Tue Oct 24 15:30:03 2006
    Copyright (c) 1982, 2002, Oracle Corporation. All rights reserved.
    IMP-00003: ORACLE error 942 encountered
    ORA-00942: table or view does not exist
    IMP-00023: Import views not installed, please notify your DBA
    IMP-00000: Import terminated unsuccessfully
    C:\>
    Any suggestions?

  • Export dump file of 8.x unable to Import in 9i

    Hi,
    I have Oracle 8.x version export dump file.
    When i try to import this file in Oracle 9i db, it fails everytime. I have tried with all possible options in import utility. When i start import it shows as a active jobs in the jobs list but after few seconds it disappears from active tab and goes to failed list. It doesnot even gives the reason for failure. No error messages also. I don't know what shoud i do as i want to upgrade my 8.x data to 9i. I have currently only 9i installed on windows 2000 Server.
    Please reply to my query as sson as possible.
    Thanks

    i beleive, u are using OEM or any other tool to do the export and import.
    FIRST make sure, u have configured OEM properly to handle the batch jobs.
    easiest way is to use the command line utilities.
    use exp of lower version to export the source database.
    create similiar users and tablespaces ( tablespace names is not a must) in target database.
    use the same lower version of imp to import into target database.

  • How-to Generate Script from the export dump file

    Hi all,
    Software : Oracle 10gR2/Windows 2003
    =============================
    Has anyone generated a Script from a export dump file ( exported via data pump export). I know it is possible to generate script from the regular export.
    I want to know how to generate the script for various objects or is it possible to get the counts of types of objects in the dump file and / or generate the script from the export dump file exported via data pump utility on Oracle 10g.
    Thanks,
    SS

    I want to know how to generate the script for various objects.what do you mean by scripts? Assuming the metadata definitions of various objects.
    First of all, you can generate the metadata definitions using import.
    Use, Impdp with SQLFILE parameter to generate the SQL file for the metadata definitions.
    Refer to
    http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm

  • Export dump file size

    i am exporting one schema.The total size of all the objects in that schema is 70GB.When i perform export on that schema the the size of the dmp file will be 70GB or less?Can any there out to help me.
    OS:Solaris
    DB Version:9.2.0.6
    Thank you..

    Hi,
    When you do an export, you are only exporting the data. Other objects, such as indexes are only exported as a description so take up virtually no space. A general rule of thumb is to work out the actual amount of used data blocks you have, remembering that tables can have allocated space that is empty, and this will give you a size that is probably slightly larger than the export, maybe by about 10%. Also, export files are highly compressible. So if you need to save space, you will usually get about 80 ~ 90% compression ratios.
    Andre

  • Size of Exported Excel file is very large

    Hi,
    I have a report that has around 10000 rows and 5 columns. When I export to excel, the excel file is very big, around 15MB.
    But If the file is opened, and saved again , the size reduces to around 13MB.
    Any suggestions on how to reduce the file size of the exported Excel file. I would prefer to export it to excel rather than CSV format.
    Any help is appreciated.
    Thanks....
    ViVa1

    Hi, actually as you have probably noticed the size grows a lot due to the use of different styles and formatting.
    If you do not require any formatting then I would recommend you to consider a CSV file format again because it will result in the smallest file size which you can achieve with exporting.
    Also for the future referencing SQL Server 2008 R2 exports reports only in XLS format, but SQL Server 2012 by default will export into a new XLSX file format which is a compressed (zipped XML based format) so the resulting files will be significantly smaller.
    For example I created a small project to test exporting of 10000 x 5 data to an Excel file. I used this
    .NET Excel component to create various spreadsheet files, it has a straightforward API for
    exporting DataTable into an Excel file in .NET so you can test this out yourself with ease. The output files, without any formatting applied to them, have the following sizes: XLS (~1MB), XLSX (~300KB) and CSV (~250KB).

  • Can we create export dump files on client side??

    Hi,
    Please see this link
    http://download-uk.oracle.com/docs/cd/B19306_01/server.102/b14215/dp_overview.htm#i1010293
    Check the "NOTE" which says that
    All Data Pump Export and Import processing, including the reading and writing of dump files, is done on the system (server) selected by the specified database connect string. This means that, for nonprivileged users, the database administrator (DBA) must create directory objects for the Data Pump files that are read and written on that server file system. For privileged users, a default directory object is available
    So does this mean, the dump files can be created only on the server???
    What if i want to create the dump file on my client c:\
    Is it possible??
    As per my knowledge, this was possible in 9i
    Thanks

    Yes in the "orignal" export ,it was possible to have a client side dump file.But the Export DataPump uses the Directory parameter which maps to a directory on the server side only and the option to give an absolute path for the dump file on the client side is removed,the files can be only on server side.
    Aman....

Maybe you are looking for