Export oracle file to dbf file

Hai,
Could I export from oracle database file to dbf file. Because I am building form and I want to use push button to export to another form but the database is dbf so I can use for another application
Regards,
Patrick

Generally speaking I use the Perl XBase module, available from CPAN (www.cpan.org), either directly or through DBI.

Similar Messages

  • Exporting of files in oracle 10g database

    hey friends,
    I installed Oracle 10g Database server on Windows XP,when i tried to take the backup through "Export to Files".I got this error.
    I have added Administrator in "Login as Batch JOb" and has given administrator along with the password in host credentials.
    I explain the procedure.I Login as Sys as SysOper as Oracle does not allow to take backup as sysdba.
    I created the job and also given the right path for the backup but some how it fails then i check the log and i find the following output.
    Output Log
    Export: Release 10.1.0.2.0 - Production on Fri Mar 4 16:39:23 2005
    Copyright (c) 1982, 2004, Oracle. All rights reserved.
    Username:
    EXP-00004: invalid username or password
    Username: Password:
    EXP-00004: invalid username or password
    Username:
    EXP-00030: Unexpected End-Of-File encountered while reading input
    EXP-00000: Export terminated unsuccessfully
    The output log says i am giving invalid username and password but i don't know where i am giving the invalid username and password.
    Please guide me as I am new to Oracle and want to learn Oracle Administration.
    Thanks & Regards
    Ankush

    Give me output of following on standby server
    SQL>select name from v$datafile;
    SQL>show parameter log_archive_dest

  • Is it possible to import an Oracle 10 (export .dmp file) into Oracle 9 db ?

    hi Oracle Guru's
    Is it possible to import a oracle 10 (export .dmp file) into an oracle 9 database ?
    I tried & got the following error ...
    IMP-00010 not a valid export file, header failed verificationperhaps their is a work-around ?
    thanks

    Hi
    You can take export from a previous version and import to a later version. Having said that, there is still a workaround to do what you want.
    On 9i server add a connect string for your 10g instance. Then use 9i export utility to export 10g data. That way your export will be compatible with 9i. You can then import the data in your 9i instance.
    To create connect string you can either use Net Configuration Assistant or manually edit TNSNAMES.ORA file in your 9i server.
    Hope it helps.
    Rgds
    Adnan

  • How to ZIP Oracle Datapump export backup file

    Hello All,
    My customer is asking to give him the production data dump to the following path \\138.90.17.56\OMNISAFE.
    I really don't understand his requirement and he also wants me to zip the export backup file. How do I do that, Do you know any unix command to zip backup files.
    thanks and regards
    cherry

    1013498 wrote:
    Well Thanks for your reply.....my oracle version is 11.2.0.3.b and if we have the compression option can you please elaborate how to do that......
    It's in the documentation.  See Data Pump Export
    let us say my expdp file is abc.dmp...should I give the command gzip abc.dmp or any different.
    Let me google that for you
    One more question what does teh customer mean by production data dump to the following path \\138.90.17.56\OMNISAFE. please explain
    How do we know what the customer means?  Why don't you ask him?
    That said, it looks like a url to an ip address and a defined folder at that ip address.  Again, if the customer wants you to send them a file, you need to be working with said customer on the mechanics of accessing their system.
    All that said ....
    Learning how to look things up in the documentation is time well spent investing in your career.  To that end, you should drop everything else you are doing and do the following:
    Go to tahiti.oracle.com.
    Locate the link for your Oracle product and version, and click on it.
    You are now at the entire documentation set for your selected Oracle product and version.
    BOOKMARK THAT LOCATION
    Spend a few minutes just getting familiar with what is available here. Take special note of the "books" and "search" tabs. Under the "books" tab (for 10.x) or the "Master Book List" link (for 11.x) you will find the complete documentation library.
    Spend a few minutes just getting familiar with what kind  of documentation is available there by simply browsing the titles under the "Books" tab.
    Open the Reference Manual and spend a few minutes looking through the table of contents to get familiar with what kind of information is available there.
    Do the same with the SQL Reference Manual.
    Do the same with the Utilities manual.
    You don't have to read the above in depth.  They are reference manuals.  Just get familiar with what is there to be referenced. Ninety percent of the questions asked on this forum can be answered in less than 5 minutes by simply searching one of the above manuals.
    Then set yourself a plan to dig deeper.
    - *Read a chapter a day from the Concepts Manual*.
    - Take a look in your alert log.  One of the first things listed at startup is the initialization parms with non-default values. Read up on each one of them (listed in your alert log) in the Reference Manual.
    - Take a look at your listener.ora, tnsnames.ora, and sqlnet.ora files. Go to the Network Administrators manual and read up on everything you see in those files.
    - *When you have finished reading the Concepts Manual, do it again*.
    Give a man a fish and he eats for a day. Teach a man to fish and he eats for a lifetime.

  • HOW TO IMPORT DATA MORE THAN ONCE FROM THE SAME EXPORT DUMP FILE?

    before asking my question i'd like to mention that i'm a french spoke...
    so my english is a little bit bad. sorry for that.
    my problem is : IMPORT
    how to import data a SECOND TIME from an export dump file within oracle?
    My Export dump file was made successfully (Full Export) and then i
    tried to import datas for the first time.
    I got this following message in my logfile: I ADDED SOME COMMENTS
    Warning: the objects were exported by L1, not by you
    . importing SYSTEM's objects into SYSTEM
    REM ************** CREATING TABLESPACES *****
    REM *********************************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "USER_DATA" DATAFILE 'E:\ORANT\DATABASE\USR1ORCL.ORA' SI"
    "ZE 3145728 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAX"
    "EXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLESPACE "ROLLBACK_DATA" DATAFILE 'E:\ORANT\DATABASE\RBS1ORCL.ORA"
    "' SIZE 10485760 DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS "
    "1 MAXEXTENTS 121 PCTINCREASE 50) ONLINE PERMANENT"
    etc........
    IMP-00017: following statement failed with ORACLE error 1119:
    "CREATE TABLESPACE "L1" DATAFILE 'E:\ORADATA\L1.DBF' SIZE 1048576000 "
    "DEFAULT STORAGE (INITIAL 10240 NEXT 10240 MINEXTENTS 1 MAXEXTENTS 121 PCTIN"
    "CREASE 50) ONLINE PERMANENT"
    IMP-00003: ORACLE error 1119 encountered
    ORA-01119: error in creating database file 'E:\ORADATA\L1.DBF'
    ORA-09200: sfccf: error creating file
    OSD-04002: unable to open file
    O/S-Error: (OS 3) The system cannot find the path specified
    --->etc..........
    the drive E: with the folder E:\ORADATA didn't exist, but after
    all that i created it.
    see below, before my IMPORT statement
    REM ********************* CREATING USER *********
    REM ********************************************
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "L1" IDENTIFIED BY VALUES 'A6E0DAA6865E7627' DEFAULT TABLESPACE"
    " "L1" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'L1' does not exist
    IMP-00017: following statement failed with ORACLE error 959:
    "CREATE USER "MLCO" IDENTIFIED BY VALUES '56AC6447B7D50467' DEFAULT TABLESPA"
    "CE "MLCO" TEMPORARY TABLESPACE "TEMPORARY_DATA""
    IMP-00003: ORACLE error 959 encountered
    ORA-00959: tablespace 'MLCO' does not exist
    ETC.......
    REM ********************* GRANTING ROLES ***********
    REM ************************************************
    IMP-00017: following statement failed with ORACLE error 1917:
    "GRANT ALTER ANY TABLE to "L1" "
    IMP-00003: ORACLE error 1917 encountered
    ORA-01917: user or role 'L1' does not exist
    ETC.........
    IMP-00017: following statement failed with ORACLE error 1918:
    "ALTER USER "L1" DEFAULT ROLE ALL"
    IMP-00003: ORACLE error 1918 encountered
    ORA-01918: user 'L1' does not exist
    -- that is normal, since the creation of the
    tablespace failed !!
    REM******************************
    IMP-00015: following statement failed because the object already exists:
    "CREATE ROLLBACK SEGMENT RB_TEMP STORAGE (INITIAL 10240 NEXT 10240 MINEXTENT"
    "S 2 MAXEXTENTS 121) TABLESPACE "SYSTEM""
    IMP-00015: following statement failed because
    . importing SCOTT's objects into SCOTT
    IMP-00015: following statement failed because the object already exists:
    "CREATE SEQUENCE "EVT_PROFILE_SEQ" MINVALUE 1 MAXVALUE 999999999999999999999"
    "999999 INCREMENT BY 1 START WITH 21 CACHE 20 NOORDER NOCYCLE"
    ETC............
    importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00003: ORACLE error 1435 encountered
    ORA-01435: user does not exist
    REM *************** IMPORTING TABLES *******************
    REM ****************************************************
    . importing SYSTEM's objects into SYSTEM
    . . importing table "AN1999_BDAT" 243 rows imported
    . . importing table "BOPD" 112 rows imported
    . . importing table "BOINFO_AP" 49
    ETC................
    . . importing table "BO_WHF" 2 rows imported
    IMP-00015: following statement failed because the object already exists:
    "CREATE TABLE "DEF$_CALL" ("DEFERRED_TRAN_DB" VARCHAR2(128),
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ROLES" FOR "SYS"."DBA_ROLES""
    IMP-00015: following statement failed because the object already exists:
    "CREATE SYNONYM "DBA_ERRORS" FOR "SYS"."DBA_ERRORS""
    IMP-00008: unrecognized statement in the export file:
    . importing L1's objects into L1
    IMP-00017: following statement failed with ORACLE error 1435:
    "ALTER SCHEMA = "L1""
    IMP-00008: unrecognized statement in the export file:
    J
    Import terminated successfully with warnings.
    -------------------------------------b]
    So after analysing this log file, i created
    the appropriate drives and folders... as the
    import statement doesn't see them.
    E:\ORADATA G:\ORDATA etc...
    And i started to [b]IMPORT ONE MORE TIME. with:
    $ IMP73 sys/pssw Full=Y FILE=c:\temp\FOLD_1\data_1.dmp BUFFER=64000
    COMMIT=Y INDEXFILE=c:\temp\FOLD_1\BOO_idx.sql
    LOG=c:\temp\FOLD_1\BOO_log.LOG DESTROY=Y IGNORE=Y;
    after that i could not see the users nor the
    tables created.
    and the following message appeared in the log file:
    Warning: the objects were exported by L1, not by you
    . . skipping table "AN1999_BDAT"
    . . skipping table "ANPK"
    . . skipping table "BOAP"
    . . skipping table "BOO_D"
    ETC.....skipping all the tables
    . . skipping table "THIN_PER0"
    . . skipping table "UPDATE_TEMP"
    Import terminated successfully without warnings.
    and only 2 new tablespaces (originally 3) were
    created without any data in ( i check that in
    the Oracle Storage manager : the tablespaces exit
    with 0.002 used space; originally 60 M for each !!)
    so,
    How to import data (with full import option) succefully
    MORE THAN ONE TIME from an exported dump file ?
    Even if we have to overwrite tablespaces , tables and users.
    thank you very much

    The Member Feedback forum is for suggestions and feedback for OTN Developer Services. This forum is not monitored by Oracle support or product teams and so Oracle product and technology related questions will not be answered. We recommend that you post this thread to the appropriate Database forum.
    The main URL is:
    http://forums.oracle.com/forums/index.jsp?cat=18

  • Error while exporting metadata file /iam-features-ldap-sync/LDAPUser.xml

    Hi All,
    i am trying to export /iam-features-ldap-sync/LDAPUser.xml metadata file with the weblogic properties mentioned below
    # Weblogic Server Name on which OIM application is running
    wls_servername=oim_server1
    # If you are importing or exporting any out of box event handlers, value is oim.
    # For rest of the out of box metadata, value is OIMMetadata.
    # If you are importing or exporting any custom data, always use application name as OIMMetadata.
    application_name=OIMMetadata
    # Directory location from which XML file should be imported.
    # Lets say I want to import User.xml and it is in the location /scratc/asmaram/temp/oim/file/User.xml,
    # I should give from location value as /scratc/asmaram/temp/oim. Make sure no other files exist
    # in this folder or in its sub folders. Import utility tries to recursively import all the files under the
    # from location folder. This property is only used by weblogicImportMetadata.sh
    metadata_from_loc=@metadata_from_loc
    # Directory location to which XML file should be exported to
    metadata_to_loc=D:/MDS
    # For example /file/User.xml to export user entity definition. You can specify multiple xml files as comma separated values.
    # This property is only used by weblogicExportMetadata.sh and weblogicDeleteMetadata.sh scripts
    metadata_files=/metadata/iam-features-ldap-sync/LDAPUser.xml
    # Application version
    application_version=11.1.1.3.0
    i get the following error
    Initializing WebLogic Scripting Tool (WLST) ...
    Welcome to WebLogic Server Administration Scripting Shell
    Type help() for help on available commands
    Starting export metadata script ....
    Please enter your username [weblogic] :weblogic
    Please enter your password [welcome1] :
    Please enter your server URL [t3://localhost:7001] :t3://hostname:7001
    Connecting to t3://hostname:7001 with userid weblogic ...
    Successfully connected to Admin Server 'AdminServer' that belongs to domain 'OIMDOMAIN'.
    Warning: An insecure protocol was used to connect to the
    server. To ensure on-the-wire security, the SSL port or
    Admin port should be used instead.
    Location changed to domainRuntime tree. This is a read-only tree with DomainMBea
    n as the root.
    For more help, use help(domainRuntime)
    Problem invoking WLST - Traceback (innermost last):
    File "C:\Oracle\Middleware\Oracle_IDM1\server\bin\weblogicExportMetadata.py";,
    line 22, in ?
    File "C:\Oracle\MIDDLE~1\ORACLE~1\common\wlst\mdsWLSTCommands.py";, line 134, i
    n exportMetadata
    File "C:\Oracle\MIDDLE~1\ORACLE~1\common\wlst\mdsWLSTCommands.py";, line 568, i
    n executeAppRuntimeMBeanOperation
    File "C:\Oracle\MIDDLE~1\ORACLE~1\common\wlst\mdsWLSTCommands.py";, line 538, i
    n getMDSAppRuntimeMBean
    UserWarning: MDS-91002: MDS Application runtime MBean for "OIMMetadata" is not available. "exportMetadata" operation failure.
    i have exported these files multiple times, it never gave an errror, but this time i see this error, please help.
    Thanks in advance

    Glad that worked.
    Working with MDS, another way is to use the EM console for exporting/importing data from/to MDS. This I find lot easier rather than working with the OOTB script.
    Steps are:
    http://ADMINSTRATION_SERVER/em
    Navigate to Identity and Access, oim. Right-click and navigate to System MBean Browser.
    Under Application Defined MBeans, navigate to oracle.mds.lcm, Server:oim_server1, Application:oim, MDSAppRuntime.
    To export the configuration files:
    •     Click the Operations tab, and then click exportMetaData.
    •     In the toLocation field, enter /tmp or the name of another directory.
    •     Select createSubDir as false.
    •     In the docs field, enter the complete file location as the Element.
    •     Also select false for excludeAllCust, excludeBaseDocs, and excludeExtendedMetadata. Then, click Invoke.
    This exports the file specified in the docs field to the directory specified in the toLocation field.
    To import the configuration files:
    •     Click importMetaData
    •     In the fromLocation field, enter /tmp or the name of the directory in which you have the configuration files.
    •     Select createSubDir as false.
    •     In the docs field, enter the complete file location as the Element. For example, /db/oim-config.xml.
    •     Also select false for excludeAllCust, excludeBaseDocs, and excludeExtendedMetadata. Then, click Invoke.
    This imports the file specified in the docs field to MDS in the toLocation field.
    HTH

  • Naming for Export dmp file in Date and Time

    Dear All,
    Please advise me how to give the file name in date and time stamp for the export dmp file and it's log file.Suggest me in Oracle 9i version.
    for example the file name get like this 020420041155.dmp and 020420041155.log where date and time are 02/04/2004 11:55 respectively...
    Kindly let me know the syntax of filename to be given along with the EXP command..
    Thanking You,
    Ragesh

    Hello
    I don't think you can get the exp command to automatically name an export for you. The only way I think you would be able to do it is through a shell script or something where you format the file name yourself and pass it to exp. If you wanted to you could generate the command line through a SQL statement and then spool it out to a file i.e.
    spool runexp.bat
    select 'exp user/pwd@tns file='||TO_CHAR(SYSDATE,'DDMMYYYYHH24MI')||'.dmp tables=(table1,table2)' from dual;
    spool offThe second option isn't as flexible as the first but if it's just a couple of exports you have to do, it will probably prove to be less work
    HTH
    David

  • Reg importing exported ccsexport files from 9.3.1 into 11.1.2.1

    Hi,
    I have exported cssimportexport.properties file of hyperion system 9.3.1
    Now i have exported .csv files with the following names(export MSAD and export Native).
    I want them to be moved into hyperion system 11.1.2.1 using LCM.
    Pl. suggest me the steps to be followed in importing the .CSV files into my system 11.1.2.1.
    Cheers
    Praveen

    The LCM documentation is available at http://download.oracle.com/docs/cd/E17236_01/epm.1112/epm_lifecycle_management/launch.htm
    Also if you are new to LCM then it is worth having a look at the following Oracle by Example which will take you through the steps - http://www.oracle.com/webfolder/technetwork/tutorials/obe/hyp/PLN11.1.2_LCM/index.html
    It may also be worth extracting some test provisioning out of your 11.1.2.1 environment first to make sure you have the format of the file correct.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Problem exporting '.txt' file size 23 KB and '.zip' file size 4 MB

    I am using Apex 3.0 version screen to upload '.txt' file and '.zip' file containing images.
    I can successfully export '.txt' file and '.zip' file containing images as long as '.txt' file size is < 23 KB and '.zip' file size < 4 MB from database table 'TBL_upload_file' to the OS directory on the server.
    processing of Larger files (sizes 35 KB and 6 MB) produce following Error Message.
    ‘ORA-21560: argument 2 is null, invalid or out of range’ error.
    Here is my code:
    I am using following code to export Documents from database table 'TBL_upload_file' to the OS directory on the server.
    create or replace procedure "PROC_LOAD_FILES_TO_FLDR_BYTES"
    (pchr_text_file IN VARCHAR2,
    pchr_zip_file IN VARCHAR2)
    is
    lzipfile varchar(100);
    lzipname varchar(100);
    sseq varchar(1000);
    ldocname varchar(100);
    lfile varchar(100);
    -- loaddoc (p_file in number) as
    l_file UTL_FILE.FILE_TYPE;
    l_buffer RAW(32000);
    l_amount NUMBER := 32000;
    l_pos NUMBER := 1;
    l_blob BLOB;
    l_blob_len NUMBER;
    l_file_name varchar(200);
    l_doc_name varchar(200);
    a_file_name varchar (200);
    end_pos NUMBER;
    begin
    -- Get LOB locator
    SELECT blob_content,doc_name
    INTO l_blob,l_file_name
    FROM tbl_upload_file
    WHERE DOC_NAME = pchr_text_file;
    --get length of blob
    l_blob_len := DBMS_LOB.getlength(l_blob);
    -- save blob length to determine end position
    end_pos:= l_blob_len;
    -- Open the destination file.
    -- l_file := UTL_FILE.fopen('BLOBS','MyImage.gif','w', 32767);
    l_file := UTL_FILE.fopen('BLOBS',l_file_name,'WB', 32760); --use write byte option supported in 10G
    -- if small enough for a single write
    IF l_blob_len < 32760 THEN
    utl_file.put_raw(l_file,l_blob);
    utl_file.fflush(l_file);
    ELSE -- write in pieces
    -- Read chunks of the BLOB and write them to the file
    -- until complete.
    WHILE l_pos < l_blob_len LOOP
    DBMS_LOB.read(l_blob, l_amount, l_pos, l_buffer);
    UTL_FILE.put_raw(l_file, l_buffer);
    utl_file.fflush(l_file); --flush pending data and write to the file
    -- set the start position for the next cut
    l_pos := l_pos + l_amount;
    -- set the end position if less than 32000 bytes, here end_pos captures length of the document
    end_pos := end_pos - l_amount;
    IF end_pos < 32000 THEN
    l_amount := end_pos;
    END IF;
    END LOOP;
    END IF;
    --- zip file
    -- Get LOB locator to locate zip file
    SELECT blob_content,doc_name
    INTO l_blob,l_doc_name
    FROM tbl_upload_file
    WHERE DOC_NAME = pchr_zip_file;
    l_blob_len := DBMS_LOB.getlength(l_blob);
    -- save blob length to determine end position
    end_pos:= l_blob_len;
    -- Open the destination file.
    -- l_file := UTL_FILE.fopen('BLOBS','MyImage.gif','w', 32767);
    l_file := UTL_FILE.fopen('BLOBS',l_doc_name,'WB', 32760); --use write byte option supported in 10G
    -- if small enough for a single write
    IF l_blob_len < 32760 THEN
    utl_file.put_raw(l_file,l_blob);
    utl_file.fflush(l_file); --flush out pending data to the file
    ELSE -- write in pieces
    -- Read chunks of the BLOB and write them to the file
    -- until complete.
    l_pos:=1;
    WHILE l_pos < l_blob_len LOOP
    DBMS_LOB.read(l_blob, l_amount, l_pos, l_buffer);
    UTL_FILE.put_raw(l_file, l_buffer);
    UTL_FILE.fflush(l_file); --flush pending data and write to the file
    l_pos := l_pos + l_amount;
    -- set the end position if less than 32000 bytes, here end_pos contains length of the document
    end_pos := end_pos - l_amount;
    IF end_pos < 32000 THEN
    l_amount := end_pos;
    END IF;
    END LOOP;
    END IF;
    -- Close the file.
    IF UTL_FILE.is_open(l_file) THEN
    UTL_FILE.fclose(l_file);
    END IF;
    exception
    WHEN NO_DATA_FOUND THEN
    RAISE_APPLICATION_ERROR(-20214,'Screen fields cannot be blank, Proc_Load_Files_To_Fldr_BYTES.');
    WHEN TOO_MANY_ROWS THEN
    RAISE_APPLICATION_ERROR(-20215,'More than one record exist in the tbl_load_file table, Proc_Load_Files_To_Fldr_BYTES.');
    WHEN OTHERS THEN
    -- Close the file if something goes wrong.
    IF UTL_FILE.is_open(l_file) THEN
    UTL_FILE.fclose(l_file);
    END IF;
    RAISE_APPLICATION_ERROR(-20216,'Some other errors occurred, Proc_Load_Files_To_Fldr_BYTES.');
    end;
    I am new to the Oracle.
    Any help to modify this scipt and resolve this problem will be greatly appreciated.
    Thank you.

    Ask this question in the Apex forums. See Oracle Application Express (APEX)
    Regards Nigel

  • Create a test DB from  the export DMP file

    Hi Gurus,
    I was able to create the DMP file by export and transfer it to the test machine. It is a Windows OS.
    Now in this machine I recently installed Oracle 11g. And I have this export *.DMP file in a dir.
    I need to import the DMP file which into this newly created database to give a test database in 11g.
    I also created the user for which the export is done.
    I need to know how to import the DMP file. I am new to this import-export business.
    Thanks
    Amitava.

    Hi,
    Use IMPDP:
    impdp username/password@dbname full=Y directory=<location of the dmp file> dumpfile=<dumpfile name> logfile=<logfile name to be created>
    Check the below link for more information:
    ORACLE-BASE - Oracle Data Pump (expdp and impdp) in Oracle Database 10g

  • TIPS(20) : MAKE A MAP OF YOUR ORACLE FILES

    제품 : SQL*PLUS
    작성날짜 : 1996-11-12
    TIPS(20) : Make a Map of Your Oracle Files
    ==========================================
    The script below generates a report on all the Oracle files
    (database, redo log, and control) in an Oracle UNIX system.
    It's like a map of your Oracle files. Moreover, it also makes available other
    information (size and physical I/Os, if applicable) related to the files.
    THE SCRIPT
    rem
    rem File: filemap.sql
    rem
    rem This script provides information on the location/path, file name,
    rem size and physical I/Os (if applicable) of all the Oracle files
    rem (database, control and redo log) in an Oracle/UNIX platform.
    rem
    rem
    set pages 999
    col path format a20 heading 'Path'
    col fname format a15 heading 'File Name'
    col fsize format 999b heading 'M bytes'
    col pr format 999999b heading 'Phy. Reads'
    col pw format 999999b heading 'Phy. Writes'
    break on path skip 1
    spool zzfilemap.rep
    select substr(name,1,instr(name, '/', -1)-1 ) path,
    substr(name,instr(name, '/', -1)+1 ) fname,
    bytes/1048576 fsize,
    phyrds pr,
    phywrts pw
    from v$datafile df, v$filestat fs
    where df.file# = fs.file#
    UNION
    select substr(name,1,instr(name, '/', -1)-1 ) path,
    substr(name,instr(name, '/', -1)+1 ) fname,
    0 fsize,
    0 pr,
    0 pw
    from v$controlfile
    UNION
    select substr(lgf.member,1,instr(lgf.member,'/', -1)-1) path,
    substr(lgf.member,instr(lgf.member, '/', -1)+1 ) fname,
    lg.bytes/1048576 fsize,
    0 pr,
    0 pw
    from v$logfile lgf, v$log lg
    where lgf.group# = lg.group#
    order by 1,2
    spool off
    exit
    --------------------------Sample Output----------------------------------
    Path File Name M bytes Phy. Reads Phy. Writes
    /u01/oradata/PDB7 control3.ctl
    rdmxf1.dbf 50 694 2781
    redo1b.log 10
    redo2b.log 10
    redo3b.log 10
    tempf1.dbf 30
    /u02/oradata/PDB7 control1.ctl
    rbsf1.dbf 40 6 2813
    redo1a.log 10
    redo2a.log 10
    redo3a.log 10
    /u03/oradata/PDB7 control2.ctl
    genf1.dbf 50 1551 396
    rdmf1.dbf 100 4403 2335
    systemf1.dbf 50 5905 1116
    toolsf1.dbf 15 2979
    16 rows selected.

    Thanks for your reply, but I'm not sure if you understood my point correctly.
    I didn't exactly mean to use my iPod as a USB memory stick.
    What I meant was that if I copied the contents of my iPod to a DVD and if something weird happened to my music player and I would have to format it...
    Could I then just copy the files (I had burned on the DVD) back to my iPod and it would be the same as before? My music would be saved?

  • Export dump files.

    hi all.
    i am using Enterprise Linux server version 5.1.19.6 /Red Hat and Oracle database 11G R2.I want
    to export dump files automatically like we are doing in windows xp with scheduled Task.
    Is it Possible in Linux?plz help me out thanks in advance.

    Mehwish wrote:
    hi all.
    i am using Enterprise Linux server version 5.1.19.6 /Red Hat and Oracle database 11G R2.I want
    to export dump files automatically like we are doing in windows xp with scheduled Task.
    Is it Possible in Linux?plz help me out thanks in advance.Yes,possible,you can use cron job refer
    http://adminschoice.com/crontab-quick-reference
    But for backup recovery operation you should be use rman(recover manager),and through enterprise manager you can schedule backup job.
    http://download.oracle.com/docs/cd/E10317_01/doc/backup.102/e05407/osb_rman_backup.htm
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10734/rcmrecov.htm

  • Export text files

    Hi
    I need to export text files delimited by '|', but our data is very large, i want to know if i can use some database utility.
    would you help me with this problem ?
    Felix Ramirez.

    The quickest and easiest approach is to use SQL*Plus.
    spool "someFile.txt"
    SELECT col1 || '|' || col2 || '|' || col3
      FROM someTablewill generate a pipe-delimited text file called someFile.txt. Depending on the size of your data and the nature of the query (i.e. do you need to do this with a bunch of different queries), and the timing constraints, you could also write a simple JDBC or OCI application to dump the data. I seem to recall Tom Kyte having a script on asktom.oracle.com that will use utl_file to dump a delimited text file, but that's going to be the slowest option.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Changes in export cmd file

    Hi Guru's,
    The offline backup job was not run last Sunday at 7am until I triggered it manually at 1pm.
    So,
    1. i Always want to keep 2 set dump files
    a. Before exporting / dumping, move the previous dump file to “Spider Dump Backup set 1” folder and delete the existing dump copy in the “Spider Dump Backup set 1” folder.
    2. Email notification to Spider Support after job completion
    Below is my export cmd file. Can someone help me to make changes accordingly.
    @echo off
    @echo ----  exp_full.cmd  ------  START  ---------------------
    REM               
    REM     Kjører full eksport av Oracle database
    REM
    REM
    REM
    SET EXPDEST=G:\
    echo on
    echo Eksport startet > %EXPDEST%\exp_full_startet.txt
    exp SYSTEM/milde5@spider FULL=Y FILE=%EXPDEST%\exp_full_spider.dmp LOG=%EXPDEST%\exp_full_spider.log BUFFER=900000 RECORDLENGTH=65536 statistics=none
    @echo ----  exp_full.cmd  ------  FERDIG ---------------------

    Pl post details of OS and database versions. Pl be aware that exports are not backups. For backups, you should be using RMAN
    http://docs.oracle.com/cd/E11882_01/backup.112/e10642/rcmintro.htm#CHDBHGGE
    HTH
    Srini

  • Getting Datapump Export Dump file to the local machine

    I apologize to everyone as this is a duplicate post.
    Re: Getting Datapump Export Dump file to the local machine
    My initial thread(started yesterday)was in 'Database General' and didn't get much response today. Where do i post questions on EXPORT/IMPORT utilities?
    Anyway, here is my problem:
    I want to take the export dump of itemrep schema in orcl database (in a remote machine). I have an Oracle server (10G Rel2) running in my local Windows machine. I have created a user john with necessary EXPORT/IMPORT privileges in my local db. Then i created a Directory object,ie a folder named datapump in my local hard drive and granted READ WRITE privileges to john.
    So john, who is a user in my local machine's oracle db is going to run the expdp utility.
    expdp john/jhendrix@my_local_db_alias SCHEMAS=itemrep directory=datapump logfile=itemrepexp.log
    The above command will fail because it will look for itemrep schema inside my local db, not the remote db where the itemprep is actually located. And you can't qualify the schemaname with its db in the SCHEMAS parameter (like SCHEMAS=itemrep@orcl).
    Can anyone provide me a solution for this?

    I think you can initiate the datapump exp utility from your client machine to export a schema in a remote database.But, Upon execution,oracle looks for the directory in the remote database and not on your local machine.
    You're inovoking expdp from a client (local DB) to export data from a remote DB.
    So, With this method, you can create the dumpfiles only on the Remote server and not on the local Machine.
    You can perform a direct import instead of export using the NETWORK_LINK option.
    Create a DBlink from your local and Remote DB and verify the connection.
    Then,Initiate the Impdp from Your local machine DB using the parameter network_link=<db_link of the Remote DB> to import the schema.
    The advantage of this option eliminates the Dumpfile creation at the Server side.
    There are no dumpfiles during the import process. the Data is imported directly to the target schema.

Maybe you are looking for

  • Wifi works at home but not when away from home

    My Wifi works at home but not when away from home, what is wrong?  I thought with a phone plan I didn't need to have a wifi connection.

  • No Audio when recording videos (iphone 4) although...

    I recorded some videos this last monday and the sound was PERFECT. I recorded one today and there's absolutely no sound, I recorded another short one just to try and again, no sound......... I tried turning it off and back on,  no change, I turned th

  • Scroll bar not working..help me.

    i have a form with vertical scroll bar on it..in browsing to some other records using the mouse it does not sroll up or down, yet if I use the down key or up-key (keyboard) it works. What is the possible way to make the vertical scroll bar works. you

  • Attachement set as INTERNAL has been sent to vendor

    Hi, We are in SRM 4.0 and using extended classic scenario. We are facing issue in PO output. Attachements which are set as "internal" in PO are being sent with Po output (through e-mail or fax) to the vendor. Can you please help me how to restrict th

  • Which table stores relationship of business roles and its associated roles.

    Hi, I want to know the table that stores the relationship between a Business role and its associated roles. Please note that i am aware of the table GRACBUSROLESNAP that stores the business role relationship corresponding to a request id and this is