Exp/imp for upgrading database

Hi,   Is it a good idea to use exp / imp to upgrade a database from 9i to 11g.To overcome parallel I plan to set ll tables degree to 8 manually and then start the export.Which is better conventional upgrade or exp/imp

HI,
Please check Oracle support id : How to Perform a Full Database Export Import during Upgrade, Migrate, Copy, or Move of a Database (Doc ID 286775.1)
Thank you

Similar Messages

  • Patch No for Upgrading Database XE

    Hii All,
    I need to patch My Oracle Express edition(10.1.0.2) to 10.1.0.4 also in my one system
    and in another ystem I need to patch 11.1.0.6.0 to 11.1.0.7.0 ,which is a prerequiste for installing OracleSOA 11g toolset.
    Can anyone plz tell me what are the exact patch no I need to down load for each case?.
    Thanks and regards,
    Deekay.

    Patches are available only when you paid for a support access to metalink, and for the non-free editions.
    Oracle XE is a free version of Oracle database 10.2.0.1.
    Since it is licence free, you are free to use for whatever you want, however, there is no patches available for this Oracle Edition.
    Regarding your 11g upgrade, again, you have to access to metalink (tab "Patches & Updates") and look for the target patchset you want to go to. Patchset depends of your plateform, help yourself.
    Nicolas.

  • Sometimes is Recommended to Exp/ Imp an production Database ?

    Hi there;
    Related to this [my yesterday post|http://forums.oracle.com/forums/message.jspa?messageID=6498780#6498780] , AWR reports I'm looking in google for it;
    Is it recommended to Export and Import an production schema ?
    This operation in others database ( not comparing then to oracle ) , sometimes is recommended, to realign indexes and reconstruct tables;
    So, may be for an not expert, solution "may be" it could be an not-expert solution !
    Thanks;
    Marcos Ortega

    The similarity between Oracle and other rdbms products (actually between any two rdbms products) ends at "SELECT * FROM EMP;". The internals are vastly different. WHat is considered best management practice for one product could be worst management practice for another.
    In the case of oracle, it generally does NOT benefit from constant re-organization. There have been many, many papers published on this. See especially Richard Foote's work on oracle indexing.
    If you have an observed performance issue, you need to track down the specific cause and address that.

  • Using exp imp on 11i apps database

    Hi,
    Does anyone have experience in using exp and imp utility on 11i apps database? Is there any special method to achieve this.
    I want to do a full export and do an import.
    Do we get any performance benefits if we do this?
    Regards,
    SA

    Hi Sa;
    Does anyone have experience in using exp and imp utility on 11i apps database? Is there any special method to achieve this. You want to take exp imp for which issue?
    I want to do a full export and do an import.Follow:
    10g Export/Import Process for Oracle Applications Release 11i [ID 331221.1]
    Export/import process for R12 using 11gR1 [ID 741818.1]
    Do we get any performance benefits if we do this?It depends many segment but finaly u will make one process on server and it has cost,answer you can have some performance issue while you are taking exp-imp
    Regard
    Helios

  • Full Database Exp & Imp

    Hi,
    I am trying to Exp & Imp a full database. I am working on Oracle 9i & 10g on Solaris 9. I am not using Data Pump. Can anyone please help me with the following:
    1. I am performing the full export using SYSTEM user.
    1.a Does a Full export include (or backups up) the DATA DICTIONARY of the database?
    1.b Does a Full export include the backup of SYS and SYSTEM objects?
    I am using the following command to export
    exp system/system@testdb file=$HOME/testdbfullexp.dmp full=y statistics=none
    I have tried importing the FULL export to another database and i did see that SYS and SYSTEM objects were also being imported ( got some errors regarding constraints and inconsistencies).
    I would like to ask like what are the ideal steps to follow to copy a database from DB1 to DB2 using EXP and IMP
    Any information will be of a great help
    Thanks,
    Harris.

    1 a) No, as the data dictionary will be automagicall recreated by implicit SQL.
    This means any non dictionary objects under SYS will be lost
    1 b) as above. SYSTEM however is a normal user.
    any %SYS user will NOT be exported (CTXSYS, MDSYS, etc)
    On import of SYSTEM there will be always errors, as SYSTEM is non-empty after initial database creation.
    Sybrand Bakker
    Senior Oracle DBA

  • HOW TO MONOTOR EXP/IMP

    Dear all,
    please send me the v$ views of exp/imp, for monitor that process.

    mithun wrote:
    http://www.dba-oracle.com/t_monitor_data_pump_import_job.htm
    http://www.oracle-base.com/articles/10g/OracleDataPump10g.php#MiscellaneousInformation
    OP doesn't use datapump, so your links aren't usefull for him.
    http://www.orafaq.com/wiki/Import_Export_FAQ

  • Database upgrade from 8i to 10g using exp/imp

    Dear Friends,
    Please provide the steps to upgrade from 8i to 10g using exp/imp.
    Thanks,
    Rathinavel

    Hi;
    Please also see cold backup option
    How to migrate from 8i to 10g to new server using cold backup [ID 742108.1]
    Also see:
    Upgrading from 8i to 10g with import utility
    http://searchoracle.techtarget.com/answer/Upgrading-from-8i-to-10g-with-import-utility
    Regard
    Helios

  • Exp/Imp alternatives for large amounts of data (30GB)

    Hi,
    I've come into a new role where various test database are to be 'refreshed' each night with cleansed copies of production data. They have been using the Imp/Exp utilities with 10g R2. The export process is ok, but what's killing us is the time it takes to transfer..unzip...and import 32GB .dmp files. I'm looking for suggestions on what we can do to reduce these times. Currently the import takes 4 to 5 hours.
    I haven't used datapump, but I've heard it doesn't offer much benefit when it comes to saving time over the old imp/exp utilities. Are 'Transportable Tablespaces' the next logical solution? I've been reading up on them and could start prototyping/testing the process next week. What else is in Oracle's toolbox I should be considering?
    Thanks
    brian

    Hi,
    I haven't used datapump, but I've heard it doesn't offer much benefit when it comes to saving time over the old imp/exp utilitiesDatapump will be faster for a couple of reasons. It uses direct path to unload the data. DataPump also supports parallel processes, so while one process is exporting metadata, the other processes can be exporting the data. In 11, you can also compress the dumpfiles as you are exporting. (Both data and metadata compression is available in 11, I think metadata compression is available in 10.2). This will remove your zip step.
    As far as transportable tablespace, yes, this is an option. There are some requirements, but if it works for you, all you will be exporting will be the metadata and no data. The data is copied from the source to the target by way of datafiles. One of the biggest requirements is that the tablespaces need to be read only while the export job is running. This is true for both exp/imp and expdp/impdp.

  • Full database exp/imp  between RAC  and single database

    Hi Experts,
    we have a RAC database oracle 10GR2 with 4 node in linux. i try to duplicate rac database into single instance window database.
    there are same version both database. during importing, I need to create 4 undo tablespace to keep imp processing.
    How to keep one undo tablespace in single instance database?
    any experience of exp/imp RAC database into single instance database to share with me?
    Thanks
    Jim
    Edited by: user589812 on Nov 13, 2009 10:35 AM

    JIm,
    I also want to know can we add the exclude=tablespace on the impdp command for full database exp/imp?You can't use exclude=tablespace on exp/imp. It is for datapump expdp/impdp only.
    I am very insteresting in your recommadition.
    But for a full database impdp, how to exclude a table during full database imp? May I have a example for this case?
    I used a expdp for full database exp. but I got a exp error in expdp log as ORA-31679: Table data object "SALE"."TOAD_PLAN_TABLE" has long columns, and longs can not >be loaded/unloaded using a network linkHaving long columns in a table means that it can't be exported/imported over a network link. To exclude this, you can use the exclude expression:
    expdp user/password exclude=TABLE:"= 'SALES'" ...
    This will exclude all tables named sales. If you have that table in schema scott and then in schema blake, it will exclude both of them. The error that you are getting is not a fatal error, but that table will not be exported/imported.
    the final message as
    Master table "SYSTEM"."SYS_EXPORT_FULL_01" successfully loaded/unloaded
    Dump file set for SYSTEM.SYS_EXPORT_FULL_01 is:
    F:\ORACLEBACKUP\SALEFULL091113.DMP
    Job "SYSTEM"."SYS_EXPORT_FULL_01" completed with 1 error(s) at 16:50:26Yes, the fact that it did not export one table does not make the job fail, it will continue on exporting all other objects.
    . I drop database that gerenated a expdp dump file.
    and recreate blank database and then impdp again.
    But I got lots of error as
    ORA-39151: Table "SYSMAN"."MGMT_ARU_OUI_COMPONENTS" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    ORA-39151: Table "SYSMAN"."MGMT_BUG_ADVISORY" exists. All dependent metadata and data will be skipped due to table_exists_action of skip
    ......ORA-31684: Object type TYPE_BODY:"SYSMAN"."MGMT_THRESHOLD" already exists
    ORA-39111: Dependent object type TRIGGER:"SYSMAN"."SEV_ANNOTATION_INSERT_TR" skipped, base object type VIEW:"SYSMAN"."MGMT_SEVERITY_ANNOTATION" >already exists
    and last line as
    Job "SYSTEM"."SYS_IMPORT_FULL_01" completed with 2581 error(s) at 11:54:57Yes, even though you think you have an empty database, if you have installed any apps or anything, it may create tables that could exist in your dumpfile. If you know that you want the tables from the dumpfile and not the existing ones in the database, then you can use this on the impdp command:
    impdp user/password table_exists_action=replace ...
    If a table that is being imported exists, DataPump will detect this, drop the table, then create the table. Then all of the dependent objects will be created. If you don't then the table and all of it's dependent objects will be skipped, (which is the default).
    There are 4 options with table_exists_action
    replace - I described above
    skip - default, means skip the table and dependent objects like indexes, index statistics, table statistics, etc
    append - keep the existing table and append the data to it, but skip dependent objects
    truncate - truncate the existing table and add the data from the dumpfile, but skip dependent objects.
    Hope this helps.
    Dean

  • Need help for full db exp-imp

    Hi,
    My database is having undo segment corruption.I have considered and tried a lot of things to come out of the situation,but didn't get any +ve result. So I have decided to take a full database export,rename the database,create a new database on the same system and import the full database in the new database.Will you plz tell me the steps to do this(full db import)? Plz include a wise syntax for full db exp-imp.
    Thanks a lot

    If your only identified problem was sudden undo corruption to an online instance and you followed the steps referenced on the Burleson site (switch to manual undo management, create a new undo tablespace, drop the old undo tablespace, and switch back to automatic undo management) and then the very next day encountered undo corruption again, I would recommend two simultaneous courses of action:
    1) Get the hardware analyzed, this could be a signal of hard drive or controller problems.
    2) Scour metalink for any bug references that may be related (or just go ahead and open an iTar).
    If you persist in importing to create the database back on this same system and if you must use the same storage, I would at least try to get the file system or volume that the corruption was on taken offline and recreated.
    Just an alternate view of things ... good luck!

  • Using exp/imp to clone a database from Solaris to NT

    Hi All,
    We have an 8.1.6 Oracle database in production under Sun Solaris.
    We actually use this database to run Siebel CRM system.
    We have a Windows NT server which we used as a QA system
    for this application.
    We have a daily full export of our production database on the
    Sun machine. Now my manager wants to import this production
    dmp file into this Win NT database on a weekly basis.
    The goal here is to maintain the state of this NT database
    as much functional and close to the production one as possible.
    I would appreciate your help on this.
    I have some other general question:
    in which case can the Oracle imp utility be run in the
    full=y mode ?
    does this make any sense at all ?
    since the Oracle import utility does not delete
    any existing objects before it imports, these objects
    should be deleted manually and that's what I do
    with all Siebel users schemas; of course I do not
    do this with SYS, SYSTEM and other special Oracle
    users; question: are there any objects in these
    special Oracle schemas that have to be carried over from
    the source database to the target database to
    make the rest of the schemas functional ?
    Somehow I failed to find Oracle documentation on how
    to use exp/imp to create a complete clone of an Oracle
    database.
    Can direct path be used with full=y export/import mode ?
    Thank you in advance,
    Yuriy.

    As an alternative to using daily/weekly exports and imports, what would probably be more effective here would be replication. The databases could synch on a nightly basis and updated tables and rows will be transferred from your OLTP system to your read-only system. More information can be found on replication from "Oracle8i DBA Handbook" by Kevin Loney and Marlene Therialt.
    LM
    PS. IGNORE=Y

  • Step by Step EXP / IMP database from Oracle 9i to Oracle 11g

    Please help on the captioned where I did the below
    1. Use EXP to export the database from a Oracle9i database (O9_DB) OLD_Windows2003_SERVER
    2. Install the Oracle 11g (default service name is ORCL) to a new Windows 2003 server and use the IMP to import the O9_DB, however, I need to create the same physical path as the OLD_Windows2003_SERVER (e.g E:\Oralce_9i\Ordata\) in order to complete the IMP
    3. I can launch the SQL Developer and obtain the records count of evey tables being imported
    However, is there any best / proper way to do the EXP/IMP, the ideal outcome would be
    1. Create a new namespace say (OLDDB) which same as the Oracle9i database name so maybe I don't need to modify much on the application side to point to the new database ?
    2. IMP the OLDDB to the new namespace OLDDB
    Thanks in advance

    Hello,
    If you create the new Database on a new Server, you could keep the same name for the Database.
    To switch from the old Database to the new one, you'll have to change the hostname on the "tnsnames.ora" or in any configuration file which define the database connexion.
    About the best way to do the Export/Import, you may have some details on the link below:
    http://download.oracle.com/docs/cd/E11882_01/server.112/e10819/expimp.htm#i262317
    Else, why do you intend to stay on a Windows 2003 Server, why not moving to Windows 2008 ?
    NB: Oracle *11.1* is certified on Windows 2008 and Oracle *11.2* is certified on Windows 2008 R2.
    Hope this help.
    Best regards,
    Jean-Valentin

  • Database Upgrade Agent not working for upgrading Oracle 9i DB to Oracle 10i

    Hi All,
    Presently I have completed my Oracle 10i installation on Solaris Server but I am facing an issue during the upgrade process of Oracle 9i to 10g. Whenever I try to upgrade the 9i databases to 10g using DBUA(Database Upgrade assistant) I am facing an issue that Data base not running from oracle home or pfile not found. The exact error message says:-
    The upgrade assistant failed in bringing up the database XIPCOTS. Oracle Home /opt/apps/home/oracle9i/product/releases/92/92 obtained from file /var/opt/oracle/oratab was used to connect to the database. Either the database is not running from Oracle Home /opt/apps/home/oracle9i/product/releases/92/92 or the correct Initialisation parameter file (pfile) was not found. To start the database , Oracle needs the pfile. Provide the full path name of the init file for the database XIPCOTS.
    In reality the XIPCOTS database is running fine in Oracle 9i and we can connect it using PL/SQL developer in windows.I have also provided the path of pfile but it is still not accepting the same.
    Path of pfile I provided :- /XIP/ofaroot/XIPCOTS/pfile/initXIPCOTS.ora
    Any help or suggestions would be greately appreciated.
    Our both Oracle 9i and 10g are installed on same Solaris server where solaris version is 9
    Installation path for 9i :- /opt/apps/home/oracle9i/product/releases/92/92
    Installation path for 10g :- /opt/apps/home/oracle/product/releases/10201
    Oracle 9i version :- 9.2.0.7
    Oracle 10g version :- 10.2.01
    Thanks in advance.
    Best Regards,
    Gourav Banerjee

    Can you please mention the specific folder name under oracle home where we need to put this pfile because there are two folder dbs and ds I have allready tried putting it inside dbs its not working. And my home path is correct I hope and is set for oracle 9i directories in the .oraenv file do we need to set it for oracle 10g before the DBUA upgrade.

  • Exp/Import database schema vs Ch 10 Exp/Imp Content

    Hi I'm using Portal 10.1.2.02. I'm a DBA tasked with migrating a Dev portal into Test, then Production.
    There seems to be a number of ways to achieve this and I was hoping to clarify which is best for me.
    I've run through note: 330391.1(copying the Portal schema) which details running a perl script to export and import onto a newly installed server. This process worked will.
    Chapter 10 of the Portal guide details a fairly complex process of creating transports sets and migrating these over and importing.
    My question is: if I want the entire Portal copied is there any difference in these processes? ie. Do you end up with the same result?
    thanks in advance of any advice :-)

    The cloning model is not a rerunnable model, its not granular and conditional migraton is not possible.
    You can do the cloning only if u want to take a copy of the entire setup and rewire to a new midtier.
    Portal exp/imp model helps u to achieve granular, conditional and rerunnable method of moving portal objects.
    Apart from that, it comes with the readily available prechecks to intimate what's going wrong during the process.

  • User/Admin guide after upgrade database for recover db provided by Oracle?

    Does anyone has a user/admin guide which is provided by oracle about using RMAN and how to use RMAN to RECOVER database tailor make for after upgrade database? I have been searching for a long time and don't find this document.
    Anyone has this, please share.
    FAN

    Look in the Backup & Recovery section:
    http://www.oracle.com/pls/db102/portal.portal_db?selected=4

Maybe you are looking for

  • Sdo.relate error

    Hello, I'm trying to select all points from POINTABLE that are within a polygon (AAAMAY00) that exist in a HDW_POLYGONS table. I'm using Oracle Spatial 8.1.5. and I'm using the following statement: select a.GEOM.SDO_point.x, a.GEOM.SDO_point.Y, a.DEP

  • How to print pdf from black to color?

    I've successfull changing the font color by set in the edit-preference-accessibility-replace color document. I changed the black to blue, but when printing, the result is still black. I want to print it in blue, not black. Should I change the default

  • Can't access site mailbox- Exchange 2013

    Hi Please I have been refered to the Exchange 2013 forum Support team by the SharePoint Team to resolve a complex issue that is unfortunately taking quite some time to resolve. I currently manage a Sharepoint 2013 environment. I have a requirement to

  • Problem with graphic card on dv9660ef

    Hi, I have hp dv9660ef laptop, and it seem that my graphic card burned. I hear that os is booting up, but there is no picture on the screen. Can, anybody tell me is the graphic card integrated on the motherboard or not. model: HP pavilion dv9500 Serv

  • DHCP client list missing devices.

    I just bought a TC and is setup as my router. I have a pc connected to one of the ethernet ports in the TC, and it works fine. My problem is that the other peripherals connected to the other two ports on the TC are not working. One of he is a printer