OS-DB migration for Oracle

Hello,
I am doing an os-db migration of ECC 5.0. I have one query,
1. My source database is on Oracle 9.2
2. I want to build my target on Oracle 10g.
After exporting the database from Oracle 9i using R3load can i import it directly into Oracle 10g ?
If yes what care I have to take, Are there any post installation steps which needs to be performed after importing the data e.g converting long column to lob etc.
Has anyone tried doing this.Please reply me if someone has performed this kind of migration.
Regards
Deepak

Hello Deepak,
>> Has someone performed the migration of LONG-LOB migration and faced performance issues ?
Yes i have done some upgrades from 4.6c to ERP 2005 (with unicode migration) and in this scenario the LONG columns will be converted to LOBs (as i already mentioned, because i have used a R3load 7.00).
A good example for the "automatic" LONG-LOB migration with R3load 7.00 is the table VBDATA. You can test it very easily by exporting and importing it with a R3load 7.00 - the column VBDATA type "LONG RAW" will be converted to BLOB.
I have not faced any performance issues relating to the LOB migration, but this depends on your tables and how they will be used. The performance problems depends on caching or if the LOBs are in-line or out-line ...
For example the column VBDATA is cached and stored in-line:
> SQL> select COLUMN_NAME, CACHE, IN_ROW from ALL_LOBS where TABLE_NAME = 'VBDATA';
> VBDATA  YES        YES
If you want to know more about LOBs and their performance factors - take a look at the oracle whitepaper:
http://www.oracle.com/technology/products/database/application_development/pdf/lob_performance_guidelines.pdf
Regards
Stefan

Similar Messages

  • Can I use SQL developer for a Sybase IQ migration to Oracle?

    Hello,
    Can I use SQL developer for a Sybase IQ migration to Oracle?
    Currently in the supported list appears only Sybase Adaptive Server 12 and 15.
    Thanks,
    Florin D.

    Sorry, we don't support IQ migrations at this time.

  • Migration Tool for Oracle 8i

    Hi Friends !
    I want to migrate Sql Server 7 database to Oracle 8i. For this I have downloaded Oracle Migration Workbench from OTN. Bt when I tried to install it I got message "The Oracle Migration Workbench 2.0.2.0.0 must be installed into an Oracle 9.0.1 Home"
    I think the tool provided on OTN is for Oracle 9i. Where can I get the tool that will help me to migrate the databases to 8i ?
    Thanks in advance.
    Abhi

    812294 wrote:
    Can anyone tell me where i can download tool for Oracle reports developer for Oracle 10g? can you give me exact link?
    Thanks muchI am not clear what software you desire.
    http://otn.oracle.com
    Edited by: sb92075 on Sep 17, 2011 12:37 PM

  • Open source tool for oracle to mysql migration

    Hi,
    Is there any open source tools available for the database migration from Oracle to MySQL.
    Please advice.

    Hi,
    This forum is for migrating to Oracle from non-Oracle databases. It would be better if you tried an Internet search or the MySQL forum -
    http://forums.mysql.com/
    for information about moving from Oracle to MySQL.
    Regards,
    Mike

  • Step for oracle database storage migration

    Hi all
    Need a help for database storage migration from old storage to new storage. Please advise and given step to migrate the database storage.
    Database version: 9i,10g(ASM),10g and 11g (ASM)

    Hi;
    Your question is not clear. Migration from what to what?
    For all migration question please review:
    Master Note For Oracle Database Upgrades and Migrations [ID 1152016.1]
    Different Upgrade Methods For Upgrading Your Database [ID 419550.1]
    Migration of Oracle Database Instances Across OS Platforms[ID 733205.1]
    How To Use RMAN CONVERT DATABASE on Source Host for Cross Platform Migration [ID 413586.1]
    PS:Please dont forget to change thread status to answered if it possible when u belive your thread has been answered, it pretend to lose time of other forums user while they are searching open question which is not answered,thanks for understanding
    Regard
    Helios

  • Migration from sunsolarise10 to red hat for oracle 9i

    hello gurus
    we are working on migration from sun solaris 10 to red hat. for oracle 9i RAC. can anyone please guide us how we should go about that and if anybody can please provide me any supporting link.
    Thanks in advance
    Message was edited by:
    closetodestination

    Before Oracle 10g, you are pretty much stuck with exp/imp option for Cross-platform migration.
    Depends on the size of your database that could be very time consuming and downtime scheduling is big concern. Therefore some people was trying the replication and switch over route. But that's not very easy to setup and you need to have in-depth knowledge of the data and the application driven.
    If you can upgrade to 10g, then you have option to use transportable tablespace which now support crossplatform migration. And you need to use RMAN to convert endianess of your datafiles.

  • Enable to open omwb.zip for Oracle Migration Workbench

    Dear Freind,
    I download omwb.zip for Oracle Migration Workbench from followinf link but i am enable to unzip/extract it, it is showing error like "It dose not appear to be a valid archive" http://www.oracle.com/technology/software/htdocs/devlic.html?http://download.oracle.com/otn/nt/omwb/101040/omwb.zip

    I have the same problem.
    Do you already have a solution?

  • ODBC errors for Oracle Migration Workbench and SQL Server 2000 on Windows

    Hi all-
    I setup a new ODBC connection for the Oracle 10g Migration Workbench to MS SQL Server 2000 on Windows XP and keep getting the following errors when I try to
    either test the new ODBC connection to MS SQL Server 2000 for use with the Oracle Migration Workbench tool. Goal is to fix these errors so that I can migrate our test SQL Server 2000 databases to Oracle 10g.
    Current configuration
    - Oracle 10.1 Migration Workbench
    - Oracle 10g Database (target)
    - MS SQL Server 2000 (source database)
    -Windows XP
    -ODBC 3.85.1117 driver for Windows XP with SQL Server and Oracle
    Errors:
    I. When I test the ODBC connection:
    Microsoft SQL Server ODBC Driver Version 03.85.1117
    Running connectivity tests...
    Attempting connection
    [Microsoft][ODBC SQL Server Driver][TCP/IP Sockets]Specified SQL server not found.
    TESTS FAILED!
    Now here is what has me totally confused:
    I can use the command line ISQL to access the database on the server and IP is all correct and port is available. So I can access the SQL Server database fine it is when I try to setup ODBC that it gives me problems.
    Any tips how to get this ODBC working?
    Thanks
    Ben Prusinski

    1. Jing Jin. If you had READ the post you would have seen that installing later versions is NOT an option.
    2. Not supported does NOT mean does NOT work, as Balmukund pointed out in the referenced thread.  The site that this databases supports is using ASP which is also NOT supported by Microsoft.  As it turns out Balmukund's solution actually turns
    out to be the correct solution.  I installed SS2K and its Service pack 4 on the machine last night.  Good answer Balmukund. as we say in the Navy, "BRAVO ZULU!!"  which means "WELL DONE!!
    3. Tom Li I just posted the question less than 48 hours ago.  My last post was less than 24 hours ago.  I think I have the right to be given the opportunity to determine whether a reply is the answer or not. I understand you don't like unanswered
    posts sitting around for weeks, months or years. (I don't like it either as I have a few that are still unanswered after several years.)  But you should not be showing replies as answers before I even get a chance to see them, much less try them out. 
    You should also not be showing replies as answers which don't address the criteria of the question.  I specifically stated that I had to install SS2K.  I went on to state, "And before anyone asks I need to install this version because that is what
    the customer is using and I need to duplicate their databases."  So the correct answer, if in fact SS2K did not work on Windows 7, would have been a virtual machine answer, as suggested on the thread Balmukund referenced.  So before you mark a reply
    as an answer read the question.
    Edward R. Joell MCSD MCDBA

  • Anybody has a working customer case on using SLT for Oracle to HANA migration?

    Hi Team,
    We had a request for Oracle to HANA realtime.
    And we know that SLT supports the same.
    Would like to know if there was any customer case or Valuable points to be taken care for the same?
    PS: Writing this thread to get a quick help if anybody has done the same already. Meanwhile we are trying to setup the same and check.
    Regards,
    Krishna Tangudu

    Hello Tobias Koebler
    Request your help with this. Can you help me in understanding if there are any active customers or case studies where they have replicated the data from Oracle to SAP HANA?
    I got only this link:
    http://events.asug.com/2013AC/Business%20Integration%20Technology%20&%20%20Infrastructure/0603%20Implementing%20SAP%20HA…
    Awaiting your response
    Regards,
    Krishna Tangudu

  • Migration from Oracle 9i 32 bit to Oracle 10G 64 bit on Windows 2003 Server

    I am working on migrating Oracle 9i 32 bit to Oracle 10G 64 bit on Windows 2003 from Old server to new server. We have db around 500 GB and 3 schemas. I installed Oracle 10G on new server, created tablespaces, schema's blah blah and now doing export and import at schema level from old server to new server.
    Import is taking more than 4 to 5 hours on new server. My manager is saying, window time is bit longer and he wants me to try possiblities to make this process faster. Can some one help me on this, to use the best possible method to complete this process faster?
    Will below step work, if I try?
    If I install, oracle 9i and 10g both on new server in different home directories and take the hot backup from old machine and restore on new machine in 9i home directory and use the upgrade configuration assistant from 10g and do the upgrade? Pls advice.
    Thanks in advance,
    Hari babu
    Edited by: user6367891 on Mar 2, 2010 5:28 AM

    It looks good.
    I have one question in below steps:
    To migrate an Oracle9i or older database to an Oracle Database 10g Release 1 (10.1) database for 64-bit Windows:
    1 Perform steps 1 - 11 in "Migrating an Oracle Database 10g Release 1 (10.1) Database".
    2 Shut down the database on the 64-bit computer:
    SQL> SHUTDOWN IMMEDIATE;
    3 Start the database migration:
    SQL> STARTUP MIGRATE;
    4 Migrate the database as described in Chapter 3, "Upgrading a Database to the New Oracle Database 10g Release" in Oracle Database Upgrade Guide.
    In step 4, it says, migrate database as described in chapter 3. In chapter 3, it says, to do the upgrade the process using dbca or manually. Pls confirm whether do the upgrade process after issuing startup migrate or not.
    Thanks in advance.

  • Error BACKINT for Oracle Connection

    hi @ maxdb gurus...
    i have a problem backing up my maxdb 7.6.0.033 (serving for a mysap.erp2005 on aix 5.3) using the backint mechanism.
    -> backup media created: two pipes and one parallel medium containing these pipes
    -> bsi.env has been created
    -> backint4sapdb.sar contains
       4 staging areas each with 4096 MB
       files per backint call 2
    -> maxdb is about 57 gb perm. data area (and 6 gb temp.)
    i use the backup wizard (dbmgui) to invoke a complete data backup which starts (pipes are created in the file system) and runs until approx. 4 gb...then the backup terminates with error "-24920 backup operation was unsuccesful. The database request failed with error -8020"
    what might be the problem? are the staging areas too small? is says that when using only one stage area it must hold the complete database...can i then conclude that 4 staging areas must hold the database as well - meaning they must be around 15 gb each?
    GreetZ, AH

    yup...but that also means that the sum of the staging areas need as much space as the database ie. around 70 gigs netto! the staging files will be filled and thus grow until the defined size, meaning that the file system(s) need to be big enough!
    everything works fine until the stage files are filled (up to the defined size) and then stops...when i change the sizes of the staging areas i can reproduce the error!
    i compared the tsm implementation of that system with another system (live cache...kind of small regarding db size...) and found nothing serious...
    -> dbm.ebp (AIX 5.3, MaxDB 7.6.0.033, ERP2005, TSM 5.3.3.0)
    2006-11-09 13:56:50
    Using environment variable 'TEMP' with value '' as directory for temporary files and pipes.
    Using environment variable 'TMP' with value '' as directory for temporary files and pipes.
    Using connection to Backint for MaxDB Interface.
    2006-11-09 13:56:50
    Checking existence and configuration of Backint for MaxDB.
        Using configuration variable 'BSI_ENV' = '/sapdb/data/wrk/TDV/bsi.env' as path of the configuration file of Backint for MaxDB.
        Setting environment variable 'BSI_ENV' for the path of the configuration file of Backint for MaxDB from '/sapdba/data/wrk/TDV/bsi.env' to configuration value '/sapdb/data/wrk/TDV/bsi.env'.
        Reading the Backint for MaxDB configuration file '/sapdb/data/wrk/TDV/bsi.env'.
            The following line of the Backint for MaxDB configuration file does not start with a proper keyword and is ignored:
            The following line of the Backint for MaxDB configuration file does not start with a proper keyword and is ignored:
            The following line of the Backint for MaxDB configuration file does not start with a proper keyword and is ignored:
            The following line of the Backint for MaxDB configuration file does not start with a proper keyword and is ignored:
            Found keyword 'BACKINT' with value '/sapdb/TDV/db/bin/backint'.
            The following line of the Backint for MaxDB configuration file does not start with a proper keyword and is ignored:
            Found keyword 'INPUT' with value '/sapdb/TDV/backup/backint4sapdb.in'.
            Found keyword 'OUTPUT' with value '/sapdb/TDV/backup/backint4sapdb.out'.
            Found keyword 'ERROROUTPUT' with value '/sapdb/TDV/backup/backint4sapdb.err'.
            The following line of the Backint for MaxDB configuration file does not start with a proper keyword and is ignored:
            Found keyword 'PARAMETERFILE' with value '/sapdb/data/wrk/TDV/backint4sapdb.par'.
            Found keyword 'TIMEOUT_SUCCESS' with value '6000'.
            Found keyword 'TIMEOUT_FAILURE' with value '300'.
        Finished reading of the Backint for MaxDB configuration file.
        Using '/sapdb/TDV/db/bin/backint' as Backint for MaxDB program.
        Using '/sapdb/TDV/backup/backint4sapdb.in' as input file for Backint for MaxDB.
        Using '/sapdb/TDV/backup/backint4sapdb.out' as output file for Backint for MaxDB.
        Using '/sapdb/TDV/backup/backint4sapdb.err' as error output file for Backint for MaxDB.
        Using '/sapdb/data/wrk/TDV/backint4sapdb.par' as parameter file for Backint for MaxDB.
        Using '6000' seconds as timeout for Backint for MaxDB in the case of success.
        Using '300' seconds as timeout for Backint for MaxDB in the case of failure.
        Using '/sapdb/data/wrk/TDV/dbm.knl' as backup history of a database to migrate.
        Using '/sapdb/data/wrk/TDV/dbm.ebf' as external backup history of a database to migrate.
        Checking availability of backups using backint's inquire function.
    Check passed successful.
    2006-11-09 13:56:50
    Checking medium.
    Check passed successfully.
    2006-11-09 13:56:50
    Preparing backup.
        Setting environment variable 'BI_CALLER' to value 'DBMSRV'.
        Setting environment variable 'BI_REQUEST' to value 'NEW'.
        Setting environment variable 'BI_BACKUP' to value 'FULL'.
        Constructed Backint for MaxDB call '/sapdb/TDV/db/bin/backint -u TDV -f backup -t file -p /sapdb/data/wrk/TDV/backint4sapdb.par -i /sapdb/TDV/backup/backint4sapdb.in -c'.
        Created temporary file '/sapdb/TDV/backup/backint4sapdb.out' as output for Backint for MaxDB.
        Created temporary file '/sapdb/TDV/backup/backint4sapdb.err' as error output for Backint for MaxDB.
        Writing '/sapdb/TDV/backup/back-tdv-data-m10.pipe #PIPE' to the input file.
        Writing '/sapdb/TDV/backup/back-tdv-data-m11.pipe #PIPE' to the input file.
    Prepare passed successfully.
    2006-11-09 13:56:51
    Creating pipes for data transfer.
        Creating pipe '/sapdb/TDV/backup/back-tdv-data-m10.pipe' ... Done.
        Creating pipe '/sapdb/TDV/backup/back-tdv-data-m11.pipe' ... Done.
    All data transfer pipes have been created.
    2006-11-09 13:56:51
    Starting database action for the backup.
        Requesting 'SAVE DATA QUICK TO '/sapdb/TDV/backup/back-tdv-data-m10.pipe' PIPE,'/sapdb/TDV/backup/back-tdv-data-m11.pipe' PIPE BLOCKSIZE 8 NO CHECKPOINT MEDIANAME 'BACK-TDV-DATA-G1'' from db-kernel.
    The database is working on the request.
    2006-11-09 13:56:51
    Waiting until database has prepared the backup.
        Asking for state of database.
        2006-11-09 13:56:51 Database is still preparing the backup.
        Waiting 1 second ... Done.
        Asking for state of database.
        2006-11-09 13:56:52 Database is still preparing the backup.
        Waiting 2 seconds ... Done.
        Asking for state of database.
        2006-11-09 13:56:54 Database is still preparing the backup.
        Waiting 3 seconds ... Done.
        Asking for state of database.
        2006-11-09 13:56:57 Database is still preparing the backup.
        Waiting 4 seconds ... Done.
        Asking for state of database.
        2006-11-09 13:57:01 Database is still preparing the backup.
        Waiting 5 seconds ... Done.
        Asking for state of database.
        2006-11-09 13:57:06 Database has finished preparation of the backup.
    The database has prepared the backup successfully.
    2006-11-09 13:57:06
    Starting Backint for MaxDB.
        Starting Backint for MaxDB process '/sapdb/TDV/db/bin/backint -u TDV -f backup -t file -p /sapdb/data/wrk/TDV/backint4sapdb.par -i /sapdb/TDV/backup/backint4sapdb.in -c >>/sapdb/TDV/backup/backint4sapdb.out 2>>/sapdb/TDV/backup/backint4sapdb.err'.
        Process was started successfully.
    Backint for MaxDB has been started successfully.
    2006-11-09 13:57:06
    Waiting for end of the backup operation.
        2006-11-09 13:57:06 The backup tool is running.
        2006-11-09 13:57:06 The database is working on the request.
        2006-11-09 13:57:11 The backup tool is running.
        2006-11-09 13:57:11 The database is working on the request.
        2006-11-09 13:57:21 The backup tool is running.
        2006-11-09 13:57:21 The database is working on the request.
        2006-11-09 13:57:36 The backup tool is running.
        2006-11-09 13:57:36 The database is working on the request.
        2006-11-09 13:57:56 The backup tool is running.
        2006-11-09 13:57:56 The database is working on the request.
        2006-11-09 13:58:21 The backup tool is running.
        2006-11-09 13:58:21 The database is working on the request.
        2006-11-09 13:58:51 The backup tool is running.
        2006-11-09 13:58:51 The database is working on the request.
        2006-11-09 13:59:21 The database has finished work on the request.
        Receiving a reply from the database kernel.
        Got the following reply from db-kernel:
            SQL-Code              :-8020
            Date                  :20061109
            Time                  :00135703
            Database              :TDV
            Server                :r4335
            KernelVersion         :Kernel    7.6.00   Build 033-123-130-873
            PagesTransfered       :377688
            PagesLeft             :6903331
            MediaName             :BACK-TDV-DATA-G1
            Location              :/sapdb/TDV/backup/back-tdv-data-m10.pipe
            Errortext             :end of file
            Label                 :DAT_000000017
            IsConsistent          :true
            FirstLogPageNo        :247812
            DBStamp1Date          :20061109
            DBStamp1Time          :00135651
            BDPageCount           :7280971
            DevicesUsed           :2
            DatabaseID            :r4335:TDV_20061109_135703
            Max Used Data Page    :0
            Converter Page Count  :5201
        2006-11-09 13:59:21 The backup tool is running.
        2006-11-09 13:59:22 The backup tool process has finished work with return code 2.
        2006-11-09 13:59:22 The backup tool is not running.
    The backup operation has ended.
    2006-11-09 13:59:22
    Filling reply buffer.
        Have encountered error -24920:
            The backup tool failed with 2 as sum of exit codes. The database request failed with error -8020.
        Constructed the following reply:
            ERR
            -24920,ERR_BACKUPOP: backup operation was unsuccessful
            The backup tool failed with 2 as sum of exit codes. The database request failed with error -8020.
    Reply buffer filled.
    2006-11-09 13:59:22
    Cleaning up.
        Removing data transfer pipes.
            Removing data transfer pipe /sapdb/TDV/backup/back-tdv-data-m11.pipe ... Done.
            Removing data transfer pipe /sapdb/TDV/backup/back-tdv-data-m10.pipe ... Done.
        Removed data transfer pipes successfully.
        Copying output of Backint for MaxDB to this file.
        ---------- Begin of output of Backint for MaxDB (/sapdb/TDV/backup/backint4sapdb.out)----------
            Reading parameter file /sapdb/data/wrk/TDV/backint4sapdb.par.
            Using staging area /sapdb/TDV/backup/stage1 with a size of 1585446912 bytes.
            Using staging area /sapdb/TDV/backup/stage2 with a size of 1585446912 bytes.
            Using 1 file per Backint for Oracle call.
            Using /sapdb/TDV/dbs/backint as Backint for Oracle.
            Using /sapdb/TDV/dbs/initTDV.utl as parameterfile of Backint for Oracle.
            Using /sapdb/data/wrk/TDV/backint4oracle.his as history file.
            Using /sapdb/TDV/backup/backint4oracle.in as input of Backint for Oracle.
            Using /sapdb/TDV/backup/backint4oracle.out as output of Backint for Oracle.
            Using /sapdb/TDV/backup/backint4oracle.err as error output of Backint for Oracle.
            Using a maximal delay for a Backint for Oracle call of 60 seconds.
            Reading input file /sapdb/TDV/backup/backint4sapdb.in.
            Backing up pipe /sapdb/TDV/backup/back-tdv-data-m10.pipe.
            Backing up pipe /sapdb/TDV/backup/back-tdv-data-m11.pipe.
            Found 2 entries in the input file.
            Starting the backup.
            Starting pipe2file program(s).
            Waiting for creation of temporary files.
            1 temporary file is available for backup.
            Calling Backint for Oracle at 2006-11-09 13:59:20.
            Calling '/sapdb/TDV/dbs/backint -u TDV -f backup -t file -p /sapdb/TDV/dbs/initTDV.utl -i /sapdb/TDV/backup/backint4oracle.in -c' .
            Backint for Oracle ended at 2006-11-09 13:59:20 with return code 2.
            Backint for Oracle output:
            Backint for Oracle output:                          Data Protection for mySAP(R)
            Backint for Oracle output:
            Backint for Oracle output:              Interface between BR*Tools and Tivoli Storage Manager
            Backint for Oracle output:              - Version 5, Release 3, Modification 2.0  for AIX LF 64-bit -
            Backint for Oracle output:                    Build: 275  compiled on Nov 20 2005
            Backint for Oracle output:         (c) Copyright IBM Corporation, 1996, 2005, All Rights Reserved.
            Backint for Oracle output:
            Backint for Oracle output: BKI2027I: Using TSM-API version 5.3.3.0 (compiled with 5.3.0.0).
            Backint for Oracle output: BKI2000I: Successfully connected to ProLE on port tdpr3ora64.
            Backint for Oracle output: BKI0005I: Start of program at: Thu Nov  9 13:59:20 MEZ 2006 .
            Backint for Oracle output: BKI5014E: Tivoli Storage Manager Error:
            Backint for Oracle output: ANS1035S (RC406)  Options file '*' could not be found.
            Backint for Oracle output:
            Backint for Oracle output: BKI0020I: End of program at: Thu Nov  9 13:59:20 MEZ 2006 .
            Backint for Oracle output: BKI0021I: Elapsed time: 00 sec .
            Backint for Oracle output: BKI0024I: Return code is: 2.
            Backint for Oracle output:
            Backint for Oracle error output:
            Finished the backup unsuccessfully.
            #ERROR /sapdb/TDV/backup/back-tdv-data-m10.pipe
            #ERROR /sapdb/TDV/backup/back-tdv-data-m11.pipe
        ---------- End of output of Backint for MaxDB (/sapdb/TDV/backup/backint4sapdb.out)----------
        Removed Backint for MaxDB's temporary output file '/sapdb/TDV/backup/backint4sapdb.out'.
        Copying error output of Backint for MaxDB to this file.
        ---------- Begin of error output of Backint for MaxDB (/sapdb/TDV/backup/backint4sapdb.err)----------
            Backint for Oracle was unsuccessful.
        ---------- End of error output of Backint for MaxDB (/sapdb/TDV/backup/backint4sapdb.err)----------
        Removed Backint for MaxDB's temporary error output file '/sapdb/TDV/backup/backint4sapdb.err'.
        Removed the Backint for MaxDB input file '/sapdb/TDV/backup/backint4sapdb.in'.
    Have finished clean up successfully.
    i invoke the backup through dbmgui, not dbmcli!
    any clues? thx in advance!
    GreetZ, AH

  • Difference between upgrdae and migration about oracle database

    Difference between upgrdae and migration about oracle database
    please give the comments

    Well, the question is almost philosophic...<br>
    In 9i, there is a Migration Guide whereas in 10g there is a Upgrade Guide.<br>
    Furthermore, in 9i, there is the command line startup migrate whereas in 10g that's startup upgrade.
    Somebody think upgrade when go to new release, and migration when go to new version.<br>
    Others think upgrade when new version replace database in place, and migration when new version include a move of database.<br>
    Another point of view is : upgrade is for technical, and migration for application/data.<br>
    <br>
    Well, after these explanations, your upgrade/migratation notion will not be more clear, but I think that is not very important, only a terminology game. The most important is to know what you need : new version or new release.<br>
    <br>
    Nicolas.

  • Migration from Oracle App Server 10.1.3.5 to WebLogic 12 - Expected issues?

    Hi all,
    the company I work for wants to migrate its Oracle Application Server 10.1.3.5 (with JRE 1.6) to WebLogic 12 and we need to guess which kind of remediation could be needed to have all web-apps still working.
    The concerned applications are using dated technologies:
    direct JDBC connections to db
    JPA 1.0
    connection pools with Oracle JDBC drivers
    some apps are compiled with Java 1.4, some others with Java 1.6
    servlet 2.3
    JSP 2.0
    We do not expect to come accross big issues but I'm not skilled on this kind of matters so I would like to receive seggestions or alerts from experts. Any remark will be appreciated.
    Thank you all.
    Regards,
    Luca

    Hi Luca,
        It would be better if you asked this question in the Application Server forum which should be at this URL -
    Application Server
    and the general forum here -
    Oracle Application Server - General
    Regards,
    Mike

  • "In-Memory Database Cache" option for Oracle 10g Enterprise Edition

    Hi,
    In one of our applications, we are using TimesTen 5.1.24 and Oracle 9i
    databases (platform - Solaris 9i).
    TimesTen holds application information which needs to be accessed quickly
    and Oracle 9i is a master application database.
    Now we are looking at an option of migrating from Oracle 9i to Oracle 10g
    database. While exploring about Oracle 10g features, came to know about
    "In-Memory Database Cache" option for Oracle Enterprise Edition. This made
    me to think about using Oracle 10g Enterprise Edition with "In-Memory
    Database Cache" option for our application.
    Following are the advantages that I could visualize by adopting the
    above-mentioned approach:
    1. Data reconciliation between Oracle and TimesTen is not required (i.e.
    data can be maintained only in Oracle tables and for caching "In-Memory
    Database Cache" can be used)
    2. Data maintenance is easy and gives one view access to data
    I have following queries regarding the above-mentioned solution:
    1. What is the difference between "TimesTen In-Memory Database" and
    "In-Memory Database Cache" in terms of features and licensing model?
    2. Is "In-Memory Database Cache" option integrated with Oracle 10g
    installable or a separate installable (i.e. TimesTen installable with only
    cache feature)?
    3. Is "In-Memory Database Cache" option same as that of "TimesTen Cache
    Connect to Oracle" option in TimesTen In-Memory Database?
    4. After integrating "In-Memory Database Cache" option with Oracle 10g, data
    access will happen only through Oracle sqlplus or OCI calls. Am I right here
    in making this statement?
    5. Is it possible to cache the result set of a join query in "In-Memory
    Database Cache"?
    In "Options and Packs" chapter in Oracle documentation
    (http://download.oracle.com/docs/cd/B19306_01/license.102/b14199/options.htm
    #CIHJJBGA), I encountered the following statement:
    "For the purposes of licensing Oracle In-Memory Database Cache, only the
    processors on which the TimesTen In-Memory Database component of the
    In-Memory Database Cache software is installed and/or running are counted
    for the purpose of determining the number of licenses required."
    We have servers with the following configuration. Is there a way to get the
    count of processors on which the Cache software could be installed and/or
    running? Please assist.
    Production box with 12 core 2 duo processors (24 cores)
    Pre-production box with 8 core 2 duo processors (16 cores)
    Development and test box with 2 single chip processors
    Development and test box with 4 single chip processors
    Development and test box with 6 single chip processors
    Thanks & Regards,
    Vijay

    Hi Vijay,
    regarding your questions:
    1. What is the difference between "TimesTen In-Memory Database" and
    "In-Memory Database Cache" in terms of features and licensing model?
    ==> Product has just been renamed and integrated better with the Oracle database - Times-Ten == In-Memory-Cache-Database
    2. Is "In-Memory Database Cache" option integrated with Oracle 10g
    installable or a separate installable (i.e. TimesTen installable with only
    cache feature)?
    ==> Seperate Installation
    3. Is "In-Memory Database Cache" option same as that of "TimesTen Cache
    Connect to Oracle" option in TimesTen In-Memory Database?
    ==> Please have a look here: http://www.oracle.com/technology/products/timesten/quickstart/cc_qs_index.html
    This explains the differences.
    4. After integrating "In-Memory Database Cache" option with Oracle 10g, data
    access will happen only through Oracle sqlplus or OCI calls. Am I right here
    in making this statement?
    ==> Please see above mentioned papers
    5. Is it possible to cache the result set of a join query in "In-Memory
    Database Cache"?
    ==> Again ... ;-)
    Kind regards
    Mike

  • What are the Issues in Migration from oracle 8i to oracle 10g

    Hi,
    Can you let me know what are the issues that i face when i migrate from oracle 8i to oracle 10g. I will be very thankful if you give me a list of issues that you face while migrating from 8i to 10g

    In addition to Max's reply, you must be aware that you have to test your applications agains 10g database before you actually upgrade the database.
    If you use CBO, it may produce different execution plans for one query when you run it in 8i and 10g. This may affect your application performances, which you don't like to happen.
    Also you have to decide how much downtime you can afford during the migration. Can you afford some downtime and how much? If you can't than this makes the situation more complicated.
    How big is your database? You may end up only with simple export/import solution using export import utility.
    The bottom line is, whatever you decide, you must test your applications running against 10g database before you actually upgrade your database.

Maybe you are looking for