RAC database Migration

Hi Experts,
I need some advice related to the Migration of RAC database, we have to migrate several RAC & Single-Instance databases from existing HP Itanium servers to Superdome2.
Initially I thought of going with "addnode" for new Servers as additional Nodes to the existing RAC database then remove old Nodes for RAC databases migration, but there are some limitations in our evironment,
1. There are no free IP addresses available to assign to new nodes from the existing IP segment.
The IP Range (1 - 255) is divided into into several small segments of 16 IPs in each segment, and the segment in which the existing 2Node RAC DB is completely occupied and there is no free IP Available.
If the new Node have IPs addresses from different segment in the same Subnet(IP Range) can it be added to the existing cluster?
and if we cannot find any IP in that Subnet then can the nodes with the IPs from other Subnet be added to the Existing Cluster?
2. To add a node to the existing RAC cluster, the new node should be exactly same, OS version, OS Patches...
Our existing rx Itanium servers have "*HP-UX 11.31 Mission Critical Operating Environment*" for One 2 Node RAC Cluster and "*HP-UX 11.31 Enterprise Operating Environment*" for for another 2 Node RAC Cluster, but the HP Superdome 2 will have the Latest "*HP-UX 11.31 Data Center Operating Environment*"
Can we proceed if there are some diferrences in OS patches...?
Please see the Existing Environment Details below:*
Prod 1 DB *2 Node RAC Cluster* :
Hardware: rx 8640 Itanium Server
OS: HP-UX 11.31 v3 Mission Critical Operating Environment
DB Version: Oracle 10.2.0.2
Prod 2 DB *2 Node RAC Cluster*:
Hardware: rx 8640 Itanium Server
OS: HP-UX 11.31 v3 Enterprise Operating Environment
DB Version: Oracle 11.2.0.1 (2 Node RAC)
Prod 3 DB *2 Node RAC Cluster*:
Hardware: rx 6600 Itanium Server
OS: HP-UX 11.31 v3 Enterprise Operating Environment
DB Version: Oracle 11.1.0.6.0 (2 Node RAC)
TARGET DB Environment for All the Databases will be;*
Hardware: HP SUperdome 2
OS: HP-UX 11.31Data Center Operating Environment
DB Version: Oracle 11.2.0.3 (2 Node RAC)
Also, all the 3 RAC databases are in Extended/Stretch RAC Cluster configuration and the new Superdome Env will also be in Extended RAC Cluster Configuration.
Kindly advice what strategy should I follow for the Migration, waiting for your kind reply.
Any help will be highly appreciated.
Thanks in Advance.

I am struck by your post with a couple of thoughts.
1. Even though high availability appears to be something you are willing to invest substantial money in obtaining ... no one think patching important.
2. Even though you've spent a substantial sum of money on hardware and software ... no one thinks networking important enough to address the lack of appropriate network infrastructure.
Personally: I'd advise refusing to do the work until management makes the very small investment necessary to doing it correctly.

Similar Messages

  • Oracle 10g RAC Database Migration from SAN to New SAN.

    Hi All,
    Our client has implemented a Two Node Oracle 10g R2 RAC on HP-UX v2. The Database is on ASM and on HP EVA 4000 SAN. The database size in around 1.2 TB.
    Now the requirement is to migrate the Database and Clusterware files to a New SAN (EVA 6400).
    SAN to SAN migration can't be done as the customer didn't get license for such storage migration.
    My immediate suggestion was to connect the New SAN and present the LUNs and add the Disks from New SAN and wait for rebalance to complete. Then drop the Old Disks which are on Old SAN. Exact Steps To Migrate ASM Diskgroups To Another SAN Without Downtime. (Doc ID 837308.1).
    Clients wants us to suggest alternate solutions as they are worried that presenting LUNs from Old SAN and New SAN at the same time may give some issues and also if re-balance fails then it may affect the database. Also they are not able to estimate the time to re-balance a 1.2 TB database across Disks from 2 different SAN. Downtime window is ony 48 hours.
    One wild suggestion was to:
    1. Connect the New SAN.
    2. Create New Diskgroups on New SAN from Oracle RAC env.
    3. Backup the Production database and restore on the same Oracle RAC servers but on New Diskgroups.
    4. Start the database from new Diskgroup location by updating the spfile/pfile
    5. Make sure everything is fine then drop the current Diskgroups from Old SAN.
    Will the above idea work in Production env? I think there is a lot of risks in doing the above.
    Customer does not have Oracle RAC on Test env so there isn't any chance of trying out any method.
    Any suggestion is appreciated.
    Rgds,
    Thiru.

    user1983888 wrote:
    Hi All,
    Our client has implemented a Two Node Oracle 10g R2 RAC on HP-UX v2. The Database is on ASM and on HP EVA 4000 SAN. The database size in around 1.2 TB.
    Now the requirement is to migrate the Database and Clusterware files to a New SAN (EVA 6400).
    SAN to SAN migration can't be done as the customer didn't get license for such storage migration.
    My immediate suggestion was to connect the New SAN and present the LUNs and add the Disks from New SAN and wait for rebalance to complete. Then drop the Old Disks which are on Old SAN. Exact Steps To Migrate ASM Diskgroups To Another SAN Without Downtime. (Doc ID 837308.1).
    Clients wants us to suggest alternate solutions as they are worried that presenting LUNs from Old SAN and New SAN at the same time may give some issues and also if re-balance fails then it may affect the database. Also they are not able to estimate the time to re-balance a 1.2 TB database across Disks from 2 different SAN. Downtime window is ony 48 hours.Adding and removing LUNs online is one of the great features of ASM. The Rebalance will be perfomed under SAN. No downtime!!!
    If your customer is not entrusting on ASM. So Oracle Support can answer all doubt.
    Any concern .. Contat Oracle Support to guide you in the best way to perform this work.
    >
    One wild suggestion was to:
    1. Connect the New SAN.
    2. Create New Diskgroups on New SAN from Oracle RAC env.
    3. Backup the Production database and restore on the same Oracle RAC servers but on New Diskgroups.
    4. Start the database from new Diskgroup location by updating the spfile/pfile
    5. Make sure everything is fine then drop the current Diskgroups from Old SAN.
    ASM Supports many Terabytes, if you need to migrate 3 Database with 20TB each using this way described above would be very laborious .. .. So add and remove Luns online is one feature that must work.
    Take the approval from Oracle support and do this work using the ASM Rebalance.
    Regards,
    Levi Pereira

  • Migration of  10g RAC database to new sever which includes 11g upgrade

    Hello all,
    I have requirement here to migrate 10g RAC database from one server to another server as part of migration i want to perform upgrade to 11g as well.
    To Help me out.Please post your inputs or way to perform the task.
    Precautions to be taken while doing the activity.
    Please find the below for more info
    Old Box:
    Current environment
    RAC -2node
    ASM
    Database Version 10.2.02
    Os:HP-UX 11.21
    New Box:
    New environment should be
    RAC -2node
    ASM
    Database Version 11.2
    Os:HP-UX 11.31
    Thanks for the help in advance
    Anand

    Pl do not post duplicate threads - Migration of  10g RAC database to new sever which includes 11g upgrade

  • Oracle E-business suite database migrate from AIX to Linux

    Oracle 11i application database migration from single Aix instance to oracle Linux cluster RAC.
    What is the easy way to migrate?
    Migrate single instance from aix to linux cluster.
    OR
    Migrate single instance from aix to linux cluster RAC.
    We have installed oracle Linux cluster.
    Please help me which way I can go with RAC.
    Thanks
    Prince

    Migrating to a single instance will be the simplest. The migration itself will not be much different between the two because you're migrating from one platform to another. The options you have available are, the traditional IMP/EXP, datapump, and RMAN to migrate to either a standalone or RAC environment. The difficulty will be in creating a RAC environment and a little a few extra steps in the RMAN restore. Working with RAC will make things harder, but the actual migration is similar. The steps to IMP/EXP, whether it's traditional or Datapump will be no different. However, standing up a RAC database will be.

  • Arivelog mode configuration proble in RAC database

    Hi,
    I'm trying to configure RAC database in archivelog mode but when try to do so get the following error, during the database startup in mount state
    LRM-00109: could not open parameter file '/u01/app/oracle/product/11.1.0/db_1/dbs/initorcl.ora'
    This error shows that the system could not find the initorcl.ora and really it does not exist. But the system at this stage is single
    instance as I chenged the value of cluster_database=false before shutdown and then start.
    when I enter the command startup mount it give me this error. What should I do to avoid it. Should I make a pfile of this from spfile in cluster mode
    and then put this pfile in the desired path to start the single instance or any other thing should I tried.
    Regards,
    Abbasi

    Perform the following steps to relocate the SPFILE from the local file system to an ASM disk group.
        * Create a text-based initialization parameter file from the current binary SPFILE located on the local file system:
          SQL> CREATE PFILE='$ORACLE_HOME/dbs/initTESTDB.ora'
            2  FROM SPFILE='$ORACLE_HOME/dbs/spfileTESTDB.ora';
          File created.
        * Create new SPFILE in an ASM disk group:
          SQL> CREATE SPFILE='+TESTDB_DATA1/TESTDB/spfileTESTDB.ora'
            2  FROM PFILE='$ORACLE_HOME/dbs/initTESTDB.ora';
          File created.
        * Shutdown the Oracle database:
          SQL> SHUTDOWN IMMEDIATE
          Database closed.
          Database dismounted.
          ORACLE instance shut down.
        * Update the text-based init<SID>.ora file with the new location of the SPFILE in ASM:
          $ echo "SPFILE='+TESTDB_DATA1/TESTDB/spfileTESTDB.ora'" > $ORACLE_HOME/dbs/initTESTDB.ora
        * Remove (actually rename) the old SPFILE on the local file system so that the new text-based init<SID>.ora will be used:
          $ mv $ORACLE_HOME/dbs/spfileTESTDB.ora $ORACLE_HOME/dbs/BACKUP_ASM.spfileTESTDB.ora
        * Open the Oracle database using the new SPFILE:
          SQL> STARTUPsource:-http://www.idevelopment.info/data/Oracle/DBA_tips/Automatic_Storage_Management/ASM_33.shtml#Migrating%20Oracle%20Database%20from%20Local%20File%20System%20to%20ASM
    if you would like, refer the link:-
    http://oracleinstance.blogspot.com/2009/12/enable-archivelog-and-flashback-in-rac.html

  • Oracle RAC Storage Migration - HP EVA to HP EVA

    Hi,
    We are in the process of migrating from an old HP EVA4000 to a new HP EVA6400. We have migrated most of the systems that had storage presented from the EVA4000 with the exception of our Oracle RAC databases and wanted to check to see if anyone knew of any gotchas before we proceeded.
    We are using CA to replicate the LUNs that are presented to the Oracle cluster. My plan is to shut down the database and all associated RAC services, disable CRS then halt the servers. I will then present the CA replicated LUNs to the servers and initiate a failover. I will then restart the servers and check that the LUNs have appeared from the 6400 before manually restarting CRS and the databases. I will then re-enable CRS before rebooting again to check that everything comes up automatically.
    Does this sound correct? Has anyone done this before?
    Thanks in advance.
    Phil

    Seems reasonable, but can't be certain unless you test. Make sure you backup OCR, VOTE, Database before you change the storage allocation.
    How big is the database ? How long will your downtime be ?
    It may be cleaner to just reinstall CRS and restore the database from backup ?
    Are you using ASM ? If so then you could just add new devices to ASM, migrate the data, then remove the old devices.

  • Running OWB on 6 RAC databases on 4 shared nodes

    Hi everyone!
    I was wondering if anyone can help with a question I have around multiple OWB installations on RAC?
    Background:
    We are migrating 6 independent data warehouses / data marts onto one Exadata box. This machine has got 4 nodes, across which we will run 6 separate databases using RAC. Each database will span across all 4 nodes, and each of the databases will contain one of the independent data mart / data warehouses.
    We understand that we will probably need to run 6 separate control center services on each node, and that there may need to be some fiddling around with log file names to get them all to start properly.
    However; the question is where to install the OWB binaries!?
    In our understanding, we need one set of OWB binaries per database, i.e. 6 in total. The actual databases will be run from just one set of local Oracle Database binaries on each RAC node, which means only four database binaries in total will be used to run 6 RAC databases across our 4 nodes.
    Problem is, this means that there won't be enough OWB binaries to go around!! (6-4 = 2 missing!)
    One alternative is to have 6 Oracle homes on each RAC node, i.e. one home per RAC database. I'm not a RAC expert - maybe this is even the preferred way of setting up the databases, regardless if we use OWB or not?
    Comments / suggestions would be much appreciated!
    Cheers
    Edited by: 817871 on 03-Dec-2010 03:25

    Hi,
    We also have a 6 node rac where we settes up owb.
    We went through a whole lot of problems but got it running finally(with service failover).
    Basically you will need the binaries on all 4 of your physical nodes.
    For the first node you can run through the normal installation (repository setu).
    Thab you need to start the repository assistant on all other nodes in ordr to register them.
    There are a lot of pitfalls with that installation and a lack od docunentation.
    We've written an onstallation guide for an OWB rac setup where with we have now setted up 9 OWB repositories spreaded over 2 racs.
    Just let me know if you require detailed information - maybe we can share our guide.

  • Export problem in RAC database (two node) setup

    i'm not able to do any backup like RMAN,export... While doing export following error is coming on the oracle 10.2.0.3.0 RAC database (two node) on windows server 2003...
    UDE-00008: operation generated ORACLE error 31623
    ORA-31623: a job is not attached to this session via the specified handle
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 2745
    ORA-06512: at "SYS.DBMS_DATAPUMP", line 3712
    ORA-06512: at line 1i checked the status of the components- but all the components are in valid status only.. but there are 137 invalid objects in the database... while running utlirp.sql in open mode, it tells to shut the database & run at startup migrate mode...
    please provide me the possible solution for this issue.. it is a very critical situation...

    below are the invalid objects....
    SQL> select count(*) from dba_objects where status='INVALID';
    COUNT(*)
    137
    SQL> select object_name,object_type,status from dba_objects where status='INVALID';
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_FILESTATXS
    SYNONYM INVALID
    DBA_HIST_SQLSTAT
    SYNONYM INVALID
    DBA_HIST_SQLBIND
    SYNONYM INVALID
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_SYSTEM_EVENT
    SYNONYM INVALID
    DBA_HIST_WAITSTAT
    SYNONYM INVALID
    DBA_HIST_LATCH
    SYNONYM INVALID
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_LATCH_MISSES_SUMMARY
    SYNONYM INVALID
    DBA_HIST_DB_CACHE_ADVICE
    SYNONYM INVALID
    DBA_HIST_ROWCACHE_SUMMARY
    SYNONYM INVALID
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_SGASTAT
    SYNONYM INVALID
    DBA_HIST_SYSSTAT
    SYNONYM INVALID
    DBA_HIST_SYS_TIME_MODEL
    SYNONYM INVALID
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_OSSTAT
    SYNONYM INVALID
    DBA_HIST_PARAMETER
    SYNONYM INVALID
    DBA_HIST_SEG_STAT
    SYNONYM INVALID
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_DLM_MISC
    SYNONYM INVALID
    DBA_HIST_INST_CACHE_TRANSFER
    SYNONYM INVALID
    DBA_HIST_ACTIVE_SESS_HISTORY
    SYNONYM INVALID
    OBJECT_NAME
    OBJECT_TYPE STATUS
    DBA_HIST_TABLESPACE_STAT
    SYNONYM INVALID
    DBA_HIST_SERVICE_STAT
    SYNONYM INVALID
    DBA_HIST_SERVICE_WAIT_CLASS
    SYNONYM INVALID
    137 rows selected.
    SQL> spool off
    some sys ojects are also invalid... please give mr the suggestion..

  • Restore single instance rman backup to rac database

    Hello all. I have a 10gR2 single instance database that I have been taking backups using the oracle's suggested backup strategy, that is, once full database backup and incremental backups here and there. That's is all fine as I have (many times) being requested to rollback the database resulting in all operations being sucessfull.
    However, since this is somehow a pre-certification database, I am now being tasked to restore this database to what is going to be our production database, which is a 10gR2 rac database. I have already set up the RAC database sucessfull.
    Now, my question is: How can I take one of the backups that I have on the single-instance database and restore onto my RAC database ?
    Some information about the infraestructure:
    Both operating systems are HP-UX. Single instance is 11.23, RAC database is 11.31.
    Both database are using ASM as datafiles storage and flash recovery area storage.
    Any help will be appreciated!
    Thanks!

    In general terms, there are several ways you can convert a single instance into a rac intance, and it involves much more activities than just adding redo threads.
    Use of DBCA
    Enterprise Manager Grid Control
    rconfig utility
    Considering your case, if you don't currently have Grid Control, the first option is rconfig.
    1. Migrate your database from the current environment to the target environment by a regular cloning procedure, this will create a single instance into your rac environment. Your regular rdbms rac installation allows you to manage both, single instance and rac instances. Use this procedure to clone your database to the target RAC environment using recovery manager --> http://madrid9999.googlepages.com/rmancloneadatabase
    2. Once you have started your database in the RAC environment proceed with the rconfig tool, this requires some files to perform the conversion, the ConvertToRAC.xml file which is located at $ORACLE_HOME/assistants/rconfig/sampleXMLs
    3. Once you copied/edited the ConvertToRAC.xml file, you have to run the rconfig. Save the ConvertToRAC.xml under a different file name.
    rconfig editedConvertToRAC.xml
    4. For further reference on the rconfig command, you may refer to this page Oracle® Real Application Clusters Installation Guide
    11g Release 1 (11.1) for Linux and UNIX
    ~ Madrid

  • Unable to startup 12c RAC database, can't open spfile in ASM

    hello,
    I'm testing 12cRAC database on RHEL5 and need your help to troubleshoot and fix the following issue  -
    DBCA fails at the end of the configuration to create and startup new RAC database with the following errors
    ORA-01078: failure in processing system parameters
    ORA-01565: error in identifying file '+DATA/TDB1/spfileTDB1.ora'
    ORA-17503: ksfdopn:2 Failed to open file +DATA/TDB1/spfileTDB1.ora
    ORA-15056: additional error message
    ORA-17503: ksfdopn:2 Failed to open file +DATA/TDB1/spfiletdb1.ora
    ORA-01017: invalid username/password; logon denied
    I also tried executing script (which was created by DBCA) to create the new DB  manually and got the same results - it happens during first attempt to start it up using SPFILE after new DB was successfully created (it was started up using pfile)
    Clusterware with ASM install was successful, binaries were also installed without issues,
    cluster seems healthy, I see correct files within ASM using asmcmd etc ..
    got stuck here and need some directions ... very confused by ORA-01017 password error
    wonder if anyone had same or similar issues ?
    Thank you !

    Fixed.
    Thank you everyone for suggestions and recommendations -
    ASM is accessible using asmcmd and I can see all the DB files incuding spfile, they were created by DBCA, alert log showed same information I posted
    After extensive troubleshooting and testing the issue was idedntified and fixed - 'dba' group needs to be the primary group for 'oracle' user, not the secondary one. I suspected that something was not right with 'oracle' user access to asm based on that strange password error. There is more I need to check as the issue and fix with swapping groups seems strange. Secondary group is not enough ti have proper rights ??? my first reaction was - seriously ? wtf ?
    Anyway - I need to move one now ...
    Thank you again !

  • How can I have a collective AWR report for RAC database in 10gR2 and 11gR1?

    Plz correct me here
    awrrpt.sql takes the snapshot at the instance level and incase we have 5 instances we have to take 5 awr reports for a particular period.correct???
    If above is true ,any wayout to collect a single collective AWR report for RAC database which includes information on all the instances in Oracle 10g R2 or 11gR1 ?
    Thanks in advance
    Gagan

    I have never come across a way for this. though I can n't say it is not there.
    But I guess it may not be feasible too becuase as we know the current AWR report contains data which is specific for one instance.
    Various hit ratios, top events, instance effeciency reports.... what not ..
    It would be really nice to see something is a new format where it lists values from each instance in a single report.. I guess such a thing Does not exiist as of now.

  • Question about DBCA generate script o create RAC database 2 node cluster

    Question about creating two node RAC database 11g after installing and configuration 11g clusterware. I've used DBCA to generate script to create a rac database. I've set
    environment variable ORACLE_SID=RAC and the creating script creates instance of RAC1 and RAC2. My understanding is that each node will represent a node, however there should only be one database with a name of 'RAC'. Please advise

    You are getting your terminology mixed up.
    You only have one database. Take a look, there are one set of datafiles on shared storage.
    You have 2 instances which are accessing one database.
    Database name is RAC. Instance names are RAC1, RAC2, etc, etc.
    Also, if you look at the listener configuration and if your tnsnames is setup properly then connecting to RAC will connect you to either one of the instances wheras connecting to RAC1 will connect you to that instance.

  • ORA-20600 error while trying to add RAC database

    I'm receiving the following error while trying to add the RAC database in grid control. Any idea?.
    ORA-20600: The specified target is in the process of being deleted
    This database doesn't exists in targets. But, I'm still unable to add.

    Check the metalink Note: Unable To Add Rac Database To Grid Control Getting Error Ora-20600 [ID 1218797.1]
    You need to wait for a while and check the status.

  • Anybody using 11g RAC Database with OIM 9.1.0.2?

    Hi,
    We have poor performance problems with 11g RAC Database with OIM 9.1.0.2.
    It looks like the issue could be because of RAC Ccluster.
    Anybody using 11g RAC Database with OIM 9.1.0.2?
    If so, What is the JDBC Driver r u using? Also what App Server (Weblogic or OAS etc) ?
    Appreciate your input.
    Regards
    Vijay Chinnasamy

    We are using weblogic 10.3. And I am sorry to tell you that we are yet to move to production.
    We are planning to have Database in RAC, so i just thought of taking inputs from you.
    can you tell me under what operations you face the poor performance issue?
    For us the max no. of users in production is around 6k. So I am wondering how this RAC availability will affect the performance.
    Thanks for sharing your inputs.

  • Can not start Rac database

    Hi,
    Oracle RAC database 10.2.0.3/RedHat4 with 2 nodes.
    In the begining we had an error ORA-600[12803] so only sys can connect to database I find the note 1026653.6 this note said that we need to create AUDSES$ sequence but befor that we have to restart the database.
    When we stop the datanbase we had another ORA-600 and it's impossible to start it!!
    Here is a coppy of our alert file:
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
    processes = 300
    sessions = 335
    sga_max_size = 524288000
    __shared_pool_size = 310378496
    __large_pool_size = 4194304
    __java_pool_size = 8388608
    __streams_pool_size = 8388608
    spfile = +DATA/osista/spfileosista.ora
    nls_language = FRENCH
    nls_territory = FRANCE
    nls_length_semantics = CHAR
    sga_target = 524288000
    control_files = DATA/osista/controlfile/control01.ctl, DATA/osista/controlfile/control02.ctl
    db_block_size = 8192
    __db_cache_size = 184549376
    compatible = 10.2.0.3.0
    log_archive_dest_1 = LOCATION=USE_DB_RECOVERY_FILE_DEST
    db_file_multiblock_read_count= 16
    cluster_database = TRUE
    cluster_database_instances= 2
    db_create_file_dest = +DATA
    db_recovery_file_dest = +FLASH
    db_recovery_file_dest_size= 68543315968
    thread = 2
    instance_number = 2
    undo_management = AUTO
    undo_tablespace = UNDOTBS2
    undo_retention = 29880
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=OSISTAXDB)
    local_listener = (address=(protocol=tcp)(port=1521)(host=132.147.160.243))
    remote_listener = LISTENERS_OSISTA
    job_queue_processes = 10
    background_dump_dest = /oracle/product/admin/OSISTA/bdump
    user_dump_dest = /oracle/product/admin/OSISTA/udump
    core_dump_dest = /oracle/product/admin/OSISTA/cdump
    audit_file_dest = /oracle/product/admin/OSISTA/adump
    db_name = OSISTA
    open_cursors = 300
    pga_aggregate_target = 104857600
    aq_tm_processes = 1
    Cluster communication is configured to use the following interface(s) for this instance
    172.16.0.2
    Wed Jun 13 11:04:30 2012
    cluster interconnect IPC version:Oracle UDP/IP (generic)
    IPC Vendor 1 proto 2
    PMON started with pid=2, OS id=8560
    DIAG started with pid=3, OS id=8562
    PSP0 started with pid=4, OS id=8566
    LMON started with pid=5, OS id=8570
    LMD0 started with pid=6, OS id=8574
    LMS0 started with pid=7, OS id=8576
    LMS1 started with pid=8, OS id=8580
    MMAN started with pid=9, OS id=8584
    DBW0 started with pid=10, OS id=8586
    LGWR started with pid=11, OS id=8588
    CKPT started with pid=12, OS id=8590
    SMON started with pid=13, OS id=8592
    RECO started with pid=14, OS id=8594
    CJQ0 started with pid=15, OS id=8596
    MMON started with pid=16, OS id=8598
    Wed Jun 13 11:04:31 2012
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=17, OS id=8600
    Wed Jun 13 11:04:31 2012
    starting up 1 shared server(s) ...
    Wed Jun 13 11:04:31 2012
    lmon registered with NM - instance id 2 (internal mem no 1)
    Wed Jun 13 11:04:31 2012
    Reconfiguration started (old inc 0, new inc 2)
    List of nodes:
    1
    Global Resource Directory frozen
    * allocate domain 0, invalid = TRUE
    Communication channels reestablished
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:04:31 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:04:31 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Post SMON to start 1st pass IR
    Wed Jun 13 11:04:31 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:04:31 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:04:31 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    Reconfiguration complete
    LCK0 started with pid=20, OS id=8877
    Wed Jun 13 11:04:43 2012
    alter database mount
    Wed Jun 13 11:04:43 2012
    This instance was first to mount
    Wed Jun 13 11:04:43 2012
    Starting background process ASMB
    ASMB started with pid=25, OS id=10068
    Starting background process RBAL
    RBAL started with pid=26, OS id=10072
    Wed Jun 13 11:04:47 2012
    SUCCESS: diskgroup DATA was mounted
    Wed Jun 13 11:04:51 2012
    Setting recovery target incarnation to 1
    Wed Jun 13 11:04:52 2012
    Successful mount of redo thread 2, with mount id 3005749259
    Wed Jun 13 11:04:52 2012
    Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
    Completed: alter database mount
    Wed Jun 13 11:05:06 2012
    alter database open
    Wed Jun 13 11:05:06 2012
    This instance was first to open
    Wed Jun 13 11:05:06 2012
    Beginning crash recovery of 1 threads
    parallel recovery started with 2 processes
    Wed Jun 13 11:05:07 2012
    Started redo scan
    Wed Jun 13 11:05:07 2012
    Completed redo scan
    61 redo blocks read, 4 data blocks need recovery
    Wed Jun 13 11:05:07 2012
    Started redo application at
    Thread 1: logseq 7924, block 3, scn 506098125
    Wed Jun 13 11:05:07 2012
    Recovery of Online Redo Log: Thread 1 Group 2 Seq 7924 Reading mem 0
    Mem# 0: +DATA/osista/onlinelog/group_2.372.742132543
    Wed Jun 13 11:05:07 2012
    Completed redo application
    Wed Jun 13 11:05:07 2012
    Completed crash recovery at
    Thread 1: logseq 7924, block 64, scn 506118186
    4 data blocks read, 4 data blocks written, 61 redo blocks read
    Switch log for thread 1 to sequence 7925
    Picked broadcast on commit scheme to generate SCNs
    db_recovery_file_dest_size of 65368 MB is 0.61% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Thread 1 advanced to log sequence 7926
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Thread 1 advanced to log sequence 7927
    Wed Jun 13 11:05:11 2012
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=31, OS id=12747
    Wed Jun 13 11:05:11 2012
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=32, OS id=12749
    Wed Jun 13 11:05:12 2012
    Thread 2 opened at log sequence 7176
    Current log# 4 seq# 7176 mem# 0: +DATA/osista/onlinelog/group_4.289.742134597
    Wed Jun 13 11:05:12 2012
    ARC1: Becoming the 'no FAL' ARCH
    ARC1: Becoming the 'no SRL' ARCH
    Wed Jun 13 11:05:12 2012
    Successful open of redo thread 2
    Wed Jun 13 11:05:12 2012
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Jun 13 11:05:12 2012
    ARC0: Becoming the heartbeat ARCH
    Wed Jun 13 11:05:12 2012
    SMON: enabling cache recovery
    Wed Jun 13 11:05:15 2012
    Successfully onlined Undo Tablespace 20.
    Wed Jun 13 11:05:15 2012
    SMON: enabling tx recovery
    Wed Jun 13 11:05:15 2012
    Database Characterset is AL32UTF8
    Wed Jun 13 11:05:16 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_9174.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Wed Jun 13 11:05:16 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_9174.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Error 600 happened during db open, shutting down database
    USER: terminating instance due to error 600
    Instance terminated by USER, pid = 9174
    ORA-1092 signalled during: alter database open...
    Wed Jun 13 11:06:16 2012
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Interface type 1 eth0 172.16.0.0 configured from OCR for use as a cluster interconnect
    Interface type 1 bond0 132.147.160.0 configured from OCR for use as a public interface
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
    processes = 300
    sessions = 335
    sga_max_size = 524288000
    __shared_pool_size = 314572800
    __large_pool_size = 4194304
    __java_pool_size = 8388608
    __streams_pool_size = 8388608
    spfile = +DATA/osista/spfileosista.ora
    nls_language = FRENCH
    nls_territory = FRANCE
    nls_length_semantics = CHAR
    sga_target = 524288000
    control_files = DATA/osista/controlfile/control01.ctl, DATA/osista/controlfile/control02.ctl
    db_block_size = 8192
    __db_cache_size = 180355072
    compatible = 10.2.0.3.0
    log_archive_dest_1 = LOCATION=USE_DB_RECOVERY_FILE_DEST
    db_file_multiblock_read_count= 16
    cluster_database = TRUE
    cluster_database_instances= 2
    db_create_file_dest = +DATA
    db_recovery_file_dest = +FLASH
    db_recovery_file_dest_size= 68543315968
    thread = 2
    instance_number = 2
    undo_management = AUTO
    undo_tablespace = UNDOTBS2
    undo_retention = 29880
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=OSISTAXDB)
    local_listener = (address=(protocol=tcp)(port=1521)(host=132.147.160.243))
    remote_listener = LISTENERS_OSISTA
    job_queue_processes = 10
    background_dump_dest = /oracle/product/admin/OSISTA/bdump
    user_dump_dest = /oracle/product/admin/OSISTA/udump
    core_dump_dest = /oracle/product/admin/OSISTA/cdump
    audit_file_dest = /oracle/product/admin/OSISTA/adump
    db_name = OSISTA
    open_cursors = 300
    pga_aggregate_target = 104857600
    aq_tm_processes = 1
    Cluster communication is configured to use the following interface(s) for this instance
    172.16.0.2
    Wed Jun 13 11:06:16 2012
    cluster interconnect IPC version:Oracle UDP/IP (generic)
    IPC Vendor 1 proto 2
    PMON started with pid=2, OS id=18682
    DIAG started with pid=3, OS id=18684
    PSP0 started with pid=4, OS id=18695
    LMON started with pid=5, OS id=18704
    LMD0 started with pid=6, OS id=18721
    LMS0 started with pid=7, OS id=18735
    LMS1 started with pid=8, OS id=18753
    MMAN started with pid=9, OS id=18767
    DBW0 started with pid=10, OS id=18788
    LGWR started with pid=11, OS id=18796
    CKPT started with pid=12, OS id=18799
    SMON started with pid=13, OS id=18801
    RECO started with pid=14, OS id=18803
    CJQ0 started with pid=15, OS id=18805
    MMON started with pid=16, OS id=18807
    Wed Jun 13 11:06:17 2012
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=17, OS id=18809
    Wed Jun 13 11:06:17 2012
    starting up 1 shared server(s) ...
    Wed Jun 13 11:06:17 2012
    lmon registered with NM - instance id 2 (internal mem no 1)
    Wed Jun 13 11:06:17 2012
    Reconfiguration started (old inc 0, new inc 2)
    List of nodes:
    1
    Global Resource Directory frozen
    * allocate domain 0, invalid = TRUE
    Communication channels reestablished
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Post SMON to start 1st pass IR
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    Reconfiguration complete
    LCK0 started with pid=20, OS id=18816
    Wed Jun 13 11:06:18 2012
    ALTER DATABASE MOUNT
    Wed Jun 13 11:06:18 2012
    This instance was first to mount
    Wed Jun 13 11:06:18 2012
    Reconfiguration started (old inc 2, new inc 4)
    List of nodes:
    0 1
    Wed Jun 13 11:06:18 2012
    Starting background process ASMB
    Wed Jun 13 11:06:18 2012
    Global Resource Directory frozen
    Communication channels reestablished
    ASMB started with pid=22, OS id=18913
    Starting background process RBAL
    * domain 0 valid = 0 according to instance 0
    Wed Jun 13 11:06:18 2012
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    RBAL started with pid=23, OS id=18917
    Reconfiguration complete
    Wed Jun 13 11:06:22 2012
    SUCCESS: diskgroup DATA was mounted
    Wed Jun 13 11:06:26 2012
    Setting recovery target incarnation to 1
    Wed Jun 13 11:06:26 2012
    Successful mount of redo thread 2, with mount id 3005703530
    Wed Jun 13 11:06:26 2012
    Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
    Completed: ALTER DATABASE MOUNT
    Wed Jun 13 11:06:27 2012
    ALTER DATABASE OPEN
    This instance was first to open
    Wed Jun 13 11:06:27 2012
    Beginning crash recovery of 1 threads
    parallel recovery started with 2 processes
    Wed Jun 13 11:06:27 2012
    Started redo scan
    Wed Jun 13 11:06:27 2012
    Completed redo scan
    61 redo blocks read, 4 data blocks need recovery
    Wed Jun 13 11:06:28 2012
    Started redo application at
    Thread 2: logseq 7176, block 3
    Wed Jun 13 11:06:28 2012
    Recovery of Online Redo Log: Thread 2 Group 4 Seq 7176 Reading mem 0
    Mem# 0: +DATA/osista/onlinelog/group_4.289.742134597
    Wed Jun 13 11:06:28 2012
    Completed redo application
    Wed Jun 13 11:06:28 2012
    Completed crash recovery at
    Thread 2: logseq 7176, block 64, scn 506138248
    4 data blocks read, 4 data blocks written, 61 redo blocks read
    Picked broadcast on commit scheme to generate SCNs
    Wed Jun 13 11:06:28 2012
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=28, OS id=19692
    Wed Jun 13 11:06:28 2012
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=29, OS id=19695
    Wed Jun 13 11:06:28 2012
    Thread 2 advanced to log sequence 7177
    Thread 2 opened at log sequence 7177
    Current log# 3 seq# 7177 mem# 0: +DATA/osista/onlinelog/group_3.291.742134597
    Successful open of redo thread 2
    Wed Jun 13 11:06:28 2012
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Jun 13 11:06:28 2012
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    Wed Jun 13 11:06:28 2012
    ARC1: Becoming the heartbeat ARCH
    Wed Jun 13 11:06:28 2012
    SMON: enabling cache recovery
    Wed Jun 13 11:06:28 2012
    db_recovery_file_dest_size of 65368 MB is 0.61% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Wed Jun 13 11:06:31 2012
    Successfully onlined Undo Tablespace 20.
    Wed Jun 13 11:06:31 2012
    SMON: enabling tx recovery
    Wed Jun 13 11:06:31 2012
    Database Characterset is AL32UTF8
    Wed Jun 13 11:06:31 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_19596.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Wed Jun 13 11:06:32 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_19596.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Error 600 happened during db open, shutting down database
    USER: terminating instance due to error 600
    Instance terminated by USER, pid = 19596
    ORA-1092 signalled during: ALTER DATABASE OPEN...
    Wed Jun 13 11:11:35 2012
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Interface type 1 eth0 172.16.0.0 configured from OCR for use as a cluster interconnect
    Interface type 1 bond0 132.147.160.0 configured from OCR for use as a public interface
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
    processes = 300
    sessions = 335
    sga_max_size = 524288000
    __shared_pool_size = 318767104
    __large_pool_size = 4194304
    __java_pool_size = 8388608
    __streams_pool_size = 8388608
    spfile = +DATA/osista/spfileosista.ora
    nls_language = FRENCH
    nls_territory = FRANCE
    nls_length_semantics = CHAR
    sga_target = 524288000
    control_files = DATA/osista/controlfile/control01.ctl, DATA/osista/controlfile/control02.ctl
    db_block_size = 8192
    __db_cache_size = 176160768
    compatible = 10.2.0.3.0
    log_archive_dest_1 = LOCATION=USE_DB_RECOVERY_FILE_DEST
    db_file_multiblock_read_count= 16
    cluster_database = TRUE
    cluster_database_instances= 2
    db_create_file_dest = +DATA
    db_recovery_file_dest = +FLASH
    db_recovery_file_dest_size= 68543315968
    thread = 2
    instance_number = 2
    undo_management = AUTO
    undo_tablespace = UNDOTBS2
    undo_retention = 29880
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=OSISTAXDB)
    local_listener = (address=(protocol=tcp)(port=1521)(host=132.147.160.243))
    remote_listener = LISTENERS_OSISTA
    job_queue_processes = 10
    background_dump_dest = /oracle/product/admin/OSISTA/bdump
    user_dump_dest = /oracle/product/admin/OSISTA/udump
    core_dump_dest = /oracle/product/admin/OSISTA/cdump
    audit_file_dest = /oracle/product/admin/OSISTA/adump
    db_name = OSISTA
    open_cursors = 300
    pga_aggregate_target = 104857600
    aq_tm_processes = 1
    Cluster communication is configured to use the following interface(s) for this instance
    172.16.0.2
    Wed Jun 13 11:11:35 2012
    cluster interconnect IPC version:Oracle UDP/IP (generic)
    IPC Vendor 1 proto 2
    PMON started with pid=2, OS id=16101
    DIAG started with pid=3, OS id=16103
    PSP0 started with pid=4, OS id=16105
    LMON started with pid=5, OS id=16107
    LMD0 started with pid=6, OS id=16110
    LMS0 started with pid=7, OS id=16112
    LMS1 started with pid=8, OS id=16116
    MMAN started with pid=9, OS id=16120
    DBW0 started with pid=10, OS id=16132
    LGWR started with pid=11, OS id=16148
    CKPT started with pid=12, OS id=16169
    SMON started with pid=13, OS id=16185
    RECO started with pid=14, OS id=16203
    CJQ0 started with pid=15, OS id=16219
    MMON started with pid=16, OS id=16227
    Wed Jun 13 11:11:36 2012
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=17, OS id=16229
    Wed Jun 13 11:11:36 2012
    starting up 1 shared server(s) ...
    Wed Jun 13 11:11:36 2012
    lmon registered with NM - instance id 2 (internal mem no 1)
    Wed Jun 13 11:11:36 2012
    Reconfiguration started (old inc 0, new inc 2)
    List of nodes:
    1
    Global Resource Directory frozen
    * allocate domain 0, invalid = TRUE
    Communication channels reestablished
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:11:36 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:11:36 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Post SMON to start 1st pass IR
    Wed Jun 13 11:11:36 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:11:36 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:11:36 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    Reconfiguration complete
    LCK0 started with pid=20, OS id=16235
    Wed Jun 13 11:11:37 2012
    ALTER DATABASE MOUNT
    Wed Jun 13 11:11:37 2012
    This instance was first to mount
    Wed Jun 13 11:11:37 2012
    Starting background process ASMB
    ASMB started with pid=22, OS id=16343
    Starting background process RBAL
    RBAL started with pid=23, OS id=16347
    Wed Jun 13 11:11:44 2012
    SUCCESS: diskgroup DATA was mounted
    Wed Jun 13 11:11:49 2012
    Setting recovery target incarnation to 1
    Wed Jun 13 11:11:49 2012
    Successful mount of redo thread 2, with mount id 3005745065
    Wed Jun 13 11:11:49 2012
    Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
    Completed: ALTER DATABASE MOUNT
    Wed Jun 13 11:22:25 2012
    alter database open
    This instance was first to open
    Wed Jun 13 11:22:26 2012
    Beginning crash recovery of 1 threads
    parallel recovery started with 2 processes
    Wed Jun 13 11:22:26 2012
    Started redo scan
    Wed Jun 13 11:22:26 2012
    Completed redo scan
    61 redo blocks read, 4 data blocks need recovery
    Wed Jun 13 11:22:26 2012
    Started redo application at
    Thread 1: logseq 7927, block 3
    Wed Jun 13 11:22:26 2012
    Recovery of Online Redo Log: Thread 1 Group 1 Seq 7927 Reading mem 0
    Mem# 0: +DATA/osista/onlinelog/group_1.283.742132543
    Wed Jun 13 11:22:26 2012
    Completed redo application
    Wed Jun 13 11:22:26 2012
    Completed crash recovery at
    Thread 1: logseq 7927, block 64, scn 506178382
    4 data blocks read, 4 data blocks written, 61 redo blocks read
    Switch log for thread 1 to sequence 7928
    Picked broadcast on commit scheme to generate SCNs
    Wed Jun 13 11:22:27 2012
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=31, OS id=13010
    Wed Jun 13 11:22:27 2012
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=32, OS id=13033
    Wed Jun 13 11:22:27 2012
    Thread 2 opened at log sequence 7178
    Current log# 4 seq# 7178 mem# 0: +DATA/osista/onlinelog/group_4.289.742134597
    Successful open of redo thread 2
    Wed Jun 13 11:22:27 2012
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Jun 13 11:22:27 2012
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    Wed Jun 13 11:22:27 2012
    ARC1: Becoming the heartbeat ARCH
    Wed Jun 13 11:22:27 2012
    SMON: enabling cache recovery
    Wed Jun 13 11:22:30 2012
    db_recovery_file_dest_size of 65368 MB is 0.61% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Wed Jun 13 11:22:31 2012
    Successfully onlined Undo Tablespace 20.
    Wed Jun 13 11:22:31 2012
    SMON: enabling tx recovery
    Wed Jun 13 11:22:32 2012
    Database Characterset is AL32UTF8
    Wed Jun 13 11:22:32 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_11751.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Wed Jun 13 11:22:33 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_11751.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Error 600 happened during db open, shutting down database
    USER: terminating instance due to error 600
    Instance terminated by USER, pid = 11751
    ORA-1092 signalled during: alter database open...
    regards,

    Hi;
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_9174.trc:Did you check trc file?
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []You are getting oracle internal error(ORA 600) which mean you could need to work wiht oracle support team. Please see below note, if its not help than i suggest log a sr:
    Troubleshoot an ORA-600 or ORA-7445 Error Using the Error Lookup Tool [ID 153788.1]
    for your future rac issue please use Forum Home » High Availability » RAC, ASM & Clusterware Installation which is RAC dedicated forum site.
    Regard
    Helios

Maybe you are looking for