Arivelog mode configuration proble in RAC database

Hi,
I'm trying to configure RAC database in archivelog mode but when try to do so get the following error, during the database startup in mount state
LRM-00109: could not open parameter file '/u01/app/oracle/product/11.1.0/db_1/dbs/initorcl.ora'
This error shows that the system could not find the initorcl.ora and really it does not exist. But the system at this stage is single
instance as I chenged the value of cluster_database=false before shutdown and then start.
when I enter the command startup mount it give me this error. What should I do to avoid it. Should I make a pfile of this from spfile in cluster mode
and then put this pfile in the desired path to start the single instance or any other thing should I tried.
Regards,
Abbasi

Perform the following steps to relocate the SPFILE from the local file system to an ASM disk group.
    * Create a text-based initialization parameter file from the current binary SPFILE located on the local file system:
      SQL> CREATE PFILE='$ORACLE_HOME/dbs/initTESTDB.ora'
        2  FROM SPFILE='$ORACLE_HOME/dbs/spfileTESTDB.ora';
      File created.
    * Create new SPFILE in an ASM disk group:
      SQL> CREATE SPFILE='+TESTDB_DATA1/TESTDB/spfileTESTDB.ora'
        2  FROM PFILE='$ORACLE_HOME/dbs/initTESTDB.ora';
      File created.
    * Shutdown the Oracle database:
      SQL> SHUTDOWN IMMEDIATE
      Database closed.
      Database dismounted.
      ORACLE instance shut down.
    * Update the text-based init<SID>.ora file with the new location of the SPFILE in ASM:
      $ echo "SPFILE='+TESTDB_DATA1/TESTDB/spfileTESTDB.ora'" > $ORACLE_HOME/dbs/initTESTDB.ora
    * Remove (actually rename) the old SPFILE on the local file system so that the new text-based init<SID>.ora will be used:
      $ mv $ORACLE_HOME/dbs/spfileTESTDB.ora $ORACLE_HOME/dbs/BACKUP_ASM.spfileTESTDB.ora
    * Open the Oracle database using the new SPFILE:
      SQL> STARTUPsource:-http://www.idevelopment.info/data/Oracle/DBA_tips/Automatic_Storage_Management/ASM_33.shtml#Migrating%20Oracle%20Database%20from%20Local%20File%20System%20to%20ASM
if you would like, refer the link:-
http://oracleinstance.blogspot.com/2009/12/enable-archivelog-and-flashback-in-rac.html

Similar Messages

  • NETCA listener configuration error for RAC database

    Hi,
    WE have a RAC setup on 2 nodes and now we want to change the default listener port for RAC database.
    Database ver: 10.2.0.4
    We now that srvctl does not work to add listener manually to CRS.
    We are now using netca to create new listener on both the RAC nodes but it gives error throwing :
    Problem in configuration : Node(S): hostname
    Try again...
    When we try to start the listener say listener_I it throws following error:
    Problem in configuration :node1
    Listener start to failed:Listener may already be running
    CRS-0215:Could not start resource
    This is urgent
    any help is appreciated.
    Regards,
    Milan

    user12073248 wrote:
    How you fix it? write to [email protected]
    Hi Mauricio,
    In this forum we help for free regardless of priority.
    First you try to help in the forum if the author of the post request your e-mail you give to him.
    Hope you understand it.
    Thanks in advance for doing your part to make this community as valuable as possible for everyone!
    Levi Pereira

  • RMAN configuration in ORACLE RAC database

    Hi there,
    We are using RMAN backup strategy for taking backups from one node(eg.node1).Database is RAC database, version 9.2.3. and OS is Solaris 5.10.
    this is the configuration which we use, as follows:
    Connecting to Prod.Database using catalog database.
    rman catalog=rman/rman@CATDB target=sys/oracle@PRODDB(node 2).
    RMAN> show all;
    RMAN configuration parameters are:
    CONFIGURE RETENTION POLICY TO REDUNDANCY 30;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '/rmanback/_%F';
    CONFIGURE DEVICE TYPE DISK PARALLELISM 5;
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE DISK FORMAT '/rmanback/CMWPROD_%s.bak';
    CONFIGURE CHANNEL 2 DEVICE TYPE DISK CONNECT 'SYS/oracle@PRODDB(node2)';
    CONFIGURE MAXSETSIZE TO 7 G;
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '/oracle/9.2.3/dbs/snapcf_CMWPROD1.f';
    As the above config shows RMAN is connecting to PROD DB (node 2). But we need to take backups (datafiles and as well as archive logs) from two nodes (i.e from node 1 and as well as node 2).
    Is there any possibility for taking backups from two nodes?
    Thanks,
    Balu.

    You will need to setup an NFS share such that both nodes can access each other's archived redo logs area. You also need to ensure your LOG_ARCHIVE_DEST_n parameter settings are properly set. Once setup correctly, either node will be able to run the backup and include each other's archived redo log files. Since this is RAC, the data files are shared by design and either node can already backup these files.
    Multiple channels would allow 'simultaneous' backup of the archived redo logs.
    I would highly recommend using a CFS (single shared file system where all nodes store their archived redo logs distinguished by thread) to store the archived redo logs, this reduces administration and configuration especially when you get ready to scale.
    Hope this helps.

  • Archive log mode  in 3 node rac database

    Sir I am using oracle 10.2.0.1 and created 3 node cluster setup on.OS i am using linux 4
    I am too much confuse in setting up this cluster in archive log mode bec why we need to false cluater_database parameter to setup cluster in archive log mode.
    I have search lot of documents but all are saying we not need setting this parameter flase perior to 10.2.But what is the actual concept behind setting this parameter to false in early release and why we need not to set it false.Please help me.
    I know how to set up in archive log mode but this parameter create confusion.
    Thanks sir in advance

    I also dont't know all the details about this but I'll try to explain what I know. Setting cluster_database to false in a rac environment is usually done when you need to mount the database from one instance exclusively. This is still needed when you upgrade the catalog, for example during release update (from 10.2.0.1 to 10.2.0.4 for example). Now with switching archivelog mode someone at oracle must have discovered that it is sufficient when the database is mounted by only one instance and you can actually save one step.
    As I (and propably most of you) don't switch archiving on or off on a daily basis, I don't worry about this a lot, simply turn this on when I create a new database and then leave it on forever.
    Bjoern

  • Enable Archivelog Mode On 2 Node RAC DATABASE

    hi,
    We are having 2 Node Rac 10g Database Installed on Windows.
    Currently the Database is in Non Archivelog Mode.
    I want to convert it to Archivelog Mode.
    I come to know the steps to convert cluster database to Archivelog mode.
    i.e.
    set archive parameters like Disk Location and Archivelog File format.
    then Convert the database in Exclusive Mode
    ALTER SYSTEM SET cluster_database=FALSE SCOPE=spfile;
    then
    C:\> srvctl stop database -d <Global_DB_SID>
    then enable Archivelog Mode.
    SQL> STARTUP MOUNT;
    SQL> ARCHIVE LOG START;
    SQL> ALTER DATABASE ARCHIVELOG;
    SQL> ALTER DATABASE SET cluster_database=TRUE SCOPE=spfile;
    SQL> SHUTDOWN IMMEDIATE
    C:\> srvctl start database -d <Global_DB_SID>
    But I am confuse with Do i need to RUN this commands in both the nodes seperatly or only on any one node.
    Please Reply,

    comments embeddedf
    Node 1: OSERP1
    Log on to OSERP1 Server
    C:\> sqlplus sys/oracle@racdb2 as sysdba
    SQL> ALTER SYSTEM SET log_archive_format='arch_%t_%s.arc' SCOPE=spfile ; -- this needs to be database wide
    SQL > ALTER SYSTEM SET log_archive_dest_1='location=\\OSERP1\archive$\ERP1 ' SCOPE=spfile sid ='ERP1'; -- required quotes added
    -- not required
    SQL> ALTER SYSTEM SET cluster_database=FALSE SCOPE=spfile;
    not required to connect to other node
    -- not required
    SQL> ALTER SYSTEM SET log_archive_format='arch_%t_%s.arc' SCOPE=spfile sid =ERP2;
    SQL > ALTER SYSTEM SET log_archive_dest_1='location=\\OSERP2\archive$\ERP2 ' SCOPE=spfile sid ='ERP2'; -- quotes added
    -- not required, and database wide setting
    SQL> ALTER SYSTEM SET cluster_database=FALSE SCOPE=spfile;
    Then following commands will be run connecting any of the instance
    -- database has to be stopped first, if you stick to switching it to a non-cluster database
    C:\> set ORACLE_SID=<local_instance_id>
    C:\> sqlplus / as sysdba
    SQL> STARTUP MOUNT;
    -- wrong sequence of commands
    SQL> ARCHIVE LOG START;
    SQL> ALTER DATABASE ARCHIVELOG;
    -- not required (see previous steps)
    SQL> ALTER DATABASE SET cluster_database=TRUE SCOPE=spfile;
    SQL> SHUTDOWN IMMEDIATE;
    6. Start the cluster database.
    C:\> srvctl start database -d <Global_DB_SID>
    -- There is no such thing as a <Global_DB_SID>. There is a database name.
    Please note your tnsnames.ora, in a standard install, needs to have the following service names.
    ERP
    ERP1
    ERP2
    I'm not sure why you have the non-standard service_name racdb2.
    Somehow it looks like you both don't understand the RAC concept, and didn't understand my answer. I had to correct several things I already corrected in my previous response. I'm not sure why I'm not getting through.
    Sybrand Bakker
    Senior Oracle DBA

  • RMAN backup in ORACLE RAC database

    Hi there,
    We are using RMAN backup strategy for taking backups from one node(eg.node1).Database is RAC database, version 9.2.3. and OS is Solaris 5.10.
    this is the configuration which we use, as follows:
    Connecting to Prod.Database using catalog database.
    rman catalog=rman/rman@CATDB target=sys/oracle@PRODDB(node 2).
    RMAN> show all;
    RMAN configuration parameters are:
    CONFIGURE RETENTION POLICY TO REDUNDANCY 30;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '/rmanback/_%F';
    CONFIGURE DEVICE TYPE DISK PARALLELISM 5;
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE CHANNEL DEVICE TYPE DISK FORMAT '/rmanback/CMWPROD_%s.bak';
    CONFIGURE CHANNEL 2 DEVICE TYPE DISK CONNECT 'SYS/oracle@PRODDB(node2)';
    CONFIGURE MAXSETSIZE TO 7 G;
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '/oracle/9.2.3/dbs/snapcf_CMWPROD1.f';
    As the above config shows RMAN is connecting to PROD DB (node 2). But we need to take backups (datafiles and as well as archive logs) from two nodes (i.e from node 1 and as well as node 2).
    Is there any possibility for taking backups from two nodes?
    Thanks,
    Balu.

    Duplicate thread
    RMAN configuration in ORACLE RAC database
    Khurram                                                                                                                                                                                                                               

  • No response from EM on 2 node RAC database 10.2.0.3

    Hi
    I have configured 2 node RAC database , version 10.2.0.3 using ASM and CRS. The database and all services are running fine . And successfully completed the EM configuration on the default ports .
    But when I am trying to login into the EM grid control , login page doesnt go to the Welcome page instead the username and password fields get blank . I checked the status for EM and its running fine on both nodes .
    See the output from the EMCA tool :
    INFO:
    **************** Current Configuration ****************
    INSTANCE NODE DBCONTROL_UPLOAD_HOST
    SITB1 ccman_rac1 ccman_rac2
    SITB2 ccman_rac2 ccman_rac2
    Getting these messages in emoms.log:
    2007-10-25 14:51:57,597 [JobWorker 87:Thread-14] ERROR jobs.pingCfmEMDDown pingHost.439 - TimeoutException caught pinging the
    host for EMD URL http://ccman_rac1:3938/emd/main
    2007-10-25 14:55:44,081 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.ias.ias.IASIntegration
    2007-10-25 14:55:44,422 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.eml.target.slb.common.SLBIntegration
    2007-10-25 14:55:44,531 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.eml.ssl.intg.SSLIntegration
    2007-10-25 14:55:44,682 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.tip.oem.central.domain.ProcessConnectDomainIntg
    2007-10-25 14:55:44,685 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.tip.oem.central.instance.ProcessConnectInstanceIntg
    2007-10-25 14:55:44,688 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.webdb.admin.em.PortalIntegration
    2007-10-25 14:55:44,691 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.webdb.admin.em.SSOIntegration
    2007-10-25 14:55:44,694 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.reports.em.RepIntg
    2007-10-25 14:55:44,717 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.ocs.mntr.target.OcsEmailIntegration
    2007-10-25 14:55:44,720 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.ocs.mntr.target.OcsOidIntegration
    2007-10-25 14:55:44,724 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.ocs.mntr.target.OcsOvfIntegration
    2007-10-25 14:55:44,727 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.ocs.mntr.target.OcsWebconfIntegration
    2007-10-25 14:55:44,729 [Orion Launcher] ERROR app.ContextInitializer contextInitialized.272 - Integration Class not found: o
    racle.sysman.ocs.mntr.target.OcsWirelessIntg
    Please suggest how to go ahead to resolve this issue.
    Thanks in advance.
    Ankur

    Is this something to do with fully qualified host names ? I've been caught with EM by that before.

  • Data Guard configuration for RAC database disappeared from Grid control

    Primary Database Environment - Three node cluster
    RAC Database 10.2.0.1.0
    Linux Red Hat 4.0 2.6.9-22 64bit
    ASM 10.2.0.1.0
    Management Agent 10.2.0.2.0
    Standby Database Environment - one Node database
    Oracle Enterprise Edition 10.2.0.1.0 Single standby
    Linux Red Hat 4.0 2.6.9-22 64bit
    ASM 10.2.0.1.0
    Management Agent 10.2.0.2.0
    Grid Control 10.2.0.1.0 - Node separate from standby and cluster environments
    Oracle 10.1.0.1.0
    Grid Control 10.2.0.1.0
    Red Hat 4.0 2.6.9-22 32bit
    After adding a logical standby database through Grid Control for a RAC database, I noticed sometime later the Data Guard configuration disappeared from Grid Control. Not sure why but it is gone. I did notice that something went wrong with the standby creation but i did not get much feedback from Grid Control. The last thing I did was to view the configuration, see output below.
    Initializing
    Connected to instance qdcls0427:ELCDV3
    Starting alert log monitor...
    Updating Data Guard link on database homepage...
    Data Protection Settings:
    Protection mode : Maximum Performance
    Log Transport Mode settings:
    ELCDV.qdx.com: ARCH
    ELXDV: ARCH
    Checking standby redo log files.....OK
    Checking Data Guard status
    ELCDV.qdx.com : ORA-16809: multiple warnings detected for the database
    ELXDV : Creation status unknown
    Checking Inconsistent Properties
    Checking agent status
    ELCDV.qdx.com
    qdcls0387.qdx.com ... OK
    qdcls0388.qdx.com ... OK
    qdcls0427.qdx.com ... OK
    ELXDV ... WARNING: No credentials available for target ELXDV
    Attempting agent ping ... OK
    Switching log file 672.Done
    WARNING: Skipping check for applied log on ELXDV : disabled
    Processing completed.
    Here are the steps followed to add the standby database in Grid Control
    Maintenance tab
    Setup and Manage Data Guard
    Logged in as sys
    Add standby database
    Create a new logical standby database
    Perform a live backup of the primary database
    Specify backup directory for staging area
    Specify standby database name and Oracle home location
    Specify file location staging area on standby node
    At the end am presented with a review of the selected options and then the standby database is created
    Has any body come across a similar issue?
    Thanks,

    Any resolution on this?
    I just created a Logical Standby database and I'm getting the same warning (WARNING: No credentials available for target ...) when I do a 'Verify Configuration' from the Data Guard page.
    Everything else seems to be working fine. Logs are being applied, etc.
    I can't figure out what credentials its looking for.

  • Configure SOA/BAM with RAC database

    Hi,
    We've configured SOA Cluster 11.1.1.3 with RAC database 11.1.7. For thiis we've create RAC multi datasources and the JDBC details of these datasources are as under:
    Datasource name: SOA-RAC-1
    =====================
    jdbc:oracle:thin:@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=host-1-vip.domain.com)(PORT=15715)))(CONNECT_DATA=(SERVICE_NAME=SID.DOMAIN.COM)(INSTANCE_NAME=SID1)))
    Datasource Name: SOA-RAC-2
    jdbc:oracle:thin:@(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=host-2-vip.domain.com)(PORT=15715)))(CONNECT_DATA=(SERVICE_NAME=SID.DOMAIN.COM)(INSTANCE_NAME=SID2)))
    Multi Datasource=rac1 & rac-2
    Everything is up and running, but when we are testing it by stopping one of the ndoe fo RAC (Say Node-1) and starting the managed server we are getting following error:
    <Nov 15, 2010 2:22:05 AM CST> <Error> <Deployer> <BEA-149205> <Failed to initialize the application 'leasing-rac0-jdbc' due to error weblogic.application.ModuleException: .
    weblogic.application.ModuleException:
    at weblogic.jdbc.module.JDBCModule.prepare(JDBCModule.java:290)
    at weblogic.application.internal.flow.ModuleListenerInvoker.prepare(ModuleListenerInvoker.java:199)
    at weblogic.application.internal.flow.DeploymentCallbackFlow$1.next(DeploymentCallbackFlow.java:508)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:41)
    at weblogic.application.internal.flow.DeploymentCallbackFlow.prepare(DeploymentCallbackFlow.java:149)
    Truncated. see log file for complete stacktrace
    ===========================================
    <Nov 15, 2010 2:22:31 AM CST> <Warning> <JDBC> <BEA-001129> <Received exception while creating connection for pool "OraSDPMDataSource-rac0": Listener refused the connection with the following error:
    ORA-12521, TNS:listener does not currently know of instance requested in connect descriptor
    >
    <Nov 15, 2010 2:22:33 AM CST> <Warning> <JDBC> <BEA-001129> <Received exception while creating connection for pool "BAMDataSource-rac0": Listener refused the connection with the following error:
    ORA-12521, TNS:listener does not currently know of instance requested in connect descriptor
    >
    <Nov 15, 2010 2:22:36 AM CST> <Warning> <JDBC> <BEA-001129> <Received exception while creating connection for pool "BAMDataSource-rac0": Listener refused the connection with the following error:
    ORA-12521, TNS:listener does not currently know of instance requested in connect descriptor
    >
    <Nov 15, 2010 2:22:36 AM CST> <Warning> <JDBC> <BEA-001129> <Received exception while creating connection for pool "OraSDPMDataSource-rac0": Listener refused the connection with the following error:
    ORA-12521, TNS:listener does not currently know of instance requested in connect descriptor
    >
    <Nov 15, 2010 2:22:37 AM CST> <Warning> <JNDI> <BEA-050007> <An attempt was made to look up non-versioned global resource "topic.oracle.bam.messaging" from an application version "oracle-bam [Version=11.1.1]". This can potentially cause conflict of the global resource usages among multiple application versions.>
    <Nov 15, 2010 2:22:38 AM CST> <Notice> <Log Management> <BEA-170027> <The Server has established connection with the Domain level Diagnostic Service successfully.>
    <Nov 15, 2010 2:22:38 AM CST> <Warning> <JDBC> <BEA-001129> <Received exception while creating connection for pool "mds-owsm-rac0": Listener refused the connection with the following error:
    ORA-12521, TNS:listener does not currently know of instance requested in connect descriptor
    >
    And managed server goes to admin mode. Its with both SOA and BAM servers.
    Please let me know if someone has come across this error or is there anything I am missing while configuring the multi datasource.
    Thanks,
    Suraj Sharma

    Hi Suraj,
    even i am facing the same issue, were you able to resolve this issue?
    Thanks,
    Deepa

  • Can we put RAC database in Archivelog mode without shutting down

    All,
    Can we put RAC database in Archivelog mode without shutting down.
    Currently our new production database (2 node RAC) is in no archive log mode, Need to enable archive log in the database...
    I believe we need to set the cluster_database=false and then put the DB in archive log mode then we need to bounce the database to take effect...
    Just curious to know in 11gR2 ...Can we put the RAC database in archive log mode without any downtime ...?

    Even RAC or non-RAC, database should bounced and enable/disable archive log mode from mount status.

  • RAC database in archivelog mode ?

    Hi all,
    may anyone provide me doc id in matalink to move a non archived RAC database in
    to archive log mode.
    THanx in advance.
    Deep

    Even RAC or non-RAC, database should bounced and enable/disable archive log mode from mount status.

  • 10G RAC Database Configuration Querry??

    Hi All,
    I have installed the 10GR3 Grid and deployed the Agent to a system which has 10GR2 RAC Database with CRS installed.
    1) When i try to configure the cluster database, should we use the VIP TNS entries as the Connect string or the Physical Node TNS Address.
    2) After i have configured it i do not see Listener shown for either monitoring nor i am able to add manually.
    3) Will any CRS components be monitored through Grid? For Me i dont find any though the CRS HOME oratab entry is enabled.
    Please share your ideas and experience.. on this..
    Regards,
    Paravasu Dev

    This is stated in the EM documentation. <br>
    Oracle Database FAQs
    </br>

  • Open 10g/11g RAC database in Non RAC Mode

    Dear Friends,
    How i open 10g and 11g RAC database in Non RAC Mode (open single instance in order to perform upgrade)
    Later i will convert the upgraded DB to RAC DB
    Regards,
    DB

    user564706 wrote:
    Dear Friends,
    How i open 10g and 11g RAC database in Non RAC Mode (open single instance in order to perform upgrade)
    Later i will convert the upgraded DB to RAC DB
    Regards,
    DBEdit spfile and set cluster_database=false before starting database in upgrade mode(optionally you can unset other cluster parameters).
    So you upgrade using only 1 instance of database.
    Once the upgrade is finished start the other instance as well and set cluster_database=true
    Also see preupgrade process - http://docs.oracle.com/cd/E11882_01/server.112/e23633/upgrade.htm#i1011979
    Post upgrade -> http://docs.oracle.com/cd/E11882_01/server.112/e23633/afterup.htm#CEGCFJJC

  • Configured TNS in oracle client to RAC database

    Hi Experts,
    I have a 4 nodes oracle RAC in radhat 5.0
    both oracle database and client as 10.2.0.4.
    I created a TNS that works in our office but it does not work in our remote brance office.
    Based on experts instruction, I set up sqlnet trace at client side. but not any file that be gererated.
    In linux RAC database, I saw listener.log file as
    28-AUG-2009 11:00:01 * (CONNECT_DATA=(SERVER=DEDICATED)(SERVICE_NAME=sale)(INSTANCE_NAME=sale1)(CID=(PR OGRAM=D:\oracle\product\10.2.0\client_1\BIN\sqlplusw.exe)(HOST=TRAC3)(USER=JIM))) * (ADDRESS=(PROTOCOL=tcp)( HOST=xxx.18.10.xx)(PORT=3289)) * establish * sale * 0
    in client pc sqlnet.ora my setting as
    SQLNET.AUTHENTICATION_SERVICES= (NTS)
    TRACE_LEVEL_CLIENT = 16
    TRACE_FILE_CLIENT = client
    TRACE_DIRECTORY_CLIENT = C:\TEMPODBC
    TRACE_TIMESTAMP_CLIENT = ON
    SQLNET.EXPIRE_TIME =10
    SQLNET.INBOUND_CONNECT_TIMEOUT = 500
    SQLNET.SEND_TIMEOUT = 500
    SQLNET.RECV_TIMEOUT = 500
    I got a error as
    ERROR:
    ORA-03135: connection lost contact
    which wrong is in my case?
    I can connect to all other window database from thsi PC.But I can not connect to RAC database from this PC.
    However, I can ping RAC node's IP and VIP from this PC
    I also can connect to RAC database with TNS from other window PC.
    What do I need to do?
    ThanKs for any help.
    JIM

    Hi Sb92075,
    Thanks for your help.
    I got same error after comment #SQLNET.AUTHENTICATION_SERVICES= (NTS) in client sqlnet.ora in client.
    the server side does not have sqlnet file in linux platform.
    I also check listener services that all instance status is ready.
    I just saw sqlnet.log file in server side that error as
    Fatal NI connect error 12170.
    VERSION INFORMATION:
    TNS for Linux: Version 10.2.0.4.0 - Production
    Oracle Bequeath NT Protocol Adapter for Linux: Version 10.2.0.4.0 - Production
    TCP/IP NT Protocol Adapter for Linux: Version 10.2.0.4.0 - Production
    Time: 28-AUG-2009 13:21:06
    Tracing not turned on.
    Tns error struct:
    ns main err code: 12535
    TNS-12535: TNS:operation timed out
    ns secondary err code: 12606
    nt main err code: 0
    nt secondary err code: 0
    nt OS err code: 0
    Client address: (ADDRESS=(PROTOCOL=tcp)(HOST=xxx.18.xx.xx)(PORT=3513))
    what do i need to do?
    Thanks
    JiM
    Edited by: user589812 on Aug 28, 2009 10:22 AM

  • Can not start Rac database

    Hi,
    Oracle RAC database 10.2.0.3/RedHat4 with 2 nodes.
    In the begining we had an error ORA-600[12803] so only sys can connect to database I find the note 1026653.6 this note said that we need to create AUDSES$ sequence but befor that we have to restart the database.
    When we stop the datanbase we had another ORA-600 and it's impossible to start it!!
    Here is a coppy of our alert file:
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
    processes = 300
    sessions = 335
    sga_max_size = 524288000
    __shared_pool_size = 310378496
    __large_pool_size = 4194304
    __java_pool_size = 8388608
    __streams_pool_size = 8388608
    spfile = +DATA/osista/spfileosista.ora
    nls_language = FRENCH
    nls_territory = FRANCE
    nls_length_semantics = CHAR
    sga_target = 524288000
    control_files = DATA/osista/controlfile/control01.ctl, DATA/osista/controlfile/control02.ctl
    db_block_size = 8192
    __db_cache_size = 184549376
    compatible = 10.2.0.3.0
    log_archive_dest_1 = LOCATION=USE_DB_RECOVERY_FILE_DEST
    db_file_multiblock_read_count= 16
    cluster_database = TRUE
    cluster_database_instances= 2
    db_create_file_dest = +DATA
    db_recovery_file_dest = +FLASH
    db_recovery_file_dest_size= 68543315968
    thread = 2
    instance_number = 2
    undo_management = AUTO
    undo_tablespace = UNDOTBS2
    undo_retention = 29880
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=OSISTAXDB)
    local_listener = (address=(protocol=tcp)(port=1521)(host=132.147.160.243))
    remote_listener = LISTENERS_OSISTA
    job_queue_processes = 10
    background_dump_dest = /oracle/product/admin/OSISTA/bdump
    user_dump_dest = /oracle/product/admin/OSISTA/udump
    core_dump_dest = /oracle/product/admin/OSISTA/cdump
    audit_file_dest = /oracle/product/admin/OSISTA/adump
    db_name = OSISTA
    open_cursors = 300
    pga_aggregate_target = 104857600
    aq_tm_processes = 1
    Cluster communication is configured to use the following interface(s) for this instance
    172.16.0.2
    Wed Jun 13 11:04:30 2012
    cluster interconnect IPC version:Oracle UDP/IP (generic)
    IPC Vendor 1 proto 2
    PMON started with pid=2, OS id=8560
    DIAG started with pid=3, OS id=8562
    PSP0 started with pid=4, OS id=8566
    LMON started with pid=5, OS id=8570
    LMD0 started with pid=6, OS id=8574
    LMS0 started with pid=7, OS id=8576
    LMS1 started with pid=8, OS id=8580
    MMAN started with pid=9, OS id=8584
    DBW0 started with pid=10, OS id=8586
    LGWR started with pid=11, OS id=8588
    CKPT started with pid=12, OS id=8590
    SMON started with pid=13, OS id=8592
    RECO started with pid=14, OS id=8594
    CJQ0 started with pid=15, OS id=8596
    MMON started with pid=16, OS id=8598
    Wed Jun 13 11:04:31 2012
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=17, OS id=8600
    Wed Jun 13 11:04:31 2012
    starting up 1 shared server(s) ...
    Wed Jun 13 11:04:31 2012
    lmon registered with NM - instance id 2 (internal mem no 1)
    Wed Jun 13 11:04:31 2012
    Reconfiguration started (old inc 0, new inc 2)
    List of nodes:
    1
    Global Resource Directory frozen
    * allocate domain 0, invalid = TRUE
    Communication channels reestablished
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:04:31 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:04:31 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Post SMON to start 1st pass IR
    Wed Jun 13 11:04:31 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:04:31 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:04:31 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    Reconfiguration complete
    LCK0 started with pid=20, OS id=8877
    Wed Jun 13 11:04:43 2012
    alter database mount
    Wed Jun 13 11:04:43 2012
    This instance was first to mount
    Wed Jun 13 11:04:43 2012
    Starting background process ASMB
    ASMB started with pid=25, OS id=10068
    Starting background process RBAL
    RBAL started with pid=26, OS id=10072
    Wed Jun 13 11:04:47 2012
    SUCCESS: diskgroup DATA was mounted
    Wed Jun 13 11:04:51 2012
    Setting recovery target incarnation to 1
    Wed Jun 13 11:04:52 2012
    Successful mount of redo thread 2, with mount id 3005749259
    Wed Jun 13 11:04:52 2012
    Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
    Completed: alter database mount
    Wed Jun 13 11:05:06 2012
    alter database open
    Wed Jun 13 11:05:06 2012
    This instance was first to open
    Wed Jun 13 11:05:06 2012
    Beginning crash recovery of 1 threads
    parallel recovery started with 2 processes
    Wed Jun 13 11:05:07 2012
    Started redo scan
    Wed Jun 13 11:05:07 2012
    Completed redo scan
    61 redo blocks read, 4 data blocks need recovery
    Wed Jun 13 11:05:07 2012
    Started redo application at
    Thread 1: logseq 7924, block 3, scn 506098125
    Wed Jun 13 11:05:07 2012
    Recovery of Online Redo Log: Thread 1 Group 2 Seq 7924 Reading mem 0
    Mem# 0: +DATA/osista/onlinelog/group_2.372.742132543
    Wed Jun 13 11:05:07 2012
    Completed redo application
    Wed Jun 13 11:05:07 2012
    Completed crash recovery at
    Thread 1: logseq 7924, block 64, scn 506118186
    4 data blocks read, 4 data blocks written, 61 redo blocks read
    Switch log for thread 1 to sequence 7925
    Picked broadcast on commit scheme to generate SCNs
    db_recovery_file_dest_size of 65368 MB is 0.61% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Thread 1 advanced to log sequence 7926
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Thread 1 advanced to log sequence 7927
    Wed Jun 13 11:05:11 2012
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=31, OS id=12747
    Wed Jun 13 11:05:11 2012
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=32, OS id=12749
    Wed Jun 13 11:05:12 2012
    Thread 2 opened at log sequence 7176
    Current log# 4 seq# 7176 mem# 0: +DATA/osista/onlinelog/group_4.289.742134597
    Wed Jun 13 11:05:12 2012
    ARC1: Becoming the 'no FAL' ARCH
    ARC1: Becoming the 'no SRL' ARCH
    Wed Jun 13 11:05:12 2012
    Successful open of redo thread 2
    Wed Jun 13 11:05:12 2012
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Jun 13 11:05:12 2012
    ARC0: Becoming the heartbeat ARCH
    Wed Jun 13 11:05:12 2012
    SMON: enabling cache recovery
    Wed Jun 13 11:05:15 2012
    Successfully onlined Undo Tablespace 20.
    Wed Jun 13 11:05:15 2012
    SMON: enabling tx recovery
    Wed Jun 13 11:05:15 2012
    Database Characterset is AL32UTF8
    Wed Jun 13 11:05:16 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_9174.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Wed Jun 13 11:05:16 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_9174.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Error 600 happened during db open, shutting down database
    USER: terminating instance due to error 600
    Instance terminated by USER, pid = 9174
    ORA-1092 signalled during: alter database open...
    Wed Jun 13 11:06:16 2012
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Interface type 1 eth0 172.16.0.0 configured from OCR for use as a cluster interconnect
    Interface type 1 bond0 132.147.160.0 configured from OCR for use as a public interface
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
    processes = 300
    sessions = 335
    sga_max_size = 524288000
    __shared_pool_size = 314572800
    __large_pool_size = 4194304
    __java_pool_size = 8388608
    __streams_pool_size = 8388608
    spfile = +DATA/osista/spfileosista.ora
    nls_language = FRENCH
    nls_territory = FRANCE
    nls_length_semantics = CHAR
    sga_target = 524288000
    control_files = DATA/osista/controlfile/control01.ctl, DATA/osista/controlfile/control02.ctl
    db_block_size = 8192
    __db_cache_size = 180355072
    compatible = 10.2.0.3.0
    log_archive_dest_1 = LOCATION=USE_DB_RECOVERY_FILE_DEST
    db_file_multiblock_read_count= 16
    cluster_database = TRUE
    cluster_database_instances= 2
    db_create_file_dest = +DATA
    db_recovery_file_dest = +FLASH
    db_recovery_file_dest_size= 68543315968
    thread = 2
    instance_number = 2
    undo_management = AUTO
    undo_tablespace = UNDOTBS2
    undo_retention = 29880
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=OSISTAXDB)
    local_listener = (address=(protocol=tcp)(port=1521)(host=132.147.160.243))
    remote_listener = LISTENERS_OSISTA
    job_queue_processes = 10
    background_dump_dest = /oracle/product/admin/OSISTA/bdump
    user_dump_dest = /oracle/product/admin/OSISTA/udump
    core_dump_dest = /oracle/product/admin/OSISTA/cdump
    audit_file_dest = /oracle/product/admin/OSISTA/adump
    db_name = OSISTA
    open_cursors = 300
    pga_aggregate_target = 104857600
    aq_tm_processes = 1
    Cluster communication is configured to use the following interface(s) for this instance
    172.16.0.2
    Wed Jun 13 11:06:16 2012
    cluster interconnect IPC version:Oracle UDP/IP (generic)
    IPC Vendor 1 proto 2
    PMON started with pid=2, OS id=18682
    DIAG started with pid=3, OS id=18684
    PSP0 started with pid=4, OS id=18695
    LMON started with pid=5, OS id=18704
    LMD0 started with pid=6, OS id=18721
    LMS0 started with pid=7, OS id=18735
    LMS1 started with pid=8, OS id=18753
    MMAN started with pid=9, OS id=18767
    DBW0 started with pid=10, OS id=18788
    LGWR started with pid=11, OS id=18796
    CKPT started with pid=12, OS id=18799
    SMON started with pid=13, OS id=18801
    RECO started with pid=14, OS id=18803
    CJQ0 started with pid=15, OS id=18805
    MMON started with pid=16, OS id=18807
    Wed Jun 13 11:06:17 2012
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=17, OS id=18809
    Wed Jun 13 11:06:17 2012
    starting up 1 shared server(s) ...
    Wed Jun 13 11:06:17 2012
    lmon registered with NM - instance id 2 (internal mem no 1)
    Wed Jun 13 11:06:17 2012
    Reconfiguration started (old inc 0, new inc 2)
    List of nodes:
    1
    Global Resource Directory frozen
    * allocate domain 0, invalid = TRUE
    Communication channels reestablished
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Post SMON to start 1st pass IR
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    Reconfiguration complete
    LCK0 started with pid=20, OS id=18816
    Wed Jun 13 11:06:18 2012
    ALTER DATABASE MOUNT
    Wed Jun 13 11:06:18 2012
    This instance was first to mount
    Wed Jun 13 11:06:18 2012
    Reconfiguration started (old inc 2, new inc 4)
    List of nodes:
    0 1
    Wed Jun 13 11:06:18 2012
    Starting background process ASMB
    Wed Jun 13 11:06:18 2012
    Global Resource Directory frozen
    Communication channels reestablished
    ASMB started with pid=22, OS id=18913
    Starting background process RBAL
    * domain 0 valid = 0 according to instance 0
    Wed Jun 13 11:06:18 2012
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Wed Jun 13 11:06:18 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:06:18 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    RBAL started with pid=23, OS id=18917
    Reconfiguration complete
    Wed Jun 13 11:06:22 2012
    SUCCESS: diskgroup DATA was mounted
    Wed Jun 13 11:06:26 2012
    Setting recovery target incarnation to 1
    Wed Jun 13 11:06:26 2012
    Successful mount of redo thread 2, with mount id 3005703530
    Wed Jun 13 11:06:26 2012
    Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
    Completed: ALTER DATABASE MOUNT
    Wed Jun 13 11:06:27 2012
    ALTER DATABASE OPEN
    This instance was first to open
    Wed Jun 13 11:06:27 2012
    Beginning crash recovery of 1 threads
    parallel recovery started with 2 processes
    Wed Jun 13 11:06:27 2012
    Started redo scan
    Wed Jun 13 11:06:27 2012
    Completed redo scan
    61 redo blocks read, 4 data blocks need recovery
    Wed Jun 13 11:06:28 2012
    Started redo application at
    Thread 2: logseq 7176, block 3
    Wed Jun 13 11:06:28 2012
    Recovery of Online Redo Log: Thread 2 Group 4 Seq 7176 Reading mem 0
    Mem# 0: +DATA/osista/onlinelog/group_4.289.742134597
    Wed Jun 13 11:06:28 2012
    Completed redo application
    Wed Jun 13 11:06:28 2012
    Completed crash recovery at
    Thread 2: logseq 7176, block 64, scn 506138248
    4 data blocks read, 4 data blocks written, 61 redo blocks read
    Picked broadcast on commit scheme to generate SCNs
    Wed Jun 13 11:06:28 2012
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=28, OS id=19692
    Wed Jun 13 11:06:28 2012
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=29, OS id=19695
    Wed Jun 13 11:06:28 2012
    Thread 2 advanced to log sequence 7177
    Thread 2 opened at log sequence 7177
    Current log# 3 seq# 7177 mem# 0: +DATA/osista/onlinelog/group_3.291.742134597
    Successful open of redo thread 2
    Wed Jun 13 11:06:28 2012
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Jun 13 11:06:28 2012
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    Wed Jun 13 11:06:28 2012
    ARC1: Becoming the heartbeat ARCH
    Wed Jun 13 11:06:28 2012
    SMON: enabling cache recovery
    Wed Jun 13 11:06:28 2012
    db_recovery_file_dest_size of 65368 MB is 0.61% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Wed Jun 13 11:06:31 2012
    Successfully onlined Undo Tablespace 20.
    Wed Jun 13 11:06:31 2012
    SMON: enabling tx recovery
    Wed Jun 13 11:06:31 2012
    Database Characterset is AL32UTF8
    Wed Jun 13 11:06:31 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_19596.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Wed Jun 13 11:06:32 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_19596.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Error 600 happened during db open, shutting down database
    USER: terminating instance due to error 600
    Instance terminated by USER, pid = 19596
    ORA-1092 signalled during: ALTER DATABASE OPEN...
    Wed Jun 13 11:11:35 2012
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Interface type 1 eth0 172.16.0.0 configured from OCR for use as a cluster interconnect
    Interface type 1 bond0 132.147.160.0 configured from OCR for use as a public interface
    Picked latch-free SCN scheme 2
    Autotune of undo retention is turned on.
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.3.0.
    System parameters with non-default values:
    processes = 300
    sessions = 335
    sga_max_size = 524288000
    __shared_pool_size = 318767104
    __large_pool_size = 4194304
    __java_pool_size = 8388608
    __streams_pool_size = 8388608
    spfile = +DATA/osista/spfileosista.ora
    nls_language = FRENCH
    nls_territory = FRANCE
    nls_length_semantics = CHAR
    sga_target = 524288000
    control_files = DATA/osista/controlfile/control01.ctl, DATA/osista/controlfile/control02.ctl
    db_block_size = 8192
    __db_cache_size = 176160768
    compatible = 10.2.0.3.0
    log_archive_dest_1 = LOCATION=USE_DB_RECOVERY_FILE_DEST
    db_file_multiblock_read_count= 16
    cluster_database = TRUE
    cluster_database_instances= 2
    db_create_file_dest = +DATA
    db_recovery_file_dest = +FLASH
    db_recovery_file_dest_size= 68543315968
    thread = 2
    instance_number = 2
    undo_management = AUTO
    undo_tablespace = UNDOTBS2
    undo_retention = 29880
    remote_login_passwordfile= EXCLUSIVE
    db_domain =
    dispatchers = (PROTOCOL=TCP) (SERVICE=OSISTAXDB)
    local_listener = (address=(protocol=tcp)(port=1521)(host=132.147.160.243))
    remote_listener = LISTENERS_OSISTA
    job_queue_processes = 10
    background_dump_dest = /oracle/product/admin/OSISTA/bdump
    user_dump_dest = /oracle/product/admin/OSISTA/udump
    core_dump_dest = /oracle/product/admin/OSISTA/cdump
    audit_file_dest = /oracle/product/admin/OSISTA/adump
    db_name = OSISTA
    open_cursors = 300
    pga_aggregate_target = 104857600
    aq_tm_processes = 1
    Cluster communication is configured to use the following interface(s) for this instance
    172.16.0.2
    Wed Jun 13 11:11:35 2012
    cluster interconnect IPC version:Oracle UDP/IP (generic)
    IPC Vendor 1 proto 2
    PMON started with pid=2, OS id=16101
    DIAG started with pid=3, OS id=16103
    PSP0 started with pid=4, OS id=16105
    LMON started with pid=5, OS id=16107
    LMD0 started with pid=6, OS id=16110
    LMS0 started with pid=7, OS id=16112
    LMS1 started with pid=8, OS id=16116
    MMAN started with pid=9, OS id=16120
    DBW0 started with pid=10, OS id=16132
    LGWR started with pid=11, OS id=16148
    CKPT started with pid=12, OS id=16169
    SMON started with pid=13, OS id=16185
    RECO started with pid=14, OS id=16203
    CJQ0 started with pid=15, OS id=16219
    MMON started with pid=16, OS id=16227
    Wed Jun 13 11:11:36 2012
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    MMNL started with pid=17, OS id=16229
    Wed Jun 13 11:11:36 2012
    starting up 1 shared server(s) ...
    Wed Jun 13 11:11:36 2012
    lmon registered with NM - instance id 2 (internal mem no 1)
    Wed Jun 13 11:11:36 2012
    Reconfiguration started (old inc 0, new inc 2)
    List of nodes:
    1
    Global Resource Directory frozen
    * allocate domain 0, invalid = TRUE
    Communication channels reestablished
    Master broadcasted resource hash value bitmaps
    Non-local Process blocks cleaned out
    Wed Jun 13 11:11:36 2012
    LMS 0: 0 GCS shadows cancelled, 0 closed
    Wed Jun 13 11:11:36 2012
    LMS 1: 0 GCS shadows cancelled, 0 closed
    Set master node info
    Submitted all remote-enqueue requests
    Dwn-cvts replayed, VALBLKs dubious
    All grantable enqueues granted
    Post SMON to start 1st pass IR
    Wed Jun 13 11:11:36 2012
    LMS 1: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:11:36 2012
    LMS 0: 0 GCS shadows traversed, 0 replayed
    Wed Jun 13 11:11:36 2012
    Submitted all GCS remote-cache requests
    Fix write in gcs resources
    Reconfiguration complete
    LCK0 started with pid=20, OS id=16235
    Wed Jun 13 11:11:37 2012
    ALTER DATABASE MOUNT
    Wed Jun 13 11:11:37 2012
    This instance was first to mount
    Wed Jun 13 11:11:37 2012
    Starting background process ASMB
    ASMB started with pid=22, OS id=16343
    Starting background process RBAL
    RBAL started with pid=23, OS id=16347
    Wed Jun 13 11:11:44 2012
    SUCCESS: diskgroup DATA was mounted
    Wed Jun 13 11:11:49 2012
    Setting recovery target incarnation to 1
    Wed Jun 13 11:11:49 2012
    Successful mount of redo thread 2, with mount id 3005745065
    Wed Jun 13 11:11:49 2012
    Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
    Completed: ALTER DATABASE MOUNT
    Wed Jun 13 11:22:25 2012
    alter database open
    This instance was first to open
    Wed Jun 13 11:22:26 2012
    Beginning crash recovery of 1 threads
    parallel recovery started with 2 processes
    Wed Jun 13 11:22:26 2012
    Started redo scan
    Wed Jun 13 11:22:26 2012
    Completed redo scan
    61 redo blocks read, 4 data blocks need recovery
    Wed Jun 13 11:22:26 2012
    Started redo application at
    Thread 1: logseq 7927, block 3
    Wed Jun 13 11:22:26 2012
    Recovery of Online Redo Log: Thread 1 Group 1 Seq 7927 Reading mem 0
    Mem# 0: +DATA/osista/onlinelog/group_1.283.742132543
    Wed Jun 13 11:22:26 2012
    Completed redo application
    Wed Jun 13 11:22:26 2012
    Completed crash recovery at
    Thread 1: logseq 7927, block 64, scn 506178382
    4 data blocks read, 4 data blocks written, 61 redo blocks read
    Switch log for thread 1 to sequence 7928
    Picked broadcast on commit scheme to generate SCNs
    Wed Jun 13 11:22:27 2012
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=31, OS id=13010
    Wed Jun 13 11:22:27 2012
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    ARC1 started with pid=32, OS id=13033
    Wed Jun 13 11:22:27 2012
    Thread 2 opened at log sequence 7178
    Current log# 4 seq# 7178 mem# 0: +DATA/osista/onlinelog/group_4.289.742134597
    Successful open of redo thread 2
    Wed Jun 13 11:22:27 2012
    MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
    Wed Jun 13 11:22:27 2012
    ARC0: Becoming the 'no FAL' ARCH
    ARC0: Becoming the 'no SRL' ARCH
    Wed Jun 13 11:22:27 2012
    ARC1: Becoming the heartbeat ARCH
    Wed Jun 13 11:22:27 2012
    SMON: enabling cache recovery
    Wed Jun 13 11:22:30 2012
    db_recovery_file_dest_size of 65368 MB is 0.61% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    SUCCESS: diskgroup FLASH was mounted
    SUCCESS: diskgroup FLASH was dismounted
    Wed Jun 13 11:22:31 2012
    Successfully onlined Undo Tablespace 20.
    Wed Jun 13 11:22:31 2012
    SMON: enabling tx recovery
    Wed Jun 13 11:22:32 2012
    Database Characterset is AL32UTF8
    Wed Jun 13 11:22:32 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_11751.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Wed Jun 13 11:22:33 2012
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_11751.trc:
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []
    Error 600 happened during db open, shutting down database
    USER: terminating instance due to error 600
    Instance terminated by USER, pid = 11751
    ORA-1092 signalled during: alter database open...
    regards,

    Hi;
    Errors in file /oracle/product/admin/OSISTA/udump/osista2_ora_9174.trc:Did you check trc file?
    ORA-00600: code d'erreur interne, arguments : [kokiasg1], [], [], [], [], [], [], []You are getting oracle internal error(ORA 600) which mean you could need to work wiht oracle support team. Please see below note, if its not help than i suggest log a sr:
    Troubleshoot an ORA-600 or ORA-7445 Error Using the Error Lookup Tool [ID 153788.1]
    for your future rac issue please use Forum Home » High Availability » RAC, ASM & Clusterware Installation which is RAC dedicated forum site.
    Regard
    Helios

Maybe you are looking for