How to connect to Oracle "real application cluster" RAC?

Hello,
I don't know how the dbUrl-Parameter should look like to connect to an Oracle RAC.
normally I use:
jdbc:oracle:thin:@(Server):(port):(servicename)
e.g.:
jdbc:oracle:thin:@frxxx7....world.com:1521:kfics
but what should I use if I want to connect to an RAC? In this case I don't have a single server?
Thanks for your help.
Regards.
Matthias

Hello again,
I found out half of the solution.
Normally you use a serverUrl like:
jdbc:oracle:thin:@(DESCRIPTION=(LOAD_BALANCE=on)
(ADDRESS=(PROTOCOL=TCP)(HOST=host1) (PORT=1521))
(ADDRESS=(PROTOCOL=TCP)(HOST=host2) (PORT=1521))
(CONNECT_DATA=(SERVICE_NAME=service)))
normally you can also add other information like failover_mode etc.
BUT: MII (12.0.6) only alows a string of 200 characters. Here is a part of the error-log:
Exception of type com.sap.sql.log.OpenSQLException caught: Cannot assign a java.lang.String object of length 309 to host variable 3 which has JDBC type VARCHAR(200)..
so if I have more than two servers or additionally information for failover_mode etc. I'm not able to use the oracle RAC, because the serverUrl is too long.
Any suggestions?
Regards
Matthias

Similar Messages

  • Configuring Oracle Real Application Cluster

    Hi,
    Is it possible to configure Real Application Cluster (RAC) Oracle 9i with Windows 2000 Server / Advanced Server ?
    If yes, what component and Oracle Software should I prepare to build this system ?
    Can anybody explain this step by step.
    Thanks in advance
    ESB

    Hi;
    For your issue i suggest close your thread here as changing thread status to answere and move it to Forum Home » System Management and Integration » Data Integrator which you can get more quick response
    Regard
    Helios

  • Hardware Architecture of Real Application Cluster (RAC)

    I'll implement Oracle 9i with Real Application Cluster (RAC). I now have two Sun SF280R servers.
    Can I use ONE common storage to build RAC?
    SF280R - Storage - SF280R
    OR is it necessary to have TWO storages?
    SF280R SF280R
    |..........\/..........|
    StorageStorage
    Thanks

    The whole idea behind RAC is that there is shared storage for 2 or more database instances. If one instance goes down, the other(s) seemlessly take over. RAC provides insurance against instance failures. Stand by databases (also known as data guard) provide insurance against database (the actual data files) failure.
    HTH.
    -Raj Suchak
    [email protected]

  • Oracle real application cluster test case

    Hello everyone! I have just finished implemented oracle real application for customer. Now I need to do UAT for them with cases. Some of the cases I am not sure whether it can make the cluster fail, or need to repair, or unrecoverable. I need to tell them what is happening if sth go wrong as the following cases:
    case 1. connection from database server to storage is down in one node, or all node in the cluster.
    case 2. private connection down
    case 3. public network down in node node or all nodes in the cluster.
    Please, help address the status and behavior of oracle rac, or suggest me some useful document regarding this cases.
    I'll appreciate your help.
    Thanks

    Hi,
    How about here site -> Oracle Clusterware
    <<http://www.datadisk.co.uk/html_docs/rac/architecture.htm>>
    <<http://onlineappsdba.com/index.php/2009/07/08/voting-disk-management-in-rac/>>
    <<http://apunhiran.blogspot.jp/2009/04/what-is-voting-disk-split-brain.html>>
    If you learned about voting disk is good.
    Regards,
    Edited by: asahideO on 2013/01/31 17:13

  • Steps Installing Oracle Database 10g R2 with Real Application Cluster (RAC)

    Hi,
    We are using Baan ERP and back end Oracle 10g R2.Now we want to use the
    Real Application Cluster technology so that if one server fail ; all of the application data remains available via another server.But i do not know how to implement RAC.
    Pl guide me.
    Regards,
    sanjeev

    stop. BAAN and Oracle will have an "implementation guide" or best practices. There is also the Oracle MAA Maximum Availability Architecture. Go Read those first.

  • Required hardware for a oracle real application cluster

    hello i´m a newbie and have a question:
    We want to migrate a DB2 database to Oracle database. This database will run in the cluster network. What hardware do I need a prerequisite to set up an Oracle Real Application Clusters (10g)?
    thank a lot for help!

    Hi,
    HW sizing depends on lot of information like the Size of the Database, Average load on the Database, Number of DB users, what application is using the database... could you please let us know all these information and oracle license depends on the number of CPU you on the server. Kindly let us know the OS on which you want to run the oracle RAC.
    Regards,
    Satya.

  • IBM HACMP Vs. Oracle Real Application Cluster

    Hello,
    We are contemplating chages in our Oracle Database on AIX. I was wondering if you can point me to a site that will help us to decide which architectiure to go with between IBM HACMP Vs. Oracle Real Aplication Cluster.
    Any information will be appreciated.
    Regards,

    Hoep this helps,
    http://www.google.com/#hl=en&q=+site:ibm.com+IBM+HACMP+Vs.+Oracle+RAC
    Regards
    http://www.oracleracexpert.com

  • Configure JDBC to Real Application Cluster (RAC)

    I'm searching for some documentation on how to setup a JDBC connection to RAC. Can anyone point me in the right direction? Specifically, we're using Weblogic 7 on a Sun Solaris box and using RAC for load balancing and failover protection. I need to know which type of driver to use (Type 2 or 4) and the specific connection information such as naming (or IP) in order to connect to the cluster.
    Thanks.

    Hi buddy,
    Let's clarify some points:
    Redhat Cluster Suite can create n-node server clusters, load balances incoming IP network requests and supports up to 16 nodes. So, do I need to configure RAC if I'm going to use Redhat Custer Suite?RAC- Real applications clusters, a clustered database so that many instances can accessing only one database, many features available like cache fusion.
    CRS- Cluster Solution from Oracle needed to run RAC databases with or without a vendor cluster solution.
    with the information above I'd say:
    1) If You want to use a cold fail over cluster, You can use a vendor cluster solution or just CRS.
    2) If You want to use RAC databases, You must use CRS, with or without a vendor cluster solution. ( since release 10G no more needed to use a vendor cluster solution)
    Hope it helps,
    Cerreia

  • Real Application Cluster (RAC) & Cluster Suite

    Redhat Cluster Suite can create n-node server clusters, load balances incoming IP network requests and supports up to 16 nodes. So, do I need to configure RAC if I'm going to use Redhat Cluster Suite?

    Hi buddy,
    Let's clarify some points:
    Redhat Cluster Suite can create n-node server clusters, load balances incoming IP network requests and supports up to 16 nodes. So, do I need to configure RAC if I'm going to use Redhat Custer Suite?RAC- Real applications clusters, a clustered database so that many instances can accessing only one database, many features available like cache fusion.
    CRS- Cluster Solution from Oracle needed to run RAC databases with or without a vendor cluster solution.
    with the information above I'd say:
    1) If You want to use a cold fail over cluster, You can use a vendor cluster solution or just CRS.
    2) If You want to use RAC databases, You must use CRS, with or without a vendor cluster solution. ( since release 10G no more needed to use a vendor cluster solution)
    Hope it helps,
    Cerreia

  • Using Tuxedo with Oracle Real Application Clusters (RAC)

    Hi all ,
    Our present Producation environment is Tuxedo8.0 with Oracle 9i as database .
    We are planning to implement oracle RAC .
    Can any one tell us , with the current tuxedo version which is 8.0 ,can we for RAC Or we need to upgrade Tuxedo to 9.1 for RAC Implementation?
    If anyone has Implemented Tuxedo and Oracle with RAC successfully
    Pls share your experience.
    Thanks In Adavance
    Nadeer .

    Subbu,
    As I mentioned earlier in this thread, support for Oracle RAC has been
    implemented in Tuxedo 9.1 and in the Tuxedo 8.1 and Tuxedo 9.0 rolling patch
    streams. There are no plans to backport this feature to Tuxedo 8.0. If you
    plan to use RAC transactionally and if it is possible for multiple services
    that could be located on different RAC instances to be accessed in a single
    transaction (which will be true for most applications), then you should
    upgrade to a version of Tuxedo with support for Oracle RAC. The
    documentation at http://e-docs.bea.com/tuxedo/tux91/ads/adorac.htm#248431
    gives an overview of RAC support in Tuxedo.
    Ed
    <Subramonian Arumugom> wrote in message news:[email protected]..
    Hello Nadeer, Hello Ed,
    Has the implementation of Tuxedo8.0 with Oracle RAC was completed
    successfully ?.
    If it is done , I have few doubts. Could you please help me ?.
    1) I hope your Tuxedo middleware connects to the oracle Database using
    Oracle XA resource manager and the TMS server built using buildtms with
    oracle as RM .As per the mail from Ed, migration of Tuxedo9.1 should be done
    before moving to Oracle RAC.
    Also from BEA documentation, Oracle RAC support was given as new feature in
    Tuxedo9.1.
    Our Application uses Oracle RM to connect to Oracle DB[ Oracle 10g] from
    Tuxedo8.0. So do we need to upgrade to Tuxedo9.1. for using Oracle RAC ?.
    Thanks for your help
    Best Regards
    Subbu
    #9880278452
    Edited by subramonian.a at 04/16/2007 12:15 AM
    Edited by subramonian.a at 04/16/2007 12:19 AM
    Edited by subramonian.a at 04/16/2007 1:37 AM
    Edited by subramonian.a at 04/16/2007 8:20 AM

  • Failed preliminary checks for Oracle Real Application Cluster 12C.

    Hi, all. Please, help to investigate and solve. Thanks.
    [grid@orac1 grid]$ ./runcluvfy.sh stage -pre crsinst -n orac1,orac2 -fixup -verbose
    Performing pre-checks for cluster services setup
    Checking node reachability...
    Check: Node reachability from node "orac1"
      Destination Node                      Reachable?            
      orac1                                 yes                   
      orac2                                 yes                   
    Result: Node reachability check passed from node "orac1"      
    Checking user equivalence...
    Check: User equivalence for user "grid"
      Node Name                             Status                
      orac2                                 passed                
      orac1                                 passed                
    Result: User equivalence check passed for user "grid"         
    Checking node connectivity...
    Checking hosts config file...
      Node Name                             Status                
      orac1                                 passed                
      orac2                                 passed                
    Verification of the hosts config file successful
    Interface information for node "orac1"
    Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU 
    eth0   10.154.137.101  10.154.137.0    0.0.0.0         10.0.4.2        08:00:27:28:D1:1F 1500
    eth1   10.154.138.101  10.154.138.0    0.0.0.0         10.0.4.2        08:00:27:51:E8:B9 1500
    eth2   10.0.4.15       10.0.4.0        0.0.0.0         10.0.4.2        08:00:27:5B:D7:29 1500
    Interface information for node "orac2"
    Name   IP Address      Subnet          Gateway         Def. Gateway    HW Address        MTU 
    eth0   10.154.137.102  10.154.137.0    0.0.0.0         10.0.4.2        08:00:27:7F:C8:70 1500
    eth1   10.154.138.102  10.154.138.0    0.0.0.0         10.0.4.2        08:00:27:9D:27:D8 1500
    eth2   10.0.4.15       10.0.4.0        0.0.0.0         10.0.4.2        08:00:27:6E:92:43 1500
    Check: Node connectivity of subnet "10.154.137.0"
      Source                          Destination                     Connected?    
      orac1[10.154.137.101]           orac2[10.154.137.102]           yes           
    Result: Node connectivity passed for subnet "10.154.137.0" with node(s) orac1,orac2
    Check: TCP connectivity of subnet "10.154.137.0"
      Source                          Destination                     Connected?    
      orac1:10.154.137.101            orac2:10.154.137.102            failed        
    ERROR:
    PRVF-7617 : Node connectivity between "orac1 : 10.154.137.101" and "orac2 : 10.154.137.102" failed
    Result: TCP connectivity check failed for subnet "10.154.137.0"                                 
    Check: Node connectivity of subnet "10.154.138.0"
      Source                          Destination                     Connected?    
      orac1[10.154.138.101]           orac2[10.154.138.102]           yes           
    Result: Node connectivity passed for subnet "10.154.138.0" with node(s) orac1,orac2
    Check: TCP connectivity of subnet "10.154.138.0"
      Source                          Destination                     Connected?    
      orac1:10.154.138.101            orac2:10.154.138.102            failed        
    ERROR:
    PRVF-7617 : Node connectivity between "orac1 : 10.154.138.101" and "orac2 : 10.154.138.102" failed
    Result: TCP connectivity check failed for subnet "10.154.138.0"                                 
    Check: Node connectivity of subnet "10.0.4.0"
      Source                          Destination                     Connected?    
      orac1[10.0.4.15]                orac2[10.0.4.15]                yes           
    Result: Node connectivity passed for subnet "10.0.4.0" with node(s) orac1,orac2 
    Check: TCP connectivity of subnet "10.0.4.0"
      Source                          Destination                     Connected?    
      orac1:10.0.4.15                 orac2:10.0.4.15                 failed        
    ERROR:
    PRVF-7617 : Node connectivity between "orac1 : 10.0.4.15" and "orac2 : 10.0.4.15" failed
    Result: TCP connectivity check failed for subnet "10.0.4.0"                           
    Interfaces found on subnet "10.0.4.0" that are likely candidates for VIP are:
    orac1 eth2:10.0.4.15                                                       
    orac2 eth2:10.0.4.15                                                       
    WARNING:
    Could not find a suitable set of interfaces for the private interconnect
    Checking subnet mask consistency...                                   
    Subnet mask consistency check passed for subnet "10.154.137.0".       
    Subnet mask consistency check passed for subnet "10.154.138.0".       
    Subnet mask consistency check passed for subnet "10.0.4.0".           
    Subnet mask consistency check passed.                                 
    ERROR:
    PRVG-1172 : The IP address "10.0.4.15" is on multiple interfaces "eth2,eth2" on nodes "orac2,orac1"
    Result: Node connectivity check failed
    Checking multicast communication...
    Checking subnet "10.154.137.0" for multicast communication with multicast group "224.0.0.251"...
    PRVG-11138 : Interface "10.154.137.101" on node "orac1" is not able to communicate with interface "10.154.137.101" on node "orac1" over multicast group "224.0.0.251"
    PRVG-11138 : Interface "10.154.137.101" on node "orac1" is not able to communicate with interface "10.154.137.102" on node "orac2" over multicast group "224.0.0.251"
    PRVG-11138 : Interface "10.154.137.102" on node "orac2" is not able to communicate with interface "10.154.137.101" on node "orac1" over multicast group "224.0.0.251"
    PRVG-11138 : Interface "10.154.137.102" on node "orac2" is not able to communicate with interface "10.154.137.102" on node "orac2" over multicast group "224.0.0.251"
    Checking subnet "10.154.138.0" for multicast communication with multicast group "224.0.0.251"...                                                                   
    PRVG-11138 : Interface "10.154.138.101" on node "orac1" is not able to communicate with interface "10.154.138.101" on node "orac1" over multicast group "224.0.0.251"
    PRVG-11138 : Interface "10.154.138.101" on node "orac1" is not able to communicate with interface "10.154.138.102" on node "orac2" over multicast group "224.0.0.251"
    PRVG-11138 : Interface "10.154.138.102" on node "orac2" is not able to communicate with interface "10.154.138.101" on node "orac1" over multicast group "224.0.0.251"
    PRVG-11138 : Interface "10.154.138.102" on node "orac2" is not able to communicate with interface "10.154.138.102" on node "orac2" over multicast group "224.0.0.251"
    Checking subnet "10.0.4.0" for multicast communication with multicast group "224.0.0.251"...                                                                       
    PRVG-11138 : Interface "10.0.4.15" on node "orac1" is not able to communicate with interface "10.0.4.15" on node "orac1" over multicast group "224.0.0.251"        
    PRVG-11138 : Interface "10.0.4.15" on node "orac1" is not able to communicate with interface "10.0.4.15" on node "orac2" over multicast group "224.0.0.251"        
    PRVG-11138 : Interface "10.0.4.15" on node "orac2" is not able to communicate with interface "10.0.4.15" on node "orac1" over multicast group "224.0.0.251"        
    PRVG-11138 : Interface "10.0.4.15" on node "orac2" is not able to communicate with interface "10.0.4.15" on node "orac2" over multicast group "224.0.0.251"        
    Checking ASMLib configuration.
      Node Name                             Status                
      orac1                                 passed                
      orac2                                 passed                
    Result: Check for ASMLib configuration passed.                
    Check: Total memory
      Node Name     Available                 Required                  Status  
      orac2         5.714GB (5991520.0KB)     4GB (4194304.0KB)         passed  
      orac1         5.714GB (5991520.0KB)     4GB (4194304.0KB)         passed  
    Result: Total memory check passed                                           
    Check: Available memory
      Node Name     Available                 Required                  Status  
      orac2         5.2293GB (5483324.0KB)    50MB (51200.0KB)          passed  
      orac1         5.1456GB (5395600.0KB)    50MB (51200.0KB)          passed  
    Result: Available memory check passed                                       
    Check: Swap space
      Node Name     Available                 Required                  Status  
      orac2         6.125GB (6422520.0KB)     5.714GB (5991520.0KB)     passed  
      orac1         6.125GB (6422520.0KB)     5.714GB (5991520.0KB)     passed  
    Result: Swap space check passed                                             
    Check: Free disk space for "orac2:/usr,orac2:/var,orac2:/etc,orac2:/sbin,orac2:/tmp"
      Path              Node Name     Mount point   Available     Required      Status    
      /usr              orac2         /             37.8398GB     1.0635GB      passed    
      /var              orac2         /             37.8398GB     1.0635GB      passed    
      /etc              orac2         /             37.8398GB     1.0635GB      passed    
      /sbin             orac2         /             37.8398GB     1.0635GB      passed    
      /tmp              orac2         /             37.8398GB     1.0635GB      passed    
    Result: Free disk space check passed for "orac2:/usr,orac2:/var,orac2:/etc,orac2:/sbin,orac2:/tmp"
    Check: Free disk space for "orac1:/usr,orac1:/var,orac1:/etc,orac1:/sbin,orac1:/tmp"
      Path              Node Name     Mount point   Available     Required      Status    
      /usr              orac1         /             32.4382GB     1.0635GB      passed    
      /var              orac1         /             32.4382GB     1.0635GB      passed    
      /etc              orac1         /             32.4382GB     1.0635GB      passed    
      /sbin             orac1         /             32.4382GB     1.0635GB      passed    
      /tmp              orac1         /             32.4382GB     1.0635GB      passed    
    Result: Free disk space check passed for "orac1:/usr,orac1:/var,orac1:/etc,orac1:/sbin,orac1:/tmp"
    Check: User existence for "grid"
      Node Name     Status                    Comment               
      orac2         passed                    exists(1100)          
      orac1         passed                    exists(1100)          
    Checking for multiple users with UID value 1100
    Result: Check for multiple users with UID value 1100 passed
    Result: User existence check passed for "grid"            
    Check: Group existence for "oinstall"
      Node Name     Status                    Comment               
      orac2         passed                    exists                
      orac1         passed                    exists                
    Result: Group existence check passed for "oinstall"             
    Check: Group existence for "dba"
      Node Name     Status                    Comment               
      orac2         passed                    exists                
      orac1         passed                    exists                
    Result: Group existence check passed for "dba"                  
    Check: Membership of user "grid" in group "oinstall" [as Primary]
      Node Name         User Exists   Group Exists  User in Group  Primary       Status    
      orac2             yes           yes           yes           yes           passed     
      orac1             yes           yes           yes           yes           passed     
    Result: Membership check for user "grid" in group "oinstall" [as Primary] passed       
    Check: Membership of user "grid" in group "dba"
      Node Name         User Exists   Group Exists  User in Group  Status        
      orac2             yes           yes           yes           passed         
      orac1             yes           yes           yes           passed         
    Result: Membership check for user "grid" in group "dba" passed               
    Check: Run level
      Node Name     run level                 Required                  Status  
      orac2         5                         3,5                       passed  
      orac1         5                         3,5                       passed  
    Result: Run level check passed                                              
    Check: Hard limits for "maximum open file descriptors"
      Node Name         Type          Available     Required      Status        
      orac2             hard          4096          65536         failed        
      orac1             hard          4096          65536         failed        
    Result: Hard limits check failed for "maximum open file descriptors"        
    Check: Soft limits for "maximum open file descriptors"
      Node Name         Type          Available     Required      Status        
      orac2             soft          1024          1024          passed        
      orac1             soft          4096          1024          passed        
    Result: Soft limits check passed for "maximum open file descriptors"        
    Check: Hard limits for "maximum user processes"
      Node Name         Type          Available     Required      Status        
      orac2             hard          46654         16384         passed        
      orac1             hard          46654         16384         passed        
    Result: Hard limits check passed for "maximum user processes"               
    Check: Soft limits for "maximum user processes"
      Node Name         Type          Available     Required      Status        
      orac2             soft          1024          2047          failed        
      orac1             soft          1024          2047          failed        
    Result: Soft limits check failed for "maximum user processes"               
    Check: System architecture
      Node Name     Available                 Required                  Status  
      orac2         x86_64                    x86_64                    passed  
      orac1         x86_64                    x86_64                    passed  
    Result: System architecture check passed                                    
    Check: Kernel version
      Node Name     Available                 Required                  Status  
      orac2         2.6.32-279.el6.x86_64     2.6.32                    passed  
      orac1         2.6.32-279.el6.x86_64     2.6.32                    passed  
    Result: Kernel version check passed                                         
    Check: Kernel parameter for "semmsl"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             250           250           250           passed                  
      orac2             250           250           250           passed                  
    Result: Kernel parameter check passed for "semmsl"                                    
    Check: Kernel parameter for "semmns"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             32000         32000         32000         passed                  
      orac2             32000         32000         32000         passed                  
    Result: Kernel parameter check passed for "semmns"                                    
    Check: Kernel parameter for "semopm"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             100           100           100           passed                  
      orac2             100           100           100           passed                  
    Result: Kernel parameter check passed for "semopm"                                    
    Check: Kernel parameter for "semmni"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             128           128           128           passed                  
      orac2             128           128           128           passed                  
    Result: Kernel parameter check passed for "semmni"                                    
    Check: Kernel parameter for "shmmax"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             4398046511104  4398046511104  3067658240    passed                
      orac2             4398046511104  4398046511104  3067658240    passed                
    Result: Kernel parameter check passed for "shmmax"                                    
    Check: Kernel parameter for "shmmni"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             4096          4096          4096          passed                  
      orac2             4096          4096          4096          passed                  
    Result: Kernel parameter check passed for "shmmni"                                    
    Check: Kernel parameter for "shmall"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             4294967296    4294967296    599152        passed                  
      orac2             4294967296    4294967296    599152        passed                  
    Result: Kernel parameter check passed for "shmall"                                    
    Check: Kernel parameter for "file-max"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             6815744       6815744       6815744       passed                  
      orac2             6815744       6815744       6815744       passed                  
    Result: Kernel parameter check passed for "file-max"                                  
    Check: Kernel parameter for "ip_local_port_range"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             between 9000 & 65500  between 9000 & 65500  between 9000 & 65535  passed        
      orac2             between 9000 & 65500  between 9000 & 65500  between 9000 & 65535  passed        
    Result: Kernel parameter check passed for "ip_local_port_range"                                     
    Check: Kernel parameter for "rmem_default"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             262144        262144        262144        passed                  
      orac2             262144        262144        262144        passed                  
    Result: Kernel parameter check passed for "rmem_default"                              
    Check: Kernel parameter for "rmem_max"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             4194304       4194304       4194304       passed                  
      orac2             4194304       4194304       4194304       passed                  
    Result: Kernel parameter check passed for "rmem_max"                                  
    Check: Kernel parameter for "wmem_default"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             262144        262144        262144        passed                  
      orac2             262144        262144        262144        passed                  
    Result: Kernel parameter check passed for "wmem_default"                              
    Check: Kernel parameter for "wmem_max"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             1048576       1048576       1048576       passed                  
      orac2             1048576       1048576       1048576       passed                  
    Result: Kernel parameter check passed for "wmem_max"                                  
    Check: Kernel parameter for "aio-max-nr"
      Node Name         Current       Configured    Required      Status        Comment   
      orac1             1048576       1048576       1048576       passed                  
      orac2             1048576       1048576       1048576       passed                  
    Result: Kernel parameter check passed for "aio-max-nr"                                
    Check: Package existence for "binutils"
      Node Name     Available                 Required                  Status  
      orac2         binutils-2.20.51.0.2-5.34.el6  binutils-2.20.51.0.2      passed  
      orac1         binutils-2.20.51.0.2-5.34.el6  binutils-2.20.51.0.2      passed  
    Result: Package existence check passed for "binutils"                            
    Check: Package existence for "compat-libcap1"
      Node Name     Available                 Required                  Status  
      orac2         compat-libcap1-1.10-1     compat-libcap1-1.10       passed  
      orac1         compat-libcap1-1.10-1     compat-libcap1-1.10       passed  
    Result: Package existence check passed for "compat-libcap1"                 
    Check: Package existence for "compat-libstdc++-33(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         compat-libstdc++-33(x86_64)-3.2.3-69.el6  compat-libstdc++-33(x86_64)-3.2.3  passed  
      orac1         compat-libstdc++-33(x86_64)-3.2.3-69.el6  compat-libstdc++-33(x86_64)-3.2.3  passed  
    Result: Package existence check passed for "compat-libstdc++-33(x86_64)"                             
    Check: Package existence for "libgcc(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         libgcc(x86_64)-4.4.7-11.el6  libgcc(x86_64)-4.4.4      passed  
      orac1         libgcc(x86_64)-4.4.7-11.el6  libgcc(x86_64)-4.4.4      passed  
    Result: Package existence check passed for "libgcc(x86_64)"                    
    Check: Package existence for "libstdc++(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         libstdc++(x86_64)-4.4.7-11.el6  libstdc++(x86_64)-4.4.4   passed  
      orac1         libstdc++(x86_64)-4.4.7-11.el6  libstdc++(x86_64)-4.4.4   passed  
    Result: Package existence check passed for "libstdc++(x86_64)"                    
    Check: Package existence for "libstdc++-devel(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         libstdc++-devel(x86_64)-4.4.7-11.el6  libstdc++-devel(x86_64)-4.4.4  passed  
      orac1         libstdc++-devel(x86_64)-4.4.7-11.el6  libstdc++-devel(x86_64)-4.4.4  passed  
    Result: Package existence check passed for "libstdc++-devel(x86_64)"                         
    Check: Package existence for "sysstat"
      Node Name     Available                 Required                  Status  
      orac2         sysstat-9.0.4-20.el6      sysstat-9.0.4             passed  
      orac1         sysstat-9.0.4-20.el6      sysstat-9.0.4             passed  
    Result: Package existence check passed for "sysstat"                        
    Check: Package existence for "gcc"
      Node Name     Available                 Required                  Status  
      orac2         gcc-4.4.7-11.el6          gcc-4.4.4                 passed  
      orac1         gcc-4.4.7-11.el6          gcc-4.4.4                 passed  
    Result: Package existence check passed for "gcc"                            
    Check: Package existence for "gcc-c++"
      Node Name     Available                 Required                  Status  
      orac2         gcc-c++-4.4.7-11.el6      gcc-c++-4.4.4             passed  
      orac1         gcc-c++-4.4.7-11.el6      gcc-c++-4.4.4             passed  
    Result: Package existence check passed for "gcc-c++"                        
    Check: Package existence for "ksh"
      Node Name     Available                 Required                  Status  
      orac2         ksh-20120801-21.el6.1     ksh-...                   passed  
      orac1         ksh-20120801-21.el6.1     ksh-...                   passed  
    Result: Package existence check passed for "ksh"                            
    Check: Package existence for "make"
      Node Name     Available                 Required                  Status  
      orac2         make-3.81-20.el6          make-3.81                 passed  
      orac1         make-3.81-20.el6          make-3.81                 passed  
    Result: Package existence check passed for "make"                           
    Check: Package existence for "glibc(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         glibc(x86_64)-2.12-1.149.el6  glibc(x86_64)-2.12        passed  
      orac1         glibc(x86_64)-2.12-1.149.el6  glibc(x86_64)-2.12        passed  
    Result: Package existence check passed for "glibc(x86_64)"                      
    Check: Package existence for "glibc-devel(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         glibc-devel(x86_64)-2.12-1.149.el6  glibc-devel(x86_64)-2.12  passed  
      orac1         glibc-devel(x86_64)-2.12-1.149.el6  glibc-devel(x86_64)-2.12  passed  
    Result: Package existence check passed for "glibc-devel(x86_64)"                      
    Check: Package existence for "libaio(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         libaio(x86_64)-0.3.107-10.el6  libaio(x86_64)-0.3.107    passed  
      orac1         libaio(x86_64)-0.3.107-10.el6  libaio(x86_64)-0.3.107    passed  
    Result: Package existence check passed for "libaio(x86_64)"                      
    Check: Package existence for "libaio-devel(x86_64)"
      Node Name     Available                 Required                  Status  
      orac2         libaio-devel(x86_64)-0.3.107-10.el6  libaio-devel(x86_64)-0.3.107  passed  
      orac1         libaio-devel(x86_64)-0.3.107-10.el6  libaio-devel(x86_64)-0.3.107  passed  
    Result: Package existence check passed for "libaio-devel(x86_64)"                          
    Check: Package existence for "nfs-utils"
      Node Name     Available                 Required                  Status  
      orac2         nfs-utils-1.2.3-26.el6    nfs-utils-1.2.3-15        passed  
      orac1         nfs-utils-1.2.3-26.el6    nfs-utils-1.2.3-15        passed  
    Result: Package existence check passed for "nfs-utils"                      
    Checking availability of ports "6200,6100" required for component "Oracle Notification Service (ONS)"
      Node Name         Port Number   Protocol      Available     Status                               
      orac2             6200          TCP           yes           successful                           
      orac1             6200          TCP           yes           successful                           
      orac2             6100          TCP           yes           successful                           
      orac1             6100          TCP           yes           successful                           
    Result: Port availability check passed for ports "6200,6100"                                       
    Checking for multiple users with UID value 0
    Result: Check for multiple users with UID value 0 passed
    Check: Current group ID
    Result: Current group ID check passed
    Starting check for consistency of primary group of root user
      Node Name                             Status                
      orac2                                 passed                
      orac1                                 passed                
    Check for consistency of root user's primary group passed
    Starting Clock synchronization checks using Network Time Protocol(NTP)...
    NTP Configuration file check started...
    The NTP configuration file "/etc/ntp.conf" is available on all nodes
    NTP Configuration file check passed                               
    Checking daemon liveness...
    Check: Liveness for "ntpd"
      Node Name                             Running?              
      orac2                                 yes                   
      orac1                                 yes                   
    Result: Liveness check passed for "ntpd"                      
    Check for NTP daemon or service alive passed on all nodes     
    Checking whether NTP daemon or service is using UDP port 123 on all nodes
    Check for NTP daemon or service using UDP port 123
      Node Name                             Port Open?            
      orac2                                 yes                   
      orac1                                 yes                   
    NTP common Time Server Check started...
    NTP Time Server "193.27.209.1" is common to all nodes on which the NTP daemon is running
    NTP Time Server "194.29.130.252" is common to all nodes on which the NTP daemon is running
    Check of common NTP Time Server passed                                                  
    Clock time offset check from NTP Time Server started...
    Checking on nodes "[orac2, orac1]"...                
    Check: Clock time offset from NTP Time Server        
    Time Server: 193.27.209.1
    Time Offset Limit: 1000.0 msecs
      Node Name     Time Offset               Status                
      orac2         -355.25                   passed                
      orac1         -124.19                   passed                
    Time Server "193.27.209.1" has time offsets that are within permissible limits for nodes "[orac2, orac1]".
    Time Server: 194.29.130.252
    Time Offset Limit: 1000.0 msecs
      Node Name     Time Offset               Status                
      orac2         -358.87                   passed                
      orac1         -109.63                   passed                
    Time Server "194.29.130.252" has time offsets that are within permissible limits for nodes "[orac2, orac1]".
    Clock time offset check passed                                                                             
    Result: Clock synchronization check using Network Time Protocol(NTP) passed
    Checking Core file name pattern consistency...
    Core file name pattern consistency check passed.
    Checking to make sure user "grid" is not in "root" group
      Node Name     Status                    Comment               
      orac2         passed                    does not exist        
      orac1         passed                    does not exist        
    Result: User "grid" is not part of "root" group. Check passed   
    Check default user file creation mask
      Node Name     Available                 Required                  Comment 
      orac2         0022                      0022                      passed  
      orac1         0022                      0022                      passed  
    Result: Default user file creation mask check passed                        
    Checking integrity of file "/etc/resolv.conf" across nodes                  
    Checking the file "/etc/resolv.conf" to make sure only one of domain and search entries is defined
    WARNING:
    PRVF-5640 : Both search and domain entries are present in file "/etc/resolv.conf" on the following nodes: orac1,orac2
    Checking if domain entry in file "/etc/resolv.conf" is consistent across the nodes...                              
    Checking file "/etc/resolv.conf" to make sure that only one domain entry is defined                                
    More than one "domain" entry does not exist in any "/etc/resolv.conf" file                                         
    All nodes have same "domain" entry defined in file "/etc/resolv.conf"                                              
    Checking if search entry in file "/etc/resolv.conf" is consistent across the nodes...                              
    Checking file "/etc/resolv.conf" to make sure that only one search entry is defined                                
    More than one "search" entry does not exist in any "/etc/resolv.conf" file                                         
    All nodes have same "search" order defined in file "/etc/resolv.conf"                                              
    Checking DNS response time for an unreachable node                                                                 
      Node Name                             Status                                                                     
      orac1                                 passed                                                                     
      orac2                                 passed                                                                     
    The DNS response time for an unreachable node is within acceptable limit on all nodes                              
    Check for integrity of file "/etc/resolv.conf" passed
    Check: Time zone consistency
    Result: Time zone consistency check passed
    Checking integrity of name service switch configuration file "/etc/nsswitch.conf" ...
    Checking if "hosts" entry in file "/etc/nsswitch.conf" is consistent across nodes...
    Checking file "/etc/nsswitch.conf" to make sure that only one "hosts" entry is defined
    More than one "hosts" entry does not exist in any "/etc/nsswitch.conf" file         
    All nodes have same "hosts" entry defined in file "/etc/nsswitch.conf"              
    Check for integrity of name service switch configuration file "/etc/nsswitch.conf" passed
    Checking daemon "avahi-daemon" is not configured and running
    Check: Daemon "avahi-daemon" not configured
      Node Name     Configured                Status                
      orac2         yes                       failed                
      orac1         yes                       failed                
    Daemon not configured check failed for process "avahi-daemon"   
    Check: Daemon "avahi-daemon" not running
      Node Name     Running?                  Status                
      orac2         yes                       failed                
      orac1         yes                      failed                
    Daemon not running check failed for process "avahi-daemon"      

    Yes, i can.
    Node1.
    [grid@orac1 grid]$ ping 10.154.137.101                                
    PING 10.154.137.101 (10.154.137.101) 56(84) bytes of data.            
    64 bytes from 10.154.137.101: icmp_seq=1 ttl=64 time=0.105 ms                                                                                                         
    64 bytes from 10.154.137.101: icmp_seq=2 ttl=64 time=0.089 ms                                                                                                         
    64 bytes from 10.154.137.101: icmp_seq=3 ttl=64 time=0.056 ms                                                                                                         
    64 bytes from 10.154.137.101: icmp_seq=4 ttl=64 time=0.086 ms                                                                                                         
    ^C                                                                                                                                                                    
    --- 10.154.137.101 ping statistics ---                                                                                                                                
    4 packets transmitted, 4 received, 0% packet loss, time 3098ms                                                                                                        
    rtt min/avg/max/mdev = 0.056/0.084/0.105/0.017 ms
    [grid@orac1 grid]$ ping 10.154.137.102
    PING 10.154.137.102 (10.154.137.102) 56(84) bytes of data.
    64 bytes from 10.154.137.102: icmp_seq=1 ttl=64 time=2.15 ms
    64 bytes from 10.154.137.102: icmp_seq=2 ttl=64 time=0.497 ms
    64 bytes from 10.154.137.102: icmp_seq=3 ttl=64 time=0.364 ms
    ^C
    --- 10.154.137.102 ping statistics ---
    3 packets transmitted, 3 received, 0% packet loss, time 2495ms
    rtt min/avg/max/mdev = 0.364/1.003/2.150/0.813 ms
    [grid@orac1 grid]$ ping 10.154.138.101
    PING 10.154.138.101 (10.154.138.101) 56(84) bytes of data.
    64 bytes from 10.154.138.101: icmp_seq=1 ttl=64 time=0.059 ms
    64 bytes from 10.154.138.101: icmp_seq=2 ttl=64 time=0.173 ms
    64 bytes from 10.154.138.101: icmp_seq=3 ttl=64 time=0.046 ms
    64 bytes from 10.154.138.101: icmp_seq=4 ttl=64 time=0.056 ms
    ^C
    --- 10.154.138.101 ping statistics ---
    4 packets transmitted, 4 received, 0% packet loss, time 3205ms
    rtt min/avg/max/mdev = 0.046/0.083/0.173/0.052 ms
    [grid@orac1 grid]$ ping 10.154.138.102
    PING 10.154.138.102 (10.154.138.102) 56(84) bytes of data.
    64 bytes from 10.154.138.102: icmp_seq=1 ttl=64 time=1.80 ms
    64 bytes from 10.154.138.102: icmp_seq=2 ttl=64 time=0.418 ms
    64 bytes from 10.154.138.102: icmp_seq=3 ttl=64 time=0.899 ms
    ^C
    --- 10.154.138.102 ping statistics ---
    3 packets transmitted, 3 received, 0% packet loss, time 2860ms
    rtt min/avg/max/mdev = 0.418/1.041/1.808/0.577 ms
    [grid@orac1 grid]$
    Node2.
    [grid@orac2 ~]$ ping 10.154.137.101
    PING 10.154.137.101 (10.154.137.101) 56(84) bytes of data.
    64 bytes from 10.154.137.101: icmp_seq=1 ttl=64 time=0.537 ms
    64 bytes from 10.154.137.101: icmp_seq=2 ttl=64 time=1.79 ms
    64 bytes from 10.154.137.101: icmp_seq=3 ttl=64 time=0.733 ms
    ^C                                                         
    --- 10.154.137.101 ping statistics ---                     
    3 packets transmitted, 3 received, 0% packet loss, time 2774ms
    rtt min/avg/max/mdev = 0.537/1.021/1.794/0.552 ms           
    [grid@orac2 ~]$ ping 10.154.137.102
    PING 10.154.137.102 (10.154.137.102) 56(84) bytes of data.
    64 bytes from 10.154.137.102: icmp_seq=1 ttl=64 time=0.043 ms
    64 bytes from 10.154.137.102: icmp_seq=2 ttl=64 time=0.106 ms
    64 bytes from 10.154.137.102: icmp_seq=3 ttl=64 time=0.039 ms
    64 bytes from 10.154.137.102: icmp_seq=4 ttl=64 time=0.039 ms
    ^C
    --- 10.154.137.102 ping statistics ---
    4 packets transmitted, 4 received, 0% packet loss, time 3282ms
    rtt min/avg/max/mdev = 0.039/0.056/0.106/0.029 ms
    [grid@orac2 ~]$ ping 10.154.138.101
    PING 10.154.138.101 (10.154.138.101) 56(84) bytes of data.
    64 bytes from 10.154.138.101: icmp_seq=1 ttl=64 time=0.718 ms
    64 bytes from 10.154.138.101: icmp_seq=2 ttl=64 time=0.515 ms
    64 bytes from 10.154.138.101: icmp_seq=3 ttl=64 time=0.504 ms
    ^C
    --- 10.154.138.101 ping statistics ---
    3 packets transmitted, 3 received, 0% packet loss, time 2468ms
    rtt min/avg/max/mdev = 0.504/0.579/0.718/0.098 ms
    [grid@orac2 ~]$ ping 10.154.138.102
    PING 10.154.138.102 (10.154.138.102) 56(84) bytes of data.
    64 bytes from 10.154.138.102: icmp_seq=1 ttl=64 time=0.058 ms
    64 bytes from 10.154.138.102: icmp_seq=2 ttl=64 time=0.087 ms
    64 bytes from 10.154.138.102: icmp_seq=3 ttl=64 time=0.052 ms
    64 bytes from 10.154.138.102: icmp_seq=4 ttl=64 time=0.090 ms
    ^C
    --- 10.154.138.102 ping statistics ---
    4 packets transmitted, 4 received, 0% packet loss, time 3755ms
    rtt min/avg/max/mdev = 0.052/0.071/0.090/0.019 ms
    [grid@orac2 ~]$

  • Differences between Oracle Real Application Cluster 11gR1 and 11gR2

    Hi Experts,
    Please provide me differences between Oracle RAC 11gR1 and 11gR2.
    Thanks in Advance,
    Jeeshan.

    Hi,
    You can also refer below Oracle notes :
    11gR2 Clusterware and Grid Home - What You Need to Know (Doc ID 1053147.1)
    Important Changes to Oracle Database Patch Sets Starting With 11.2.0.2 (Doc ID 1189783.1)
    Regards,
    Neha

  • Oracle real application cluster

    Hi,Evry body hase good documentation for install clustware step by step ,
    thank..

    Hi,
    you can find in www.otn.oracle.com
    cheers

  • Real Application Cluster on Sun Solaris 8 and Sun Cluster 3

    Hello,
    we want to install Oracle 9i Enterprise Edition in combination with Oracle Real Application Cluster-Option on 2 Nodes. Every node (12-CPU-SMP-Machine) should running Sun Solaris 8 and Sun Cluster 3 Service.
    Does this configuration work with ORAC? I found nowhere informations about. Is there anything I have to pay special attention for during installation?
    Thank you for helping and best regards from Berlin/Germany
    Michael Wuttke

    Forms and report services work fine on solaris 8.
    My problem is on the client side.
    I have to use solaris 8 with netscape like forms clients
    and I wasn't able to make it work with java plugins.
    Any solution?
    Mauro

  • How to configure SAP 4.7 to Oracle 9i Real Application Cluster ?

    We are using SAP 4.7 running on Oracle 9i with HP-UNIX clustering.
    If we change our existing SAP 4.7 for Oracle 9i Real Application Cluster, is it very complicated ?
    Do we need to re-install SAP and Oracle ?

    Both 4.7 and Oracle 9 are already out of support.
    If you plan to use RAC then I'd first upgrade the application and the database and then implement RAC. You won't have support for your environment without additional costs.
    Markus

Maybe you are looking for