Diff hardwares for 2 node rac cluster

Dear All,
Can i implement RAC 2 node cluster on 2 nodes with diff hardwares like diff server models?
Regards

hungry_dba wrote:
Dear All,
Can i implement RAC 2 node cluster on 2 nodes with diff hardwares like diff server models?
Hi,
Also you can read tech note below:
*RAC: Frequently Asked Questions [ID 220970.1]*
Can I have different servers in my Oracle RAC? Can they be from different vendors? Can they be different sizes?
Oracle RAC does support a cluster with nodes that have different hardware configurations. An example is a cluster with 3 nodes with 4 CPUs and another node with 6 CPUs. This can easily occur when adding a new node after the cluster has been in production for a while. For this type of configuration, customers must consider some additional features to get the optimal cluster performance. The servers used in the cluster can be from different vendors; this is fully supported as long as they run the same binaries. Since many customers implement Oracle RAC for high availability, you must make sure that your hardware vendor will support the configuration. If you have a failure, will you get support for the hardware configuration?
Regards,
Levi Pereira

Similar Messages

  • Any idea about the pricing of the hardwares for Oracle 10g RAC?

    Hi,
    If I go for 2 nodes RAC using Dell servers and Disk array, any ideas about the prices? (just for the hardwares.)
    Thanks.
    Regards,
    Jason

    It's all about the shared disk(s).
    If you are only ever going to use two (2) nodes a SCSI JBOD solution is cheap. (practical too with ASM as the cluster filesystem / volume manager)
    More than 2 nodes, you step up to certified NFS, iSCSI or FC/AL connected storage. More expensive, faster, more flexible.
    The computers are the cheapest portion of the cost especially with multiple core multiple cpu 64bit linux based servers being so inexpensive.

  • Can any one say that list of hardware for 11gr2 2node rac installation?  thank in advance

    can any one say that list of hardware for 11gr2 2node rac installation?  thank in advance

    Hi ,
    Hardware Requirements
    Physical memory (at least 1.5 gigabyte (GB) of RAM)
    An amount of swap space equal to the amount of RAM
    Temporary space (at least 1 GB) available in /tmp
    A processor type (CPU) that is certified with the release of the Oracle software being installed
    A minimum of 1024 x 786 display resolution, so that Oracle Universal Installer (OUI) displays correctly
    All servers that are used in the cluster must have the same chip architecture, for example, all 32-bit processors or all 64-bit processors
    Adequate disk space in the software installation locations to store the Oracle software
    You need at least 5.5 GB of available disk space for the Grid home directory, which includes both the binary files for Oracle Clusterware and Oracle Automatic Storage Management (Oracle ASM) and their associated log files, and at least 4 GB of available disk space for the Oracle Database home directory, or Oracle home directory.
    About Shared Storage
    These Oracle Clusterware components require the following disk space on a shared file system:
    Three Oracle Clusterware Registry (OCR) files, 300 MB each, or 900 MB total disk space
    Three voting disk files, 300 MB each, or 900 MB total disk space
    If you are not using Oracle ASM for storing Oracle Clusterware files, then for best performance and protection, you should use multiple disks, each using a different disk controller for voting disk file placement. Ensure that each voting disk is configured so that it does not have share any hardware device or have a single point of failure.
    Network Hardware Requirements
    When you configure the network for Oracle RAC and Oracle Clusterware, each node in the cluster must meet the following requirements:
    Each node must have at least two network interface cards (NIC), or network adapters. One adapter is for the public network interface and the other adapter is for the private network interface (the interconnect). Install additional network adapters on a node if that node meets either of the following conditions:
    Does not have at least two network adapters
    Has two network interface cards but is using network attached storage (NAS). You should have a separate network adapter for NAS.
    Has two network cards, but you want to use redundant interconnectsIn previous releases, to make use of redundant networks for the interconnect, bonding, trunking, teaming, or similar technology was required. Oracle Grid Infrastructure for a cluster and Oracle RAC can now make use of redundant network interconnects, without the use of other network technology, to enhance optimal communication in the cluster. This functionality is available starting with Oracle Database 11g Release 2 (11.2.0.2).Redundant Interconnect Usage enables load-balancing and high availability across multiple (up to 4) private networks (also known as interconnects).
    However, When you install Oracle software, Oracle Universal Installer (OUI) automatically performs hardware prerequisite checks and notifies you if they are not met.
    Regards,
    Pradeep. V

  • Multiple databases/instances on 4-node RAC Cluster including Physical Stand

    OS: Windows 2003 Server R2 X64
    DB: 10.2.0.4
    Virtualization: NONE
    Node Configuration: x64 architecture - 4-Socket Quad-Core (16 CPUs)
    Node Memory: 128GB RAM
    We are planning the following on the above-mentioned 4-node RAC cluster:
    Node 1: DB1 with instanceDB11 (Active-Active: Load-balancing & Failover)
    Node 2: DB1 with instanceDB12 (Active-Active: Load-balancing & Failover)
    Node 3: DB1 with instanceDB13 (Active-Passive: Failover only) + DB2 with instanceDB21 (Active-Active: Load-balancing & Failover) + DB3 with instanceDB31 (Active-Active: Load-balancing & Failover) + DB4 with instance41 (Active-Active: Load-balancing & Failover)
    Node 4: DB1 with instanceDB14 (Active-Passive: Failover only) + DB2 with instanceDB22 (Active-Active: Load-balancing & Failover) + DB3 with instanceDB32 (Active-Active: Load-balancing & Failover) + DB4 with instance42 (Active-Active: Load-balancing & Failover)
    Note: DB1 will be the physical primary PROD OLTP database and will be open in READ-WRITE mode 24x7x365.
    Note: DB2 will be a Physical Standby of DB1 and will be open in Read-Only mode for reporting purposes during the day-time, except for 3 hours at night when it will apply the logs.
    Note: DB3 will be a Physical Standby of a remote database DB4 (not part of this cluster) and will be mounted in Managed Recovery mode for automatic failover/switchover purposes.
    Note: DB4 will be the physical primary Data Warehouse DB.
    Note: Going to 11g is NOT an option.
    Note: Data Guard broker will be used across the board.
    Please answer/advise of the following:
    1. Is the above configuration supported and why so? If not, what are the alternatives?
    2. Is the above configuration recommended and why so? If not, what are the recommended alternatives?

    Hi,
    As far as i understand, there's nothing wrong in configuration except you need to consider below points while implementing final design.
    1. No of CPU on each servers
    2. Memory on each servers
    3. If you've RAC physical standby then apply(MRP0) will run on only one instance.
    4. Since you are configuring physical standby for on 3rd and 4th nodes of DB1 4 node cluster where DB13 and DB14 instances are used only for failver, if you've a disaster at data center or power failure in entire data center, you are losing both primary and secondary with an assumption that your primary and physical standby reside in same data center so it may not be highly available architecture. If you are going to use extended RAC for this configuration then it makes sense where Node 1 and Node 2 will reside in Datacenter A and Node 3 ,4 will reside in Datacenter B.
    Thanks,
    Keyur

  • Automatic restart of services on a 1 node rac cluster with Clusterware

    How do we enable a service to automaticly start-up when the db starts up?
    Thanks,
    Dave

    srvctl enable service -d DBThanks for your reply M. Nauman. I researched that command and found we do have it enabled and that it only works if the database instance was previously taken down. Since the database does not go down on an Archiver Hung error as we are using FRA with an alt location, this never kicks in and brings up the service. What we are looking for something that will trigger off of when the archive logs error and switch from FRA(Flash Recovery Area) to our Alternate disk location. Or more presicely, when it goes back to a Valid status(on the FRA - after we've run an archive log backup to clear it).
    I found out from our 2 senior dba's that our other 2 node rac environment does not suffer from this problem, only the newly created 1 node rac cluster environment. The problem is we don't know what that is(a parameter on the db or cluster or what) and how do we set it?
    Anyone know?
    Thanks,
    Gib
    Message was edited by:
    Gib2008
    Message was edited by:
    Gib2008

  • Install mulutple RAC databases on 2-node RAC cluster

    I am installing 5 RAC databases on a 2-node RAC cluster. I have setup SCAN using 3 IP addresses.
    Do I have to use SCAN listener for all databases?
    When installing the 3 database, I get ORA-12537: TNS connection closed error.
    ENV: 11gR2 2-node RH5.x
    Thanks!

    I have setup SCAN using 3 IP addresses.
    Do I have to use SCAN listener for all databases?These 3 scan ip will work for youe all database running under this cluster setup.
    you may also use VIP to make connection like 10g.
    I get ORA-12537: TNS connection closed error.Appear some connectivity/configuration isue,please try MOS doc contain detail on this.
    How to Troubleshoot Connectivity Issue with 11gR2 SCAN Name [ID 975457.1]

  • All connections are connecting to 2nd node only in a 2 Node RAC Cluster

    Hello,
    I have a 10.2.0.3 database on a two node RAC Cluster with only one service configured. This service set to be preferred on both nodes.
    However, all the connections are falling on Node2 only. Any idea where to look.
    $> srvctl config service -d PSDB
    psdbsrv1 PREF: psdb1 psdb2 AVAIL:
    Thanks,
    MM

    Application is using the following connection string.
    jdbc:oracle:thin:@(DESCRIPTION =(ADDRESS = (PROTOCOL = TCP)(HOST = PQ2-PS-db-01-vip)(PORT = 1521))(ADDRESS = (PROTOCOL = TCP)(HOST = PQ2-PS-db-02-vip)(PORT = 1521)) (LOAD_BALANCE = yes) (CONNECT_DATA =(SERVER = DEDICATED)(SERVICE_NAME = PSDBSRV1)(FAILOVER_MODE =(TYPE = SELECT)(METHOD = BASIC)(RETRIES = 180)(DELAY = 5))))
    --MM                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Data guard setup for 2 node RAC primary to 2 node RAC standby

    Hi All,
    I am going to setup data guard for 2 node RAC primary to 2 node RAC standby on Oracle 10.2.0.4. in AIX5L.
    Can you please provide the document on the above setup which is having all the steps (details).
    Also, the documents on different scenarios like
    1) If one node of standby goes down, how the redo logs will be applied. IS there any problem?
    2) If both nodes of standby are failed, how to reciver them?
    3) If one node of primary fails, is there any issue?
    4) If two nodes of primary fails, is there any issue?
    Thanks in advance,
    Mahi

    Have a look at the following location, you may find some similar documents:
    http://www.oracle.com/technology/deploy/availability/htdocs/maa.htm
    By
    http://www.oraxperts.com

  • Daily checklist maintainace activities for two node rac

    i need queries for two node rac which i can check everyday maintainace
    10.2.0.4
    windows (os)
    asm
    please provide what do i do everyday for maintainace
    please give me queries and check the health check
    thanks in advance

    Hi;
    pelase check below notes, you can see notes and can understand which process is important and what you need to check.
    RAC Assurance Support Team: RAC and Oracle Clusterware Starter Kit and Best Practices (Generic) [ID 810394.1]
    Master Note for Real Application Clusters (RAC) Oracle Clusterware and Oracle Grid Infrastructure [ID 1096952.1]
    11gR2 Clusterware and Grid Home - What You Need to Know [ID 1053147.1]
    Also see:
    daily check list for rac
    daily check list for rac
    Regard
    Helios

  • How many Scan Listeners, we have to configure for 5 Node RAC?

    Dear Professionals,
    How many scan Listeners we need to configure for 5 Node RAC? Oracle will not allow us to create More than 3 scan's? What If i have one scan Listener per each for 1st 3 node's? what about remaining two Node's? How application user will connect to 4 or 5 Node? Can you please explain..? Forgive me, If am totally wrong?
    Thanks
    Sagar

    Each of the 5 instances will register itself with the scan listener (using the instance parameter remote_listener).  Thus, the scan listener is "aware" of the database instances on the other two nodes where it is not running.  It can still redirect incoming connections to the local listeners on these nodes (registered as local_listener).
    Hemant K Chitale

  • Routing all connections through a one node in a 2 node RAC cluster

    Hi everyone
    My client has the following requirement: an active/active RAC cluster (eg node1/node2), but with only one of the nodes being used (node1) and the other sitting there just in case.
    For things like services, i'm sure this is straightforward enough - just have them set to preferred on node1 and available on node 2.
    For connections, I imagine I would just have the vips in order in the tns file, but with LOAD_BALANCING=OFF, so they go through the tns entries in order (i.e node 1, then node 2), so this would still allow the vip to failover if node 1 is down.
    Does that sound about right? Have I missed anything?
    Many thanks
    Rup

    user573914 wrote:
    My client has the following requirement: an active/active RAC cluster (eg node1/node2), but with only one of the nodes being used (node1) and the other sitting there just in case.Why? What is the reason for a "+just in case+" node - and when and how is is "enabled" when that just-in-case situation occurs?
    This does not many any kind of sense from a high availability or redundancy view.
    For connections, I imagine I would just have the vips in order in the tns file, but with LOAD_BALANCING=OFF, so they go through the tns entries in order (i.e node 1, then node 2), so this would still allow the vip to failover if node 1 is down.
    Does that sound about right? Have I missed anything?Won't work on 10g - may not work on 11g. The Listener can and does handoff connections, depending on what the TNS connection string say. If you do not connect via a SID entry but via a SERVICE entry, and that service is available on multiple nodes, you may not (and often will not) be connected to instance on the single IP that you used in your TNS connection.
    Basic example:
    // note that this TEST-RAC alias refers to a single specific IP of a cluster, and use
    // SERVICE_NAME as the request
    /home/billy> tnsping test-rac
    TNS Ping Utility for Linux: Version 10.2.0.1.0 - Production on 18-JAN-2011 09:06:33
    Copyright (c) 1997, 2005, Oracle.  All rights reserved.
    Used parameter files:
    /usr/lib/oracle/xe/app/oracle/product/10.2.0/server/network/admin/sqlnet.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS=(PROTOCOL=TCP)(HOST= 196.1.83.116)(PORT=1521)) (LOAD_BALANCE=no) (CONNECT_DATA=(SERVER=shared)(SERVICE_NAME=myservicename)))
    OK (50 msec)
    // now connecting to the cluster using this TEST-RAC TNS alias - and despite we listing a single
    // IP in our TNS connection, we are handed off to a different RAC node (as the service is available
    // on all nodes)
    // and this also happens despite our TNS connection explicitly requesting no load balancing
    /home/billy> sqlplus scott/tiger@test-rac
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jan 18 09:06:38 2011
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, Real Application Clusters, Data Mining and Real Application Testing options
    SQL> !lsof -n -p $PPID | grep TCP
    sqlplus 5432 billy    8u  IPv4 2199967      0t0     TCP 10.251.93.58:33220->196.1.83.127:37031 (ESTABLISHED)
    SQL> So we connected to RAC node 196.1.83.116 - and that listener handed us off to RAC node 196.1.83.127. The 11gr2 Listener seems to behave differently - it does not do a handoff (from a quick test I did on a 11.2.0.1 RAC) in the above scenario.
    This issue aside - how do you deal with just-in-case situation? How do you get clients to connect to node 2 when node 1 is down? Do you rely on the virtual IP of node 1 to be switched to node 2? Is this a 100% safe and guaranteed method?
    It can take some time (minutes, perhaps more) for a virtual IP address to fail over to another node. During that time, any client connection using that virtual IP will fail. Is this acceptable?
    I dunno - I dislike this concept of your client of treating the one RAC node as some kind of standby database for a just-in-case situation. I fail to see any logic in that approach.

  • For 4 node RAC database on Exadata x3-2, how many scan IP I should assign?

    We want to build a 4 node RAC 11gr23 on exadata x3-2.  Oracle only recommend 3 ip for scan listener. by previous observation of non-exadata RAC, seems each node will spread randomly one Scan Ip address. So if we assign 3 ip to one scan listener, how that will work on a 4-node RAC?
    Thanks

    You're welcome
    by Load balancing
    Oracle RAC databases increase load balancing by using the single client access name (SCAN) for the cluster to balance connections among all instances in a cluster database. SCAN also adds location independence for the databases, so that client configuration does not have to depend on which nodes are running a particular database or database instance.
    http://docs.oracle.com/cd/E11882_01/install.112/e48195/undrstnd.htm#RIWIN607
    Based on the environment, the following actions occur when you use SCAN to connect to an Oracle RAC database using a service name. The numbered actions correspond to the arrows shown in Figure 7-1.
    http://docs.oracle.com/cd/E11882_01/install.112/e48195/undrstnd.htm#CIHIBEIH
    Regards

  • How many SCAN listners for 2 node rac ?

    Hi,
    Oracle recommends min 3 SCAN listeners , but what if I've got 2 node RAC ? :)
    Regards
    GregG

    Hi Friend
    About SCAN :
    1. The single client access name (SCAN) is the address used by all clients connecting to the cluster.
    2. The SCAN name is a domain name registered to three IP addresses, either in the domain name service (DNS) or the Grid Naming Service (GNS).
    3. The SCAN name eliminates the need to change clients when nodes are added to or removed from the cluster.
    4. Clients using SCAN names can also access the cluster using EZCONNECT.
    5. The Single Client Access Name (SCAN) is a domain name that resolves to all the addresses allocated for the SCAN name.
    6. Three IP addresses should be provided (in DNS) to use for SCAN name mapping as this ensures high availability.
    7. During Oracle Grid Infrastructure installation, listeners are created for each of the SCAN addresses, and Oracle Grid Infrastructure controls which server responds to a
    SCAN address request.
    8. The SCAN addresses need to be on the same subnet as the VIP addresses for nodes in the cluster.
    Note 1 : SCAN IPs must NOT be in the /etc/hosts file, the SCAN name must be resolved by DNS.
    Note 2 : When you are installing Grid Infrastructure it will ask for the following things :
    *1. Cluster Name :*
    *2. Scan Name :*
    *3. Scan Port :*
    Thanks
    LaserSoft

  • Theoretical question on having "stand by" instance on 4 node RAC cluster.

    We have been having discussions on how we can improve the overall stability and performance of our RAC cluster. Currently we are running multiple databases on a 2 node cluster. The idea is that we will add two more nodes to the cluster and then spread the instances across the 4 nodes but only have two active and 1 as a standby in the event of node failure.
    In other words:
    Node1, Node2, and Node3 would run DB1
    Node 4 Node3 and Node2 would run DB2
    Additionally DB1_Instance3 would be shutdown on Node3 and only started if Node1 or 2 failed.
    This would be the same for DB2 where DB2_Instance3 on Node2 would only be started if Node3 or Node4 failed.
    Underneath them all would be the same ASM database shared for all 4 nodes.
    Has anyone seen such a configuration? Is anyone aware of a white paper that discusses it.

    the parameter active_instance_count is a 9i parameter and should only be set if you want to run RAC with a 2 node cluster - one instance active (all users connected here) and the second instance up but no one actively using it. IF instance one goes away, all work is move to instance 2. This is hot failover .
    If you are adding a new instance to a rac database, you are recommended to use DBCA or EM to add the instance. If you do not want to have users connected to the instance all the time, you can either leave it running but do not enable any services on it until you need it or stop the instance with srvctl stop instance. The cluster will not try to restart the instance automatically.
    The services option, when you create a service, you define which instances you wish it to run on. You can dynamically change where it runs. For the end user, they just connect to the service, the listener directs the connection to the least loaded instance providing the service at connect time. See the workload management with RAC 10g white paper on otn (otn.oracle.com/rac) or Chapter 6 of admin & deploy guide

  • SOA with Oracle 2 node RAC cluster

    Hi All,
    Just a simple doubt, I have successfully installed and configured SOA suite 11.1.1.3 & BAM in one wls Domain 10.3.3 in a linux box and could access all the application like BAM console, BPEL console etc .... also could see all my data-sources deployed in the Data Sources with a single node database.
    1. Now I have to RE-configure this whole SOA suite with RAC (2 node database cluster) what changes or configuration needed to implement SOA suite with RAC database?
    2. Do I need to create a "Multi Data Source" to configure RAC with SOA suite..?
    Thanks
    Sam

    DB wrote:
    This is regarding Oracle RAC..so if there is a specific category..please let me know..
    I have installed OEL linux 5.6 as guest OS (using virtualbox) in two laptops.
    I want to install 2 node oracle 10gR2 RAC with the OEL linux as OS and each laptop as one node.
    Read docs and understood that there must be shared storage for oracle clusterware and oracle ASM for oracle RAC to work.
    Please let me know the steps to create shared storage for oracle clusterware and oracle ASM (considering virtualbox OEL) and to configure public,private and virtual IPs.
    I already have document to create 2node oracle RAC using virtualbox with two nodes on the same laptop.so please dont suggest that doc.
    Thanks,
    DBMay be my step by step RAC installation guide can help you somehow?
    http://kamranagayev.wordpress.com/2011/04/05/step-by-step-installing-oracle-10g-rac-on-vmware/

Maybe you are looking for