Oracle 11gR2 RAC setup with ASM option running on two nodes . Seeking help

Hi ,
Recently we have migrated to 11gR2 extended RAC for one of our application using ASM . Its a two node RAC setup . We are using ASM Normal redundancy feature maintaining two failure group .
I need few clarification regarding our new DR set up .
We are planning to have EMC SRDF replication from any of node storage ( Either 1 or 2) to the DR disks .
Its stright to use when we use the normal file systems . In case of ASM , I am not sure how this replication will work ? and on which level the replication should happen ?
Do i need an ASM instance up and running on DR server , Which will recognize the disks as it is once DR disks are separated from prod replication and assigned to the server ?
What should be prerequisite on DR server before we can start the srdf replication (EMC) and make it work .
Thanks for All your Valuble suggestion .

Before you decide to use EMC replication give considerable thought to the fact that you are introducing a single point of failure.
Hardware snapshot replication is very fast. That is all it is. If you database crashes because of corruption on disk ... your replica will be equally corrupt.
What is fast can not do what DataGuard does ... verify quality. If the point is DR. Go with DataGuard unless EMC is willing to, in writing (not verbally) give you a guarantee (in hard cash) to cover all costs if they replicate bad blocks and it causes an outage.
And just so no one thinks I'm picking o EMC ... the same weakness exists with all of the other storage vendor's products.

Similar Messages

  • Reinstalling Oracle 11gR2 RAC Grid Problem - ASM Disks Group

    Folks,
    Hello.
    I am installing Oracle 11gR2 RAC using 2 Virtual Machines (rac1 and rac2 whose OS are Oracle Linux 5.6) in VMPlayer.
    I have been installing Grid Infrastructure using runInstaller in the first VM rac1 from step 1 to step 9 of 10.
    On the step 9 of 10 in the Wizard, accidentally, I touch the Mouse, and the Wizard is gone.
    The directory for installing Grid in the 2 VMs is the same: /u01
    In order to make sure everything is correct, I delete entire directory /u01 in the 2 VMs and install Grid in rac1 again.
    I have understood it's not the right way to delete /u01. The right way is to follow the tutorial
    http://docs.oracle.com/cd/E11882_01/install.112/e22489/rem_orcl.htm#CBHEFHAC
    But I have deleted /u01 and need to fix one by one. I install Grid again and get the error message on step 5 of 9 as follows:
    [INS - 30516] Please specify unique disk groups.
    [INS-3050] Empty ASM disk group.
    Cause - Installer has detected the disk group name provided already exists on the system.
    Action - Specify different disk group.
    In Wizard, the previous Disk Group name is "DATA" and its Candidate disks (5 ASMDISKs) are gone. I try to use a different name "DATA2", but no ASMDISKs come up under "Candidate disks". For "ALL Disks", all ASMDISKs cannot be selected.
    I want to use the same ASM disk group "DATA" and don't want to create a new disk group.
    My question is:
    How to have the previous ASM disks and its group "DATA" come up under "Candidate Disks" so that can use it again ?
    Thanks.

    Hi, in case this helps anyone else. I got this INS-30516 error too was stumped for little while. I have 2 x 2-node RAC which are hitting same SAN. The first-built RAC has a DATA diskgroup. When went to build second RAC on new ASM disk new DIskgroup (but same diskgroup name DATA) got INS-30516 about diskgroup name already in use etc. Finally figured out all that was required was to restrict diskstring using button in installer to only retrieve the LUNS for this RAC (this was quick and dirty - all LUNS for both RAC being presented to both RAC). Once diskstring only searched for the LUNS required for this RAC only, e.g.
    ORCL:DATA_P* (for DATA_PD and FRA_PD)
    the error went away.
    I also have DATA_DR and FRA_DR presenting to both RAC. Apparently it scans the header and if it finds a diskgroup name that is already in use based on the diskstring scan it will not allow reuse of the diskgroup name since it has no way of knowing that the other ASM disks are for a different RAC.
    HTH

  • Oracle 11gR2 RAC problem with resource state

    Hi all,
    I installed Oracle 11gR2 grid infrastructure with 2 nodes and I installed DB 11gR2.
    S.O: HP-UX
    I actived both DB instance in each node.
    For an hardware problem node 1 become unstable (continuos auto reboot).
    I found that the problem was RAM.
    However I note that database resource is in a particular state and i don't able to reset it.
    Performing command crsctl status resource ora.orcl.db this is the result
    ora.orcl.db
    1 OFFLINE UNKNOWN node1 Startup Initiated
    2 ONLINE ONLINE node2 Open
    That UNKNOWN state is really abstruse.
    I tryed to perform crsctl stop resource ora.orcl.db -n node1 and the result is
    CRS-2679: Attempting to clean 'ora.orcl.db' on 'node1'
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    HPUX-ia64 Error: 2: No such file or directory
    Process ID: 0
    Session ID: 0 Serial number: 0
    CRS-2680: Clean of 'ora.orcl.db' on 'node1' failed
    CRS-4000: Command Stop failed, or completed with errors.
    I tryed to perform crsctl start resource ora.orcl.db -n node1 and the result is
    CRS-2662: Resource 'ora.orcl.db' is disabled on server 'node1'
    CRS-4000: Command Start failed, or completed with errors.
    How do I do to reset that UNKNOWN state?
    Thanks in advance.
    Bye
    Alessandro

    I tryed srvctl enable instance -d orcl -i ORCL_1
    but the results is
    srvctl enable instance command is not supported for configuration using server pool.
    I tryed to delete service ora.orcl.db and recreate it.
    Now I have
    NAME=ora.orcl.db
    TYPE=ora.database.type
    TARGET=ONLINE , ONLINE
    STATE=UNKNOWN on node1, OFFLINE
    So Targets are both ONLINE, but if I write crsctl start resource ora.orcl.db
    the result is
    CRS-2679: Attempting to clean 'ora.orcl.db' on 'node1'
    CRS-2672: Attempting to start 'ora.orcl.db' on 'node2'
    CRS-5003: Invalid attribute value: '' for attribute DB_UNIQUE_NAME
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    HPUX-ia64 Error: 2: No such file or directory
    Process ID: 0
    Session ID: 0 Serial number: 0
    CRS-2674: Start of 'ora.orcl.db' on 'node2' failed
    CRS-2679: Attempting to clean 'ora.orcl.db' on 'node2'
    CRS-2681: Clean of 'ora.orcl.db' on 'node2' succeeded
    CRS-2632: There are no more servers to try to place resource 'ora.orcl.db' on that would satisfy its placement policy
    CRS-2680: Clean of 'ora.orcl.db' on 'node1' failed
    CRS-4000: Command Start failed, or completed with errors.
    Where DB_UNIQUE_NAME attribute must be set?
    Any other suggest?
    Thanks in advance.
    Regards.
    Alessandro
    Edited by: Alessandro Zenoni on 21-giu-2010 11.26

  • Oracle 10g RAC design with ASM and OCFS

    Hi all,
    I have a question about a proposed Oracle 10g Release 2 RAC design for a 2 node cluster.
    ASM can store database files but not Oracle binaries nor OCR and voting disk. As such, OCFS version 1 does not support a shared Oracle Home. We plan to use OCFS version 2 with ASM version 2 on Red Hat Linux Enteprrise Server 4 with Oracle 10g Release 2 (10.2.0.1).
    For OCFS v2, a shared Oracle home and shared OCR and voting disk are supported. My question is does the following proposed architecture make sense for OCFS v2 with ASM v2 on Red Hat Linux 4?
    Oracle 10g Release 2 on Red Hat Enterprise Linux Server 4:
    OCFS V2:
    - shared Oracle home and binaries
    - shared OCR and vdisk files
    - CRS software shared OCFS v2 filesystem
    - spfile
    - controlfiles
    - tnsnames.ora
    ASM v2 with ASMLib v2:
    Proposed ASM disk groups:
    - data_dg for application data
    - backupdg for flashback and archivelogs
    - undo_rac1dg ASM diskgroup for undo tablespace for racnode1
    - undo_rac2dg ASM diskgroup for undo tablespace for racnode2
    - redo_rac1dg ASM diskgroup to hold redo logs for racnode1
    - redo_rac2dg ASM diskgroup to hold redo logs for racnode2
    - temp1dg temp tablespace for racnode1
    - temp2dg temp tablespace for racnode2
    Does this sound like a good initial design?
    Ben Prusinski, Senior DBA

    OK Tim, thanks for advices.
    I think Netbackup can be integrated with RMAN but I don't want to loose time on this (political).
    To summarize:
    ORACLE_HOME and CRS_HOME on each node (RAID1 and NTFS)
    Shared storage:
    Disk1 and disk 2: RAID1: - Raw partition 1 for OCR
    - Raw partition 2 for VotingDisk
    - OCFS for FLASH_RECOVERY_AREA
    Disk3, disk4 and disk5: RAID 0 - Raw with ASM redundancy normal 1 diskgroup for database files.
    This is a running project here, will start testing the design on VMware and then go for production setup.
    Regards

  • Oracle 11gR2 RAC installation with DNS

    Dear Guru's,
    I want to configure RAC on two node cluster with DNS(For SCAN) in my testing server.
    O.S: RHEL5.4(64 bit)
    Oracle Version: 11gR2(11.2.0.2)
    Note: Am creating nodes in VMware ESXi server.
    Just I want to know whether DNS should be configure in any one of those NODES or in separate machine.
    Bala :)

    Hi Bala,
    I want to configure my Database with & without DNS. (Will try both installation in different servers)I think, you have already received a link on very good article about Oracle RAC installation without DNS and DHCP (I mean Jeffrey Hunter's article).
    There is one more about Oracle RAC installation with DNS and DHCP (http://gjilevski.com/2011/10/05/build-two-node-oracle-rac-11gr2-11-2-0-3-with-gns-dns-dhcp-and-haip/).
    And want to know how it forwards the client connection if a node fails.. Can anyone update the URL for this..There are a lot of information about it:
    - http://www.oracle.com/technetwork/database/features/oci/taf-10-133239.pdf
    - http://docs.oracle.com/cd/E14072_01/java.112/e10589/apxracfan.htm
    - http://www.oracle.com/technetwork/database/app-failover-oracle-database-11g-173323.pdf
    Very good videos:
    - http://www.dsvolk.ru/oracle/racdd4d/demos/video/taf/session/session_viewlet_swf.html
    - http://www.dsvolk.ru/oracle/racdd4d/demos/video/taf/select/select_viewlet_swf.html
    - http://www.dsvolk.ru/oracle/racdd4d/demos/video/taf/callback/callback_viewlet_swf.html
    - http://www.dsvolk.ru/oracle/racdd4d/demos/video/fan/fan_viewlet_swf.html
    Hope it helps,
    Best regards,
    Gena

  • Oracle 11gR2 RAC with Sun clsuter 3.3 clsetup not showing  ASMDiskGroup RS

    Hello,
    I am installing Oracle 11gR2 RAC on Solaris cluster 3.3 with hardware RAID and ASM
    I was able to register rac-famework-rg and crs-framework-rg.
    Oracle RAC is up and running (confirmed from crs_stat -t -v)
    However, while creating option 5. Automatic Storage Management (ASM) on clsetup, I was getting ERROR: ASM is either not installed or the installation is invalid!
    I have created the asm-inst-rg with resource asm-inst-rs and asm-dg-rg with resources asm-dg-fra, asm-dg-data and asm-dg-crs using CLI. Please find the complete clrg and clrs status output below:
    root@rac1 # clrg status
    === Cluster Resource Groups ===
    Group Name Node Name Suspended Status
    rac-framework-rg rac1 No Online
    rac2 No Online
    asm-inst-rg rac1 No Online
    rac2 No Online
    asm-dg-rg rac1 No Online
    rac2 No Online
    root@rac1 # clrs status
    === Cluster Resources ===
    Resource Name Node Name State Status Message
    crs_framework-rs rac1 Online Online
    rac2 Online Online
    rac-udlm-rs rac1 Online Online
    rac2 Online Online
    rac-framework-rs rac1 Online Online
    rac2 Online Online
    asm-inst-rs rac1 Online Online - +ASM1 is UP and ENABLED
    rac2 Online Online - +ASM2 is UP and ENABLED
    asm-dg-fra rac1 Online Online - Mounted: FRA
    rac2 Online Online - Mounted: FRA
    asm-dg-crs rac1 Online Online - Mounted: CRS
    rac2 Online Online - Mounted: CRS
    asm-dg-data rac1 Online Online - Mounted: DATA
    rac2 Online Online - Mounted: DATA
    Once I reach clsetup option 6: Resources for Oracle Real Application Clusters Database Instances, I was able to select cluster db name, oracle home, SID for both nodes, and it listed the asm-inst-rg and asm-inst-rs for selection of "Specify Oracle ASM Instance Resource".
    The problem is the the next selection which is "Specify Oracle ASM Disk Group Resources" which does NOT list any asm-dg-rg or asm-dg-fra or anything, instead, it throws error "No existing disk-group resources available for selection. Run ASM subwizard create new resource(s) and come back to this panel for the selection."
    Kindly guide me to create Resources for Oracle Real Application Clusters Database Instances either in CLI or clsetup
    Thanks and Regards
    Ushas Symon

    Solaris Cluster 3.3 u1 solved the issue.
    The problem right now I am facing is IPMP on the public interface.
    When I remove the cable on the bge0, the ip on bge0 and the SCAN IP's are plumbed to bge1 which is in sc_ipmp0, but NOT the virtual IP!!! and the listeners on both nodes go offiline!!!
    any inputs on the configuration will be highly appreciated..
    Thanks and Regards
    Ushas Symon

  • Oracle 11gR2 RAC: Running the script root.sh problem

    Folks,
    Hello. I am installing Oracle 11gR2 RAC using 2 Virtual Machines (rac1 and rac2 whose OS are Oracle Linux 5.6) in VMPlayer and according to the tutorial
    http://appsdbaworkshop.blogspot.com/2011/10/11gr2-rac-on-linux-56-using-vmware.html
    I have been installing Grid Infrastructure using runInstaller from step 1 to step 9 of 10 in both VM rac1 and rac2.
    Now, I am running the script root.sh in VM rac1 and rac2 as below:
    [root@rac1 /]# /u01/app/grid/root.sh
    Output:
    CRS-4123: ohasd is starting
    FATAL: Module oracleoks not found
    FATAL: Module oracleadvm not found
    FATAL: Module oracleacfs not found
    ACFS-9121: Failed to detect /dev/asm/.asm_ctl_spec
    ACFS-9310: ADVM/ACFS installation failed.
    ACFS-9311: not all components were detected after installation
    'UpdateNodeList' was successful.
    [root@rac1 /]#
    [root@rac2 /]# /u01/app/grid/root.sh
    Output:
    CRS-4123: ohasd is starting
    FATAL: Module oracleoks not found
    FATAL: Module oracleadvm not found
    FATAL: Module oracleacfs not found
    ACFS-9121: Failed to detect /dev/asm/.asm_ctl_spec
    ACFS-9310: ADVM/ACFS installation failed.
    ACFS-9311: not all components were detected after installation
    Start of resource "ora.asm_init" failed.
    Failed to start ASM.
    Failed to start Oracle Clusterware Stack.
    [root@rac2 /]#
    As we see the output above, rac1 and rac2 get the same problems. in addition, rac2 fails to start ASM and Clusterware stack. Thus, I have 2 questions:
    First, the common problem for rac1 and rac2 :
    1)Module "oracleoks, oracleadvm, oracleacfs" not found
    2)/dev/asm/.asm_ctl_spec not detected
    3)ADVM/ACFS installation failed
    Do the above 3 factors affect Grid and DataBase Installation later ? If yes, how to solve these problems ?
    Second, how to start ASM and Clusterware stack in rac2 ?
    Thanks.

    you have 2 options:
    1. OEL 5.6 comes with 2 kernels, the original Red Hat, and Oracle UEK, you can choose not to use UEK
    or the one i would recommend:
    2. Upgrade Grid Infrastructure,
    *11.2.0.3.1* (patch 13348650, released about a 1 month ago) supports ACFS on UEK (starting with 2.6.32.200 IIRC)

  • Oracle 11gR2 RAC and DR setup

    Environment:
    Oracle 11gR2 RAC (2 node) with ASM
    HP-UX B.11.31 U ia64
    We have configured two ASM disk group for Data & FRA, and we configured RMAN backup into FRA location. Now our Database, Log file, Archive logs and RMAN backup files all are in ASM on primary site.
    Our network team is using Recover Point to mirror all the disks into DR site; we have no setup from Oracle side. What is the possible configuration that we need to have in DR site from Oracle side in order to bring-up the database in-case of primary site failure (disaster).
    I am new to Oracle RAC can some one give me a suggestion.
    - Mano

    What is the possible configuration that we need to have in DR site from Oracle side in order to bring-up the database in-case of primary site failure ORACLE DATAGUARD is what you need; not RAC

  • Oracle 10g R2 installation with ASM+RAC

    Gurus,
    Need some suggestuon on Oracle 10g R2 installation with ASM and RAC option.
    We have found many documents on the Oracle, HP, HP-Oracle CTC and third party web sites, but nothing that is specific to this particular combination of separate
    ORACLE_HOMEs, ASM and 10g RAC CRS. It is unclear for me from the documentation how this combination of ASM and 10.2g RAC may best be installed.
    The high level steps i got after reading lot of docs as follows - but i am not sure whether these are correct or not. if they are correct, can any one share their experience/notes please?
    1) Install CRS
    2) Install RDBMS for ASM HOME - create separater oracle home for ASM instance using OUI
    3) Install RDBMS for RAC Database Home - create separater oracle home for RAC database using OUI
    4) Create ASM database using DBCA -
    5) Use dbca to create database.

    Oracle provides 'paint by numbers' tutorials called 'Oracle By Example'. (Go to OTN, check under the Training tab)
    They have one for a Windows based ASM/RAC that you might want to review. Not your specific environment, but the steps will be dag-nabbed close.
    I recommend walking the path (http://otn.oracle.com >> training:OBE >> Database 10g Release 1:VMWare:Installation
    http://www.oracle.com/technology/obe/obe10gdb_vmware/install/racinstallwin2k/racinstallwin2k.htm

  • Oracle 11gR2 RAC clusterware not picking up ASM diskgroups

    Hello: I had to deinstall a failed oracle 11gR2 rac cuslterware. But, not the asm disks are not in the OUI pick list. How can I resolve that?
    The storage is on SAN partitions.

    Probably the install already put something on the disks.
    Clean them with dd if=/dev/zero of=<your asm disks> bs=1M count=10
    Regards
    Sebastian

  • The best option to create  a shared storage for Oracle 11gR2 RAC in OEL 5?

    Hello,
    Could you please tell me the best option to create a shared storage for Oracle 11gR2 RAC in Oracel Enterprise Linux 5? in production environment? And could you help to create shared storage? Because there is no additional step in Oracle installation guide. There are steps for only asm disk creation.
    Thank you.

    Here are names of partitions and permissions. Partitions which have 146 GB, 438 GB, 438 GB of capacity are my storage. Two of three disks which are 438 GB were configured as RAID 5 and remaining disk was configured as RAID 0. My storage is Dell MD 3000i and connected to nodes through ethernet.
    Node 1
    [root@rac1 home]# ll /dev/sd*
    brw-r----- 1 root disk 8, 0 Aug 8 17:39 /dev/sda
    brw-r----- 1 root disk 8, 1 Aug 8 17:40 /dev/sda1
    brw-r----- 1 root disk 8, 16 Aug 8 17:39 /dev/sdb
    brw-r----- 1 root disk 8, 17 Aug 8 17:39 /dev/sdb1
    brw-r----- 1 root disk 8, 32 Aug 8 17:40 /dev/sdc
    brw-r----- 1 root disk 8, 48 Aug 8 17:41 /dev/sdd
    brw-r----- 1 root disk 8, 64 Aug 8 18:26 /dev/sde
    brw-r----- 1 root disk 8, 65 Aug 8 18:43 /dev/sde1
    brw-r----- 1 root disk 8, 80 Aug 8 18:34 /dev/sdf
    brw-r----- 1 root disk 8, 81 Aug 8 18:43 /dev/sdf1
    brw-r----- 1 root disk 8, 96 Aug 8 18:34 /dev/sdg
    brw-r----- 1 root disk 8, 97 Aug 8 18:43 /dev/sdg1
    [root@rac1 home]# fdisk -l
    Disk /dev/sda: 72.7 GB, 72746008576 bytes
    255 heads, 63 sectors/track, 8844 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sda1 * 1 8844 71039398+ 83 Linux
    Disk /dev/sdb: 72.7 GB, 72746008576 bytes
    255 heads, 63 sectors/track, 8844 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sdb1 * 1 4079 32764536 82 Linux swap / Solaris
    Disk /dev/sdd: 20 MB, 20971520 bytes
    1 heads, 40 sectors/track, 1024 cylinders
    Units = cylinders of 40 * 512 = 20480 bytes
    Device Boot Start End Blocks Id System
    Disk /dev/sde: 146.2 GB, 146278449152 bytes
    255 heads, 63 sectors/track, 17784 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sde1 1 17784 142849948+ 83 Linux
    Disk /dev/sdf: 438.8 GB, 438835347456 bytes
    255 heads, 63 sectors/track, 53352 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sdf1 1 53352 428549908+ 83 Linux
    Disk /dev/sdg: 438.8 GB, 438835347456 bytes
    255 heads, 63 sectors/track, 53352 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sdg1 1 53352 428549908+ 83 Linux
    Node 2
    [root@rac2 ~]# ll /dev/sd*
    brw-r----- 1 root disk 8, 0 Aug 8 17:50 /dev/sda
    brw-r----- 1 root disk 8, 1 Aug 8 17:51 /dev/sda1
    brw-r----- 1 root disk 8, 2 Aug 8 17:50 /dev/sda2
    brw-r----- 1 root disk 8, 16 Aug 8 17:51 /dev/sdb
    brw-r----- 1 root disk 8, 32 Aug 8 17:52 /dev/sdc
    brw-r----- 1 root disk 8, 33 Aug 8 18:54 /dev/sdc1
    brw-r----- 1 root disk 8, 48 Aug 8 17:52 /dev/sdd
    brw-r----- 1 root disk 8, 64 Aug 8 17:52 /dev/sde
    brw-r----- 1 root disk 8, 65 Aug 8 18:54 /dev/sde1
    brw-r----- 1 root disk 8, 80 Aug 8 17:52 /dev/sdf
    brw-r----- 1 root disk 8, 81 Aug 8 18:54 /dev/sdf1
    [root@rac2 ~]# fdisk -l
    Disk /dev/sda: 145.4 GB, 145492017152 bytes
    255 heads, 63 sectors/track, 17688 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sda1 * 1 8796 70653838+ 83 Linux
    /dev/sda2 8797 12875 32764567+ 82 Linux swap / Solaris
    Disk /dev/sdc: 146.2 GB, 146278449152 bytes
    255 heads, 63 sectors/track, 17784 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sdc1 1 17784 142849948+ 83 Linux
    Disk /dev/sdd: 20 MB, 20971520 bytes
    1 heads, 40 sectors/track, 1024 cylinders
    Units = cylinders of 40 * 512 = 20480 bytes
    Device Boot Start End Blocks Id System
    Disk /dev/sde: 438.8 GB, 438835347456 bytes
    255 heads, 63 sectors/track, 53352 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sde1 1 53352 428549908+ 83 Linux
    Disk /dev/sdf: 438.8 GB, 438835347456 bytes
    255 heads, 63 sectors/track, 53352 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sdf1 1 53352 428549908+ 83 Linux
    [root@rac2 ~]#
    Thank you.
    Edited by: user12144220 on Aug 10, 2011 1:10 AM
    Edited by: user12144220 on Aug 10, 2011 1:11 AM
    Edited by: user12144220 on Aug 10, 2011 1:13 AM

  • Installing Oracle 11gR2 RAC Problem: ASM disks

    Folks,
    Hello. I am installing Oracle 11gR2 RAC using 2 Virtual Machines (rac1 and rac2 whose OS are Oracle Linux 5.6) in VMPlayer and according to the tutorial
    http://appsdbaworkshop.blogspot.com/2011/10/11gr2-rac-on-linux-56-using-vmware.html
    At the first time, I create VM rac1. While its OS Oracle Linux 5.6 is booting, initializing ASMLib Driver OK. I create 5 ASM disks successfully using the command:
    [root@rac1 /]# /etc/init.d/oracleasm createdisk ASMDISK1 /dev/sdb1
    Output: Marking disk "ASMDISK1" as an ASM disk: OK
    Because the hard disk space of rac1 is not enough to install Database, I create VM rac1 again with bigger hard disk space (30GB). I do everything the same with that at the first time. But this time, While OS Oracle Linux 5.6 is booting, initializing ASMLib Driver failed. I create 5 ASM disks using the same command:
    [root@rac1 /]# /etc/init.d/oracleasm createdisk ASMDISK1 /dev/sdb1
    Output: Marking disk "ASMDISK1" as an ASM disk: failed
    But when I use the command:
    [root@rac1 /]# /etc/init.d/oracleasm listdisks
    Output: ASMDISK1 ASMDISK2 ASMDISK3 ASMDISK4 ASMDISK5
    My questions are:
    First, can the 5 disks "ASMDISK1 ASMDISK2 ASMDISK3 ASMDISK4 ASMDISK5" be used correctly in spite of Marking disk "ASMDISK1(2 3 4 5)" as an ASM disk failed ?
    Second, how to fix OS Oracle Linux 5.6 so that initializing ASMLib Driver OK while booting ?
    Thanks.

    Folks,
    Hello. The issue is solved by myself. Thanks.

  • Oracle 11gR2 RAC VM and SCAN and DNS and /etc/hosts (two) setup questions

    Hi,
    I am looking forward to setting up two Oracle 11gR2 RAC instances
    on my Oracle VM test machine.
    I plan on using the Oracle 11gR2 RAC VM template.
    I want the final Oracle 11gR2 RAC instances to have SCAN that uses DNS.
    The DNS will be pre-installed in the JeOS.
    My first simple question about the setup is the following.
    In my DNS name file, for example,
    /var/named/chroot/var/named/milkyway.univ.db
    do I need to provide the racnode1 and racnode2 information,
    for example,
    # DNS name file (snippet)
    myjeos IN A 192.168.1.150
    racnode1 IN A 192.168.1.161
    racnode1-vip IN A 192.168.1.163
    racnode2 IN A 192.168.1.162
    racnode2-vip IN A 192.168.1.164
    rac-scan IN A 192.168.1.131
    rac-scan IN A 192.168.1.132
    rac-scan IN A 192.168.1.133
    Or, can I just provide only the rac-scan information
    # DNS name file alternate (snippet)
    myjeos IN A 192.168.1.150
    rac-scan IN A 192.168.1.131
    rac-scan IN A 192.168.1.132
    rac-scan IN A 192.168.1.133
    What I am getting at is the following.
    Within the install process, will racnode1, racnode1-vip, racnode2,
    and racnode2-vip host names and their IP address be written
    to the RAC instances /etc/hosts files? (So I should not bother
    to put them in the DNS name file like '# DNS name file alternate (snippet)'?)
    Or, should I put the racnode and racnode-vip host names and IP addresses
    in the DNS name file like '# DNS name file (snippet)'?
    The second question is the following.
    Are the cluster name and the scan name allowed to be different?
    Currently, I would plan them to be different,
    for example, rac-cluster and rac-scan.
    Or, are they required to be the same,
    for example, rac-cluster and rac-cluster.
    Thank you.
    AIM

    AIM wrote:
    do I need to provide the racnode1 and racnode2 information,
    Or, can I just provide only the rac-scan information You need to provide all of it in DNS, because other hosts in your network will need to be able to resolve all of the normal, VIP and SCAN addresses for your RAC nodes. We write this data out to /etc/hosts just to reduce the amount of round-trip DNS requests the cluster nodes make for themselves.
    Are the cluster name and the scan name allowed to be different?They can be different.

  • Oracle 11gR2 RAC in LDOM Network issue

    Hi, Requesting your expert advise regarding this configuration.
    We are implementing LDOM 2.2 on two SPARC T4-4 for Oracle 11gR2 RAC; Solaris 10 U10 on both control and guest domain. The setup for each primary/control domain is: Two 10g links aggregated and have four VLAN trunked on the aggregate. vSwitch created using the aggr as the device as following per T4-4:
    NOTE: VLAN 1501 is for data connection and VLAN 10 is for heartbeat for one RAC cluster and VL 1601 and 11 is for another RAC. all together four LDOMS.
    ldm add-vswitch vid=1501,1601,10,11 net-dev=aggr1 primary-vsw0 primary
    ldm add-vnet pvid=1501 vnetprod primary-vsw0 guest1
    ldm add-vnet pvid=10 vnethb primary-vsw0 guest1
    ldm add-vnet pvid=1601 vnetprod primary-vsw0 guest2
    ldm add-vnet pvid=11 vnethb primary-vsw0 guest2
    vnet inside the LDOM are not tagged:
    vnet1: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 3
    inet 10.220.128.20 netmask ffffff80 broadcast 10.220.128.127
    ether 0:14:4f:f9:ec:7f
    vnet2: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 4
    inet 192.168.2.11 netmask ffffff80 broadcast 192.168.2.127
    ether 0:14:4f:fb:2b:8f
    Here is the whole configuration:
    root@gp-cpu-suh004 # ldm -V
    Logical Domains Manager (v 2.2.0.0)
    Hypervisor control protocol v 1.9
    Using Hypervisor MD v 1.4
    System PROM:
    Hostconfig v. 1.2.0. @(#)Hostconfig 1.2.0.a 2012/05/11 07:34
    Hypervisor v. 1.11.0. @(#)Hypervisor 1.11.0.a 2012/05/11 05:28
    OpenBoot v. 4.34.0 @(#)OpenBoot 4.34.0 2012/04/30 14:26
    root@gp-cpu-suh004 # ldm ls
    NAME STATE FLAGS CONS VCPU MEMORY UTIL UPTIME
    primary active -n-cv- UART 32 16G 2.5% 48m
    oidrac1 active -n---- 5000 32 16G 0.0% 27m
    root@gp-cpu-suh004 # ldm ls -l
    NAME STATE FLAGS CONS VCPU MEMORY UTIL UPTIME
    primary active -n-cv- UART 32 16G 3.1% 48m
    SOFTSTATE
    Solaris running
    UUID
    e73421fe-7003-e748-be7e-801fee5bfcc7
    MAC
    00:21:28:f1:95:26
    HOSTID
    0x85f19526
    CONTROL
    failure-policy=ignore
    extended-mapin-space=off
    cpu-arch=native
    DEPENDENCY
    master=
    CORE
    VCPU
    MEMORY
    RA PA SIZE
    0x20000000 0x20000000 16G
    CONSTRAINT
    threading=max-throughput
    VARIABLES
    auto-boot-on-error?=true
    auto-boot?=true
    boot-device=/pci@400/pci@1/pci@0/pci@0/LSI,sas@0/disk@w5000cca0251e7a29,0:a
    keyboard-layout=US-English
    nvramrc=." ChassisSerialNumber 1207BDYFFE " cr
    use-nvramrc?=true
    IO
    DEVICE PSEUDONYM OPTIONS
    VCC
    NAME PORT-RANGE
    primary-vcc0 5000-5100
    VSW
    NAME MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    primary-vsw-mgmt 00:14:4f:fb:75:c0 igb1 0 switch@0 1 1 1500 on
    primary-vsw0 00:14:4f:fa:33:8b aggr1 1 switch@1 1 1 1501,1601,10,11 1500 on
    VDS
    NAME VOLUME OPTIONS MPGROUP DEVICE
    primary-vds0 rootoid /dev/dsk/c14t50060E8005BFAA04d1s2
    data_oid /dev/dsk/c14t50060E8005BFAA04d2s2
    ocr_oid /dev/dsk/c14t50060E8005BFAA04d3s2
    VCONS
    NAME SERVICE PORT
    UART
    NAME STATE FLAGS CONS VCPU MEMORY UTIL UPTIME
    oidrac1 active -n---- 5000 32 16G 0.0% 27m
    SOFTSTATE
    Solaris running
    UUID
    0fcbbf21-14a2-eb21-f544-d4424212f3ef
    MAC
    00:14:4f:f9:1b:d4
    HOSTID
    0x84f91bd4
    CONTROL
    failure-policy=ignore
    extended-mapin-space=off
    cpu-arch=native
    DEPENDENCY
    master=
    CORE
    CID CPUSET
    VCPU
    VID PID CID UTIL STRAND
    MEMORY
    RA PA SIZE
    0x20000000 0x420000000 16G
    CONSTRAINT
    threading=max-throughput
    VARIABLES
    auto-boot?=true
    boot-device=disk:a
    keyboard-layout=US-English
    NETWORK
    NAME SERVICE ID DEVICE MAC MODE PVID VID MTU LINKPROP
    vnet1 primary-vsw-mgmt@primary 0 network@0 00:14:4f:fa:61:77 1 1500
    vnetprod primary-vsw0@primary 1 network@1 00:14:4f:f9:ec:7f 1501 1500
    vnethb primary-vsw0@primary 2 network@2 00:14:4f:fb:2b:8f 10 1500
    DISK
    NAME VOLUME TOUT ID DEVICE SERVER MPGROUP
    oneidrootdisk rootoid@primary-vds0 0 disk@0 primary
    oid_data data_oid@primary-vds0 1 disk@1 primary
    oid_ocr ocr_oid@primary-vds0 2 disk@2 primary
    VCONS
    NAME SERVICE PORT
    oidrac1 primary-vcc0@primary 5000
    root@gp-cpu-suh004 # ldm ls-services
    VCC
    NAME LDOM PORT-RANGE
    primary-vcc0 primary 5000-5100
    VSW
    NAME LDOM MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    primary-vsw-mgmt primary 00:14:4f:fb:75:c0 igb1 0 switch@0 1 1 1500 on
    primary-vsw0 primary 00:14:4f:fa:33:8b aggr1 1 switch@1 1 1 1501,1601,10,11 1500 on
    VDS
    NAME LDOM VOLUME OPTIONS MPGROUP DEVICE
    primary-vds0 primary rootoid /dev/dsk/c14t50060E8005BFAA04d1s2
    data_oid /dev/dsk/c14t50060E8005BFAA04d2s2
    ocr_oid /dev/dsk/c14t50060E8005BFAA04d3s2
    root@gp-cpu-suh004 # dladm show-link
    vsw0 type: non-vlan mtu: 1500 device: vsw0
    vsw1 type: non-vlan mtu: 1500 device: vsw1
    vsw1501001 type: vlan 1501 mtu: 1500 device: vsw1
    igb0 type: non-vlan mtu: 1500 device: igb0
    igb1 type: non-vlan mtu: 1500 device: igb1
    qlge0 type: non-vlan mtu: 1500 device: qlge0
    qlge1 type: non-vlan mtu: 1500 device: qlge1
    qlge2 type: non-vlan mtu: 1500 device: qlge2
    qlge3 type: non-vlan mtu: 1500 device: qlge3
    aggr1 type: non-vlan mtu: 1500 aggregation: key 1
    root@gp-cpu-suh004 # ifconfig -a
    lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu 8232 index 1
    inet 127.0.0.1 netmask ff000000
    igb0: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 2
    inet 10.223.12.14 netmask ffffff00 broadcast 10.223.12.255
    ether 0:21:28:f1:95:26
    vsw1501001: flags=201000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4,CoS> mtu 1500 index 3
    inet 10.220.128.9 netmask ffffff80 broadcast 10.220.128.127
    ether 0:14:4f:fa:33:8b
    root@gp-cpu-suh004 # netstat -nr
    Routing Table: IPv4
    Destination Gateway Flags Ref Use Interface
    default 10.220.128.1 UG 1 7
    10.220.128.0 10.220.128.9 U 1 5 vsw1501001
    10.223.0.0 10.223.12.1 UG 1 2
    10.223.12.0 10.223.12.14 U 1 1 igb0
    224.0.0.0 10.220.128.9 U 1 0 vsw1501001
    127.0.0.1 127.0.0.1 UH 8 261 lo0

    Yes, I can connect to the vswitch interface on the control domain. I didn't specify any PVID because my understanding is that PVID will tag any frame with the PVID VLAN by default. basically the PVID for this interface is 1.
    Here's the VNET config for the other LDOM in the RAC cluster:
    VSW
    NAME MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    primary-vsw-mgmt 00:14:4f:f9:91:fa igb1 0 switch@0 1 1 1500 on
    primary-vsw0 00:14:4f:fa:8e:cf aggr1 1 switch@1 1 1 1501,1601,10,11 1500 on
    NETWORK
    NAME SERVICE ID DEVICE MAC MODE PVID VID MTU LINKPROP
    vnet1 primary-vsw-mgmt 0 00:14:4f:fb:65:6d 1
    vnetprod primary-vsw0 1 00:14:4f:fa:2b:02 1501
    vnethb primary-vsw0 2 00:14:4f:f8:12:c1 10
    Thanks for reviewing my configuration.

  • Oracle 11gR2 RAC Backup & Recovery

    I am new to Oracle RAC, I have installed Oracle 11gR2 RAC (2 node) on HP-UX with ASM, configured FRA location on ASM.
    Now I would like to configure RMAN backup, Can someone give some idea or document to setup RMAN backup for Oracle 11gR2 RAC.
    - Mano

    ManoRangasamy wrote:
    I am new to Oracle RAC, I have installed Oracle 11gR2 RAC (2 node) on HP-UX with ASM, configured FRA location on ASM.
    Now I would like to configure RMAN backup, Can someone give some idea or document to setup RMAN backup for Oracle 11gR2 RAC. There won't be any different settings for RMAN for RAC except for where and how you would be configuring the Archive logs and place your backups. Both, ideally speaking, should be on the shared storage like ASM but there are other combinations also possible. Review the above given links for the same. That should provide a good starting point.
    Aman....

Maybe you are looking for