Shared storage check failed on nodes

hi friends,
I am installing rac 10g on vmware and os is OEL4.i completed all the prerequisites but when i run the below command
./runclufy stage -post hwos -n rac1,rac2, i am facing below error.
node connectivity check failed.
Checking shared storage accessibility...
WARNING:
Unable to determine the sharedness of /dev/sde on nodes:
rac2,rac2,rac2,rac2,rac2,rac1,rac1,rac1,rac1,rac1
Shared storage check failed on nodes "rac2,rac1"
please help me anyone ,it's urgent
Thanks,
poorna.
Edited by: 958010 on 3 Oct, 2012 9:47 PM

Hello,
It seems that your storage is not accessible from both the nodes. If you want you can follow these steps to configure 10g RAC on VMware.
Steps to configure Two Node 10 RAC on RHEL-4
Remark-1: H/W requirement for RAC
a) 4 Machines
1. Node1
2. Node2
3. storage
4. Grid Control
b) 2 switchs
c) 6 straight cables
Remark-2: S/W requirement for RAC
a) 10g cluserware
b) 10g database
Both must have the same version like (10.2.0.1.0)
Remark-3: RPMs requirement for RAC
a) all 10g rpms (Better to use RHEL-4 and choose everything option to install all the rpms)
b) 4 new rpms are required for installations
1. compat-gcc-7.3-2.96.128.i386.rpm
2. compat-gcc-c++-7.3-2.96.128.i386.rpm
3. compat-libstdc++-7.3-2.96.128.i386.rpm
4. compat-libstdc++-devel-7.3-2.96.128.i386.rpm
------------ Start Machine Preparation --------------------
1. Prepare 3 machines
i. node1.oracle.com
etho (192.9.201.183) - for public network
eht1 (10.0.0.1) - for private n/w
gateway (192.9.201.1)
subnet (255.255.255.0)
ii. node2.oracle.com
etho (192.9.201.187) - for public network
eht1 (10.0.0.2) - for private n/w
gateway (192.9.201.1)
subnet (255.255.255.0)
iii. openfiler.oracle.com
etho (192.9.201.182) - for public network
gateway (192.9.201.1)
subnet (255.255.255.0)
NOTE:-
-- Here eth0 of all the nodes should be connected by Public N/W using SWITCH-1
-- eth1 of all the nodes should be connected by Private N/W using SWITCH-2
2. network Configuration
#vim /etc/host
192.9.201.183 node1.oracle.com node1
192.9.201.187 node2.oracle.com node2
192.9.201.182 openfiler.oracle.com openfiler
10.0.0.1 node1-priv.oracle.com node1
10.0.0.2 node2-priv.oracle.com node2-priv
192.9.201.184 node1-vip.oracle.com node1-vip
192.9.201.188 node2-vip.oracle.com node2-vip
2. Prepare Both the nodes for installation
a. Set Kernel Parameters (/etc/sysctl.conf)
kernel.shmall = 2097152
kernel.shmmax = 2147483648
kernel.shmmni = 4096
kernel.sem = 250 32000 100 128
fs.file-max = 65536
net.ipv4.ip_local_port_range = 1024 65000
net.core.rmem_default = 262144
net.core.rmem_max = 262144
net.core.wmem_default = 262144
net.core.wmem_max = 262144
b. Configure /etc/security/limits.conf file
oracle soft nproc 2047
oracle hard nproc 16384
oracle soft nofile 1024
oracle hard nofile 65536
c. Configure /etc/pam.d/login file
session required /lib/security/pam_limits.so
d. Create user and groups on both nodes
# groupadd oinstall
# groupadd dba
# groupadd oper
# useradd -g oinstall -G dba oracle
# passwd oracle
e. Create required directories and set the ownership and permission.
# mkdir –p /u01/crs1020
# mkdir –p /u01/app/oracle/product/10.2.0/asm
# mkdir –p /u01/app/oracle/product/10.2.0/db_1
# chown –R oracle:oinstall /u01/
# chmod –R 755 /u01/
f. Set the environment variables
$ vi .bash_profile
ORACLE_BASE=/u01/app/oracle/; export ORACLE_BASE
ORA_CRS_HOME=/u01/crs1020; export ORA_CRS_HOME
#LD_ASSUME_KERNEL=2.4.19; export LD_ASSUME_KERNEL
#LANG=”en_US”; export LANG
3. storage configuration
PART-A Open-filer Set-up
Install openfiler on a machine (Leave 60GB free space on the hdd)
a) Login to root user
b) Start iSCSI target service
# service iscsi-target start
# chkconfig –level 345 iscsi-target on
PART –B Configuring Storage on openfiler
a) From any client machine open the browser and access openfiler console (446 ports).
https://192.9.201.182:446/
b) Open system tab and update the local N/W configuration for both nodes with netmask (255.255.255.255).
c) From the Volume tab click "create a new physical volume group".
d) From "block Device managemrnt" click on "(/dev/sda)" option under 'edit disk' option.
e) Under "Create a partition in /dev/sda" section create physical Volume with full size and then click on 'CREATE'.
f) Then go to the "Volume Section" on the right hand side tab and then click on "Volume groups"
g) Then under the "Create a new Volume Group" specify the name of the volume group (ex- racvgrp) and click on the check box and then click on "Add Volume Group".
h) Then go to the "Volume Section" on the right hand side tab and then click on "Add Volumes" and then specify the Volume name (ex- racvol1) and use all space and specify the "Filesytem/Volume type" as ISCSI and then click on CREATE.
i) Then go to the "Volume Section" on the right hand side tab and then click on "iSCSI Targets" and then click on ADD button to add your Target IQN.
j) then goto the 'LUN Mapping" and click on "MAP".
k) then goto the "Network ACL" and allow both node from there and click on UPDATE.
Note:- To create multiple volumes with openfiler we need to use Multipathing that is quite complex that’s why here we are going for a single volume. Edit the property of each volume and change access to allow.
f) install iscsi-initiator rpm on both nodes to acces iscsi disk
#rpm -ivh iscsi-initiator-utils-----------
g) Make entry in iscsi.conf file about openfiler on both nodes.
#vim /etc/iscsi.conf (in RHEL-4)
and in this file you will get a line "#DiscoveryAddress=192.168.1.2" remove comment and specify your storage ip address here.
OR
#vim /etc/iscsi/iscsi.conf (in RHEL-5)
and in this file you will get a line "#ins.address = 192.168.1.2" remove comment and specify your storage ip address here.
g) #service iscsi restart (on both nodes)
h) From both Nodes fire this command to access volume of openfiler-
# iscsiadm -m discovery -t sendtargets -p 192.2.201.182
i) #service iscsi restart (on both nodes)
j) #chkconfig –level 345 iscsi on (on both nodes)
k) make the partition 3 primary and 1 extended and within extended make 11 logical partition
A. Prepare partitions
1. #fdisk /dev/sdb
:e (extended)
Part No. 1
First Cylinder:
Last Cylinder:
:p
:n
:l
First Cylinder:
Last Cylinder: +1024M
2. Note the /dev/sdb* names.
3. #partprobe
4. Login as root user on node2 and run partprobe
B. On node1 login as root user and create following raw devices
# raw /dev/raw/raw5 /dev/sdb5
#raw /dev/raw/taw6 /dev/sdb6
# raw /dev/raw/raw12 /dev/sdb12
Run ls –l /dev/sdb* and ls –l /dev/raw/raw* to confirm the above
-Repeat the same thing on node2
C. On node1 as root user
# vi .etc/sysconfig/rawdevices
/dev/raw/raw5 /dev/sdb5
/dev/raw/raw6 /dev/sdb6
/dev/raw/raw7 /dev/sdb7
/dev/raw/raw8 /dev/sdb8
/dev/raw/raw9 /dev/sdb9
/dev/raw/raw10 /dev/sdb10
/dev/raw/raw11 /dev/sdb11
/dev/raw/raw12 /dev/sdb12
/dev/raw/raw13 /dev/sdb13
/dev/raw/raw14 /dev/sdb14
/dev/raw/raw15 /dev/sdb15
D. Restart the raw service (# service rawdevices restart)
#service rawdevices restart
Assigning devices:
/dev/raw/raw5 --> /dev/sdb5
/dev/raw/raw5: bound to major 8, minor 21
/dev/raw/raw6 --> /dev/sdb6
/dev/raw/raw6: bound to major 8, minor 22
/dev/raw/raw7 --> /dev/sdb7
/dev/raw/raw7: bound to major 8, minor 23
/dev/raw/raw8 --> /dev/sdb8
/dev/raw/raw8: bound to major 8, minor 24
/dev/raw/raw9 --> /dev/sdb9
/dev/raw/raw9: bound to major 8, minor 25
/dev/raw/raw10 --> /dev/sdb10
/dev/raw/raw10: bound to major 8, minor 26
/dev/raw/raw11 --> /dev/sdb11
/dev/raw/raw11: bound to major 8, minor 27
/dev/raw/raw12 --> /dev/sdb12
/dev/raw/raw12: bound to major 8, minor 28
/dev/raw/raw13 --> /dev/sdb13
/dev/raw/raw13: bound to major 8, minor 29
/dev/raw/raw14 --> /dev/sdb14
/dev/raw/raw14: bound to major 8, minor 30
/dev/raw/raw15 --> /dev/sdb15
/dev/raw/raw15: bound to major 8, minor 31
done
E. Repeat the same thing on node2 also
F. To make these partitions accessible to oracle user fire these commands from both Nodes.
# chown –R oracle:oinstall /dev/raw/raw*
# chmod –R 755 /dev/raw/raw*
F. To make these partitions accessible after restart make these entry on both nodes
# vi /etc/rc.local
Chown –R oracle:oinstall /dev/raw/raw*
Chmod –R 755 /dev/raw/raw*
4. SSH configuration (User quivalence)
On node1:- $ssh-keygen –t rsa
$ssh-keygen –t dsa
On node2:- $ssh-keygen –t rsa
$ssh-keygen –t dsa
On node1:- $cd .ssh
$cat *.pub>>node1
On node2:- $cd .ssh
$cat *.pub>>node2
On node1:- $scp node1 node2:/home/oracle/.ssh
On node2:- $scp node2 node2:/home/oracle/.ssh
On node1:- $cat node*>>authowized_keys
On node2:- $cat node*>>authowized_keys
Now test the ssh configuration from both nodes
$ vim a.sh
ssh node1 hostname
ssh node2 hostname
ssh node1-priv hostname
ssh node2-priv hostname
$ chmod +x a.sh
$./a.sh
first time you'll have to give the password then it never ask for password
5. To run cluster verifier
On node1 :-$cd /…/stage…/cluster…/cluvfy
$./runcluvfy stage –pre crsinst –n node1,node2
First time it will ask for four New RPMs but remember install these rpms by double clicking because of dependancy. So better to install these rpms in this order (rpm-3, rpm-4, rpm-1, rpm-2)
1. compat-gcc-7.3-2.96.128.i386.rpm
2. compat-gcc-c++-7.3-2.96.128.i386.rpm
3. compat-libstdc++-7.3-2.96.128.i386.rpm
4. compat-libstdc++-devel-7.3-2.96.128.i386.rpm
And again run cluvfy and check that "It should given a clean cheat" then start clusterware installation.

Similar Messages

  • Runcluvfy.bat comp ssa - Shared storage check failed

    I've run the cluvfy on Windows 2003 64bits on SAN with 3 nodes and found that it's unsuccessful for the shared storage checking. (whereas it's successful on Windows2003 32 bits.)
    C:\Documents and Settings\Administrator>C:\_Ly1\102010_win64_x64_clusterware\clusterware\cluvfy\runcluvfy.bat comp ssa -n rac1,rac2,rac3
    The system cannot find the file specified.
    Verifying shared storage accessibility
    Checking shared storage accessibility...
    Shared storage check failed on nodes "rac2,rac1,rac3".
    Verification of shared storage accessibility was unsuccessful on all the nodes.
    C:\Documents and Settings\Administrator>
    I'm not sure that it may result in the Clusterware Installation fail or not.
    Here, I captured the failure screen :
    http://alkaspace.com/is.php?i=30223&img=clusterware-fai.jpg
    Please help me out. Thank you!!

    I just ran into this myself while building an enterprise system on Win Server 2003. The answers here did not sit well with me and I needed to be sure of the shared storage prior to proceeding further. Researching Metalink did not uncover any relevant information either. I then opened an SR with Oracle and I did get back a satisfactory response which allowed me to verify my shared storage. The entire text of their solution can be found at http://www.webofwood.com/rac/oracle-response-to-shared-storage-check-failed-on-nodes/. Basically, it is a method of using another utility to identify the storage device names used by Windows and then writing and reading to them from each node to verify each node 'sees' what is written by the other node(s). If this check is successful, you can then proceed.

  • Shared Storage Check

    Hi all,
    We are planning to add a node to our existing RAC deployment (Database: 10gr2 and Sun Solaris 5.9 OS). Currently the shared storage is IBM SAN.
    When i run shared storage check using cluvfy, it fails to detect any shared storage. Given that i can ignore this error message (since cluvfy doesn't work wth SAN i beleive), how can i check whether the storage is shared or not?
    Note
    When i see partition table from both servers, it looks same (for the SAN drive, of course) but the name/label of the storages are different (For example: In existing node it show c6t0d0 but in the new node, which is to be added, it shows something different. Is it ok?).
    regards,
    Muhammad Riaz

    Never mind. I found solution from http://www.idevelopment.info.
    (1) Create following directory structure on second node (same as first node) with the same permissions on existins node:
    /asmdisks
    - crs
    -disk1
    -disk2
    - vote
    (2) use ls -lL /dev/rdsk/<Disk> to find out major and minor ids of shared disk and attach those ids to relveant direcotries above using mknod command:
    # ls -lL /dev/rdsk/c4t0d0*
    crw-r-----   1 root     sys       32,256 Aug  1 11:16 /dev/rdsk/c4t0d0s0
    crw-r-----   1 root     sys       32,257 Aug  1 11:16 /dev/rdsk/c4t0d0s1
    crw-r-----   1 root     sys       32,258 Aug  1 11:16 /dev/rdsk/c4t0d0s2
    crw-r-----   1 root     sys       32,259 Aug  1 11:16 /dev/rdsk/c4t0d0s3
    crw-r-----   1 root     sys       32,260 Aug  1 11:16 /dev/rdsk/c4t0d0s4
    crw-r-----   1 root     sys       32,261 Aug  1 11:16 /dev/rdsk/c4t0d0s5
    crw-r-----   1 root     sys       32,262 Aug  1 11:16 /dev/rdsk/c4t0d0s6
    crw-r-----   1 root     sys       32,263 Aug  1 11:16 /dev/rdsk/c4t0d0s7
    mknod /asmdisks/crs      c 32 257
    mknod /asmdisks/disk1      c 32 260
    mknod /asmdisks/disk2      c 32 261
    mknod /asmdisks/vote      c 32 259
    # ls -lL /asmdisks
    total 0
    crw-r--r--   1 root     oinstall  32,257 Aug  3 09:07 crs
    crw-r--r--   1 oracle   dba       32,260 Aug  3 09:08 disk1
    crw-r--r--   1 oracle   dba       32,261 Aug  3 09:08 disk2
    crw-r--r--   1 oracle   oinstall  32,259 Aug  3 09:08 vote

  • 11gR2 Verification of shared storage accessibility

    Friends,
    I do not understand how is this possible. I am trying to apply 11.2.0.2.5 psu on a 2-node cluster running on RHEL 5.5 vms. Followed ORACLE-BASE examples when installed this laptop RAC.
    I am not using any ACFS and none of GI and DB homes are shared. But, on node 2, cluvfy THINKS database home is shared.
    [oracle@rac1 trace]$ /oracle_grid/product/11.2.0.2/bin/cluvfy comp ssa -t software -s /oracle/product/11.2.0.2/db_1 -n rac1,rac2 -display_status
    Verifying shared storage accessibility
    Checking shared storage accessibility...
    "/oracle/product/11.2.0.2/db_1" is not shared
    Shared storage check failed on nodes "rac2,rac1"
    Verification of shared storage accessibility was unsuccessful on all the specified nodes.
    NODE_STATUS::rac2:VFAIL
    NODE_STATUS::rac1:VFAIL
    OVERALL_STATUS::VFAIL
    [oracle@rac1 trace]$ /oracle_grid/product/11.2.0.2/bin/cluvfy comp ssa -t software -s /oracle_grid/product/11.2.0.2 -n rac1,rac2 -display_status
    Verifying shared storage accessibility
    Checking shared storage accessibility...
    "/oracle_grid/product/11.2.0.2" is not shared
    Shared storage check failed on nodes "rac2,rac1"
    Verification of shared storage accessibility was unsuccessful on all the specified nodes.
    NODE_STATUS::rac2:VFAIL
    NODE_STATUS::rac1:VFAIL
    OVERALL_STATUS::VFAIL
    [oracle@rac1 trace]$ hostname
    rac1
    [oracle@rac1 trace]$ echo $ORACLE_HOSTNAME
    rac1
    [oracle@rac1 trace]$
    [oracle@rac2 trace]$ /oracle_grid/product/11.2.0.2/bin/cluvfy comp ssa -t software -s /oracle/product/11.2.0.2/db_1 -n rac1,rac2 -display_status
    Verifying shared storage accessibility
    Checking shared storage accessibility...
    "/oracle/product/11.2.0.2/db_1" is shared
    Shared storage check was successful on nodes "rac2,rac1"
    Verification of shared storage accessibility was successful.
    NODE_STATUS::rac2:SUCC
    NODE_STATUS::rac1:SUCC
    OVERALL_STATUS::SUCC
    [oracle@rac2 trace]$ /oracle_grid/product/11.2.0.2/bin/cluvfy comp ssa -t software -s /oracle_grid/product/11.2.0.2/ -n rac1,rac2 -display_status
    Verifying shared storage accessibility
    Checking shared storage accessibility...
    "/oracle_grid/product/11.2.0.2/" is not shared
    Shared storage check failed on nodes "rac2,rac1"
    Verification of shared storage accessibility was unsuccessful on all the specified nodes.
    NODE_STATUS::rac2:VFAIL
    NODE_STATUS::rac1:VFAIL
    OVERALL_STATUS::VFAIL
    [oracle@rac2 trace]$ hostname
    rac2
    [oracle@rac2 trace]$ echo $ORACLE_HOSTNAME
    rac2
    [oracle@rac2 trace]$
    I can not determine any reasons and do not know how to fix.
    Any help?
    Thank you.

    Hi,
    CLUFVY COMP SSA checks if the used storage/location is shared.
    If you do "cluvfy comp ssa -t software" cluvfy checks if your software home is shared.
    It tells you it is not. Hence the checks fails (which is correct, because you said DB Home is not shared).
    So where is the problem?
    CLUVFY COMP SSA only makes sense to check the sharedness. If it is not shared, then there is no sense in testing for it.
    Regards
    Sebastian

  • When installing clusterware, shared storage trouble

    I was trying to install clusterware. When I typed location of OCR, I got error below:
    Oracle Cluster Registry (OCR) is not shared across all the nodes in the cluster
    Then, I found I can not mount ocfs2 on both nodes at the same time. But I can mount it on any one of nodes if it is umounted on the other one.
    Can you anyone give me a hand?
    Environment is as following:
    - OS: Oracle Linux 5 (update 4)
    - Openfiler 3 + ocfs2
    - Oracle 10gR2

    Hi;
    Please see:
    http://kr.forums.oracle.com/forums/thread.jspa?messageID=4254569
    Oracle 10g RAC install- OPEN FAIL ON DEV
    Oracle Cluster Registry (OCR) is not shared across all the nodes . . .
    After OCFS2 install/configure ~ Shared storage check check fails
    Problem of using OCFS2 as shared storage to install RAC 10g on VMware
    Regard
    Helios

  • Runcluvy on shared storage

    Hey, I just might forget anything.
    I am going to install a rac cluster on SLES11 SP1.
    I am using multipathing where shared storage are provided by two SAN´s.
    I see all LUNs and did a raw device mapping via disk by-id ( all LUNs hava a partion on it)
    I am able to write with dd on the raw devices from both rac nodes at the same time.
    I set the permissions to the raw devices
    Installed cluvy.rpm on both nodes.
    When I start cluvy it passes the first checks successful.
    ./runcluvfy.sh stage -post hwos -n raca,racb -verbose
    But when its starts checking the shared storages - it fails.
    But prereq needs to be configured for a successful check of the storage devices ?
    Chris
    cluvy comes out of the installation package from 10.2.0.1

    Christian wrote:
    Hey, I just might forget anything.
    I am going to install a rac cluster on SLES11 SP1.
    I am using multipathing where shared storage are provided by two SAN´s.
    I see all LUNs and did a raw device mapping via disk by-id ( all LUNs hava a partion on it)
    I am able to write with dd on the raw devices from both rac nodes at the same time.
    I set the permissions to the raw devices
    Installed cluvy.rpm on both nodes.
    When I start cluvy it passes the first checks successful.
    ./runcluvfy.sh stage -post hwos -n raca,racb -verbose
    But when its starts checking the shared storages - it fails.
    But prereq needs to be configured for a successful check of the storage devices ?
    Chris
    cluvy comes out of the installation package from 10.2.0.1do you want to do pre-check or post-check?
    if you would like to do pre-check, check with
    ./runcluvfy.sh stage -pre crsinst -n raca,racb -verbose
    HTH,
    Refer installation guide for reference.
    Good Luck.

  • User equivalence check failed for user "oracle" while running cluvfy

    Hi All,
    I am getting User equivalence error as follows while executing cluvfy. I want to install CRS on this server.
    oracle@node200$ ./runcluvfy.sh stage -pre crsinst -n node200,node201,node202
    Performing pre-checks for cluster services setup
    Checking node reachability...
    Node reachability check passed from node "node200".
    Checking user equivalence...
    User equivalence check failed for user "oracle".
    Check failed on nodes:
    node202,node201,node200
    ERROR:
    User equivalence unavailable on all the nodes.
    Verification cannot proceed.
    Pre-check for cluster services setup was unsuccessful on all the nodes.
    I have sucessfully configured ssh and able to do ssh from all the nodes without prompting for the password. Also oracle uid and gid on all the servers are same.
    I also checked in .kshrc file and there is no tty entry in that file. and I have also setup /etc/hosts.equiv file on all the node.
    I still get the user equivalance error. Can anybody help me in this ?
    Regards,
    Raj.

    Hi,
    thanks.
    i need some more help because first time i am working on RAC,
    i am also having User equivalence check failed for user "oracle"
    i got two machines, one is IBM machine and one is DELL machine installed with solais 10 (AIX installation). both machines are not connected by private network now. and no shared external HDD available now. it will take one week time to set up the HW. the user equivalence check depends the above two factors?
    now public network is available and node reachability test "passed"
    could you please tell me what may be the reason for the user equivalence check in my case. details are given below
    i have oracle user in both machine. user creation done through SMC. home directory is /export/home/oracle
    the following are the details.
    1) $ ssh pc000905 id -a oracle
    Enter passphrase for key '/home/oracle/.ssh/id_rsa':
    uid=100(oracle) gid=300(oinstall) groups=301(dba)
    $ ssh pc026376 id -a oracle
    Enter passphrase for key '/home/oracle/.ssh/id_rsa':
    uid=100(oracle) gid=300(oinstall) groups=301(dba)
    2) user nobody exists.
    3) all files n .ssh directory have same permissions.
    4) ssh and scp is working with out password(rsa and dsa keys generated).
    5) oracle user is using bourne shell
    6)soft link of /usr/bin/ssh and /usr/bin/scp are created to /export/home/oracle/bin
    7)ForwardX11 is "yes" in sshd_config
    8) created a config file in .ssh directory and ForwardX11 is "no" in that file
    8) umask =022 in both nodes.
    error displaying is
    $/export/gome/cluvfy/runcluvfy.sh stage -pre crsinst -n pc026376,pc000905
    Performing pre-checks for cluster services setup
    Checking node reachability...
    Node reachability check passed from node "PC026376".
    Checking user equivalence...
    User equivalence check failed for user "oracle".
    Check failed on nodes:
    pc026376,pc000905
    ERROR:
    User equivalence unavailable on all the nodes.
    Verification cannot proceed.
    Pre-check for cluster services setup was unsuccessful on all the nodes.
    $

  • DFSr supported cluster configurations - replication between shared storage

    I have a very specific configuration for DFSr that appears to be suffering severe performance issues when hosted on a cluster, as part of a DFS replication group.
    My configuration:
    3 Physical machines (blades) within a physical quadrant.
    3 Physical machines (blades) hosted within a separate physical quadrant
    Both quadrants are extremely well connected, local, 10GBit/s fibre.
    There is local storage in each quadrant, no storage replication takes place.
    The 3 machines in the first quadrant are MS clustered with shared storage LUNs on a 3PAR filer.
    The 3 machines in the second quadrant are also clustered with shared storage, but on a separate 3PAR device.
    8 shared LUNs are presented to the cluster in the first quadrant, and an identical storage layout is connected in the second quadrant. Each LUN has an associated HAFS application associated with it which can fail-over onto any machine in the local cluster.
    DFS replication groups have been set up for each LUN and data is replicated from an "Active" cluster node entry point, to a "Passive" cluster node that provides no entry point to the data via DFSn and a Read-Only copy on it's shared cluster
    storage.
    For the sake of argument, assume that all HAFS application instances in the first quadrant are "Active" in a read/write configuration, and all "Passive" instances of the HAFS applications in the other quadrants are Read-Only.
    This guide: http://blogs.technet.com/b/filecab/archive/2009/06/29/deploying-dfs-replication-on-a-windows-failover-cluster-part-i.aspx defines
    how to add a clustered service to a replication group. It clearly shows using "Shared storage" for the cluster, which is common sense otherwise there effectively is no application fail-over possible and removes the entire point of using a resilient
    cluster.
    This article: http://technet.microsoft.com/en-us/library/cc773238(v=ws.10).aspx#BKMK_061 defines the following:
    DFS Replication in Windows Server 2012 and Windows Server 2008 R2 includes the ability to add a failover cluster
    as a member of a replication group. The DFS Replication service on versions of Windows prior to Windows Server 2008 R2
    is not designed to coordinate with a failover cluster, and the service will not fail over to another node.
    It then goes on to state, quite incredibly: DFS Replication does not support replicating files on Cluster Shared Volumes.
    Stating quite simply that DFSr does not support Cluster Shared Volumes makes absolutely no sense at all after stating clusters
    are supported in replication groups and a technet guide is provided to setup and configure this configuration. What possible use is a clustered HAFS solution that has no shared storage between the clustered nodes - none at all.
    My question:  I need some clarification, is the text meant to read "between" Clustered
    Shared Volumes?
    The storage configuration must to be shared in order to form a clustered service in the first place. What
    we am seeing from experience is a serious degradation of
    performance when attempting to replicate / write data between two clusters running a HAFS configuration, in a DFS replication group.
    If for instance, as a test, local / logical storage is mounted to a physical machine the performance of a DFS replication group between the unshared, logical storage on the physical nodes is approaching 15k small files per minute on initial write and even high
    for file amendments. When replicating between two nodes in a cluster, with shared clustered storage the solution manages a weak 2,500 files per minute on initial write and only 260 files per minute when attempting to update data / amend files.
    By testing various configurations we have effectively ruled out the SAN, the storage, drivers, firmware, DFSr configuration, replication group configuration - the only factor left that makes any difference is replicating from shared clustered storage, to another
    shared clustered storage LUN.
    So in summary:
    Logical Volume ---> Logical Volume = Fast
    Logical Volume ---> Clustered Shared Volume = ??
    Clusted Shared Volume ---> Clustered Shared Volume = Pitifully slow
    Can anyone explain why this might be?
    The guidance in the article is in clear conflict with all other evidence provided around DFSr and clustering, however it seems to lean towards why we may be seeing a real issue with replication performance.
    Many thanks for your time and any help/replies that may be received.
    Paul

    Hello Shaon Shan,
    I am also having the same scenario at one of my customer place.
    We have two FileServers running on Hyper-V 2012 R2 as guest VM using Cluster Shared Volume.  Even the data partition drive also a part of CSV.
    It's really confusing whether the DFS replication on CSV are supported or not, then what would be consequence if using.
    In my knowledge we have some customers they are using Hyper-V 2008 R2 and DFS is configured and running fine on CSV since more than 4 years without any issue.
    Appreciate if you can please elaborate and explain in details about the limitations on using CSV.
    Thanks in advance,
    Abul

  • Shared Storage RAC

    Hello,
    This is my Oracle RAC 11gR2 real world installation, I need to configure the shared storage for RAC 2 nodes on redhat enterprise linux 5.
    could please send me a step by step how to do it? I want to use Device Mapper Multipath for that and ASM for Storage.
    Thank you

    899660 wrote:
    Hello,
    This is my Oracle RAC 11gR2 real world installation, I need to configure the shared storage for RAC 2 nodes on redhat enterprise linux 5.
    could please send me a step by step how to do it? I want to use Device Mapper Multipath for that and ASM for Storage.
    Thank youHi,
    Shared storage is a hardware device, which cant be created by you. Ofcourse you can use that shared device to configure ASM.
    Check the below links
    http://martincarstenbach.wordpress.com/2010/11/16/configuration-device-mapper-multipath-on-oel5-update-5/
    http://www.oracle.com/technetwork/database/device-mapper-udev-crs-asm.pdf

  • Configure shared storage (RAC setup in Windows)

    Hi,
    Can I use mapped drive instead of direct connect to the external drive while configuring the shared storage?
    My problem is:
    I have a LaCie Group SA LaCie Big Disk Extreme LUN 0 IEEE 1394 SBP2 Device (500g). It can only be directly attached to one node because there is only one FW400 port. When I use switch and tried to connect both nodes directly to the device by switch, only one node show it as a disk drive. On another node, the device is utomatically stopped by windows and unaccessible.
    Then I tried to map it, but the shared storage test failed.
    Any idea?
    Thanks
    Laura

    Hi.
    I struggled with this problem, but then I gave up and used VMware to solve the problem:
    http://www.oracle-base.com/articles/10g/OracleDB10gR2RACInstallationOnWindows2003UsingVMware.php
    http://www.oracle-base.com/articles/10g/OracleDB10gR2RACInstallationOnCentos4UsingVMware.php
    Howard Rogers has some stuff that might give you a clue also:
    http://www.dizwell.com/oracle/articles/laptop_rac.html
    Cheers
    Tim...

  • 11g R2 RAC - Grid Infrastructure installation - "root.sh" fails on node#2

    Hi there,
    I am trying to create a two node 11g R2 RAC on OEL 5.5 (32-bit) using VMWare virtual machines. I have correctly configured both nodes. Cluster Verification utility returns on following error \[which I believe can be ignored]:
    Checking daemon liveness...
    Liveness check failed for "ntpd"
    Check failed on nodes:
    rac2,rac1
    PRVF-5415 : Check to see if NTP daemon is running failed
    Clock synchronization check using Network Time Protocol(NTP) failed
    Pre-check for cluster services setup was unsuccessful on all the nodes.
    While Grid Infrastructure installation (for a Cluster option), things go very smooth until I run "root.sh" on node# 2. orainstRoot.sh ran OK on both. "root.sh" run OK on node# 1 and ends with:
    Checking swap space: must be greater than 500 MB.   Actual 1967 MB    Passed
    The inventory pointer is located at /etc/oraInst.loc
    The inventory is located at /u01/app/oraInventory
    *'UpdateNodeList' was successful.*
    *[root@rac1 ~]#*
    "root.sh" fails on rac2 (2nd node) with following error:
    CRS-2672: Attempting to start 'ora.evmd' on 'rac2'
    CRS-2676: Start of 'ora.evmd' on 'rac2' succeeded
    Timed out waiting for the CRS stack to start.
    *[root@rac2 ~]#*
    I know this info may not be enough to figure out what the problem may be. Please let me know what should I look for to find the issue and fix it. Its been like almost two weeks now :-(
    Regards
    Amer

    Hi Zheng,
    ocssd.log is HUGE. So I am putting few of the last lines in the log file hoping they may give some clue:
    2011-07-04 19:49:24.007: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 2180 > margin 1500  cur_ms 36118424 lastalive 36116244
    2011-07-04 19:49:26.005: [    CSSD][2997803920]clssnmvSchedDiskThreads: DiskPingThread for voting file ORCL:DATA sched delay 4150 > margin 1500 cur_ms 36120424 lastalive 36116274
    2011-07-04 19:49:26.006: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 4180 > margin 1500  cur_ms 36120424 lastalive 36116244
    2011-07-04 19:49:27.997: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:49:27.997: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:49:33.001: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:49:33.001: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:49:37.996: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:49:37.996: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:49:43.000: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:49:43.000: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:49:48.004: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:49:48.005: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:50:12.003: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:50:12.008: [    CSSD][2997803920]clssnmvSchedDiskThreads: DiskPingThread for voting file ORCL:DATA sched delay 1660 > margin 1500 cur_ms 36166424 lastalive 36164764
    2011-07-04 19:50:12.009: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 1660 > margin 1500  cur_ms 36166424 lastalive 36164764
    2011-07-04 19:50:15.796: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 2130 > margin 1500  cur_ms 36170214 lastalive 36168084
    2011-07-04 19:50:16.996: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:50:16.996: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:50:17.826: [    CSSD][2997803920]clssnmvSchedDiskThreads: DiskPingThread for voting file ORCL:DATA sched delay 1540 > margin 1500 cur_ms 36172244 lastalive 36170704
    2011-07-04 19:50:17.826: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 1570 > margin 1500  cur_ms 36172244 lastalive 36170674
    2011-07-04 19:50:21.999: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:50:21.999: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:50:26.011: [    CSSD][2997803920]clssnmvSchedDiskThreads: DiskPingThread for voting file ORCL:DATA sched delay 1740 > margin 1500 cur_ms 36180424 lastalive 36178684
    2011-07-04 19:50:26.011: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 1620 > margin 1500  cur_ms 36180424 lastalive 36178804
    2011-07-04 19:50:27.004: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:50:27.004: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:50:28.002: [    CSSD][2997803920]clssnmvSchedDiskThreads: DiskPingThread for voting file ORCL:DATA sched delay 1700 > margin 1500 cur_ms 36182414 lastalive 36180714
    2011-07-04 19:50:28.002: [    CSSD][2997803920]clssnmvSchedDiskThreads: KillBlockThread for voting file ORCL:DATA sched delay 1790 > margin 1500  cur_ms 36182414 lastalive 36180624
    2011-07-04 19:50:31.998: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:50:31.998: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    2011-07-04 19:50:37.001: [    CSSD][2901298064]clssnmSendingThread: sending status msg to all nodes
    2011-07-04 19:50:37.002: [    CSSD][2901298064]clssnmSendingThread: sent 5 status msgs to all nodes
    *<end of log file>*And the alertrac2.log contains:
    *[root@rac2 rac2]# cat alertrac2.log*
    Oracle Database 11g Clusterware Release 11.2.0.1.0 - Production Copyright 1996, 2009 Oracle. All rights reserved.
    2011-07-02 16:43:51.571
    [client(16134)]CRS-2106:The OLR location /u01/grid/oracle/product/11.2.0/grid/cdata/rac2.olr is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/client/ocrconfig_16134.log.
    2011-07-02 16:43:57.125
    [client(16134)]CRS-2101:The OLR was formatted using version 3.
    2011-07-02 16:44:43.214
    [ohasd(16188)]CRS-2112:The OLR service started on node rac2.
    2011-07-02 16:45:06.446
    [ohasd(16188)]CRS-2772:Server 'rac2' has been assigned to pool 'Free'.
    2011-07-02 16:53:30.061
    [ohasd(16188)]CRS-2302:Cannot get GPnP profile. Error CLSGPNP_NO_DAEMON (GPNPD daemon is not running).
    2011-07-02 16:53:55.042
    [cssd(17674)]CRS-1713:CSSD daemon is started in exclusive mode
    2011-07-02 16:54:38.334
    [cssd(17674)]CRS-1707:Lease acquisition for node rac2 number 2 completed
    [cssd(17674)]CRS-1636:The CSS daemon was started in exclusive mode but found an active CSS daemon on node rac1 and is terminating; details at (:CSSNM00006:) in /u01/grid/oracle/product/11.2.0/grid/log/rac2/cssd/ocssd.log
    2011-07-02 16:54:38.464
    [cssd(17674)]CRS-1603:CSSD on node rac2 shutdown by user.
    2011-07-02 16:54:39.174
    [ohasd(16188)]CRS-2765:Resource 'ora.cssdmonitor' has failed on server 'rac2'.
    2011-07-02 16:55:43.430
    [cssd(17945)]CRS-1713:CSSD daemon is started in clustered mode
    2011-07-02 16:56:02.852
    [cssd(17945)]CRS-1707:Lease acquisition for node rac2 number 2 completed
    2011-07-02 16:56:04.061
    [cssd(17945)]CRS-1605:CSSD voting file is online: ORCL:DATA; details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/cssd/ocssd.log.
    2011-07-02 16:56:18.350
    [cssd(17945)]CRS-1601:CSSD Reconfiguration complete. Active nodes are rac1 rac2 .
    2011-07-02 16:56:29.283
    [ctssd(18020)]CRS-2403:The Cluster Time Synchronization Service on host rac2 is in observer mode.
    2011-07-02 16:56:29.551
    [ctssd(18020)]CRS-2407:The new Cluster Time Synchronization Service reference node is host rac1.
    2011-07-02 16:56:29.615
    [ctssd(18020)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 16:56:29.616
    [ctssd(18020)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 16:56:29.641
    [ctssd(18020)]CRS-2401:The Cluster Time Synchronization Service started on host rac2.
    [client(18052)]CRS-10001:ACFS-9327: Verifying ADVM/ACFS devices.
    [client(18056)]CRS-10001:ACFS-9322: done.
    2011-07-02 17:01:40.963
    [ohasd(16188)]CRS-2757:Command 'Start' timed out waiting for response from the resource 'ora.asm'. Details at (:CRSPE00111:) in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ohasd/ohasd.log.
    [client(18590)]CRS-10001:ACFS-9327: Verifying ADVM/ACFS devices.
    [client(18594)]CRS-10001:ACFS-9322: done.
    2011-07-02 17:27:46.385
    [ctssd(18020)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 17:27:46.385
    [ctssd(18020)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 17:46:48.717
    [crsd(22519)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:46:49.641
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:46:51.459
    [crsd(22553)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:46:51.776
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:46:53.928
    [crsd(22574)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:46:53.956
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:46:55.834
    [crsd(22592)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:46:56.273
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:46:57.762
    [crsd(22610)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:46:58.631
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:00.259
    [crsd(22628)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:47:00.968
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:02.513
    [crsd(22645)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:47:03.309
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:05.081
    [crsd(22663)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:47:05.770
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:07.796
    [crsd(22681)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:47:08.257
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:10.733
    [crsd(22699)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:47:11.739
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:13.547
    [crsd(22732)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 17:47:14.111
    [ohasd(16188)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 17:47:14.112
    [ohasd(16188)]CRS-2771:Maximum restart attempts reached for resource 'ora.crsd'; will not restart.
    2011-07-02 17:58:18.459
    [ctssd(18020)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 17:58:18.459
    [ctssd(18020)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    [client(26883)]CRS-10001:ACFS-9200: Supported
    2011-07-02 18:13:34.627
    [ctssd(18020)]CRS-2405:The Cluster Time Synchronization Service on host rac2 is shutdown by user
    2011-07-02 18:13:42.368
    [cssd(17945)]CRS-1603:CSSD on node rac2 shutdown by user.
    2011-07-02 18:15:13.877
    [client(27222)]CRS-2106:The OLR location /u01/grid/oracle/product/11.2.0/grid/cdata/rac2.olr is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/client/ocrconfig_27222.log.
    2011-07-02 18:15:14.011
    [client(27222)]CRS-2101:The OLR was formatted using version 3.
    2011-07-02 18:15:23.226
    [ohasd(27261)]CRS-2112:The OLR service started on node rac2.
    2011-07-02 18:15:23.688
    [ohasd(27261)]CRS-8017:location: /etc/oracle/lastgasp has 2 reboot advisory log files, 0 were announced and 0 errors occurred
    2011-07-02 18:15:24.064
    [ohasd(27261)]CRS-2772:Server 'rac2' has been assigned to pool 'Free'.
    2011-07-02 18:16:29.761
    [ohasd(27261)]CRS-2302:Cannot get GPnP profile. Error CLSGPNP_NO_DAEMON (GPNPD daemon is not running).
    2011-07-02 18:16:30.190
    [gpnpd(28498)]CRS-2328:GPNPD started on node rac2.
    2011-07-02 18:16:41.561
    [cssd(28562)]CRS-1713:CSSD daemon is started in exclusive mode
    2011-07-02 18:16:49.111
    [cssd(28562)]CRS-1707:Lease acquisition for node rac2 number 2 completed
    2011-07-02 18:16:49.166
    [cssd(28562)]CRS-1605:CSSD voting file is online: ORCL:DATA; details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/cssd/ocssd.log.
    [cssd(28562)]CRS-1636:The CSS daemon was started in exclusive mode but found an active CSS daemon on node rac1 and is terminating; details at (:CSSNM00006:) in /u01/grid/oracle/product/11.2.0/grid/log/rac2/cssd/ocssd.log
    2011-07-02 18:17:01.122
    [cssd(28562)]CRS-1603:CSSD on node rac2 shutdown by user.
    2011-07-02 18:17:06.917
    [ohasd(27261)]CRS-2765:Resource 'ora.cssdmonitor' has failed on server 'rac2'.
    2011-07-02 18:17:23.602
    [mdnsd(28485)]CRS-5602:mDNS service stopping by request.
    2011-07-02 18:17:36.217
    [gpnpd(28732)]CRS-2328:GPNPD started on node rac2.
    2011-07-02 18:17:43.673
    [cssd(28794)]CRS-1713:CSSD daemon is started in clustered mode
    2011-07-02 18:17:49.826
    [cssd(28794)]CRS-1707:Lease acquisition for node rac2 number 2 completed
    2011-07-02 18:17:49.865
    [cssd(28794)]CRS-1605:CSSD voting file is online: ORCL:DATA; details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/cssd/ocssd.log.
    2011-07-02 18:18:03.049
    [cssd(28794)]CRS-1601:CSSD Reconfiguration complete. Active nodes are rac1 rac2 .
    2011-07-02 18:18:06.160
    [ctssd(28861)]CRS-2403:The Cluster Time Synchronization Service on host rac2 is in observer mode.
    2011-07-02 18:18:06.220
    [ctssd(28861)]CRS-2407:The new Cluster Time Synchronization Service reference node is host rac1.
    2011-07-02 18:18:06.238
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 18:18:06.239
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 18:18:06.794
    [ctssd(28861)]CRS-2401:The Cluster Time Synchronization Service started on host rac2.
    [client(28891)]CRS-10001:ACFS-9327: Verifying ADVM/ACFS devices.
    [client(28895)]CRS-10001:ACFS-9322: done.
    2011-07-02 18:18:33.465
    [crsd(29020)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:33.575
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:35.757
    [crsd(29051)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:36.129
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:38.596
    [crsd(29066)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:39.146
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:41.058
    [crsd(29085)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:41.435
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:44.255
    [crsd(29101)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:45.165
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:47.013
    [crsd(29121)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:47.409
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:50.071
    [crsd(29136)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:50.118
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:51.843
    [crsd(29156)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:52.373
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:54.361
    [crsd(29171)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:54.772
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:56.620
    [crsd(29202)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:57.104
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:58.997
    [crsd(29218)]CRS-1013:The OCR location in an ASM disk group is inaccessible. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/crsd/crsd.log.
    2011-07-02 18:18:59.301
    [ohasd(27261)]CRS-2765:Resource 'ora.crsd' has failed on server 'rac2'.
    2011-07-02 18:18:59.302
    [ohasd(27261)]CRS-2771:Maximum restart attempts reached for resource 'ora.crsd'; will not restart.
    2011-07-02 18:49:58.070
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 18:49:58.070
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 19:21:33.362
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 19:21:33.362
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 19:52:05.271
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 19:52:05.271
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 20:22:53.696
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 20:22:53.696
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 20:53:43.949
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 20:53:43.949
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 21:24:32.990
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 21:24:32.990
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 21:55:21.907
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 21:55:21.908
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 22:26:45.752
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 22:26:45.752
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 22:57:54.682
    [ctssd(28861)]CRS-2412:The Cluster Time Synchronization Service detects that the local time is significantly different from the mean cluster time. Details in /u01/grid/oracle/product/11.2.0/grid/log/rac2/ctssd/octssd.log.
    2011-07-02 22:57:54.683
    [ctssd(28861)]CRS-2409:The clock on host rac2 is not synchronous with the mean cluster time. No action has been taken as the Cluster Time Synchronization Service is running in observer mode.
    2011-07-02 23:07:28.603
    [cssd(28794)]CRS-1612:Network communication with node rac1 (1) missing for 50% of timeout interval.  Removal of this node from cluster in 14.020 seconds
    2011-07-02 23:07:35.621
    [cssd(28794)]CRS-1611:Network communication with node rac1 (1) missing for 75% of timeout interval.  Removal of this node from cluster in 7.010 seconds
    2011-07-02 23:07:39.629
    [cssd(28794)]CRS-1610:Network communication with node rac1 (1) missing for 90% of timeout interval.  Removal of this node from cluster in 3.000 seconds
    2011-07-02 23:07:42.641
    [cssd(28794)]CRS-1632:Node rac1 is being removed from the cluster in cluster incarnation 205080558
    2011-07-02 23:07:44.751
    [cssd(28794)]CRS-1601:CSSD Reconfiguration complete. Active nodes are rac2 .
    2011-07-02 23:07:45.326
    [ctssd(28861)]CRS-2407:The new Cluster Time Synchronization Service reference node is host rac2.
    2011-07-04 19:46:26.008
    [ohasd(27261)]CRS-8011:reboot advisory message from host: rac1, component: mo155738, with time stamp: L-2011-07-04-19:44:43.318
    [ohasd(27261)]CRS-8013:reboot advisory message text: clsnomon_status: need to reboot, unexpected failure 8 received from CSS
    *[root@rac2 rac2]#* This log file start with complaint that OLR is not accessible. Here is what I see (rca2):
    -rw------- 1 root oinstall 272756736 Jul  2 18:18 /u01/grid/oracle/product/11.2.0/grid/cdata/rac2.olrAnd I guess rest of the problems start with this.

  • Runcluvfy.sh stage pre fails with node reachability on 1 node only

    Having a frustrating problem. 2 node RAC system on RHEL 5.2 installing 11.2.0.1 grid/clusterware. Performing the following pre check command from node 1:
    ./runcluvfy.sh stage -pre crsinst -n node1,node2 -verboseI'm getting the following error and it cannot write the trace information
    [grid@node1 grid]$ sudo chmod -R 777 /tmp
    [grid@node1 grid]$ ./runcluvfy.sh stage -pre crsinst -n node1,node2 -verbose
    WARNING:
    Could not access or create trace file path "/tmp/bootstrap/cv/log". Trace information could not be collected
    Performing pre-checks for cluster services setup
    Checking node reachability...
    node1.mydomain.com: node1.mydomain.com
    Check: Node reachability from node "null"
      Destination Node                      Reachable?
      node2                       no
      node1                       no
    Result: Node reachability check failed from node "null"
    ERROR:
    Unable to reach any of the nodes
    Verification cannot proceed
    Pre-check for cluster services setup was unsuccessful on all the nodes.
    [grid@node1 grid]$
    [grid@node1 grid]$ echo $CV_DESTLOC
    /home/grid/software/grid/11gr2/gridI've verified the following:
    1) there is user equivalence between the nodes for user grid
    2) /tmp is read/writable by user grid on both nodes
    3) Setting the CV_DESTLOC appears to do nothing - it seems to go back to wanting to write to /tmp
    4) ./runcluvfy comp nodecon -n node1,node2-verbose succeeds no problem
    And the weirdest thing of all, when I run ./runcluvfy.sh stage -pre crsinst -n node1,node2 -verbose from node 2, it succeeds without errors.
    What am I missing? And TIA..

    I made a copy of the runcluvfy.sh and commented out all rm -rf commands so that it would at least save the trace files. Re-ran, and the following trace output - not entirely helpful to me, but any gurus out there see anything?
    [main] [ 2010-04-20 15:48:38.275 CDT ] [TaskNodeConnectivity.performTask:354]  _nw_:Performing Node Reachability verification task...
    [main] [ 2010-04-20 15:48:38.282 CDT ] [ResultSet.traceResultSet:341]
    Target ResultSet BEFORE Upload===>
            Overall Status->UNKNOWN
    [main] [ 2010-04-20 15:48:38.283 CDT ] [ResultSet.traceResultSet:341]
    Source ResultSet ===>
            Overall Status->OPERATION_FAILED
            node2-->OPERATION_FAILED
            node1-->OPERATION_FAILED
    [main] [ 2010-04-20 15:48:38.283 CDT ] [ResultSet.traceResultSet:341]
    Target ResultSet AFTER Upload===>
            Overall Status->OPERATION_FAILED
            node2-->OPERATION_FAILED
            node1-->OPERATION_FAILED
    [main] [ 2010-04-20 15:48:38.284 CDT ] [ResultSet.getSuccNodes:556]  Checking for Success nodes from the total list of nodes in the resultset
    [main] [ 2010-04-20 15:48:38.284 CDT ] [ReportUtil.printReportFooter:1553]  stageMsgID: 8302
    [main] [ 2010-04-20 15:48:38.284 CDT ] [CluvfyDriver.main:299]  ==== cluvfy exiting normally.I'm still baffled why the precheck is successful from the second node. And, in fact, all other cluvfy checks that I've run succeed form both nodes.

  • User equivalence check failed for user "oracle".

    Hi,
    I am trying to install the Oracle Clusterware 10g as part of my RAC setup.
    I have configured RHEL 4 in 2 nodes (rac and rac2). But when I run the runcluvfy utility, I get the below error:
    [oracle@rac cluvfy]$ ./runcluvfy.sh stage -pre crsinst -n rac,rac2
    Performing pre-checks for cluster services setup
    Checking node reachability...
    Node reachability check passed from node "rac".
    Checking user equivalence...
    User equivalence check failed for user "oracle".
    Check failed on nodes:
    rac
    WARNING:
    User equivalence is not set for nodes:
    rac
    Verification will proceed with nodes:
    rac2
    The problem is only with the first node(rac). The second node passes succesfully.
    ssh works fine. Not sure what is wrong!
    Thanks!

    Did you do Configure SSH for User Equivalence ?
    Or
    Did you Establish User Equivalence ?
    If yes.. please check...
    $ ssh rac hostname
    rac
    $ ssh rac2 hostname
    rac2
    You should Establish User Equivalence before...
    exec /usr/bin/ssh-agent $SHELL
    /usr/bin/ssh-add
    If You don't configure ssh for user equivalence
    http://www.oracle.com/technology/pub/articles/smiley_rac10g_install.html
    Or
    read SSH User equivalence is not set up on the remote nodes

  • Free diskspace check failed foe ecah node

    Hi all
    I am tring to install oracle RAC in Windows 2003 srver.I have creared 2 virtual pc.I follwed every step correctly coz the clusterware software installs upto vipca.After that i sudelny see a blue screen stating that the hardware failure .Now i am trng eith 2 virtual pcs.when i run runcluvfy.bat stage -pre crsinst -n RAC1,RAC2 -verbose
    The error is
    Free disk space in
    Check: Free disk space in "C:\DOCUME~1\ADMINI~1.RAC\LOCALS~1\Temp" dir
    Node Name Available Required Comment
    RAC2 unknown 400MB (409600KB) failed
    RAC1 10.79GB (11313436KB) 400MB (409600KB) passed
    Result: Free disk space check failed.
    System requirement failed for 'crs'.
    what should i do.Thanks
    Edited by: user12119634(bobs) on Dec 13, 2009 9:02 PM

    Define virtual PCs. Provide full description of the technology, vendor, and version number. If this is another attempt to use unsupported technology like VMWare then you are going against what Oracle recommends and you are on your own.
    Since you do not have a valid clusterware installation as indicated by the cluster verify tool this is a really good time to stop the installation and think through these two questions:
    1. What is your shared storage solution? With Windows you only have a couple of possible choices none of which you have mentioned.
    2. What is your cache fusion interconnect strategy?
    If you don't get these right you are wasting your time.
    SB ... I have no idea why you think this question is in the wrong forum. Can you explain?

  • Oracle RAC with QFS shared storage going down when one disk fails

    Hello,
    I have an oracle RAC on my testing environment. The configuration follows
    nodes: V210
    Shared Storage: A5200
    #clrg status
    Group Name Node Name Suspended Status
    rac-framework-rg host1 No Online
    host2 No Online
    scal-racdg-rg host1 No Online
    host2 No Online
    scal-racfs-rg host1 No Online
    host2 No Online
    qfs-meta-rg host1 No Online
    host2 No Offline
    rac_server_proxy-rg host1 No Online
    host2 No Online
    #metastat -s racdg
    racdg/d200: Concat/Stripe
    Size: 143237376 blocks (68 GB)
    Stripe 0:
    Device Start Block Dbase Reloc
    d3s0 0 No No
    racdg/d100: Concat/Stripe
    Size: 143237376 blocks (68 GB)
    Stripe 0:
    Device Start Block Dbase Reloc
    d2s0 0 No No
    #more /etc/opt/SUNWsamfs/mcf
    racfs 10 ma racfs - shared
    /dev/md/racdg/dsk/d100 11 mm racfs -
    /dev/md/racdg/dsk/d200 12 mr racfs -
    When the disk /dev/did/dsk/d2 failed (I have failed it by removing from the array), the oracle RAC went offline on both nodes, and then both nodes paniced and rebooted. Now the #clrg status shows below output.
    Group Name Node Name Suspended Status
    rac-framework-rg host1 No Pending online blocked
    host2 No Pending online blocked
    scal-racdg-rg host1 No Online
    host2 No Online
    scal-racfs-rg host1 No Online
    host2 No Pending online blocked
    qfs-meta-rg host1 No Offline
    host2 No Offline
    rac_server_proxy-rg host1 No Pending online blocked
    host2 No Pending online blocked
    crs is not started in any of the nodes. I would like to know if anybody faced this kind of a problem when using QFS on diskgroup. When one disk is failed, the oracle is not supposed to go offline as the other disk is working, and also my qfs configuration is to mirror these two disks !!!!!!!!!!!!!!
    Many thanks in advance
    Ushas Symon

    I'm not sure why you say QFS is mirroring these disks!?!? Shared QFS has no inherent mirroring capability. It relies on the underlying volume manager (VM) or array to do that for it. If you need to mirror you storage, you do it at the VM level by creating a mirrored metadevice.
    Tim
    ---

Maybe you are looking for

  • Getting OU of users into CSV created by script

    Hi. I've found this script that let's us analyze mailbox Growth over time. http://gallery.technet.microsoft.com/scriptcenter/Generate-report-of-user-e4e9afca  (script shown below) But I need to add the OU of the user to the script as we are hosting d

  • Newbie: How can I store a previous value??

    I am trying to figure out how to do the following: int xcoor=0, ycoor=0, hstep, vstep; while true do update_step(hstep, vstep); // essentially have a big while loop to generate the necessary hstep/vsteps xcoor = xcoor + hstep; // I need to add a new

  • Can't group slices in pie chart

    I used to be able to shift click on a number of pie slices and group them and then drag (explode) the group from the rest of the pie. Doesn't seem to work now. What am I doing wrong?

  • Question about random function

    Is there a way to get 6 random numbers that not the same? I need this because I'm making a lottery from numbers 1-20. I need to make sure I don't get duplicate numbers. Will I need to make a nested loop statement? Thx for whoever answers my questions

  • OBIEE 11.1.1.1.7

    Hi, I'm having problems exporting a logo in an analysis or dashboard to excel but fine works for PDF. The logo is referenced using fmap. The files are located under two directories and i've restarted /u01/app/oracle/product/fmw/Oracle_BI1/bifoundatio