Recover OCR and VOTE disk after complete corruption of ASM disk groups.

Hi Gurus,
I am simulating a recovery situation to perform recover of OCR and Vote files after complete corruption of ASM related disks and diskgroups. I have setup my environment as follows:\
Environment: RAC
OS: OEL 5.5 32-bit
GI Version: 11.2.0.2.0
ASM Disk groups: +OCR, +DATA
OCR, Vote Files location: +OCR
ASM Redundancy: External
ASM Disks: /dev/asm-disk1, /dev/asm-disk2
/dev/asm-disk1 - mapped on +OCR
/dev/asm-disk2 - mapped on +DATA
With the above configuration in place I have manually corrupted +OCR, +DATA diskgroups with dd command. I used this command to completely corrupt +OCR disk group.
dd if=/dev/zero of=/dev/asm-disk1. I have manual backups as well as automatic backups of OCR and Vote disk. I am not using ASMLib.
I followed this link:
http://docs.oracle.com/cd/E11882_01/rac.112/e17264/adminoc.htm#TDPRC237
When I tried to recover OCR file, I could not do so as there is no such diskgroup which ASM can restore the OCR, Voting disk to. I could not Re-create OCR and DATA diskgroups as I cannot connect to ASM instance. If you have a solution or workaround for my situation please describe it. That will be greatly appreciated.
Thanks and Regards,
Suresh.

Please go through the following document which have the detailed steps to restore the OCR
How to restore ASM based OCR after complete loss of the CRS diskgroup on Linux/Unix systems [ID 1062983.1]

Similar Messages

  • About OCR and Vote Disk size

    Hallo,
    I have a simple question about OCR and Vote Disk in a 10g/11g RAC setup:
    official documentation speaks about sizing OCR and Vote Disk about 280 MB, but do they mean each OCR and Vote disk LUN ?
    In other words, for 2 OCR and 3 Vote disk I need 280*5 MB ( in multiple o single LUN)
    Is it so?
    thx in advance

    ottocolori wrote:
    Hallo,
    I have a simple question about OCR and Vote Disk in a 10g/11g RAC setup:
    official documentation speaks about sizing OCR and Vote Disk about 280 MB, but do they mean each OCR and Vote disk LUN ?
    In other words, for 2 OCR and 3 Vote disk I need 280*5 MB ( in multiple o single LUN)
    Is it so?
    thx in advanceYes. Also see Re: Reduce size of OCR file

  • Move OCR and Vote Disk to a new SAN

    Hi,
    Can you guys help me with some links/documents for moving OCR and Vote disk files from old SAN to new SAN.
    Both SAN will have ocfs2 formatted 3 partitions on which ocr and votedisk will be placed (normal redundancy).
    Oracle version 11g R2,
    OS linux 5.5
    RAC nodes : 2
    Datafiles on ASM
    OCR/Votedisk on ocfs2
    Regards,
    Ajay

    user7443016 wrote:
    Thanks for the reply, but all the examples were showing either ocfs2 to ASM or raw to ocfs2. could you just point out the high level steps for moving 3 ocr & votedisk files from one san to another(ocfs to ocfs).
    Hi Ajay,
    Steps will be same . It doesnt matter whether it is raw to ocfs2 or ocfs to ocfs.
    There will be change only how do you assign storage for OCR/voting disk.
    I have another doubt that, i already have 3 ocr files, can i add more ocr files pointing to another mount point? the same goes with votedisk.
    You can use replace option so there is no need to add 4th OCR disk.
    i found steps in one of the website: i have few doubts in it, could any one clarify it...All the steps seems to be fine.
    Moving OCR
    1) create mount point(new san - /u05) in all nodes
    2) precreate ocr files in new mount points (i am not sure how to do it? is it just touch and filename, its a ocfs2 file system)on ocfs you need to create file as below :
    dd if=/dev/zero of=/ocfs/OCR/ocrnew.dbf bs=1M count=128
    Also I would suggest to have a look at :
    'ocrconfig -replace ocr' Fails With PROT-16 [ID 444757.1]
    RAC Assurance Support Team: RAC Starter Kit and Best Practices (Linux) [ID 811306.1]
    Regards
    Rajesh

  • OCR and vote disk Allocation for shared raw storage with Solaris 10 questio

    Hi all,
    Current environment is Solaris 10 SPARC 64 bit OS with Hitachi SAN for shared storage and Sun E6900 servers.
    For Oracle 10g RAC (10.2) and ASM, I am setting up the vote disk and OCR files on shared raw storage area network.
    Assume that I have a 35Gb LUN carved out on raw device for the vote disk and OCR for a 2-node Oracle 10g RAC cluster and since the vote disk and OCR only require 120MB of storage, is there a way to use only a 120MB slice from the LUN or do I need to allocate the entire LUN/raw device to the vote disk?
    I am looking for a way to avoid wasting space for the OCR and vote disk with my 10g RAC cluster and we are using raw devices and Oracle 10g RAC Clusterware with Veritas filesystem for the disk management. Can I have several slices on Solaris 10 with raw shared storage (Hitachi) to triple mirror my OCR and vote disk files?
    It seems odd to have to use an entire 35GB LUN just for 2 small files: vote disk and OCR. Is there a way to partition a 120MB slice on one of the devices and allocate the rest to ASM?

    Hi
    In our RAC env.
    we have 1GB LUN for OCR and Voting Disk and 100GB LUN for ASM disk. we have 4 Node RAC env on AIX with 2 SAN storage on Hitachi system for our Application.
    we keept 1st Voting diks on SAN-1 2nd Voting Disk on SAN-2 and third voting on SAN-3.
    OCR disk on SAN1 and OCR_MIRROR Disk on SAN-2.
    we have ASM DIsk with Normal Redundancy . Failure Group 1 on SAN-1 and
    Failure Group 2 on SAN-2
    Regards
    Bharat

  • OCR and VOT disk, in the same asm disk or different disks ?! and how ?!

    What Oracle recommends regarding the OCR and Voting disks in Rac 11gR2 on ASM ?!
    Making them under the same disk or different disks ?!
    and how ?!
    Thanks a lot,
    Regards,
    Gehad.

    Best practice is that OCR/Voting should be under ASM. Making them under the same disk or different disks depends upon the ASM diskgroup. If we want to keep it on seperate failover groups then we should create muliple failover group.
    For external : 1 ; Normal : 3 and higher : 5 asm diskgroups.
    I hope i would be clear ith my answers.
    Regards,
    Dheeraj Vaish

  • SAN for ocr and vote file  for Linux RAC

    Hi Experts,
    Based on Oracle document, i have to create 2 devices for the OCR and 3 for the VOTE in RAC.
    However, my boss think SAN maybe mirroe OCR and Vote file. It is enough for 1 location in SAN storage. It does not matter how many driver sidplay for user.
    Actually it is only one SAN in back side.
    Any Explain for OCR/VOTE file in SAN?
    Thnaks
    JIm

    The idea to have these multiplex copy is to reduce the single point of failure, non the less, that's reason you build RAC in the first place isn't?
    So if it's possible you'd better have redundancy on physical disk level (LUN etc) then logic level (volume, partition etc) then file level.
    Well I agree modern SAN are pretty reliable and had built in redundancy on physical and logic level, but if you only have single vote disk and OCR, your RAC is going to fail right away when vote disk and OCR has problem. The chances are if your OCR and Vote disk has logical corruption, the corruption is going to populated to your mirrors. Even if your mirror is good you still need to get mirror copy to replace working copy and restart cluster during the time your RAC is offline.
    With more than one OCR and voting disk, your RAC is likely to survive single OCR/vote disk failure, so you can fix the problem while RAC online.

  • OCR and voting disks on ASM, problems in case of fail-over instances

    Hi everybody
    in case at your site you :
    - have an 11.2 fail-over cluster using Grid Infrastructure (CRS, OCR, voting disks),
    where you have yourself created additional CRS resources to handle single-node db instances,
    their listener, their disks and so on (which are started only on one node at a time,
    can fail from that node and restart to another);
    - have put OCR and voting disks into an ASM diskgroup (as strongly suggested by Oracle);
    then you might have problems (as we had) because you might:
    - reach max number of diskgroups handled by an ASM instance (63 only, above which you get ORA-15068);
    - experiment delays (especially in case of multipath), find fake CRS resources, etc.
    whenever you dismount disks from one node and mount to another;
    So (if both conditions are true) you might be interested in this story,
    then please keep reading on for the boring details.
    One step backward (I'll try to keep it simple).
    Oracle Grid Infrastructure is mainly used by RAC db instances,
    which means that any db you create usually has one instance started on each node,
    and all instances access read / write the same disks from each node.
    So, ASM instance on each node will mount diskgroups in Shared Mode,
    because the same diskgroups are mounted also by other ASM instances on the other nodes.
    ASM instances have a spfile parameter CLUSTER_DATABASE=true (and this parameter implies
    that every diskgroup is mounted in Shared Mode, among other things).
    In this context, it is quite obvious that Oracle strongly recommends to put OCR and voting disks
    inside ASM: this (usually called CRS_DATA) will become diskgroup number 1
    and ASM instances will mount it before CRS starts.
    Then, additional diskgroup will be added by users, for DATA, REDO, FRA etc of each RAC db,
    and will be mounted later when a RAC db instance starts on the specific node.
    In case of fail-over cluster, where instances are not RAC type and there is
    only one instance running (on one of the nodes) at any time for each db, it is different.
    All diskgroups of db instances don't need to be mounted in Shared Mode,
    because they are used by one instance only at a time
    (on the contrary, they should be mounted in Exclusive Mode).
    Yet, if you follow Oracle advice and put OCR and voting inside ASM, then:
    - at installation OUI will start ASM instance on each node with CLUSTER_DATABASE=true;
    - the first diskgroup, which contains OCR and votings, will be mounted Shared Mode;
    - all other diskgroups, used by each db instance, will be mounted Shared Mode, too,
    even if you'll take care that they'll be mounted by one ASM instance at a time.
    At our site, for our three-nodes cluster, this fact has two consequences.
    One conseguence is that we hit ORA-15068 limit (max 63 diskgroups) earlier than expected:
    - none ot the instances on this cluster are Production (only Test, Dev, etc);
    - we planned to have usually 10 instances on each node, each of them with 3 diskgroups (DATA, REDO, FRA),
    so 30 diskgroups each node, for a total of 90 diskgroups (30 instances) on the cluster;
    - in case one node failed, surviving two should get resources of the failing node,
    in the worst case: one node with 60 diskgroups (20 instances), the other one with 30 diskgroups (10 instances)
    - in case two nodes failed, the only node survived should not be able to mount additional diskgroups
    (because of limit of max 63 diskgroup mounted by an ASM instance), so all other would remain unmounted
    and their db instances stopped (they are not Production instances);
    But it didn't worked, since ASM has parameter CLUSTER_DATABASE=true, so you cannot mount 90 diskgroups,
    you can mount 62 globally (once a diskgroup is mounted on one node, it is given a number between 2 and 63,
    and other diskgroups mounted on other nodes cannot reuse that number).
    So as a matter of fact we can mount only 21 diskgroups (about 7 instances) on each node.
    The second conseguence is that, every time our CRS handmade scripts dismount diskgroups
    from one node and mount it to another, there are delays in the range of seconds (especially with multipath).
    Also we found inside CRS log that, whenever we mounted diskgroups (on one node only), then
    behind the scenes were created on the fly additional fake resources
    of type ora*.dg, maybe to accomodate the fact that on other nodes those diskgroups were left unmounted
    (once again, instances are single-node here, and not RAC type).
    That's all.
    Did anyone go into similar problems?
    We opened a SR to Oracle asking about what options do we have here, and we are disappointed by their answer.
    Regards
    Oscar

    Hi Klaas-Jan
    - best practises require that also online redolog files are in a separate diskgroup, in case of ASM logical corruption (we are a little bit paranoid): in case DATA dg gets corrupted, you can restore Full backup plus Archived RedoLog plus Online Redolog (otherwise you will stop at the latest Archived).
    So we have 3 diskgroups for each db instance: DATA, REDO, FRA.
    - in case of fail-over cluster (active-passive), Oracle provide some templates of CRS scripts (in $CRS_HOME/crs/crs/public) that you edit and change at your will, also you might create additionale scripts in case of additional resources you might need (Oracle Agents, backups agent, file systems, monitoring tools, etc)
    About our problem, the only solution is to move OCR and voting disks from ASM and change pfile af all ASM instance (parameter CLUSTER_DATABASE from true to false ).
    Oracle aswers were a litlle bit odd:
    - first they told us to use Grid Standalone (without CRS, OCR, voting at all), but we told them that we needed a Fail-over solution
    - then they told us to use RAC Single Node, which actually has some better features, in csae of planned fail-over it might be able to migreate
    client sessions without causing a reconnect (for SELECTs only, not in case of a running transaction), but we already have a few fail-over cluster, we cannot change them all
    So we plan to move OCR and voting disks into block devices (we think that the other solution, which needs a Shared File System, will take longer).
    Thanks Marko for pointing us to OCFS2 pros / cons.
    We asked Oracle a confirmation that it supported, they said yes but it is discouraged (and also, doesn't work with OUI nor ASMCA).
    Anyway that's the simplest approach, this is a non-Prod cluster, we'll start here and if everthing is fine, after a while we'll do it also on Prod ones.
    - Note 605828.1, paragraph 5, Configuring non-raw multipath devices for Oracle Clusterware 11g (11.1.0, 11.2.0) on RHEL5/OL5
    - Note 428681.1: OCR / Vote disk Maintenance Operations: (ADD/REMOVE/REPLACE/MOVE)
    -"Grid Infrastructure Install on Linux", paragraph 3.1.6, Table 3-2
    Oscar

  • OCR and Voting Disk Mirror

    Hi,
    We have oracle 10.2.0.4 RAC database. Now we are planning to mirror the OCR and Voting disk. Currently we are using SAN mirror copy. SAN admin assures that we don't need mirror copy as it is being mirrored at the disk level. Can any one please suggest me if we need any mirror copy of these. If so what advantages I get from the regular SAN copy.

    You use SAN or High Availability Storage + RAID. You don't need to mirror OCR/VOTING .
    Anyway , If you can do it... Mirror OCR + Multi VOTING... , that's good thing oracle recommend.
    By the way, If you can not mirror OCR + multi VOTE on 10g, You should backup VOTE File ("dd" linux command) after setup completed/add node/delete node.
    http://download.oracle.com/docs/cd/B19306_01/rac.102/b28759/adminoc.htm#sthref148
    And Check OCR file automatic backup:
    $ ocrconfig -showbackup
    http://download.oracle.com/docs/cd/B19306_01/rac.102/b14197/ocrsyntax.htm

  • Confusion with OCFS2 File system for OCR and Voting disk RHEL 5, Oracle11g,

    Dear all,
    I am in the process of installing Oracle 11g 3 Node RAC database
    The environment on which i have to do this implementation is as follows:
    Oracle 11g.
    Red Hat Linux 5 x86
    Oracle Clusterware
    ASM
    EMC Storage
    250 Gb of Storage drive.
    SAN
    As of now i am in the process of installing Oracle Clusterware on the 3 nodes.
    I have performed these tasks for the cluster installs.
    1. Configure Kernel Parameters
    2. Configure User Limits
    3. Modify the /etc/pam.d/login file
    4. Configure Operating System Users and Groups for Oracle Clusterware
    5. Configure Oracle Clusterware Owner Environment
    6. Install CVUQDISK rpm package
    7. Configure the Hosts file
    8. Verify the Network Setup
    9. Configure the SSH on all Cluster Nodes (User Equivalence)
    9. Enable the SSH on all Cluster Nodes (User Equivalence)
    10. Install Oracle Cluster File System (OCFS2)
    11.Verify the Installation of Oracle Cluster File System (OCFS2)
    12. Configure the OCFS2 (/etc/ocfs2/cluster.conf)
    13. Configure the O2CB Cluster Stack for OCFS2
    BUT, here after i am a little bit confused on how to proceed further. The next step is to Format the disk and mount the OCFS2, Create Software Directories... so and so forth.
    I asked my system admin to provide me two partitions so that i could format them with OCFS2 file system.
    He wrote back to me saying.
    *"Is what you want before I do it??*
    */dev/emcpowera1 is 3GB and formatted OCFS2.*
    */dev/emcpowera2 is 3GB and formatted OCFS2.*
    *Are those big enough for you? If not, I can re-size and re-format them*
    *before I mount them on the servers.*
    *the SAN is shared storage. /dev/emcpowera is one of three LUNs on*
    *the shared storage, and it's 214GB. Right now there are only two*
    *partitions on it- the ones I listed below. I can repartition the LUN any*
    *way you want it.*
    *Where do you want these mounted at:*
    */dev/emcpowera1*
    */dev/emcpowera2*
    *I was thinking if this mounting techique would work like so:*
    *emcpowera1: /u01/shared_config/OCR_config*
    *emcpowera2: /u01/shared_config/voting_disk*
    *Let me know how you'd like them mounted."*
    Please recommend me what i should convey to him so that i can ask him to the exact same thing.
    My second question is, as we are using ASM, for which i am gonna configure ASM after clusterware installation, should i install Openfiler??
    Pls refer the enviroment information i provided above and make recommendations.
    As of now i am using Jeffery Hunters guide to install the entire setup. You think the entire install guide goes well with my enviroment??
    http://www.oracle.com/technology/pub/articles/hunter_rac11gr1_iscsi.html?rssid=rss_otn_articles
    Kind regards
    MK

    Thanks for ur reply Mufalani,
    You have managed to solve half part of my query. But still i am stuck with what kind of mount point i should ask the system admin to create for OCR and Voting disk. Should i go with the mount point he is mentioning??
    Let me put forth few more questions here.
    1. Is 280 MB ok for OCR and voting disks respectively??
    2. Should i ask the system admin to create 4 voting disk mount points and two for ocr??
    3. As mentioned by the system admin.
    */u01/shared_config/OCR_config*
    */u01/shared_config/voting_disk*
    Is this ok for creating the ocr and voting disks?
    4. Can i use OCFS2 file system for formating the disk instead of using them as RAW device!!?
    5. As u mentioned that Openfiler is not needed for Configuring ASM... Could you provide me the links which will guide me to create partition disks, voting disks and ocr disks!! I could not locate them on the doc or else were. I did find a couple of them, but was unable to identify a suitable one for my envoirement.
    Regards
    MK

  • OCR and Voting Disk partition

    Dear all
    I uninstall clusterware. And i want to install again that but when i made partition in windows 2003 then CVU fail the stage command.
    I did the following things..
    1) diskpart --->select Disk 1----> clean -->exit (both the nodes)
    2) diskpart -->automount enable--->exit ---> restart (both the nodes)
    3) Then i make four partition in one node. And see after restart node 2 have that partition with drive letter. I remove all the drive letter.
    But CVU stage command failed. And if i run clusterware then it can get partition which has not previous OCR and voting. I think previous OCR and Voting disk could not clean.
    Could u pls tell me how can i get solution.
    Moinul

    Check Metalink for this problem area ! Raise TAR

  • Ocr and voting disk recreate

    Hi,
    I'm recreating ocr and voting disk as for a lab test on RHEL4(10gr2),for server side ons configuration for racgons ,what should be my port? below is the output from ons.config file .Should it be 6200 or 6101 for both nodes?Please let me know .
    cat ons.config
    localport=6101
    remoteport=6200
    loglevel=3
    useocr=on
    CRS_home/bin/racgons add_config hostname1:port hostname2:port
    Thanks

    Hi
    You may run this command with user root
    root@yourracnode1# $CRS_HOME/bin/racgons add_config yourracnode1:6200 yourracnode2:6200
    After run,this command helps you to ping your ons
    root@yourracnode1# $ORA_CRS_HOME/bin/onsctl ping
    I hope this helps you
    BR,
    Hasan

  • Help needed : How to move OCR and Voting Disk

    Hi all experts...
    I am Mahesh. Working in DXB as DBA. We have serveral RAC installed clients. Among those one major client wants their OCR and Voting Disk to be moved from existing SAN to a new SAN which will be configured. Only OCR and Voting Disks needs to be moved.
    Details:
    =====
    Oracle Clusterware,ASM and DB version 10.2.0.3.0
    Windows 2003 Server Enterprise Edition
    2-Node RAC
    OCR and Voting disks are on RAW devices (not on CFS or NFS)
    I have read the clusterware administration document and also Metalink Note 428681.1 for the same.
    But it does not contain information how to move it to another new RAW in a new SAN device. (we need to keep it in raw device itself)
    How to identify the new disk path from windows operating system ? Then only we can move the OCR and VD right ???
    Help me for the same please...
    Thanks & regards,
    Mahesh.

    Hi Mashes,
    Could you list out the steps you did after you create the link on the new SAN using GUIObjectManager.exe?
    I tried on my VM and also in customer's test environment but I only successful using the following steps
    1. shutdown asm service, database service, listener service, CRS, EVM and CSS
    2. Object Manager Service is kept running
    3. from both nodes I issued the following command
    ocrconfig -repair \\.\OCRPRIMARY
    ocrconfig -repair \\.\OCRMIRROR
    4. according to metalink id 428681.1, I should do the following
    ocrconfig -overwrite
    However I was not successful because that command returns an error message (PROT-1: Failed to initialize ocr config).
    Therefore, I ran the following command
    ocrconfig -restore D:\oracle\1020\CRS\cdata\crs\backup00.ocr
    5. after that command, ocrcheck giving me the result and the ocr primary and mirror have been moved to the new partition.
    6. start the whole services starting from CSS, EVM, CRS, ASM, Listerner and Database service.
    Although what I experienced seems a bit weird and not the same as what was written in metalink, the RAC is running fine until now.
    What I would like to know is your experience on this. May be you can share it here.
    Thank you,
    Adhika

  • Advantage of having OCR and Voting disk on ASM

    What are advantages of putting OCR and Voting disk on ASM from 11g

    well, other than the sharing thing, you dont have to go RAIDing an additional shared disk either. If you have properly configured ASM, redundancy should be built in as well, either soft or hard.
    not sure what other advantages you may need. Theres the IO thing with ASM, buit thats not really an advantage persay with ocr and vote. I may be contradicted by others but Ive never seen performance hit of any kind attributed to ocr and vote on non-ASM disk.

  • Where to put OCR and Voting disks?

    Hi,
    I will be installing Clusterware 11.2.0.2 on a 2 node cluster. I have read several whitepapers where a third asm disk group DATA, FRA, and CRS was created to hold the OCR and Voting disks. In the interest of keeping things simple can I just put the OCR and Voting disks in the DATA DG? I am using external SAN storage.
    Thanks
    Tom Cullen

    Hi,
    yes that is absolutely possible.
    There are several best practices. Some recommend a separate diskgroup, some recommend to put OCR and Voting in the same diskgroup like your data.
    So if there is no special use case it is absolutely valid to only use 2 diskgroups. 1 DATA (including OCR and Vote) and 1 FRA (Fast Recovery Area for Database Backups/Archivelogs/1 Controlfile/1 RedologMember and a second copy of OCR).
    Regards
    Sebastian

  • OCR and Votinging Disk File Permissions for NFS 2 node RAC cluster

    This is a fresh reinstall of Oracle 11gR2 Clusterware on Red Hat Enterprise Linux 64bit and I'm using NFS as my shared storage.
    I've create three file systems /ocrVote01 /ocrVote02 and /ocrVote03 as NFS share and exported them to two nodes that will use them as shared OCR and Voting Disk. Currently I'm getting errors when I run root.sh
    I've done the following:
    chown rac:dba -R /ocrVote01 /ocrVote02 /ocrVote03
    It appears that the root.sh script is trying to create a directory /ocrVote01/storage and getting permissions denied.
    Any idea why Root can not create this directory?
    root.sh is also getting error permission denied at /u01/gridinfra/11.2.0/GI/crs/install/crsconfiglib.pm at line 4478
    Any Ideas on what file permissions should be set for /ocrVote01 /ocrVote02 and /ocrVote03 ?
    Thanks
    32352 close(3) = 0
    32352 write(2, "mkdir /ocrVote01/storage/: Permi"..., 112) = 112
    | 00000 6d 6b 64 69 72 20 2f 6f 63 72 56 6f 74 65 30 31 **mkdir /o crVote01** |
    | 00010 2f 73 74 6f 72 61 67 65 2f 3a 20 50 65 72 6d 69 /storage /: Permi |
    | 00020 73 73 69 6f 6e 20 64 65 6e 69 65 64 20 61 74 20 ssion de nied at |
    | 00030 2f 75 30 31 2f 67 72 69 64 69 6e 66 72 61 2f 31 /u01/gri dinfra/1 |
    | 00040 31 2e 32 2e 30 2f 47 49 2f 63 72 73 2f 69 6e 73 1.2.0/GI /crs/ins |
    | 00050 74 61 6c 6c 2f 63 72 73 63 6f 6e 66 69 67 5f 6c tall/crs config_l |
    | 00060 69 62 2e 70 6d 20 6c 69 6e 65 20 34 34 37 38 0a ib.pm li ne 4478. |
    32352 exit_group(25) = ?

    After reading OTN feedback, which is very much appreciated. I decided to look deeper into my NFS configuration and found that the /etc/exports had the incorrect configuration for shared disk where in root can make appropriate changes to voting disk and crs files. The correct configuration for /etc/fstab using NFS for shared storage for voting disk and CRS:
    192.168.1.21:/ocrVote01 /ocrVote01 nfs rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp,vers=3,timeo=600,actimeo=0 1 2
    192.168.1.21:/ocrVote02 /ocrVote02 nfs rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp,vers=3,timeo=600,actimeo=0 1 2
    192.168.1.21:/ocrVote03 /ocrVote03 nfs rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp,vers=3,timeo=600,actimeo=0 1 2

Maybe you are looking for