About OCR and Vote Disk size

Hallo,
I have a simple question about OCR and Vote Disk in a 10g/11g RAC setup:
official documentation speaks about sizing OCR and Vote Disk about 280 MB, but do they mean each OCR and Vote disk LUN ?
In other words, for 2 OCR and 3 Vote disk I need 280*5 MB ( in multiple o single LUN)
Is it so?
thx in advance

ottocolori wrote:
Hallo,
I have a simple question about OCR and Vote Disk in a 10g/11g RAC setup:
official documentation speaks about sizing OCR and Vote Disk about 280 MB, but do they mean each OCR and Vote disk LUN ?
In other words, for 2 OCR and 3 Vote disk I need 280*5 MB ( in multiple o single LUN)
Is it so?
thx in advanceYes. Also see Re: Reduce size of OCR file

Similar Messages

  • Move OCR and Vote Disk to a new SAN

    Hi,
    Can you guys help me with some links/documents for moving OCR and Vote disk files from old SAN to new SAN.
    Both SAN will have ocfs2 formatted 3 partitions on which ocr and votedisk will be placed (normal redundancy).
    Oracle version 11g R2,
    OS linux 5.5
    RAC nodes : 2
    Datafiles on ASM
    OCR/Votedisk on ocfs2
    Regards,
    Ajay

    user7443016 wrote:
    Thanks for the reply, but all the examples were showing either ocfs2 to ASM or raw to ocfs2. could you just point out the high level steps for moving 3 ocr & votedisk files from one san to another(ocfs to ocfs).
    Hi Ajay,
    Steps will be same . It doesnt matter whether it is raw to ocfs2 or ocfs to ocfs.
    There will be change only how do you assign storage for OCR/voting disk.
    I have another doubt that, i already have 3 ocr files, can i add more ocr files pointing to another mount point? the same goes with votedisk.
    You can use replace option so there is no need to add 4th OCR disk.
    i found steps in one of the website: i have few doubts in it, could any one clarify it...All the steps seems to be fine.
    Moving OCR
    1) create mount point(new san - /u05) in all nodes
    2) precreate ocr files in new mount points (i am not sure how to do it? is it just touch and filename, its a ocfs2 file system)on ocfs you need to create file as below :
    dd if=/dev/zero of=/ocfs/OCR/ocrnew.dbf bs=1M count=128
    Also I would suggest to have a look at :
    'ocrconfig -replace ocr' Fails With PROT-16 [ID 444757.1]
    RAC Assurance Support Team: RAC Starter Kit and Best Practices (Linux) [ID 811306.1]
    Regards
    Rajesh

  • OCR and vote disk Allocation for shared raw storage with Solaris 10 questio

    Hi all,
    Current environment is Solaris 10 SPARC 64 bit OS with Hitachi SAN for shared storage and Sun E6900 servers.
    For Oracle 10g RAC (10.2) and ASM, I am setting up the vote disk and OCR files on shared raw storage area network.
    Assume that I have a 35Gb LUN carved out on raw device for the vote disk and OCR for a 2-node Oracle 10g RAC cluster and since the vote disk and OCR only require 120MB of storage, is there a way to use only a 120MB slice from the LUN or do I need to allocate the entire LUN/raw device to the vote disk?
    I am looking for a way to avoid wasting space for the OCR and vote disk with my 10g RAC cluster and we are using raw devices and Oracle 10g RAC Clusterware with Veritas filesystem for the disk management. Can I have several slices on Solaris 10 with raw shared storage (Hitachi) to triple mirror my OCR and vote disk files?
    It seems odd to have to use an entire 35GB LUN just for 2 small files: vote disk and OCR. Is there a way to partition a 120MB slice on one of the devices and allocate the rest to ASM?

    Hi
    In our RAC env.
    we have 1GB LUN for OCR and Voting Disk and 100GB LUN for ASM disk. we have 4 Node RAC env on AIX with 2 SAN storage on Hitachi system for our Application.
    we keept 1st Voting diks on SAN-1 2nd Voting Disk on SAN-2 and third voting on SAN-3.
    OCR disk on SAN1 and OCR_MIRROR Disk on SAN-2.
    we have ASM DIsk with Normal Redundancy . Failure Group 1 on SAN-1 and
    Failure Group 2 on SAN-2
    Regards
    Bharat

  • Recover OCR and VOTE disk after complete corruption of ASM disk groups.

    Hi Gurus,
    I am simulating a recovery situation to perform recover of OCR and Vote files after complete corruption of ASM related disks and diskgroups. I have setup my environment as follows:\
    Environment: RAC
    OS: OEL 5.5 32-bit
    GI Version: 11.2.0.2.0
    ASM Disk groups: +OCR, +DATA
    OCR, Vote Files location: +OCR
    ASM Redundancy: External
    ASM Disks: /dev/asm-disk1, /dev/asm-disk2
    /dev/asm-disk1 - mapped on +OCR
    /dev/asm-disk2 - mapped on +DATA
    With the above configuration in place I have manually corrupted +OCR, +DATA diskgroups with dd command. I used this command to completely corrupt +OCR disk group.
    dd if=/dev/zero of=/dev/asm-disk1. I have manual backups as well as automatic backups of OCR and Vote disk. I am not using ASMLib.
    I followed this link:
    http://docs.oracle.com/cd/E11882_01/rac.112/e17264/adminoc.htm#TDPRC237
    When I tried to recover OCR file, I could not do so as there is no such diskgroup which ASM can restore the OCR, Voting disk to. I could not Re-create OCR and DATA diskgroups as I cannot connect to ASM instance. If you have a solution or workaround for my situation please describe it. That will be greatly appreciated.
    Thanks and Regards,
    Suresh.

    Please go through the following document which have the detailed steps to restore the OCR
    How to restore ASM based OCR after complete loss of the CRS diskgroup on Linux/Unix systems [ID 1062983.1]

  • OCR and VOT disk, in the same asm disk or different disks ?! and how ?!

    What Oracle recommends regarding the OCR and Voting disks in Rac 11gR2 on ASM ?!
    Making them under the same disk or different disks ?!
    and how ?!
    Thanks a lot,
    Regards,
    Gehad.

    Best practice is that OCR/Voting should be under ASM. Making them under the same disk or different disks depends upon the ASM diskgroup. If we want to keep it on seperate failover groups then we should create muliple failover group.
    For external : 1 ; Normal : 3 and higher : 5 asm diskgroups.
    I hope i would be clear ith my answers.
    Regards,
    Dheeraj Vaish

  • SAN for ocr and vote file  for Linux RAC

    Hi Experts,
    Based on Oracle document, i have to create 2 devices for the OCR and 3 for the VOTE in RAC.
    However, my boss think SAN maybe mirroe OCR and Vote file. It is enough for 1 location in SAN storage. It does not matter how many driver sidplay for user.
    Actually it is only one SAN in back side.
    Any Explain for OCR/VOTE file in SAN?
    Thnaks
    JIm

    The idea to have these multiplex copy is to reduce the single point of failure, non the less, that's reason you build RAC in the first place isn't?
    So if it's possible you'd better have redundancy on physical disk level (LUN etc) then logic level (volume, partition etc) then file level.
    Well I agree modern SAN are pretty reliable and had built in redundancy on physical and logic level, but if you only have single vote disk and OCR, your RAC is going to fail right away when vote disk and OCR has problem. The chances are if your OCR and Vote disk has logical corruption, the corruption is going to populated to your mirrors. Even if your mirror is good you still need to get mirror copy to replace working copy and restart cluster during the time your RAC is offline.
    With more than one OCR and voting disk, your RAC is likely to survive single OCR/vote disk failure, so you can fix the problem while RAC online.

  • OCR and voting disks on ASM, problems in case of fail-over instances

    Hi everybody
    in case at your site you :
    - have an 11.2 fail-over cluster using Grid Infrastructure (CRS, OCR, voting disks),
    where you have yourself created additional CRS resources to handle single-node db instances,
    their listener, their disks and so on (which are started only on one node at a time,
    can fail from that node and restart to another);
    - have put OCR and voting disks into an ASM diskgroup (as strongly suggested by Oracle);
    then you might have problems (as we had) because you might:
    - reach max number of diskgroups handled by an ASM instance (63 only, above which you get ORA-15068);
    - experiment delays (especially in case of multipath), find fake CRS resources, etc.
    whenever you dismount disks from one node and mount to another;
    So (if both conditions are true) you might be interested in this story,
    then please keep reading on for the boring details.
    One step backward (I'll try to keep it simple).
    Oracle Grid Infrastructure is mainly used by RAC db instances,
    which means that any db you create usually has one instance started on each node,
    and all instances access read / write the same disks from each node.
    So, ASM instance on each node will mount diskgroups in Shared Mode,
    because the same diskgroups are mounted also by other ASM instances on the other nodes.
    ASM instances have a spfile parameter CLUSTER_DATABASE=true (and this parameter implies
    that every diskgroup is mounted in Shared Mode, among other things).
    In this context, it is quite obvious that Oracle strongly recommends to put OCR and voting disks
    inside ASM: this (usually called CRS_DATA) will become diskgroup number 1
    and ASM instances will mount it before CRS starts.
    Then, additional diskgroup will be added by users, for DATA, REDO, FRA etc of each RAC db,
    and will be mounted later when a RAC db instance starts on the specific node.
    In case of fail-over cluster, where instances are not RAC type and there is
    only one instance running (on one of the nodes) at any time for each db, it is different.
    All diskgroups of db instances don't need to be mounted in Shared Mode,
    because they are used by one instance only at a time
    (on the contrary, they should be mounted in Exclusive Mode).
    Yet, if you follow Oracle advice and put OCR and voting inside ASM, then:
    - at installation OUI will start ASM instance on each node with CLUSTER_DATABASE=true;
    - the first diskgroup, which contains OCR and votings, will be mounted Shared Mode;
    - all other diskgroups, used by each db instance, will be mounted Shared Mode, too,
    even if you'll take care that they'll be mounted by one ASM instance at a time.
    At our site, for our three-nodes cluster, this fact has two consequences.
    One conseguence is that we hit ORA-15068 limit (max 63 diskgroups) earlier than expected:
    - none ot the instances on this cluster are Production (only Test, Dev, etc);
    - we planned to have usually 10 instances on each node, each of them with 3 diskgroups (DATA, REDO, FRA),
    so 30 diskgroups each node, for a total of 90 diskgroups (30 instances) on the cluster;
    - in case one node failed, surviving two should get resources of the failing node,
    in the worst case: one node with 60 diskgroups (20 instances), the other one with 30 diskgroups (10 instances)
    - in case two nodes failed, the only node survived should not be able to mount additional diskgroups
    (because of limit of max 63 diskgroup mounted by an ASM instance), so all other would remain unmounted
    and their db instances stopped (they are not Production instances);
    But it didn't worked, since ASM has parameter CLUSTER_DATABASE=true, so you cannot mount 90 diskgroups,
    you can mount 62 globally (once a diskgroup is mounted on one node, it is given a number between 2 and 63,
    and other diskgroups mounted on other nodes cannot reuse that number).
    So as a matter of fact we can mount only 21 diskgroups (about 7 instances) on each node.
    The second conseguence is that, every time our CRS handmade scripts dismount diskgroups
    from one node and mount it to another, there are delays in the range of seconds (especially with multipath).
    Also we found inside CRS log that, whenever we mounted diskgroups (on one node only), then
    behind the scenes were created on the fly additional fake resources
    of type ora*.dg, maybe to accomodate the fact that on other nodes those diskgroups were left unmounted
    (once again, instances are single-node here, and not RAC type).
    That's all.
    Did anyone go into similar problems?
    We opened a SR to Oracle asking about what options do we have here, and we are disappointed by their answer.
    Regards
    Oscar

    Hi Klaas-Jan
    - best practises require that also online redolog files are in a separate diskgroup, in case of ASM logical corruption (we are a little bit paranoid): in case DATA dg gets corrupted, you can restore Full backup plus Archived RedoLog plus Online Redolog (otherwise you will stop at the latest Archived).
    So we have 3 diskgroups for each db instance: DATA, REDO, FRA.
    - in case of fail-over cluster (active-passive), Oracle provide some templates of CRS scripts (in $CRS_HOME/crs/crs/public) that you edit and change at your will, also you might create additionale scripts in case of additional resources you might need (Oracle Agents, backups agent, file systems, monitoring tools, etc)
    About our problem, the only solution is to move OCR and voting disks from ASM and change pfile af all ASM instance (parameter CLUSTER_DATABASE from true to false ).
    Oracle aswers were a litlle bit odd:
    - first they told us to use Grid Standalone (without CRS, OCR, voting at all), but we told them that we needed a Fail-over solution
    - then they told us to use RAC Single Node, which actually has some better features, in csae of planned fail-over it might be able to migreate
    client sessions without causing a reconnect (for SELECTs only, not in case of a running transaction), but we already have a few fail-over cluster, we cannot change them all
    So we plan to move OCR and voting disks into block devices (we think that the other solution, which needs a Shared File System, will take longer).
    Thanks Marko for pointing us to OCFS2 pros / cons.
    We asked Oracle a confirmation that it supported, they said yes but it is discouraged (and also, doesn't work with OUI nor ASMCA).
    Anyway that's the simplest approach, this is a non-Prod cluster, we'll start here and if everthing is fine, after a while we'll do it also on Prod ones.
    - Note 605828.1, paragraph 5, Configuring non-raw multipath devices for Oracle Clusterware 11g (11.1.0, 11.2.0) on RHEL5/OL5
    - Note 428681.1: OCR / Vote disk Maintenance Operations: (ADD/REMOVE/REPLACE/MOVE)
    -"Grid Infrastructure Install on Linux", paragraph 3.1.6, Table 3-2
    Oscar

  • Confusion with OCFS2 File system for OCR and Voting disk RHEL 5, Oracle11g,

    Dear all,
    I am in the process of installing Oracle 11g 3 Node RAC database
    The environment on which i have to do this implementation is as follows:
    Oracle 11g.
    Red Hat Linux 5 x86
    Oracle Clusterware
    ASM
    EMC Storage
    250 Gb of Storage drive.
    SAN
    As of now i am in the process of installing Oracle Clusterware on the 3 nodes.
    I have performed these tasks for the cluster installs.
    1. Configure Kernel Parameters
    2. Configure User Limits
    3. Modify the /etc/pam.d/login file
    4. Configure Operating System Users and Groups for Oracle Clusterware
    5. Configure Oracle Clusterware Owner Environment
    6. Install CVUQDISK rpm package
    7. Configure the Hosts file
    8. Verify the Network Setup
    9. Configure the SSH on all Cluster Nodes (User Equivalence)
    9. Enable the SSH on all Cluster Nodes (User Equivalence)
    10. Install Oracle Cluster File System (OCFS2)
    11.Verify the Installation of Oracle Cluster File System (OCFS2)
    12. Configure the OCFS2 (/etc/ocfs2/cluster.conf)
    13. Configure the O2CB Cluster Stack for OCFS2
    BUT, here after i am a little bit confused on how to proceed further. The next step is to Format the disk and mount the OCFS2, Create Software Directories... so and so forth.
    I asked my system admin to provide me two partitions so that i could format them with OCFS2 file system.
    He wrote back to me saying.
    *"Is what you want before I do it??*
    */dev/emcpowera1 is 3GB and formatted OCFS2.*
    */dev/emcpowera2 is 3GB and formatted OCFS2.*
    *Are those big enough for you? If not, I can re-size and re-format them*
    *before I mount them on the servers.*
    *the SAN is shared storage. /dev/emcpowera is one of three LUNs on*
    *the shared storage, and it's 214GB. Right now there are only two*
    *partitions on it- the ones I listed below. I can repartition the LUN any*
    *way you want it.*
    *Where do you want these mounted at:*
    */dev/emcpowera1*
    */dev/emcpowera2*
    *I was thinking if this mounting techique would work like so:*
    *emcpowera1: /u01/shared_config/OCR_config*
    *emcpowera2: /u01/shared_config/voting_disk*
    *Let me know how you'd like them mounted."*
    Please recommend me what i should convey to him so that i can ask him to the exact same thing.
    My second question is, as we are using ASM, for which i am gonna configure ASM after clusterware installation, should i install Openfiler??
    Pls refer the enviroment information i provided above and make recommendations.
    As of now i am using Jeffery Hunters guide to install the entire setup. You think the entire install guide goes well with my enviroment??
    http://www.oracle.com/technology/pub/articles/hunter_rac11gr1_iscsi.html?rssid=rss_otn_articles
    Kind regards
    MK

    Thanks for ur reply Mufalani,
    You have managed to solve half part of my query. But still i am stuck with what kind of mount point i should ask the system admin to create for OCR and Voting disk. Should i go with the mount point he is mentioning??
    Let me put forth few more questions here.
    1. Is 280 MB ok for OCR and voting disks respectively??
    2. Should i ask the system admin to create 4 voting disk mount points and two for ocr??
    3. As mentioned by the system admin.
    */u01/shared_config/OCR_config*
    */u01/shared_config/voting_disk*
    Is this ok for creating the ocr and voting disks?
    4. Can i use OCFS2 file system for formating the disk instead of using them as RAW device!!?
    5. As u mentioned that Openfiler is not needed for Configuring ASM... Could you provide me the links which will guide me to create partition disks, voting disks and ocr disks!! I could not locate them on the doc or else were. I did find a couple of them, but was unable to identify a suitable one for my envoirement.
    Regards
    MK

  • OCR and Voting Disk Migration to new media

    Hello,
    We need to move OCR,voting disk and ASM from one storage media to another .I think ASM related files are very easy to move/re map but how OCR and Voting Disk files will be migrated to the new storage media?
    Thanks,
    S.Mann

    Hi,
    when you said:
    "Move Voting Disk : run as root.
    crsctl query css votedisk
    crsctl add css votedisk new_backup_san_device
    crsctl query css votedisk - Make sure the new device is added
    crsctl delete css votedisk current_san_device_to_be_encrypted
    crsctl query css votedisk - Make sure that the old one is removed "
    I supose the new_backup_san_device could be a raw device??
    In my system I have a 2 nodes cluster, oracle 10gR2, RAC and ASM
    the ocr and voting disk are stored in /dev/raw/raw1 and /dev/raw/raw2
    thoses 2 raw devices are mapped with LUNS in the first storade media.
    We are about to change it..
    So what must I do at this step? What are the logical steps I must done?
    add the new LUNS in the nodes and map it as raw devices?
    I have this at the moment:
    crw-r----- 1 root oinstall 162, 1 Feb 8 17:33 /dev/raw/raw1
    crw-r----- 1 root oinstall 162, 2 Feb 8 17:33 /dev/raw/raw2
    crw-r----- 1 oracle oinstall 162, 3 Feb 8 17:33 /dev/raw/raw3
    in the case of migration, I must add 3 new row devices?
    as this?
    crw-r----- 1 root oinstall 162, 1 Feb 8 17:33 /dev/raw/raw1
    crw-r----- 1 root oinstall 162, 2 Feb 8 17:33 /dev/raw/raw2
    crw-r----- 1 oracle oinstall 162, 3 Feb 8 17:33 /dev/raw/raw3
    crw-r----- 1 root oinstall 162, 1 Feb 8 17:33 /dev/raw/raw4
    crw-r----- 1 root oinstall 162, 2 Feb 8 17:33 /dev/raw/raw5
    crw-r----- 1 oracle oinstall 162, 3 Feb 8 17:33 /dev/raw/raw6
    That's it?
    Thanks to help

  • OCR and Voting Disk Mirror

    Hi,
    We have oracle 10.2.0.4 RAC database. Now we are planning to mirror the OCR and Voting disk. Currently we are using SAN mirror copy. SAN admin assures that we don't need mirror copy as it is being mirrored at the disk level. Can any one please suggest me if we need any mirror copy of these. If so what advantages I get from the regular SAN copy.

    You use SAN or High Availability Storage + RAID. You don't need to mirror OCR/VOTING .
    Anyway , If you can do it... Mirror OCR + Multi VOTING... , that's good thing oracle recommend.
    By the way, If you can not mirror OCR + multi VOTE on 10g, You should backup VOTE File ("dd" linux command) after setup completed/add node/delete node.
    http://download.oracle.com/docs/cd/B19306_01/rac.102/b28759/adminoc.htm#sthref148
    And Check OCR file automatic backup:
    $ ocrconfig -showbackup
    http://download.oracle.com/docs/cd/B19306_01/rac.102/b14197/ocrsyntax.htm

  • Install issue w/ 10.2.0.1, OEL5 using raw for ocr and vote

    Hello,
    I am having a challenge getting the raw devices setup for ocr and vote, 10.2.0.1 and OEL5.
    The raw devices are not getting created and not sure why. I have followed the steps in note 443996.1 and 465001.1 and not having any luck. What am I missing here?
    Here is the contents of the 55-oracle-naming.rules:
    # cat /etc/udev/rules.d/55-oracle-naming.rules
    # Configure persistent, user-defined Oracle Clusterware device file names
    KERNEL=="sda1", BUS=="scsi", PROGRAM=="/sbin/scsi_id", RESULT=="360001fe1000f3270000904609091002c", NAME="ocr1", OWNER="root", GROUP="oinstall", MODE="0640"
    KERNEL=="sdb1", BUS=="scsi", PROGRAM=="/sbin/scsi_id", RESULT=="360001fe1000f3270000904609091002f", NAME="vote1", OWNER="oracle", GROUP="oinstall", MODE="0640"
    and 60-raw.rules
    # This file and interface are deprecated.
    # Applications needing raw device access should open regular
    # block devices with O_DIRECT.
    # Enter raw device bindings here.
    # An example would be:
    # ACTION=="add", KERNEL=="sda", RUN+="/bin/raw /dev/raw/raw1 %N"
    # to bind /dev/raw/raw1 to /dev/sda, or
    # ACTION=="add", ENV{MAJOR}=="8", ENV{MINOR}=="1", RUN+="/bin/raw /dev/raw/raw2 %M %m"
    # to bind /dev/raw/raw2 to the device with major 8, minor 1.
    ACTION=="add", KERNEL=="/dev/sda1", RUN+="/bin/raw /dev/raw/raw1 %N"
    ACTION=="add", KERNEL=="/dev/sdb1", RUN+="/bin/raw /dev/raw/raw2 %N"
    ACTION=="add", KERNEL=="/dev/sdc1", RUN+="/bin/raw /dev/raw/raw3 %N"
    When I use udevtest, udevtest /block/sda/sda1, I get the following results;
    [root@pdrn01 ~]# udevtest /block/sda/sda1
    main: looking at device '/block/sda/sda1' from subsystem 'block'
    run_program: '/bin/bash -c '/sbin/lsmod | /bin/grep ^dm_multipath''
    run_program: '/bin/bash' (stdout) 'dm_multipath 21577 0 '
    run_program: '/bin/bash' returned with status 0
    udev_db_get_device: no db file to read /dev/.udev/db/block@sda: No such file or directory
    udev_db_get_device: no db file to read /dev/.udev/db/block@sda: No such file or directory
    run_program: '/lib/udev/vol_id --export /dev/.tmp-8-1'
    run_program: '/lib/udev/vol_id' returned with status 4
    run_program: '/sbin/scsi_id'
    run_program: '/sbin/scsi_id' returned with status 1
    udev_rules_get_name: no node name set, will use kernel name 'sda1'
    udev_db_get_device: no db file to read /dev/.udev/db/block@sda@sda1: No such file or directory
    This should be getting labeled as /dev/ocr1 as noted in note 465001.1 and the output shows /dev/sda1
    udev_node_add: creating device node '/dev/sda1', major = '8', minor = '1',mode = '0640', uid = '0', gid = '6'
    main: run: '/sbin/multipath -v0 8:1'
    main: run: 'socket:/org/kernel/udev/monitor'
    main: run: '/lib/udev/udev_run_devd'
    main: run: 'socket:/org/freedesktop/hal/udev_event'
    main: run: '/sbin/pam_console_apply /dev/sda1 '
    [root@pdrn01 ~]#
    And 99-raw.rules output;
    KERNEL=="raw[1-2]*", GROUP="oinstall", MODE="640"
    KERNEL=="raw[3-5]*", OWNER="oracle", GROUP="oinstall", MODE="660"
    So, any ideas or suggestions would be HUGELY appreciated!
    TIA, Cheers!

    Here is one note about raw device issue in release 5 and workaround.
    Raw device technology deprecated in Release 5.
    http://oraclepitstop.wordpress.com/2008/02/15/raw-devices-on-rhel-5-or-oel-5/
    Ashok

  • Ocr and vote on raw for 10.2.0.1 install

    Hello,
    I am having a challenge getting the raw devices setup for ocr and vote, 10.2.0.1 and OEL5.
    The raw devices are not getting created and not sure why. I have followed the steps in note 443996.1 and 465001.1 and not having any luck. What am I missing here?
    Here is the contents of the 55-oracle-naming.rules:
    # cat /etc/udev/rules.d/55-oracle-naming.rules
    # Configure persistent, user-defined Oracle Clusterware device file names
    KERNEL=="sda1", BUS=="scsi", PROGRAM=="/sbin/scsi_id", RESULT=="360001fe1000f3270000904609091002c", NAME="ocr1", OWNER="root", GROUP="oinstall", MODE="0640"
    KERNEL=="sdb1", BUS=="scsi", PROGRAM=="/sbin/scsi_id", RESULT=="360001fe1000f3270000904609091002f", NAME="vote1", OWNER="oracle", GROUP="oinstall", MODE="0640"
    and 60-raw.rules
    # This file and interface are deprecated.
    # Applications needing raw device access should open regular
    # block devices with O_DIRECT.
    # Enter raw device bindings here.
    # An example would be:
    # ACTION=="add", KERNEL=="sda", RUN+="/bin/raw /dev/raw/raw1 %N"
    # to bind /dev/raw/raw1 to /dev/sda, or
    # ACTION=="add", ENV{MAJOR}=="8", ENV{MINOR}=="1", RUN+="/bin/raw /dev/raw/raw2 %M %m"
    # to bind /dev/raw/raw2 to the device with major 8, minor 1.
    ACTION=="add", KERNEL=="/dev/sda1", RUN+="/bin/raw /dev/raw/raw1 %N"
    ACTION=="add", KERNEL=="/dev/sdb1", RUN+="/bin/raw /dev/raw/raw2 %N"
    ACTION=="add", KERNEL=="/dev/sdc1", RUN+="/bin/raw /dev/raw/raw3 %N"
    When I use udevtest, udevtest /block/sda/sda1, I get the following results;
    [root@pdrn01 ~]# udevtest /block/sda/sda1
    main: looking at device '/block/sda/sda1' from subsystem 'block'
    run_program: '/bin/bash -c '/sbin/lsmod | /bin/grep ^dm_multipath''
    run_program: '/bin/bash' (stdout) 'dm_multipath 21577 0 '
    run_program: '/bin/bash' returned with status 0
    udev_db_get_device: no db file to read /dev/.udev/db/block@sda: No such file or directory
    udev_db_get_device: no db file to read /dev/.udev/db/block@sda: No such file or directory
    run_program: '/lib/udev/vol_id --export /dev/.tmp-8-1'
    run_program: '/lib/udev/vol_id' returned with status 4
    run_program: '/sbin/scsi_id'
    run_program: '/sbin/scsi_id' returned with status 1
    udev_rules_get_name: no node name set, will use kernel name 'sda1'
    udev_db_get_device: no db file to read /dev/.udev/db/block@sda@sda1: No such file or directory
    This should be getting labeled as /dev/ocr1 as noted in note 465001.1 and the output shows /dev/sda1
    udev_node_add: creating device node '/dev/sda1', major = '8', minor = '1',mode = '0640', uid = '0', gid = '6'
    main: run: '/sbin/multipath -v0 8:1'
    main: run: 'socket:/org/kernel/udev/monitor'
    main: run: '/lib/udev/udev_run_devd'
    main: run: 'socket:/org/freedesktop/hal/udev_event'
    main: run: '/sbin/pam_console_apply /dev/sda1 '
    [root@pdrn01 ~]#
    And 99-raw.rules output;
    KERNEL=="raw[1-2]*", GROUP="oinstall", MODE="640"
    KERNEL=="raw[3-5]*", OWNER="oracle", GROUP="oinstall", MODE="660"
    So, any ideas or suggestions would be HUGELY appreciated!
    TIA, Cheers!

    Here is one note about raw device issue in release 5 and workaround.
    Raw device technology deprecated in Release 5.
    http://oraclepitstop.wordpress.com/2008/02/15/raw-devices-on-rhel-5-or-oel-5/
    Ashok

  • Backup of OCR and voting disk

    Hello,
    Couple of quick questions related to backup of OCR(via ocrconfig -export) and Voting disk(via dd).
    a) I can backup OCR and Voting disk while the database is up and running.
    b) Do I need to login as root to the backup of OCR and Voting disk.
    Thanks,
    C.

    a) I can backup OCR and Voting disk while the database is up and running.
    Yes, the OCR/Voting is at the Clusterware layer, it is not really care about if database is running or not BECAUSE OCR is just a Registry of Clusterware and is static unless you Add/Remove Node/Resources into clusterware.
    OCR/Voting are managed by CRS Stack (CRSCTL), hence can be backed up at any given time, as a matter of fact the backup happens automatically every 4 hours or so ( not sure exact frequency).
    b) Do I need to login as root to the backup of OCR and Voting disk.
    Yes, take the backup as root.
    HTH,

  • Advantage of having OCR and Voting disk on ASM

    What are advantages of putting OCR and Voting disk on ASM from 11g

    well, other than the sharing thing, you dont have to go RAIDing an additional shared disk either. If you have properly configured ASM, redundancy should be built in as well, either soft or hard.
    not sure what other advantages you may need. Theres the IO thing with ASM, buit thats not really an advantage persay with ocr and vote. I may be contradicted by others but Ive never seen performance hit of any kind attributed to ocr and vote on non-ASM disk.

  • Where to put OCR and Voting disks?

    Hi,
    I will be installing Clusterware 11.2.0.2 on a 2 node cluster. I have read several whitepapers where a third asm disk group DATA, FRA, and CRS was created to hold the OCR and Voting disks. In the interest of keeping things simple can I just put the OCR and Voting disks in the DATA DG? I am using external SAN storage.
    Thanks
    Tom Cullen

    Hi,
    yes that is absolutely possible.
    There are several best practices. Some recommend a separate diskgroup, some recommend to put OCR and Voting in the same diskgroup like your data.
    So if there is no special use case it is absolutely valid to only use 2 diskgroups. 1 DATA (including OCR and Vote) and 1 FRA (Fast Recovery Area for Database Backups/Archivelogs/1 Controlfile/1 RedologMember and a second copy of OCR).
    Regards
    Sebastian

Maybe you are looking for