RAC - Clusterd file system

Hi All,
This is very basic question but if you can shed some light. I have understanding that Clustered file system gives the ability to read and write same time for a file from multinodes. Is that true ? Can you pls. explain with an example ?
Please confirm if my understanding is correct -
For ( database related files and VOTING Disk and OCR )
Either we can use raw devices
Either we can use CFS ( any other or oracle given CFS )
ASM only for database files any of the above option for VOTING,OCR ..
Now back to orignal question - If we really need cluster file system so that both the nodes can access database files same time then how can ASM has this capability .. Is ASM another clustered file system ?
your response would be appreciated.
Thanks
Lyxx

coming to original questions: Basic concept for allowing concurrent read/write on a file is to allow the operations on block wise locking mechanism, while usual file systems takes file lock for read write. so in cluster file system two nodes can modify two different blocks (not the same).
ASM implements the same through its own locking mechanism, using lock manger processes and CRS architecture.
And its absolutely true that only one ASM instance will suffice for multiple DB but we can create multiple instances if needed (probably for better grouping/isolation of storage)
Experts please correct me if I am missing out on anything.
Thanks
Rajesh

Similar Messages

  • Question about a file system storage option for RAC on 10g

    Hello everyone,
    I am in the beginning of connecting our storage and switches, and building RAC on them but there is a little argument between our specialists.
    We have two database servers(10g with OEL 5) to be clustered and two visible disk groups to each of those nodes. So question is can we choose only one disk group as a shared storage leaving the rest one as the redundant copy during database a creation window while installing the database.  Because some of us argue that oracle database has a built-in capability to decide on what level of RAID we store our data.
    Thank you for your help.

    "some of us argue that oracle database has a built-in capability to decide on what level of RAID we store our data". 
    That statement is not true.  Oracle has optional multiplexing for control files, redo logs, and archive logs but this is not enabled by default and Oracle will not automatically enable it.  If you want redundancy of tables, indexes, temp, and undo you must provide this because Oracle does not offer it standard or as an option.  You can achieve redundancy with RAID at the array level, or host based mirroring (like ASM redundancy groups or Linux mdadm).  This can also depend on your file system because, I think, OCFS2 does not support host based mirroring (so you cannot use mdadm or lvm to mirror the storage if you are using OCFS2).
    Redundancy is not required, but it is recommended if you are using hard disks because they are prone to failures.  You can configure RAID 10 across all disks in the array and present this as one big LUN to the database server.  If you have two storage arrays and you want to mirror the data across the two arrays, then present all of the devices as JBOD and use Linux mdadm to create your RAID group.
    RAC requires shared storage.  Maybe you have a NAS or SAN device, and you will present LUNs to the Oracle database servers.  That is no problem.  The problem is making those LUNs usable by Oracle RAC.  When I used Oracle 10g RAC, I used the Linux raw device facility to manage these LUNs and make them ready for Oracle RAC.  However, raw has been desupported.  Today I would use either ASM or OCFS2.  This has nothing to do with redundancy, this is just because you are using RAC.

  • Does /sapmnt need in cluster file system(SAP ECC 6.0 with oracle RAC)

    We are going to be installing SAP with Oracle 10.2.0.4 RAC on Linux SuSE 10 and OCFS2. The Oracle RAC documentation states:
    You must store the following components in the cluster file system when you use RAC
    in the SAP environment:
    - Oracle Clusterware (CRS) Home
    - Oracle RDBMS Home
    - SAP Home (also /sapmnt)
    - Voting Disks
    - OCR
    - Database
    What I want to ask is if I really need put SAP Home(also /sapmnt) on cluster file system? I will build a two nodes oracel 10g RAC and I also have another two nodes to install SAP CI and DI. My orginial think is sapmnt is a NFS share, and mount to all four nodes(RAC node and CI/DI), and all oracle stuff was on OCFS2(only two rac nodes are OCFS), anybody can tell me if SAP Home(also /sapmnt) can be NFS mount not OCFS2, thanks.
    Best regards,
    Peter

    Hi Peter,
    I don't think you need to keep /sapmnt in  ocfs2 . Reason any file system  need  to be in cluster is,in RAC environment, data stored in the cache of one Oracle instance to be accessed by any other instance by transferring it across the private network  and preserves data integrity and cache coherency by transmitting locking and other synchronization information across cluster nodes.
    AS this applies to redo files, datafiles and control files only ,  you should be fine with nfs mount of /sapmnt sharing across and not having ocfs2.
    -SV

  • Linux Cluster File System partitions for 10g RAC

    Hi Friends,
    I planned to install 2 Node Oracle 10g RAC on RHEL and I planned to use Linux File system itself for OCR,Voting Disk and datafiles (no OCFS2/RAW/ASM)
    I am having SAN storage.
    I would like to know how do i create shared/cluster partitions for OCR,Voting Disk and Datafiles (common storage on SAN).
    Do i need to install any Linux cluster file system for creating these shared partitions (as we have sun cluster in solaris)?
    If so let me know what versions are supported and provide the necessary Note / Link
    Regards,
    DB

    Hi ,
    Below link may be useful to you:
    ORACLE-BASE - Oracle 10g RAC On Linux Using NFS

  • Local file system for archive destination in RAC with ASM

    hi Gurus .
    I need an info.
    I have ASM file system in 2 node RAC.
    The client does not want to use flash recovery area.
    Can we use the archive destination as local file system rather than ASM.
    like /xyzlog1 for archvies coming from node 1 and /xyzlog2 for archive logs coming from node 2.
    Imortant thing is these two destinations are anot shared with each nodes.
    OS is solaris sparc 10.
    version is 10.2.0.2

    There is huge space in the storage.
    Pls tell me in general ho wdo you do this.
    Do we take and one disk from the storage and format it with local file system and share it among the 2 nodes?
    If so then that mount point will have same mount point name if we see it from other node... ..ryt
    In this scenario if one instance is down then from the same shared mount point which is on the same node(down) can apply archves?Here, Earlier you are using any ASM shared location for ARCHIVES ?
    if so you can add a CANDIDATE disk to your existing ARHCIVE disk group(shared).
    if not from LUN's after format, I.e. candidates based on that you can create a new DISK group(shared) according to your space requirements. THen you can assign to log_archive_dest_1 for both the nodes to single shared location(disk group)

  • Single node file system to 3 node rac and asm migration

    hi,
    we have several utl_file and external table applications running on 10.2 single node veritas file system. and we want to migrate to 3 node RAC ASM environment. what is the best practices in order to succeed this migration during this migration. thanks.

    1. Patch to 10.2.0.3 or 10.2.0.4 if not already there.
    2. Dump Veritas from any future consideration.
    3. Build and validate the new RAC environment and then plug in your data using transportable tablespaces.
    Do not expect the first part of step 3 to work perfectly the first time if you do not have experience building RAC clusters.
    This means have appropriate hardware in place for perfecting your skills.
    Be sure, too, that you are not trying to do this with blade or 1U servers. You need a minimum of 2U servers to be able
    to plug in sufficient hardware to have redundant paths to storage and for cache fusion and public access (a minimum of 6 ports).
    And don't let any network admin try to convince you that they can virtualize the network paths: They can not do so successfully
    for RAC.

  • RAC 10gr2 using ASM for RMAN a cluster file system or a Local directory

    The environment is composed of a RAC with 2 nodes using ASM. I have to determine what design is better for Backup and Recovery with RMAN. The backups are going to be saved to disk only. The database is only transactional and small in size
    I am not sure how to create a cluster file system or if it is better to use a local directory. What's the benefit of having a recovery catalog that is optional to the database?
    I very much appreciate your advice and recommendation, Terry

    Arf,
    I am new to RAC. I analyzed Alejandro's script. He is main connection is to the first instance; then through sql*plus, he gets connected to the second instance. he exits the second instance and starts with RMAN backup to the database . Therefore the backup to the database is done from the first instance.
    I do not see where he setenv again to change to the second instance to run RMAN to backup the second instance. It looks to me that the backup is only done to the first instance, but not to the second instance. I may be wrong, but I do not see the second instance backup.
    Kindly, I request your assistance on the steps/connection to backup the second instance. Thank you so much!! Terry

  • Uninstall Oracle Culster (RAC) & Oracle Database with Raw File System (LVM)

    Hello,
    I have configured "Oracle RAC" with "Database" on Hp Unix 11.23 Platform. The shared storage is based on "raw" file system,which has mounted through "SAN" Storage. We have used total 18 LV (logical volumes) for oracle & clusterware.
    By mistaken, I have configured wrong "Public, Private IP Address & Virtual IP Addresses" during RAC installation,thus my oracle "crs" is not working perfectly. So, is it possible to modify the "public,private & Virtual" IP Addresses? If Yes, provide me the step by step modification link.
    Moreover, I would like to know certain steps to uninstall entire "Oracle Clusterware & Database." Kindly do the needful the "uninstallation" steps & is it necessary to remove & recreate LVM partitions, on which "oracle crs & database" has installed?
    Advanced Thanks,
    Nishith Vyas

    Hi,
    You haven't told anything about version...
    By mistaken, I have configured wrong "Public, Private IP Address & Virtual IP Addresses" during RAC installation,thus my oracle "crs" is not working perfectly. So, is it possible to modify the "public,private & Virtual" IP Addresses? If Yes, provide me the step by step modification link.BTW, this thread has information that can drive You to this change:
    Re: Is it possible?
    Moreover, I would like to know certain steps to uninstall entire "Oracle Clusterware & Database." Kindly do the needful the "uninstallation" steps &Again... No versions, but You this thread can drive You to this
    How to cleanup / remove RAC installation (CRS,RDBMS,ASM)
    is it necessary to remove & recreate LVM partitions, on which "oracle crs & database" has installed?I don't think so. If all steps on uninstalling procedure don't work, You'll can try "zero" the information on the raw device using dd.
    Hope it helps,
    Cerreia

  • Oracle RAC binaries on vxfs shared file system

    Hi,
    Is it possible to install oracle binaries on vxfs cluster file system for oracle RAC under sun cluster? Because as I know we can not use vxfs cluster file system for our oracle datafiles.
    TIA

    The above post is incorrect. You can have a cluster (global) file system using VxVM and VxFS. You do not need to have VxVM/CVM for this. A cluster file system using VxVM+VxFS can be used for Oracle binaries but cannot be used for Oracle RAC data files where they are updated from both nodes simultaneously.
    If further clarification is needed, please post.
    Tim
    ---

  • R12 shared application file system with Rac cloning help

    Hello,
    We have R12 with two application servers(Shared File system) and two db servers(on RAC). The same setup we have on test as well. We are using linux servers. Can anyone help me with cloning steps? Shared file system to Shared file system and RAC to RAC cloning. I am waiting for your replies.
    Thanks
    Mehmood

    Hi,
    Please refer to:
    Note: 783188.1 - Certified RAC Scenarios for E-Business Suite Cloning
    Note: 559518.1 - Cloning Oracle E-Business Suite Release 12 RAC-Enabled Systems with Rapid Clone
    Note: 799735.1 - Rapid Clone Documentation Resources, Release 11i and 12
    Note: 756050.1 - Troubleshooting Autoconfig issues with Oracle Applications RAC Databases
    Thanks,
    Hussein

  • Raw devices versus Cluster File Systems in RAC 10gR2

    Hi,
    Does anyone using cluster file systems in a RAC 10gR2 installation, specifically IBM’s GPFS?
    I’ve visited a company that is running RAC 10gR2 in AIX over raw devices. Why someone would choose to use raw devices , with all the problems to administer , when all the modern file systems are so powerful? Is there any issues when using cluster file systems + RAC? Is there considerable performance benefits when using raw devices with RAC ?
    I´ve always used Oracle stand alone instances over file systems (since version 7) , and performance was always very good. I´ve tested raw devices almost 10 years ago , and even in that time (the hardware today is much better - SAN , 15K rpm disks , huge caches - and the file systems software today is much better) the cost to administer it does not compensate the benefits (only 5% more faster than file systems in Oracle 7).
    So , besides any limitations imposed by RAC , why use raw devices nowadays ?
    Regards,
    Antonio Belloni

    Hi,
    spontaneously, my question would be: How did you eliminate the influence of the Linux File System Cache on ext3? OCFS2 is accessed with the o_direct flag - there will be no caching. The same holds true for RAW devices. This could have an influence on your test and I did not see a configuration step to avoid it.
    What I saw, though, is "counter test": "I have tried comparing simple file copies from an OS level and the speed differences are not apparent - so the issue only manifests itself when the data is read via an oracle db." and I have no good answer to that one.
    Maybe this paper has: http://www.oracle.com/technology/tech/linux/pdf/Linux-FS-Performance-Comparison.pdf - it's a bit older, but explains some of the interdependencies.
    Last question: While you spent a lot of effort on proving that this one query is slower on OCFS2 or RAW than on ext3 for the initial read (that's why you flushed the buffer cache before each run), how realistic is this scenario when this system goes into production? I mean, how many times will this query be read completely from disk as opposed to use some block from the buffer? If you consider that, what impact does the "IO read time from disk" have on the overall performance of the system? If you do not isolate the test to just a read, how do writes compare?
    Just some questions. Thanks.

  • File system in 9i RAC

    ASM is recommended for file system and oracle clusterware in 10g RAC, what is the recommended filesystem/cluster in 9i RAC, and which one is commenly used in 9i RAC Environment,

    which filesystem and clusterware is recommended while installing 9i RAC?
    what would be the best practice for using raw devices, as shared storage. Like keeping different file like controlfile, logfile , datafile and backups of different raw devices?
    What could be best practice for LUN configuration for raw devices, like how many LUNs for How many raw devices?
    What is the alternative of OCR, VOTING DISK in 9i RAC?
    Can i manage 9i RAC while i am connect to OEM as standalone?
    lol...too many questions i am asking........

  • Oracle´s File system over GPFS on Oracle RAC

    Hi all,
    I would like to know which Oracle´s File systems should be over GPFS on a Oracle RAC implementation.
    Regards,
    Sebastian.
    Edited by: user12549343 on 02-feb-2010 9:10

    You didn't provide info about DB relase and OS.
    But probably you are talking about AIX OS (I can't see any reason to use GPFS on Linux for RAC and I don't remember if this FS was/is supported for RAC on Linux (check Certification Matrix for this information)). Oracle is provides only one shared "filesystem" for AIX and it is ASM.
    Now it depends on DB version what options you can use.
    Till 11g R2 you have to store OCR and Voting disks outside of ASM (So you needed to use either RAW devices or Cluster Filesystem which GPFS is). Since 11g R2 you can store these disks in ASM directly.
    On Linux if you don't want to use ASM to store database files you can use OCFS2 clusterfile system.

  • Backup RAC with RMAN..(on diff file system)..URGENT..HELP NEEDED

    hello all,
    I have not work a lot on SAN storages. One of my client has implementated 9i RAC. Now he wants to Add two more disk in SAN storage (Implemented RAID). So sun engineers will do this but before this i have to take full database backup(80GB database) throught RMAN. The problem or confussion from my side is that the database is on Sun SAN storage and i have to put the Full database backup taken by RMAN on local hard disk on the node (node 1 of rac). Is the possible since the SAN storage is RAW file system (as i guess) and i am putting the backup on local system.
    Please help me out ...i have do this in couple of days..
    Please tell me prosedure too to how to change the path of backup in RMAN if above is possible..
    Its urgent
    Thanks and Regards!!
    Pankaj Rawat

    You best bet is to get your san folks to give you some disk space for a mount point on the last node in the cluster. Once they assign the lun (not raw) have your SAs make it a mounted file system such as /dbbackup.
    Run all of your rman backups on your last node on the cluster. The only draw back is, if that node dies, then you will not be able to run backups. but your backups would be safe as they are on the san.
    Only way to have a mounted file system on all nodes in your cluser is to have some sort of third party file system cluster ware .. such as veritas or OCFS2 - then all nodes could see it.
    I feel the first node in the cluster (the primary node) is the most busy out of the rest .. so thats why the suggestion of running rman on the last node in the cluster.

  • OPS/RAC Vs Raw Device/File System

    Hello, firts to all, I want to know what is OPS/RAC and Raw Device/File System.
    And then know the diferens between both.
    I apologize for my english.
    Thanks.

    OPS is Oracle parallel server was available up to Oracle 8i and RAC is Real application cluster is available from Oracle 9i.
    Raw device is just presenting the disk to server and using it. File system is presenting disk more transparent and managable form using OS or third party software.
    You need to refer respective documents to know them in detail.
    Ashok

Maybe you are looking for