Cloning ASM disks question?

I have a customer who has been running oracle 10.2.0.4 two node rac on HPUX Itanium platform and using raw devices for the datafile locations. They use EMC clraion for the storage platfom.
The customer uses EMC snapshot/cloning technology on the storage side to create a clone of the database luns for their staging database environment. This works well and they are able to simply snap the luns on the storage side and then simply mount them on another host and bring the database copy up an another host.
Well, now that they are in the process of 11g upgrade and the fact that raw devices will no longer be supported going forward for the most part the cusomter would like to start to use ASM for the clustered file system for the RAC databases.
I am trying to find out if there are any white papers or some documentation on how we could make clones when the disks are ASM disks? I know that EMC can snapshot the luns as they have in the past but not sure how we can get them mounted up in another ASM instance on a separate server?
Just wanting to know if anyone has some experience on cloning ASM disk or disk groups between different hosts and maybe some supporting documentation to go along with it.
Thanks.

YOU CAN. You can use the same EMC copy for the ASM Lun, attach it to another server and can start the ASM instance up. It is quiet same as you were doing it with raw devices.
ASM files has ASM headers on it. When you copy the disks over, asm instance can recognize those headers and be able to come online. ASm instance can identify the copied disks from their disk headers.

Similar Messages

  • Oracle ASM Disks Question

    Hi,
    We have two Linux hosts accessing EMC SAN LUNs and we have Oracle ASM Disks created on those devices. Both the Linux Hosts (server1/2)are in once Dell Blade Chassis1 and one of the Server we are planning to move to a different Dell Chassis(2) to have redundency in case of Chassis1 faliure. But if we move one server2 to Chassis2 the HBA WWN Number for the host on the new Dell Chassis2 will change ( WWNs are assigned by Chassis on Dell) and the server2 will loose all the SAN LUNs from EMC Array. I will have to rezone the new server WWNs to storage and pesent the same disks again and rescan the disks which will take couple of reboots. During this period EMC LUNs will not be visible to server2 and I would like to what will happen to ASM disks on the server2 and how to I accomlish this goal? Any detailed steps please? This is a production system with oracle 11g running.
    This is the output of emc powerpath devices (LUNs) which is same on both the hosts.
    [root@server2 ~]# powermt display dev=all | grep emc
    Pseudo name=emcpowerl
    Pseudo name=emcpowerj
    Pseudo name=emcpowerf
    Pseudo name=emcpowerk
    Pseudo name=emcpowerg
    Pseudo name=emcpowerh
    Pseudo name=emcpowerm
    Pseudo name=emcpoweri
    [root@server2 ~]#
    This are the ASM disks on both the hosts:
    [root@server2 ~]# oracleasm listdisks
    DATA01_F1
    DATA01_G1
    DATA01_H1
    DATA01_I1
    FRA01_L1
    LOG01_J1
    LOG01_K1
    [root@server2 ~]#
    Thanks in advance.

    808306 wrote:
    But if we move one server2 to Chassis2 the HBA WWN Number for the host on the new Dell Chassis2 will change ( WWNs are assigned by Chassis on Dell) and the server2 will loose all the SAN LUNs from EMC Array.This does not sound correct to me. WWNs are global - not local. If the WWN is foo123 on server1 for EMC LUN1, why would it not also be foo123 on server2 for EMC LUN1? When you move to server3 or install server3, why would EMC LUN1 not have a WWN of foo123 on server3?
    I will have to rezone the new server WWNs to storage and pesent the same disks again and rescan the disks which will take couple of reboots. Using Powerpath perhaps - using multipath, not. New devices (LUNs) can be made visible by the kernel my refreshing multipath - no reboot required. If existing LUNs are going to be changed - yes, a single reboot will be needed. But not multiple (unless you have a driver issue with stale connections not being reset on the storage server side).
    This is the output of emc powerpath devices (LUNs) which is same on both the hosts.We dropped Powerpath a long time ago. There were compatibility issues between it and AsmLib kernel drivers. Powerpath taints the Linux kernel and limits kernel upgrade flexibility. I refuse to run it on our new clusters. I use multipath instead - the very same s/w used by some of the largest clusters (with Pentabyte storage) in the world. The comfort factor using Open Source drivers (and certified for that kernel version) is significantly higher than installing 3rd party kernel driver binaries - and having to deal with a 2nd party when it comes to o/s and kernel support.

  • Cloned startup disk question

    This past June I cloned my startup hard drive, the original 320GB HDD that came with my Mac Pro new in March 2008. I used Disk Utility's Restore and erased the new drive first, selecting Mac OS Extended, Journaled. This drive, a WD 1TB Caviar Black, is recognized when I boot from another drive because it shows up as a choice using System Preferences > Startup Disk.
    The question I have is that a second, more recent 1TB drive, a Hitachi 1TB drive, was sent through the same process. My Mac Pro will start up fine using this drive, but when I look at System Preferences > Startup Disk, this newly-cloned drive is not an option. The first cloned drive that I mentioned above is showing, and a second option shows, Network Startup. Why does the Hitachi drive I cloned a couple of days ago not show up in System Preferences > Startup Disk even though my system is booted from this drive?
    The drives were cloned using an eSata dock, and no errors were reported during the 2 to 3 hour process.

    Since I see no way to edit this question, I am adding some more info.
    When I restart my Mac and hold down the option key, both cloned drives (mentioned above) are shown as options for startup - the Hitachi and the WD drives, both clones of the original 320GB drive. Also, a third option is Windows. Since I chose the Hitachi drive as my startup drive, the Mount Point shows it is the primary drive.
    When I select System Preferences > Startup Disk, the Hitachi drive still does not show up as an option.
    One more observation is that when I used About This Mac > More Info, this Hitachi drive has
    Partition Map Type:          MBR (Master Boot Record)
    even though I selected Mac OS Extended (Journaled) when I erased the new drive.
    All the other drives (four of them) show
    Partition Map Type:          GPT (GUID Partition Table)
    Do I need to re-format the drive to ensure it is partitioned as GPT?
    Message was edited by: Leon Smith

  • Questions on asm disk discovery:

    Questions on asm disk discovery:
    1)What is the relationship btween asm_diskstring in the init.ora and DiscoveryString in the GPNP profile.xml?
    2)  Which one of the above two finally accounts for the disk discovery process?
    3)  We know that asmlib disks are self describing at the disk header. This overcomes the disk name/path persistency issue as we no long need to rely on the path to discover the asm disks, by setting asm_diskstring='ORCL:*' , ASM instance will identify the right disks automatically. However, I am not sure if setting asm_diskstring='ORCL:*' is the most economic way to do the discovery as I am not sure if Oracle will have to probe all the disks on the OS to determine the right disks. If Oracle has to screen all the disks in this way, then I think setting asm_diskstring='<path_to_asmlib_disk>' will be much faster, although this will be open to the persistent problem. Is my understanding correct?
    Thanks.

    From my understanding all disk you see in /dev/oracleasm/disks are the disks in your system that been discovered by asmlib at discovery stage.
    Currently, due to bug 13465545, ASM instance will discover disks from both locations, ASM_DISKSTRING and gpnp profile, which can cause some mess in disk representation for asm. You can check the settings using asmcmd command: dsget, and set to be the same using dsset.
    I think its more secure to set ASM_DISKSTRING to only the disks used by asm instance.
    ASMCMD> dsget
    Regards
    Ed

  • Questions on asm disk discover:

    Questions on asm disk discover:
    1)What is the relationship btween asm_diskstring in the init.ora and DiscoveryString in the GPNP profile.xml?
    2) which one finally accounts for the disk discovery process?
    3) We know that asmlib disks are self describing at the disk header. This overcomes the disk name/path persistency issue as we do not rely on the path the discover the asmlib
    disks. asm_diskstring='ORCL:*' will identify the right disks. I am not sure if setting 'ORCL:*' is the most economic way as I am not sure if Oracle will have to scan all the disks
    on the OS and probe the disks that it has rigths to determine which disks belong to ASM. If Oracle has to screen all the disks in this way, then I think setting
    asm_diskstring='<path_to_asmlib_disk>' will be much faster. However, this will be open to the persistent problem. Is my understanding correct?
    Thanks.

    Questions on asm disk discovery:
    1)What is the relationship btween asm_diskstring in the init.ora and DiscoveryString in the GPNP profile.xml?
    2)  Which one of the above two finally accounts for the disk discovery process?
    3)  We know that asmlib disks are self describing at the disk header. This overcomes the disk name/path persistency issue as we no long need to rely on the path to discover the asm disks, by setting asm_diskstring='ORCL:*' , ASM instance will identify the right disks automatically. However, I am not sure if setting asm_diskstring='ORCL:*' is the most economic way to do the discovery as I am not sure if Oracle will have to probe all the disks on the OS to determine the right disks. If Oracle has to screen all the disks in this way, then I think setting asm_diskstring='<path_to_asmlib_disk>' will be much faster, although this will be open to the persistent problem. Is my understanding correct?
    Thanks.

  • Asm disk removed taking too much time to boot

    hi
    a local machine is configured with ASM by oracleasm with a total disk.As it was a training purpose so i have removed the disk eg /dev/sdb bluntly ie. just removing it from box.now i tried fsck -c -c -f /dev/ in rescue mode its didnot worked even not mounting the /mnt/sysimage.It claims for ext2fs error and mounting the fs etc.
    Then after lots of days has gone.i fed up with this issues and reinstalled OS.
    but my question is what exactly to do?obviously re-installation is not the exact way to do.
    regards

    If the +/dev/sdb+ was an ASM disk, then it should not impact the o/s when it is removed. ASM instance itself will fail with an error saying something like it was not able to mount the disk group.
    If your system failed to boot correctly after this disk was removed, then +/dev/sdb+ contained more than just ASM data.
    We dynamically add and remove ASM (multipath'ed) disks via kpartx - while o/s is running. No reboot. No problems.
    I fail to see how a ASM disk could cause the type of problems you describe - unless it was more than just a disk used by ASM alone.

  • Problem with create asm disk group

    Hi all
    I am about configuring ASM, so I have downloaded the Grid infrastructure 11g (32 bit), I have configured and created parameters and directories.
    I runned the installer but get stack at the 3 step where I have to change the discovery path. I have taped as path /dev where I have created 3 partitions sdb1, sdc1 and sdd1.
    Is there any thing should I perform on partitions may be or parameters to set before I go through the installation?
    Thanks for help

    You can use the below link to install ASMLIB:
    http://gssdba.wordpress.com/category/asm/
    REFERANCE : Doc ID 580153.1
    There are two different methods to configure ASM on Linux:
    ASM with ASMLib I/O: This method creates all Oracle database files on raw block devices managed by ASM using ASMLib calls. RAW devices are not required with this method as ASMLib works with block devices.
    ASM with Standard Linux I/O: This method creates all Oracle database files on raw character devices managed by ASM using standard Linux I/O system calls. You will be required to create RAW devices for all disk partitions used by ASM.
    You can download the ASMLIB rpm’s from below URL:
    http://www.oracle.com/technetwork/server-storage/linux/downloads/rhel5-084877.html
    STEP 01: LOG IN AS ROOT USER AND INSTALL THE RPMS
    [root@node1 ASMLIB]# rpm -Uvh oracleasm-2.6.18-164.el5-2.0.5-1.el5.i686.rpm \
    > oracleasmlib-2.0.4-1.el5.i386.rpm \
    > oracleasm-support-2.1.8-1.el5.i386.rpm
    warning: oracleasm-2.6.18-164.el5-2.0.5-1.el5.i686.rpm: Header V3 DSA signature: NOKEY, key ID 1e5e0159
    Preparing… ########################################### [100%]
    1:oracleasm-support ########################################### [ 33%]
    2:oracleasm-2.6.18-164.el########################################### [ 67%]
    3:oracleasmlib ########################################### [100%]
    STEP 02: CONFIGURE ASMLIB
    [root@node1 ASMLIB]# /etc/init.d/oracleasm configure
    Configuring the Oracle ASM library driver.
    This will configure the on-boot properties of the Oracle ASM library
    driver. The following questions will determine whether the driver is
    loaded on boot and what permissions it will have. The current values
    will be shown in brackets (‘[]‘). Hitting <ENTER> without typing an
    answer will keep that current value. Ctrl-C will abort.
    Default user to own the driver interface []: oracle
    Default group to own the driver interface []: dba
    Start Oracle ASM library driver on boot (y/n) [n]: y
    Scan for Oracle ASM disks on boot (y/n) [y]: y
    Writing Oracle ASM library driver configuration: done
    Initializing the Oracle ASMLib driver: [ OK ]
    Scanning the system for Oracle ASMLib disks: [ OK ]
    STEP 03 :CREATE ASM DISK
    [root@node1 ASMLIB]# /etc/init.d/oracleasm listdisks
    [root@node1 ASMLIB]#
    [root@node1 ~]# /etc/init.d/oracleasm createdisk VOL1 /dev/sdb1
    Marking disk “VOL1″ as an ASM disk: [ OK ]
    [root@node1 ~]# /etc/init.d/oracleasm createdisk VOL2 /dev/sdc1
    Marking disk “VOL2″ as an ASM disk: [ OK ]
    [root@node1 ~]# /etc/init.d/oracleasm createdisk VOL3 /dev/sdd1
    Marking disk “VOL3″ as an ASM disk: [ OK ]
    [root@node1 ~]# /etc/init.d/oracleasm createdisk VOL4 /dev/sde1
    Marking disk “VOL4″ as an ASM disk: [ OK ]
    [root@node1 ~]# /etc/init.d/oracleasm createdisk VOL5 /dev/sdf1
    Marking disk “VOL5″ as an ASM disk: [ OK ]
    [root@node1 ~]# /etc/init.d/oracleasm listdisks
    VOL1
    VOL2
    VOL3
    VOL4
    VOL5
    [root@node1 ~]#

  • Soft I/O error in ASM disk volume when using iostat commmand

    Dear All,
    We are having 2 node 11g rac/sun solaris 10/ASM storage, when we using iostat in node 2 getting soft I/O error keep on increasing.
    s/w h/w trn tot device
    *82 0 0 82 c4t600A0B80006766CD000003DF4B20A7FCd0*
    s/w h/w trn tot device
    0 0 0 0 md/d10
    0 0 0 0 md/d11
    0 0 0 0 md/d12
    0 0 0 0 md/d20
    0 0 0 0 md/d21
    0 0 0 0 md/d22
    0 0 0 0 md/d30
    0 0 0 0 md/d31
    0 0 0 0 md/d32
    0 0 0 0 md/d40
    0 0 0 0 md/d41
    0 0 0 0 md/d42
    0 0 0 0 md/d50
    4 0 2 6 c0t0d0
    0 0 0 0 c1t2d0
    0 0 0 0 c1t0d0
    0 0 0 0 c1t4d0
    0 0 0 0 c1t1d0
    2 0 0 2 c3t203200A0B86766CDd31
    0 0 0 0 c4t600A0B80006766CD000003E14B20C891d0
    40 0 0 40 c4t600A0B80006766CD000003DF4B20A7FCd0
    0 0 0 0 c4t600A0B80006766CD000003DD4B20A57Fd0
    0 0 0 0 c4t600A0B80006766CD000003C34B0D2D31d0
    2 0 0 2 c4t600C0FF0000000000859635DB0097B00d0
    2 0 0 2 c4t600C0FF000000000085963724A889B00d0
    2 0 0 2 c4t600C0FF0000000000859635AFC5A7F02d0
    2 0 0 2 c4t600C0FF0000000000859635AFC5A7F01d0
    2 0 0 2 c4t600C0FF0000000000859635AFC5A7F00d0
    2 0 0 2 c2t202200A0B86766CDd31
    0 0 0 0 c4t600A0B80006766CD000052274FD0B057d0
    1 0 0 1 c4t600C0FF0000000000859635DB0097B02d0
    2 0 0 2 c4t600C0FF000000000085963565767D300d0
    2 0 0 2 c4t600C0FF00000000008596326C5F42901d0
    1 0 0 1 c4t600C0FF00000000008596326C5F42900d0
    1 0 0 1 c2t203300A0B86766CDd31
    0 0 0 0 proddb2.zakathouse.org:vold(pid672)
    please advice me how to troubshoot and fix this error
    Thanks & Regards

    Hi,
    I suggest to have a look at following script which is provided by oracle to monitor IO.
    ASMIOSTAT Script to collect iostats for ASM disks [ID 437996.1]
    Further this i have questions
    1) Are you logged in asm sysasm? Make sure that you have proper rights
    2) Did you check the liveness of asm instance? Make sure its up and functional also check the asm alert.log for awareness.
    Hope this helps

  • ASM disks in RAC need to be shared disks?

    Hi all,
    sorry for the very basic question but I'm getting a lot of confusion...
    I have clusterware installed on node1 and node2 and I'm going to create ASM instances on both nodes.
    Does the ASM disks need to be physically shared among the two nodes?
    Many thanks for help
    aldo

    ...I've brutally tried both ways and found the answer. Yes the disks need to be physically shared
    aòdp

  • ASM Disk Sizes in a Disk Group

    I have a Diskgroup that contains one ASMDisk that is 500gb. It needs to grow and my question is - Does the next disk added to the group need to be of the same size?
    For example -
    ASM_Disk1 = 500gb and is in ASM_Diskgrp1
    ASM_Disk2 - 250gb >> can it be added to ASM_Diskgrp1
    There is concern about balancing, etc. done by ASM
    Wondering what experiences anyone has had with this?
    Thanks!

    Hi,
    It's not mandatory, You can use disks with diferent sizes, but is extremally recomended that You put disks with the same size in a disk group.
    you'd better use disks with the same storage capacity and performance.
    The main reason is that ASM balances the I/O from all disks in a disk group due to this, put disks with the same same storage capacity and performance helps to improve the I/O and storage efficiently within the ASM disk group.
    Regards,
    Cerreia

  • Share an ASM disk group among multiple nodes

    According to Oracle documentation:
    *“To share an ASM disk group among multiple nodes, you must install Oracle Cluster ware on all of the nodes, regardless of whether you install Oracle RAC on the nodes”.*
    And if I understand it right to share the same ASM storage group from multiple nodes from separate RACs or multiple non-RAC nodes ASM instances in those nodes need to communicate to synchronize ASM related metadata using same technique like cache fusions.
    My question is how this ASM communication take place among different ASM instances located in different RACs and standalone servers. Do we have to have some kind of Interconnect settings among the nodes?

    Hi,
    ASM and database instances require shared access to the disks in a disk group. ASM instances manage the metadata of the disk group and provide file layout information to the database instances.
    ASM instances can be clustered using Oracle Clusterware; there is one ASM instance for each cluster node. If there are several database instances for different databases on the same node, then the database instances share the same single ASM instance on that node.
    If the ASM instance on a node fails, then all of the database instances on that node also fail. Unlike a file system failure, an ASM instance failure does not require restarting the operating system. In an Oracle RAC environment, the ASM and database instances on the surviving nodes automatically recover from an ASM instance failure on a node.
    see this link
    http://download.oracle.com/docs/cd/B28359_01/server.111/b31107/asmcon.htm :)

  • ASM Disk Migration

    HI There,
    Im working on an 11.2.0.3 GI and RDBMS running on AIX 7.1
    I have a situation where one of the one of the ASM disk groups on a particular cluster consists of 20 disks, 10 of which need to be swapped out for 10 new disks.
    As far as I am aware (correct me if im wrong) the following procedure should work, can be done online and should have no impact to any databases using the diskgroup (apart from the asm rebalance):
    -Start with 20 disks (lets call them 'disk1 to disk20' allocated to diskgroup '+DISKGROUP_A'
    - I have ten new disks (lets call them disk21 - disk30'
    - I want to swap out 'disk11 to 'disk20' for new disks 'disk21 to disk30'
    - First Add disks 'disk21 to disk30' to ASM diskgroup '+DISKGROUP_A'
    - Allow rebalance to finish
    - Once rebalance finished drop 'disk11 to disk20' from '+DISKGROUP_A'
    - Allow rebalance to finish again
    - Confirm that disks 'disk11 to disk20' are now shown as 'FORMER'
    - Disks can then physically be removed.
    My Question, is there a better way of doing what I am trying to achieve here, the above process worries me a little, simply because Im putting my self totally in the hands of oracle, crossing my fingers and hoping it works.
    - Any better suggestions?
    - Any suggestions around safeguarding the above approach?
    Any opinions would be appreciated
    Thanks

    Hi
    The best way is to allow the ASM to do the migration:
    Create the new ASM disks
    Add those disks to your ASM diskgroup
    Wait to the re balance processes finish
    Remove old disks from your diskgroup
    Wait to the re balance processes finish
    Drop the old ASM disks
    Disconnect physical disks
    I migrate all the storage (old SAN to new SAN ) for one RAC including OCR_Voting with zero down time, without problems.
    This is the post:
    OCR_Voting Disk Migration
    Regards

  • ASM disk busy 99% only on one cluster node

    Hello,
    We have a three node Oracle RAC cluster. Our dba(s) called us and said they are getting OEM critical alers for an asm disk on one node only. I checked and the SAN attached drive does not show the same high utilization on either of the other two nodes. I checked the hardware and it seems fine. If the issue was with the SAN attached disk, we would be seeing the same errors on all three nodes since they share the same disks. The system crashed last week(alert dump in the +asm directories), and at the disk has been busy ever since. I asked if the dba reviewed the ADDM reports and he said he had and that there were no suspicious looking entries that would lead us to the root cause based on those reports. CPU utilization is fine. I am not sure where to look at this point and any help pointing me in the right direction would be appreciated. They do use RMAN, could there be a backup running using those disks only on one node? Has anyone ever seen this before?
    Thank you,
    Benita Ulisano
    Unix/SAN Team
    Chicago Public Schools
    [email protected]

    Hi Harish,
    Thank you for responding. To answer your question, yes, the disks are all of the same spec and are shared among the three cluster node. The asm disk sdw1 is the one with the issue.
    Problem Node: coefsdb02
    three nodes in RAC cluster
    coefsdb01, coefsdb02, coefsdb03
    iostat results for all three nodes - same disk
    coefsdb01
    Device: rrqm/s wrqm/s r/s w/s rsec/s wsec/s avgrq-sz avgqu-sz await svctm %util
    sdw1 0.00 1.71 0.12 0.58 1.27 18.78 28.63 0.01 13.38 1.75 0.12
    coefsdb02
    sdw1 0.11 0.02 4.00 0.62 305.84 21.72 70.93 2.96 12.58 211.95 97.88
    coefdb03
    sdw1 0.21 0.01 4.70 0.33 224.05 13.52 47.22 0.05 10.11 6.15 3.09
    The dba(s) run RMAN backups, but only on coefsdb01.
    Benita

  • How reInstall Gride Infrastructure and Use old ASM disk groups

    <pre>Hello to all
    I installed Grig Infrastructure 11gR2 on a standalone server (OS is Linux)
    and I configured ASM and my database created on ASM
    Conceive that my OS disk corrupted and OS doesn't start and the Gride Home is on that disk,
    and I have to install OS again
    My ASM disks are safe , Now how can I install Grig Infrastructure again somehow that it can use previous ASM disks
    and disk groups and I don't oblige to create my database again ?
    In the step 2 of installing Gride Infrastructure it has four options
    <pre>
    1.Install and configure Oracle Grid Infrastructure for a Cluster
    2.Configure Oracle Grid Infrastructure for a Standalone Server
    3.Upgrade Oracle Gride Infrastructure or Oracle Automatice Storage Management
    4.Install Oracle Gride Infrastructure Software Only
    </pre>
    If I select the option 2 it wants to create a disk group again
    I guess that I need to select option 4 and then do some configuration but I don't know what I must configure
    Do you know answer of my question , if yes please explain it's stages
    Thank you so much
    </pre>

    Hi,
    no you are not obliged to recreate your database again. However there is a small flaw in the installation procedure, which does not make it 100% easy...
    When you installed the Oracle Restart (Standalone GI), your ASM diskgroup will contain the SPFILE of the ASM instance. And this is exactly the small flaw you will be encountering. So you have 2 options for "recovery":
    1.) Do a software only install (4), and run roothas.pl. This however will not create any ASM entries. You would have to add it manually (using srvctl) and you can specify the ASM Spfile with the srvctl command. Problem here is however to have to know where your ASM spfile has been. If you have a backup of your OLR and a backup of the GPNP profile, this might be easier to find out.
    2.) Do a new installation (2) and configure a new diskgroup (with a "spare" disk or small lun and a new name), that Oracle restart creates ASM instance and the new ASMSpfile for you.
    Then you can simply mount the diskgroup containing your database additionally. You then shoudl however move your new ASMSpfile to the new diskgroup (or simply exchange it with the existing one). In this case it is easier to find out where it was - however you will need a spare (though small) LUN for the new spfile (temporarily, until you exchange it).
    In either case after you have your ASM instance back (and access to your old diskgroup), you have to reregister your database and services - if you do not have an OLR backup.
    Again => It is doable and you can simply mount the ASM diskgroup containing your database. However I suggest you try this one time to know what really needs to be done in this case.
    Regards
    Sebastian

  • ASM DISK groups set up at O/S level

    Hi
    I am having couple of questions regarding the ASM.
    I was given LINUX Server and asked me to create the ASM DISK groups at O/S level.
    Oracle version: Oracle 11g realease 2 with ASM and Stand alone server
    O/S: Redhat Lilnux with 64 bit
    1.Unix admin has already given the server and gave me the root access also.Now I need to map/set up the ASM DISK groups manually and create them at the O/S level and then INSTALL the ASM software.
    a)We are using the asmlib for mapping at o/s level.Can any body give the command to know whether asmlib was installed on the Linux server.
    b)How to see the free Partitions on the Linux server and allot them to ASM disks at O/S level and to create it.
    2)Should I use two separate users ie . one for Grid infrastructure(Asm) and other for RDBMS Installation.
    Say GRID user for Grid Infrastructure(Asm) and ORACLE user for RDBMS Installation.
    If so what is the advantage in using different users for ASM and RDBMS installation respectively.
    The grid user is a member of these OS groups :asmdba,asmoper, and asmadmin and not dba group
    The oracle user is a member of these OS groups :dba and asmdba groups.
    ofcourse grid user and oracle user is also a part of oinstall group.
    Please find the url which I picked the above information.
    http://levipereira.wordpress.com/2010/11/21/installing-the-grid-infrastructure-for-a-standalone-server/
    Thanks in advance
    Thanks
    Veera

    First of all you should be using installation guide from oracle which you can find at ..http://www.oracle.com/pls/db112/portal.portal_db?selected=11&frame=
    If so what is the advantage in using different users for ASM and RDBMS installation respectively.There are two things, Grid & Database, Oracle wants to separate the administartion for both software, where users who are DBAs should not be allowed to change ASM stuff and vice-versa.
    You can use fdisk command to partition the disks, and use # /usr/sbin/oracleasm createdisk command to create the disk.

Maybe you are looking for