ASM Disk Remove/Add Question

Hello All,
I have a quick question -
We have an ASM Diskgroup that had 8 Disks in it. I removed 2 Diskss and the Diskgroup has been rebalanced. I have not done anything with the removed Disks.
They both appear in the ASM target as eligble to add back into the Diskgroup.
Do I need to re-format/partition/create disk on them before adding them back in? Would existing data on the Diskl cause an issue in the Diskgroup?
This is the first time I will be adding back Disks that have been previously in the Diskgroup.
Thanks in advance for the advice.
Michele

Hi,
Disks eligible to be assigned to a diskgroup must have the status "CANDIDATE" or "FORMER" or "PROVISIONED".
· CANDIDATE - Disk is not part of a disk group and may be added to a disk group with the ALTER DISKGROUP statement
· PROVISIONED - Disk is not part of a disk group and may be added to a disk group with the ALTER DISKGROUP statement. The PROVISIONED header status is different from the
CANDIDATE header status in that PROVISIONED implies that an additional platform-specific action has been taken by an administrator to make the disk available for ASM.
· FORMER - Disk was once part of a disk group but has been dropped cleanly from the group. It may be added to a new disk group with the ALTER DISKGROUP statement.
The ALTER DISKGROUP...DROP DISK without WAIT option statement returns before the drop and rebalance operations are complete. Do not reuse, remove, or disconnect the dropped disk until the HEADER_STATUS column for this disk in the V$ASM_DISK view changes to FORMER.
You don't need change anything on OS Level. Oracle will reuse asmdisk dropped without needs perform any administrative task on OS Level.
Regards,
Levi Pereira

Similar Messages

  • Asm disk removed taking too much time to boot

    hi
    a local machine is configured with ASM by oracleasm with a total disk.As it was a training purpose so i have removed the disk eg /dev/sdb bluntly ie. just removing it from box.now i tried fsck -c -c -f /dev/ in rescue mode its didnot worked even not mounting the /mnt/sysimage.It claims for ext2fs error and mounting the fs etc.
    Then after lots of days has gone.i fed up with this issues and reinstalled OS.
    but my question is what exactly to do?obviously re-installation is not the exact way to do.
    regards

    If the +/dev/sdb+ was an ASM disk, then it should not impact the o/s when it is removed. ASM instance itself will fail with an error saying something like it was not able to mount the disk group.
    If your system failed to boot correctly after this disk was removed, then +/dev/sdb+ contained more than just ASM data.
    We dynamically add and remove ASM (multipath'ed) disks via kpartx - while o/s is running. No reboot. No problems.
    I fail to see how a ASM disk could cause the type of problems you describe - unless it was more than just a disk used by ASM alone.

  • Questions on asm disk discovery:

    Questions on asm disk discovery:
    1)What is the relationship btween asm_diskstring in the init.ora and DiscoveryString in the GPNP profile.xml?
    2)  Which one of the above two finally accounts for the disk discovery process?
    3)  We know that asmlib disks are self describing at the disk header. This overcomes the disk name/path persistency issue as we no long need to rely on the path to discover the asm disks, by setting asm_diskstring='ORCL:*' , ASM instance will identify the right disks automatically. However, I am not sure if setting asm_diskstring='ORCL:*' is the most economic way to do the discovery as I am not sure if Oracle will have to probe all the disks on the OS to determine the right disks. If Oracle has to screen all the disks in this way, then I think setting asm_diskstring='<path_to_asmlib_disk>' will be much faster, although this will be open to the persistent problem. Is my understanding correct?
    Thanks.

    From my understanding all disk you see in /dev/oracleasm/disks are the disks in your system that been discovered by asmlib at discovery stage.
    Currently, due to bug 13465545, ASM instance will discover disks from both locations, ASM_DISKSTRING and gpnp profile, which can cause some mess in disk representation for asm. You can check the settings using asmcmd command: dsget, and set to be the same using dsset.
    I think its more secure to set ASM_DISKSTRING to only the disks used by asm instance.
    ASMCMD> dsget
    Regards
    Ed

  • Questions on asm disk discover:

    Questions on asm disk discover:
    1)What is the relationship btween asm_diskstring in the init.ora and DiscoveryString in the GPNP profile.xml?
    2) which one finally accounts for the disk discovery process?
    3) We know that asmlib disks are self describing at the disk header. This overcomes the disk name/path persistency issue as we do not rely on the path the discover the asmlib
    disks. asm_diskstring='ORCL:*' will identify the right disks. I am not sure if setting 'ORCL:*' is the most economic way as I am not sure if Oracle will have to scan all the disks
    on the OS and probe the disks that it has rigths to determine which disks belong to ASM. If Oracle has to screen all the disks in this way, then I think setting
    asm_diskstring='<path_to_asmlib_disk>' will be much faster. However, this will be open to the persistent problem. Is my understanding correct?
    Thanks.

    Questions on asm disk discovery:
    1)What is the relationship btween asm_diskstring in the init.ora and DiscoveryString in the GPNP profile.xml?
    2)  Which one of the above two finally accounts for the disk discovery process?
    3)  We know that asmlib disks are self describing at the disk header. This overcomes the disk name/path persistency issue as we no long need to rely on the path to discover the asm disks, by setting asm_diskstring='ORCL:*' , ASM instance will identify the right disks automatically. However, I am not sure if setting asm_diskstring='ORCL:*' is the most economic way to do the discovery as I am not sure if Oracle will have to probe all the disks on the OS to determine the right disks. If Oracle has to screen all the disks in this way, then I think setting asm_diskstring='<path_to_asmlib_disk>' will be much faster, although this will be open to the persistent problem. Is my understanding correct?
    Thanks.

  • Please Help - When I try to add ASM Disk to ASM Diskgroup it crashes Server

    We are using a Pillar SAN and have LUNS Created and are using the following multipath device: (I'm a DBA more then anything else... but I am rather familiar with linux .... SAN Hardware not so much)
    Device Size Mount Point
    /dev/dpda1 11G /u01
    The Above device is working fine... Below are the ASM Disks being Created
    Device Size Oracle ASM Disk Name
    /dev/dpdb1 198G ORCL1
    /dev/dpdc1 21G SIRE1
    /dev/dpdd1 21G CART1
    /dev/dpde1 21G SRTS1
    /dev/dpdf1 21G CRTT1
    I try to create to the first ASM Disk
    /etc/init.d/oracleasm createdisk ORCL1 /dev/dpdb1
    Marking disk "ORCL1" as an ASM disk: [FAILED]
    So I check the oracleasm log:
    #cat /var/log/oracleasm
    Device "/dev/dpdb1" is not a partition
    I did some research and found that this is a common problem with multipath devices and to work around it you have to use asmtool
    # /usr/sbin/asmtool -C -l /dev/oracleasm -n ORCL1 -s /dev/dpdb1 -a force=yes
    asmtool: Device "/dev/dpdb1" is not a partition
    asmtool: Continuing anyway
    now I scan and list the disks
    # /etc/init.d/oracleasm scandisks
    Scanning the system for Oracle ASMLib disks: [  OK  ]
    # /etc/init.d/oracleasm listdisks
    ORCL1
    Here is whats going on in /var/log/messages when I run the oracleasm scandisks command
    # date
    Fri Aug 14 13:51:58 MST 2009
    # /etc/init.d/oracleasm scandisks
    Scanning the system for Oracle ASMLib disks: [  OK  ]
    cat /var/log/messages | grep "Aug 14 13:5"
    Aug 14 13:52:06 seer kernel: dpdb: dpdb1
    Aug 14 13:52:06 seer kernel: dpdc: dpdc1
    Aug 14 13:52:06 seer kernel: dpdd: dpdd1
    Aug 14 13:52:06 seer kernel: dpde: dpde1
    Aug 14 13:52:06 seer kernel: dpdf: dpdf1
    Aug 14 13:52:06 seer kernel: dpdg: dpdg1
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: printk: 30 messages suppressed.
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: sda : READ CAPACITY failed.
    Aug 14 13:52:06 seer kernel: sda : status=1, message=00, host=0, driver=08
    Aug 14 13:52:06 seer kernel: sd: Current: sense key: Illegal Request
    Aug 14 13:52:06 seer kernel: Add. Sense: Logical unit not supported
    Aug 14 13:52:06 seer kernel:
    Aug 14 13:52:06 seer kernel: sda: test WP failed, assume Write Enabled
    Aug 14 13:52:06 seer kernel: sda: asking for cache data failed
    Aug 14 13:52:06 seer kernel: sda: assuming drive cache: write through
    Aug 14 13:52:06 seer kernel: sda:end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: Dev sda: unable to read RDB block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sda, logical block 0
    Aug 14 13:52:06 seer kernel: unable to read partition table
    Aug 14 13:52:06 seer kernel: SCSI device sdb: 21502464 512-byte hdwr sectors (11009 MB)
    Aug 14 13:52:06 seer kernel: sdb: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdb: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdb: sdb1
    Aug 14 13:52:06 seer kernel: SCSI device sdc: 421476864 512-byte hdwr sectors (215796 MB)
    Aug 14 13:52:06 seer kernel: sdc: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdc: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdc: sdc1
    Aug 14 13:52:06 seer kernel: SCSI device sdd: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdd: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdd: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdd: sdd1
    Aug 14 13:52:06 seer kernel: SCSI device sde: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sde: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sde: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sde: sde1
    Aug 14 13:52:06 seer kernel: SCSI device sdf: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdf: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdf: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdf: sdf1
    Aug 14 13:52:06 seer kernel: SCSI device sdg: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdg: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdg: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdg: sdg1
    Aug 14 13:52:06 seer kernel: SCSI device sdh: 2107390464 512-byte hdwr sectors (1078984 MB)
    Aug 14 13:52:06 seer kernel: sdh: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdh: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdh: sdh1
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdi, sector 0
    Aug 14 13:52:06 seer kernel: Buffer I/O error on device sdi, logical block 0
    Aug 14 13:52:06 seer kernel: sdi : READ CAPACITY failed.
    Aug 14 13:52:06 seer kernel: sdi : status=1, message=00, host=0, driver=08
    Aug 14 13:52:06 seer kernel: sd: Current: sense key: Illegal Request
    Aug 14 13:52:06 seer kernel: Add. Sense: Logical unit not supported
    Aug 14 13:52:06 seer kernel:
    Aug 14 13:52:06 seer kernel: sdi: test WP failed, assume Write Enabled
    Aug 14 13:52:06 seer kernel: sdi: asking for cache data failed
    Aug 14 13:52:06 seer kernel: sdi: assuming drive cache: write through
    Aug 14 13:52:06 seer kernel: sdi:end_request: I/O error, dev sdi, sector 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdi, sector 0
    Aug 14 13:52:06 seer last message repeated 4 times
    Aug 14 13:52:06 seer kernel: Dev sdi: unable to read RDB block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdi, sector 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdi, sector 0
    Aug 14 13:52:06 seer kernel: unable to read partition table
    Aug 14 13:52:06 seer kernel: SCSI device sdj: 21502464 512-byte hdwr sectors (11009 MB)
    Aug 14 13:52:06 seer kernel: sdj: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdj: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdj: sdj1
    Aug 14 13:52:06 seer kernel: SCSI device sdk: 421476864 512-byte hdwr sectors (215796 MB)
    Aug 14 13:52:06 seer kernel: sdk: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdk: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdk: sdk1
    Aug 14 13:52:06 seer kernel: SCSI device sdl: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdl: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdl: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdl: sdl1
    Aug 14 13:52:06 seer kernel: SCSI device sdm: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdm: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdm: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdm: sdm1
    Aug 14 13:52:06 seer kernel: SCSI device sdn: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdn: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdn: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdn: sdn1
    Aug 14 13:52:06 seer kernel: SCSI device sdo: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdo: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdo: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdo: sdo1
    Aug 14 13:52:06 seer kernel: SCSI device sdp: 2107390464 512-byte hdwr sectors (1078984 MB)
    Aug 14 13:52:06 seer kernel: sdp: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdp: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdp: sdp1
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdq, sector 0
    Aug 14 13:52:06 seer kernel: sdq : READ CAPACITY failed.
    Aug 14 13:52:06 seer kernel: sdq : status=1, message=00, host=0, driver=08
    Aug 14 13:52:06 seer kernel: sd: Current: sense key: Illegal Request
    Aug 14 13:52:06 seer kernel: Add. Sense: Logical unit not supported
    Aug 14 13:52:06 seer kernel:
    Aug 14 13:52:06 seer kernel: sdq: test WP failed, assume Write Enabled
    Aug 14 13:52:06 seer kernel: sdq: asking for cache data failed
    Aug 14 13:52:06 seer kernel: sdq: assuming drive cache: write through
    Aug 14 13:52:06 seer kernel: sdq:end_request: I/O error, dev sdq, sector 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdq, sector 0
    Aug 14 13:52:06 seer last message repeated 5 times
    Aug 14 13:52:06 seer kernel: Dev sdq: unable to read RDB block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdq, sector 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdq, sector 0
    Aug 14 13:52:06 seer kernel: unable to read partition table
    Aug 14 13:52:06 seer kernel: SCSI device sdr: 21502464 512-byte hdwr sectors (11009 MB)
    Aug 14 13:52:06 seer kernel: sdr: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdr: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdr: sdr1
    Aug 14 13:52:06 seer kernel: SCSI device sds: 421476864 512-byte hdwr sectors (215796 MB)
    Aug 14 13:52:06 seer kernel: sds: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sds: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sds: sds1
    Aug 14 13:52:06 seer kernel: SCSI device sdt: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdt: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdt: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdt: sdt1
    Aug 14 13:52:06 seer kernel: SCSI device sdu: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdu: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdu: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdu: sdu1
    Aug 14 13:52:06 seer kernel: SCSI device sdv: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdv: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdv: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdv: sdv1
    Aug 14 13:52:06 seer kernel: SCSI device sdw: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdw: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdw: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdw: sdw1
    Aug 14 13:52:06 seer kernel: SCSI device sdx: 2107390464 512-byte hdwr sectors (1078984 MB)
    Aug 14 13:52:06 seer kernel: sdx: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdx: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdx: sdx1
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdy, sector 0
    Aug 14 13:52:06 seer kernel: sdy : READ CAPACITY failed.
    Aug 14 13:52:06 seer kernel: sdy : status=1, message=00, host=0, driver=08
    Aug 14 13:52:06 seer kernel: sd: Current: sense key: Illegal Request
    Aug 14 13:52:06 seer kernel: Add. Sense: Logical unit not supported
    Aug 14 13:52:06 seer kernel:
    Aug 14 13:52:06 seer kernel: sdy: test WP failed, assume Write Enabled
    Aug 14 13:52:06 seer kernel: sdy: asking for cache data failed
    Aug 14 13:52:06 seer kernel: sdy: assuming drive cache: write through
    Aug 14 13:52:06 seer kernel: sdy:end_request: I/O error, dev sdy, sector 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdy, sector 0
    Aug 14 13:52:06 seer last message repeated 5 times
    Aug 14 13:52:06 seer kernel: Dev sdy: unable to read RDB block 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdy, sector 0
    Aug 14 13:52:06 seer kernel: end_request: I/O error, dev sdy, sector 0
    Aug 14 13:52:06 seer kernel: unable to read partition table
    Aug 14 13:52:06 seer kernel: SCSI device sdz: 21502464 512-byte hdwr sectors (11009 MB)
    Aug 14 13:52:06 seer kernel: sdz: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdz: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdz: sdz1
    Aug 14 13:52:06 seer kernel: SCSI device sdaa: 421476864 512-byte hdwr sectors (215796 MB)
    Aug 14 13:52:06 seer kernel: sdaa: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdaa: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdaa: sdaa1
    Aug 14 13:52:06 seer kernel: SCSI device sdab: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdab: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdab: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdab: sdab1
    Aug 14 13:52:06 seer kernel: SCSI device sdac: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdac: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdac: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdac: sdac1
    Aug 14 13:52:06 seer kernel: SCSI device sdad: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdad: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdad: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdad: sdad1
    Aug 14 13:52:06 seer kernel: SCSI device sdae: 43006464 512-byte hdwr sectors (22019 MB)
    Aug 14 13:52:06 seer kernel: sdae: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdae: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdae: sdae1
    Aug 14 13:52:06 seer kernel: SCSI device sdaf: 2107390464 512-byte hdwr sectors (1078984 MB)
    Aug 14 13:52:06 seer kernel: sdaf: Write Protect is off
    Aug 14 13:52:06 seer kernel: SCSI device sdaf: drive cache: write through w/ FUA
    Aug 14 13:52:06 seer kernel: sdaf: sdaf1
    Aug 14 13:52:06 seer kernel: scsi_wr_disk: unknown partition table
    Aug 14 13:52:07 seer kernel: end_request: I/O error, dev sda, sector 0
    Aug 14 13:52:07 seer kernel: end_request: I/O error, dev sdi, sector 0
    Aug 14 13:52:07 seer kernel: end_request: I/O error, dev sdq, sector 0
    Aug 14 13:52:07 seer kernel: end_request: I/O error, dev sdy, sector 0
    Here's some extra info:
    # /sbin/blkid | grep asm
    /dev/sdc1: LABEL="ORCL1" TYPE="oracleasm"
    /dev/sdk1: LABEL="ORCL1" TYPE="oracleasm"
    /dev/sds1: LABEL="ORCL1" TYPE="oracleasm"
    /dev/sdaa1: LABEL="ORCL1" TYPE="oracleasm"
    /dev/dpdb1: LABEL="ORCL1" TYPE="oracleasm"
    I have learned that by excluding devices in the oracleasm configuration file I eliminate those I/O errors in /var/log/messages
    # cat /etc/sysconfig/oracleasm
    # This is a configuration file for automatic loading of the Oracle
    # Automatic Storage Management library kernel driver. It is generated
    # By running /etc/init.d/oracleasm configure. Please use that method
    # to modify this file
    # ORACLEASM_ENABELED: 'true' means to load the driver on boot.
    ORACLEASM_ENABLED=true
    # ORACLEASM_UID: Default user owning the /dev/oracleasm mount point.
    ORACLEASM_UID=oracle
    # ORACLEASM_GID: Default group owning the /dev/oracleasm mount point.
    ORACLEASM_GID=oinstall
    # ORACLEASM_SCANBOOT: 'true' means scan for ASM disks on boot.
    ORACLEASM_SCANBOOT=true
    # ORACLEASM_SCANORDER: Matching patterns to order disk scanning
    ORACLEASM_SCANORDER="dp sd"
    # ORACLEASM_SCANEXCLUDE: Matching patterns to exclude disks from scan
    ORACLEASM_SCANEXCLUDE="sdc sdk sds sdaa sda"
    # ls -la /dev/oracleasm/disks/
    total 0
    drwxr-xr-x 1 root root 0 Aug 14 10:47 .
    drwxr-xr-x 4 root root 0 Aug 13 15:32 ..
    brw-rw---- 1 oracle oinstall 251, 33 Aug 14 13:46 ORCL1
    Now I can go into dbca to create the ASM instance, which starts up fine...  create a new diskgroup, I see ORCL1 as a provision ASM disk I select it ...  Click OK
    CRASH!!!  Box hangs have to reboot it....
    I have gotten myself to exactly the same point right before clicking OK and here is what is in the ASM alertlog so far
    Fri Aug 14 14:42:02 2009
    Starting ORACLE instance (normal)
    LICENSE_MAX_SESSION = 0
    LICENSE_SESSIONS_WARNING = 0
    Picked latch-free SCN scheme 3
    Using LOG_ARCHIVE_DEST_1 parameter default value as /u01/app/oracle/product/11.1.0/db_1/dbs/arch
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =0
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    Starting up ORACLE RDBMS Version: 11.1.0.6.0.
    Using parameter settings in server-side spfile /u01/app/oracle/product/11.1.0/db_1/dbs/spfile+ASM.ora
    System parameters with non-default values:
    large_pool_size = 12M
    instance_type = "asm"
    diagnostic_dest = "/u01/app/oracle"
    Fri Aug 14 14:42:04 2009
    PMON started with pid=2, OS id=3300
    Fri Aug 14 14:42:04 2009
    VKTM started with pid=3, OS id=3302 at elevated priority
    VKTM running at (20)ms precision
    Fri Aug 14 14:42:04 2009
    DIAG started with pid=4, OS id=3306
    Fri Aug 14 14:42:04 2009
    PSP0 started with pid=5, OS id=3308
    Fri Aug 14 14:42:04 2009
    DSKM started with pid=6, OS id=3310
    Fri Aug 14 14:42:04 2009
    DIA0 started with pid=7, OS id=3312
    Fri Aug 14 14:42:04 2009
    MMAN started with pid=8, OS id=3314
    Fri Aug 14 14:42:04 2009
    DBW0 started with pid=9, OS id=3316
    Fri Aug 14 14:42:04 2009
    LGWR started with pid=6, OS id=3318
    Fri Aug 14 14:42:04 2009
    CKPT started with pid=10, OS id=3320
    Fri Aug 14 14:42:04 2009
    SMON started with pid=11, OS id=3322
    Fri Aug 14 14:42:04 2009
    RBAL started with pid=12, OS id=3324
    Fri Aug 14 14:42:04 2009
    GMON started with pid=13, OS id=3326
    ORACLE_BASE from environment = /u01/app/oracle
    Fri Aug 14 14:42:04 2009
    SQL> ALTER DISKGROUP ALL MOUNT
    Fri Aug 14 14:42:41 2009
    At this point I don't want to click the OK until I am sure someone is in the office to reboot the machine manually if I do hang it again....  I hung it twice yesterday, however I did not have the devices excluded in the oracleasm configuration file as i do now
    Edited by: user10193377 on Aug 14, 2009 3:23 PM
    Well Clicking OK hun it again and I am waiting to get back into it, to see what new information might be gleened
    Does anyone have any ideas on what to check or where to look?????    Will update more once I can log back in

    Hi Mark,
    It looks like something is not correct with your raw device partition based on the error messages:
    Aug 14 13:52:06 seer kernel: Add. Sense: Logical unit not supported
    Aug 14 13:52:06 seer kernel:
    Aug 14 13:52:06 seer kernel: sda: test WP failed, assume Write Enabled
    Aug 14 13:52:06 seer kernel: sda: asking for cache data failed
    Aug 14 13:52:06 seer kernel: sda: assuming drive cache: write through
    Aug 14 13:52:06 seer kernel: sda:end_request: I/O error, dev sda, sector 0
    It could be a number of things. I would check with your vendor and Oracle support to see if the multipath software drive is supported and if there is a potential workaround for ASM. Sorry this is not quite the solution, but its what jumps to mind based on issues with multipath software and storage vendors for ASM with Linux and Oracle. Have you checked the validation matrix available on Metalink?
    Cheers,
    Ben

  • Add ASM disk in RAC

    Hello everyone,
    We are on 10gR2 ASM.
    The system admin has provided one disk as follows:
    On node-1 (tphsmsd1)
    /dev/rdisk/disk83 CANDIDATE
    On node-2 (tphsmsd2)
    /dev/rdisk/disk77 CANDIDATE
    I've read that paths of ASM disks in RAC may be different.
    When I give "ls -l /dev/rdisk/disk83" gives no such files and vice-versa for disk on node2.
    I don't understand how ASM handles this ? If I add disk (disk83) on node 1 how it's going to get to that disk on node 2 ?
    thanks
    Jitu Keshwani

    Jitu Keshwani wrote:
    I've read that paths of ASM disks in RAC may be different.
    When I give "ls -l /dev/rdisk/disk83" gives no such files and vice-versa for disk on node2.
    I don't understand how ASM handles this ? If I add disk (disk83) on node 1 how it's going to get to that disk on node 2 ?The header label of the disk, identifies the ASM disk name and diskgroup it belongs to. You can hexdump or octaldump the 1st 128 bytes of the device as ascii chars to view the label. The string "+ORCLDISK+" in the header identifies it as an ASM disk. This is then followed by the ASM diskname and then the AM diskgroup it belongs to.
    But there's no reason for not having a consistent and static device name layer across all cluster nodes. Each scsi device has a WWID (World Wide ID)- a unique identifier. This means that the kernel (and kernel drivers) can uniquely identify a device.
    On Linux, Multipath is used to map a logical device name to a WWID - and using the same +/etc/multipath.conf+ configuration file on all cluster nodes, ensures that the same device names are used across the cluster.
    But from the the device file entry you listed, you're likely not running Linux? In that case, depending on the Unix flavour used and the type of cluster storage, there can be similar options to Linux's Multipath. One such option is EMC's PowerPath - but that of course is specific to EMC SANs and requires additional licensing fees.

  • Removal of ASM Disks

    Hi I am using Oracle 10gR2 on Solaris 10. I have some 12 ASM disks in a Disk Group. I want to remove 3 ASM Disks. Can I do this operation online? Will my data be affected? How should I proceed about it? When I will remove the ASM Disks, can I safely detach those disks physically? please help.
    regards

    ahsen.javaid wrote:
    Hi I am using Oracle 10gR2 on Solaris 10. I have some 12 ASM disks in a Disk Group. I want to remove 3 ASM Disks. Can I do this operation online? As already commented, yes. I would like to give you a practical example of how utterly neat ASM is in this regard and why it is IMO a major mistake not to use ASM.
    We had to swap storage systems. This means that we have to migrate an entire Oracle database (over 1TB in size) from one set of LUNs to a completely different set of LUNs (on a different storage system).
    The new LUNs are added to the existing diskgroup. The old LUNs are dropped (the drop is of course not immediate). A rebalance is issued. The next morning the entire database has been moved and restriped on the new set of LUNs and the old LUNs can now be removed from the system. This while the database was up and running and normal processing continued uninterrupted (this is a 24x7 database).
    Then I get comments from some customers (like from a major local financial institution recently) that "+No, we do not want to use ASM as we rather use Veritas+". People like that are missing the point of what ASM is. Totally.

  • How to add ASM Disk to Disk Group that is shown as MEMBER?

    HI,
    We have a Production Oracle RAC on HP-UX. We have Two Disk Groups one for Arcive (ARC_DISK) and other for Database(DATA_DISK).
    Today I wanted to add another 200 GB of Disk Space to the DATA_DISK Group.
    I opened DBCA and could not find the new disk in the 'Show Candidates' option. But after the Unix admin changed the ownership and permission of the Disk, it was shown in 'Show Candidates' option. I selected this Disk and continued but got an error later as we didnt change the ownership and permission from the 2nd Unix server. After doing this, , when I open DBCA to add disk, nothing is shown in 'Show Candidates' option but when we click on 'Show All', the new disk is listed as MEMBER (header status) but not allocated to any Disk Group.
    I would like to know how to allocate this Disk (which is already shown as MEMBER) to the DATA_DISK - ASM diskgroup. This is a production database system.
    Rgds,
    Thiru

    user1983888 wrote:
    Hi,
    Due to huge difference in Disk Size which I am trying to add to the existing Disk Group which has 1024 GB of Disk, would it be a better idea to create a new Disk Group with 200g GB Disk? Will this command work as the header_status is already MEMBER?
    SQL>CREATE DISKGROUP DATA_DISK1
    EXTERNAL REDUNDANCY
    DISK 'disk path';
    Thanks.Yes, you can create new diskgroup (DATA_DISK1). Just make sure all datafile in DATA_DISK autoextend off, then create all your new datafile to the new diskgroup DATA_DISK1. The command will not work as the header_status is MEMBER. You'll receive ORA-15018 & ORA-15033 error. As I mentioned before, you need to use FORCE option or format the disk using dd* command before perform above command.
    1st Option : (Format the raw disk using dd command creating new diskgroup. Header_status MEMBER -> CANDIDATE). E.g.
    1. dd if=/dev/zero of=/dev/rdsk/c2t1d0 bs=8192 count=1000 //any rac nodes
    2. # chown oracle:oinstall /dev/rdsk/c2t1d0 //perform on both node. This apply if using LVM or direct access to raw
    3. SQL> create diskgroup data_disk1 external redundancy disk '/dev/rdsk/c2t1d0'; //any RAC node in ASM instance
    4. SQL> alter diskgroup data_disk1 mount; //try to mount on the other node
    2nd Option : (Using FORCE option when creating the new diskgroup. Header_Status = MEMBER)
    1. # chown oracle:oinstall /dev/rdsk/c2t1d0 //perform on both node. This apply if using LVM or direct access to raw
    2. SQL> create diskgroup data_disk1 external redundancy disk '/dev/rdsk/c2t1d0' force; //any RAC node in ASM instance
    3. SQL> alter diskgroup data_disk1 mount; //try to mount on the other node
    Just be careful when using FORCE option or DD command. If you perform this operation on existing ASM disk groups disk, it may destroy existing disk groups.

  • Cloning ASM disks question?

    I have a customer who has been running oracle 10.2.0.4 two node rac on HPUX Itanium platform and using raw devices for the datafile locations. They use EMC clraion for the storage platfom.
    The customer uses EMC snapshot/cloning technology on the storage side to create a clone of the database luns for their staging database environment. This works well and they are able to simply snap the luns on the storage side and then simply mount them on another host and bring the database copy up an another host.
    Well, now that they are in the process of 11g upgrade and the fact that raw devices will no longer be supported going forward for the most part the cusomter would like to start to use ASM for the clustered file system for the RAC databases.
    I am trying to find out if there are any white papers or some documentation on how we could make clones when the disks are ASM disks? I know that EMC can snapshot the luns as they have in the past but not sure how we can get them mounted up in another ASM instance on a separate server?
    Just wanting to know if anyone has some experience on cloning ASM disk or disk groups between different hosts and maybe some supporting documentation to go along with it.
    Thanks.

    YOU CAN. You can use the same EMC copy for the ASM Lun, attach it to another server and can start the ASM instance up. It is quiet same as you were doing it with raw devices.
    ASM files has ASM headers on it. When you copy the disks over, asm instance can recognize those headers and be able to come online. ASm instance can identify the copied disks from their disk headers.

  • Oracle ASM Disks Question

    Hi,
    We have two Linux hosts accessing EMC SAN LUNs and we have Oracle ASM Disks created on those devices. Both the Linux Hosts (server1/2)are in once Dell Blade Chassis1 and one of the Server we are planning to move to a different Dell Chassis(2) to have redundency in case of Chassis1 faliure. But if we move one server2 to Chassis2 the HBA WWN Number for the host on the new Dell Chassis2 will change ( WWNs are assigned by Chassis on Dell) and the server2 will loose all the SAN LUNs from EMC Array. I will have to rezone the new server WWNs to storage and pesent the same disks again and rescan the disks which will take couple of reboots. During this period EMC LUNs will not be visible to server2 and I would like to what will happen to ASM disks on the server2 and how to I accomlish this goal? Any detailed steps please? This is a production system with oracle 11g running.
    This is the output of emc powerpath devices (LUNs) which is same on both the hosts.
    [root@server2 ~]# powermt display dev=all | grep emc
    Pseudo name=emcpowerl
    Pseudo name=emcpowerj
    Pseudo name=emcpowerf
    Pseudo name=emcpowerk
    Pseudo name=emcpowerg
    Pseudo name=emcpowerh
    Pseudo name=emcpowerm
    Pseudo name=emcpoweri
    [root@server2 ~]#
    This are the ASM disks on both the hosts:
    [root@server2 ~]# oracleasm listdisks
    DATA01_F1
    DATA01_G1
    DATA01_H1
    DATA01_I1
    FRA01_L1
    LOG01_J1
    LOG01_K1
    [root@server2 ~]#
    Thanks in advance.

    808306 wrote:
    But if we move one server2 to Chassis2 the HBA WWN Number for the host on the new Dell Chassis2 will change ( WWNs are assigned by Chassis on Dell) and the server2 will loose all the SAN LUNs from EMC Array.This does not sound correct to me. WWNs are global - not local. If the WWN is foo123 on server1 for EMC LUN1, why would it not also be foo123 on server2 for EMC LUN1? When you move to server3 or install server3, why would EMC LUN1 not have a WWN of foo123 on server3?
    I will have to rezone the new server WWNs to storage and pesent the same disks again and rescan the disks which will take couple of reboots. Using Powerpath perhaps - using multipath, not. New devices (LUNs) can be made visible by the kernel my refreshing multipath - no reboot required. If existing LUNs are going to be changed - yes, a single reboot will be needed. But not multiple (unless you have a driver issue with stale connections not being reset on the storage server side).
    This is the output of emc powerpath devices (LUNs) which is same on both the hosts.We dropped Powerpath a long time ago. There were compatibility issues between it and AsmLib kernel drivers. Powerpath taints the Linux kernel and limits kernel upgrade flexibility. I refuse to run it on our new clusters. I use multipath instead - the very same s/w used by some of the largest clusters (with Pentabyte storage) in the world. The comfort factor using Open Source drivers (and certified for that kernel version) is significantly higher than installing 3rd party kernel driver binaries - and having to deal with a 2nd party when it comes to o/s and kernel support.

  • Move ASM disks with database from one server to another?

    I have a 11.2.0 ASM with a 11.2.0 database on two internal disks on server1. The disks are not in any raid or volume manager configuration, i.e. they are just two disks. The disks were physically removed from server1 and installed on server2, which is the same hardware, OS, patch level etc, in the same target position. Installed the 11.2.0 rdbms and grid infrastructure binaries on server2; changed the raw disk partition ownership to oracle and started asmca. asmca does not see the disks.
    My question, is this possible and if so, what am I missing?

    I did not manually create an ASM instance and try to add the diskgoup. My understanding is, if you use asmca then it starts an ASM instance, if one is not already running, and lets you configure your diskgroup. This is what I am trying to do, with no luck (asmca does not see the disks)I may be totally wrong here but afaik dbca creates the ASM instance. ASMCA is just for adding disks or diskgroups......
    Did you try to create an ASM instance with dbca? When doing so do not put your disks in there - this will create a new disk group.....
    Ronny Egner
    My Blog: http://blog.ronnyegner-consulting.de

  • ASM Disk Migration

    HI There,
    Im working on an 11.2.0.3 GI and RDBMS running on AIX 7.1
    I have a situation where one of the one of the ASM disk groups on a particular cluster consists of 20 disks, 10 of which need to be swapped out for 10 new disks.
    As far as I am aware (correct me if im wrong) the following procedure should work, can be done online and should have no impact to any databases using the diskgroup (apart from the asm rebalance):
    -Start with 20 disks (lets call them 'disk1 to disk20' allocated to diskgroup '+DISKGROUP_A'
    - I have ten new disks (lets call them disk21 - disk30'
    - I want to swap out 'disk11 to 'disk20' for new disks 'disk21 to disk30'
    - First Add disks 'disk21 to disk30' to ASM diskgroup '+DISKGROUP_A'
    - Allow rebalance to finish
    - Once rebalance finished drop 'disk11 to disk20' from '+DISKGROUP_A'
    - Allow rebalance to finish again
    - Confirm that disks 'disk11 to disk20' are now shown as 'FORMER'
    - Disks can then physically be removed.
    My Question, is there a better way of doing what I am trying to achieve here, the above process worries me a little, simply because Im putting my self totally in the hands of oracle, crossing my fingers and hoping it works.
    - Any better suggestions?
    - Any suggestions around safeguarding the above approach?
    Any opinions would be appreciated
    Thanks

    Hi
    The best way is to allow the ASM to do the migration:
    Create the new ASM disks
    Add those disks to your ASM diskgroup
    Wait to the re balance processes finish
    Remove old disks from your diskgroup
    Wait to the re balance processes finish
    Drop the old ASM disks
    Disconnect physical disks
    I migrate all the storage (old SAN to new SAN ) for one RAC including OCR_Voting with zero down time, without problems.
    This is the post:
    OCR_Voting Disk Migration
    Regards

  • ASM Disk offline - State hung

    Hey in my ASM configuration I setup this
    Disk Path ASM Name Failure Group
    /dev/raw/raw4 Data1 FG1_Data [on SAN1]
    /dev/raw/raw5 Data2 FG2_Data [on SAN2]
    /dev/raw/raw6 Reco1 FG1_Reco [on SAN1]
    /dev/raw/raw7 Reco2 FG2_Reco [on SAN2]
    all good.
    My I switched off SAN1. A query on v$asm_disk shows disk DATA1 & RECO1 as offline - hung state.
    I did in the asm instance in sqlplus:
    ALTER DISKGROUP DATA DROP DISK data1; -> success
    ALTER DISKGROUP RECO DROP DISK reco1; -> success
    ALTER DISKGROUP DATA ADD failgroup DATA3 '/dev/raw/raw4'; -> success
    ALTER DISKGROUP RECO ADD failgroup RECO3 '/dev/raw/raw6'; -> success
    The query on v$asm_disk stills show the DATA1 and RECO1 - but DATA3 and RECO3 as well.
    Is it right, that the "old" entries will disappear, once the ASM instance is restarted ? I can´t do this at the moment.
    In generell - why went the ASM disks to offline ?
    SAN1 was switched off for about 5 minutes - and was then switched back on.
    Is this expected behaviour ?

    Hi Chris,
    as long as the information is still partially valid, the devices will be displayed in v$asm_disk. Main reason is: Some Oracle processes still have handles open on these devices, and as long as they exist the devices can still be seen.
    A Restart of ASM + database will definitely remove these entries (if they cannot be seen by the operating system).
    Regarding your other question: Oracle can only detect if a device is going offline (errors in the database and asm alter log). If a device goes online again, Oracle does not check this (this would be a major overhead to rescan lets say every minute to see if a disk is back online).
    Hence you have to reinclude the dropped/offlined disks yourself.
    Depending on the version (10g or 11g) and depending on the diskgroup compatibility level for rdbms a disk can simply be returned into the diskgroup (11g - alter diskgroup <dg> set disk <disk> online) and resilvered or have to be reincluded into the ASM instance (no resilvering, but a complete rebalance 10g: alter diskgroup <dg> add disk <disk> to failgroup <fg> [force])
    Sebastian

  • ORA-15042: ASM disk "2" is missing from group number "1"

    Hi,
    I'm working on an Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production With the Automatic Storage Management option.
    Into the ASM I had 3 diskgroups:
    - ARCHIVELOG (4 disks)
    - ONLINELOG (1 disks)
    - DATA (10 disks)
    When I try to startup the ASM instance I got:
    A-15042: ASM disk "2" is missing from group number "1"The diskgroup won't be mounted.
    I would like to remove that disk and later add a new one.
    I can I do that?
    I'm not able to mount the ARCHIVELOG diskgroup.
    I tried the command
    SQL> alter diskgroup archivelog drop disk ARCH3 force;
    alter diskgroup archivelog drop disk ARCH3 force
    ERROR at line 1:
    ORA-15032: not all alterations performed
    ORA-15001: diskgroup "ARCHIVELOG" does not exist or is not mountedThanks in advance,
    Samuel
    Edited by: Samuel Rabini on Jan 10, 2012 4:11 PM

    As that database is on AWS, I tried this:
    - drop diskgroup archivelog
    - detach of those 4 disks
    - create new 4 disks
    - attach new disks
    - assign those disks to ASM with oracleasm utilty
    - create diskgroup archivelog
    It worked.
    But because I was on AWS and more because it was the ARCHIVELOG diskgroup.
    What would I had to do if it was the DATA diskgroup?
    Thanks

  • How reInstall Gride Infrastructure and Use old ASM disk groups

    <pre>Hello to all
    I installed Grig Infrastructure 11gR2 on a standalone server (OS is Linux)
    and I configured ASM and my database created on ASM
    Conceive that my OS disk corrupted and OS doesn't start and the Gride Home is on that disk,
    and I have to install OS again
    My ASM disks are safe , Now how can I install Grig Infrastructure again somehow that it can use previous ASM disks
    and disk groups and I don't oblige to create my database again ?
    In the step 2 of installing Gride Infrastructure it has four options
    <pre>
    1.Install and configure Oracle Grid Infrastructure for a Cluster
    2.Configure Oracle Grid Infrastructure for a Standalone Server
    3.Upgrade Oracle Gride Infrastructure or Oracle Automatice Storage Management
    4.Install Oracle Gride Infrastructure Software Only
    </pre>
    If I select the option 2 it wants to create a disk group again
    I guess that I need to select option 4 and then do some configuration but I don't know what I must configure
    Do you know answer of my question , if yes please explain it's stages
    Thank you so much
    </pre>

    Hi,
    no you are not obliged to recreate your database again. However there is a small flaw in the installation procedure, which does not make it 100% easy...
    When you installed the Oracle Restart (Standalone GI), your ASM diskgroup will contain the SPFILE of the ASM instance. And this is exactly the small flaw you will be encountering. So you have 2 options for "recovery":
    1.) Do a software only install (4), and run roothas.pl. This however will not create any ASM entries. You would have to add it manually (using srvctl) and you can specify the ASM Spfile with the srvctl command. Problem here is however to have to know where your ASM spfile has been. If you have a backup of your OLR and a backup of the GPNP profile, this might be easier to find out.
    2.) Do a new installation (2) and configure a new diskgroup (with a "spare" disk or small lun and a new name), that Oracle restart creates ASM instance and the new ASMSpfile for you.
    Then you can simply mount the diskgroup containing your database additionally. You then shoudl however move your new ASMSpfile to the new diskgroup (or simply exchange it with the existing one). In this case it is easier to find out where it was - however you will need a spare (though small) LUN for the new spfile (temporarily, until you exchange it).
    In either case after you have your ASM instance back (and access to your old diskgroup), you have to reregister your database and services - if you do not have an OLR backup.
    Again => It is doable and you can simply mount the ASM diskgroup containing your database. However I suggest you try this one time to know what really needs to be done in this case.
    Regards
    Sebastian

Maybe you are looking for

  • Adobe Flash Player and IE8

    I am using Internet Explorer 8. I have tried to install Flash Player 10.2.152.32. It goes through the Installation Manager & appears to have done all the installation with no problems. I get no error messages & get an installation done confirmation,

  • IMac 20" vs 17"

    HI im currently thinking of buying an imac but i need to know if there is any major graphics performance hit in choosing the 20" model over the 17"? both have the same 2.0ghz G5 and same graphics card? but its just the fact that you have to power the

  • Actual price calculation configuration

    Dear experts,          What are the configuration settings required for actual price calculation in product costing? Regards Jay

  • I  wants to change the quantity in billing after PGI? plz cive me idea

    i wants to change the quantity in billing doc after pgi plz give me the ans

  • How to integrate iPayment module with external Payment System

    Hi, I have a requirement where I need to integrate Oracle iPayments in Release 11i with external payment system which is different than standard payment system given by Oracle,so can anybody please guide me on the steps which need to be performed to