WSA RAID Status: Degraded

Hi,
   I Have the next status on WSA
RAID Status:
Degraded
Any idea how to get Optimal ?

Hello Oscar,
Please try rebooting the appliance and leave it for a day for the raid to rebuild. If the raid still does not re-build there might be  a possible disk issue, kindly open a TAC case.
Regards,
Puja

Similar Messages

  • RAID status: degraded or failed? is there help?

    We are using mac os sever 10.4.3. When we recently restarted the servers we noticed that 2 drives on each of the servers has yellow warning lights showing on the moitor software but not on the actual drives themselves. We have 2 servers one a G4 and the other a G5. The G4 has 2 drives and both are registering yellow, the G5 has 4 drivers with 2 registering yellow. The monitor says the drives are "degraded" but everything seems to be transferring and backing up fine. Should we do a backup and use Disk Utility to repair the degraded/damaged drives?
    Mac OS Server   Mac OS X (10.4.3)   Processor 2 GHz G5 Memory 2 GB

    You must have RAID 1 mirrored drives setup then and are running from the working disk of the mirrored pair.
    Ofcourse you should fix these types of problems/failing disks.
    Having recent backups is always a good thing. A RAID doesn't save you from human errors but it should prevent "downed" servers.
    Your supposed to be able to repair a degraded mirror from within Disk Utility but I have only done it from the CLI (Terminal) diskutil.
    If you do a : sudo diskutil checkraid
    What do you get?
    If the servers are updated from 10.3.x you can (if not done already) also update the RAID version from the older ver. 1 to ver. 2 available from 10.4.0-> using the command:
    sudo diskutil convertraid

  • Mars Raid-10 status degraded

    Hi
    My CS-Mars 110 had crashed for some unknown reason. After reload the device works fine but there is an error with the raid-10 which shows a status of "Degraded", please see attached file.
    Is there a way to solve this problem?
    Thanks

    Hi
    [pnadmin]$ model
    mars110
    local
    extension
    [pnadmin]$ pndbusage
    Current partition started on Tue Dec 18 19:10:01 CET 2007 and uses 6.93% of its available capacity.
    Switching to next partition is estimated for Tue Apr 19 14:00:51 CET 2022.
    9 empty partitions are available for storage
    [pnadmin]$ raidstatus
    Adapter Information:
    Product Name : Intel(R) RAID Controller SROMBSAS18E
    Firmware Version : 1.03.00-0211
    BIOS Version : MT30
    Adapter RaidType Status Stripe Size Cache
    a0 Raid-10 Degraded 64kB 1427649MB Enabled
    PD Status Size & Block Model Serial#
    p0 Unconfigured(good) 476940MB [0x3a386030 Sectors] ATA ST3500630NS K 9QG0SASJThis Drive IsForeign.
    p1 Online 476940MB [0x3a386030 Sectors] ATA ST3500630NS K 9QG0SA5M
    p2 Online 476940MB [0x3a386030 Sectors] ATA ST3500630NS K 9QG69XW1
    p3 Online 476940MB [0x3a386030 Sectors] ATA ST3500630NS K 9QG0QTT9
    p4 Unconfigured(good) 476940MB [0x3a386030 Sectors] ATA ST3500630NS K 9QG0H8R3This Drive IsForeign.
    p5 Online 476940MB [0x3a386030 Sectors] ATA ST3500630NS K 9QG677HW

  • Has anyone upgraded the Ironport ESA to 8.5.6-074 and had the issues of Raid status showing unknown?

    Has anyone upgraded the Ironport ESA to 8.5.6-074 and had the issues of Raid status showing unknown? After we upgraded our appliances we are having issues with our ESA appliances showing the RAID status as unknown. When we reported the issue to CISCO we were updated there were no issues reported at all. Could anyone please confirm if you have experienced the same issue. 

    You should see OPTIMAL - meaning the drives in the C170 are in good health/status:
    myc680.local> version
    Current Version
    ===============
    UDI: C680 V FCH1611V0B2
    Name: C680
    Product: Cisco IronPort C680 Messaging Gateway(tm) Appliance
    Model: C680
    Version: 8.5.6-074
    Build Date: 2014-07-21
    Install Date: 2014-07-29 11:16:34
    Serial #: xxx-yyy1611Vzzz
    BIOS: C240M3.1.4.5.2.STBU
    RAID: 3.220.75-2196, 5.38.00_4.12.05.00_0x05180000
    RAID Status: Optimal
    RAID Type: 10
    BMC: 1.05
    There are times post-reboot, that you'll see and get notification of RAID sub-optimal --- meaning that the appliance is running through on a health-check of the appliance's RAID.  You should be getting a notification once RAID status has returned to OPTIMAL, or as per the older OS revisions, READY:
    myc170.local> version
    Current Version
    ===============
    UDI: C170 V01 FCH1428V06A
    Name: C170
    Description: Cisco IronPort C170
    Product: Cisco IronPort C170 Messaging Gateway(tm) Appliance
    Model: C170
    Version: 7.6.3-019
    Build Date: 2013-06-09
    Install Date: 2014-09-12 13:52:24
    Serial #: xxxxxxD87B39-yyyyyy8V06A
    BIOS: 9B1C115A
    RAID: 02
    RAID Status: READY
    RAID Type: 1
    BMC: 2.01

  • My Duo 12TB says RAID Status: Cannot Access Data

    When I started up my computer today, My Duo 12TB (Raid 1 Mirrored) came up with a RAID error.  When I ran WD Drive Utilities program, it gives status:  RAID Status Cannot Access Data   Drive 1 Status  OnlineDrive 2 Status  Online I have tried unplugging both the power and the USB connections.  Restarted computer. What needs to be done to fix this without lossing any Data on the HDs.Thanks,

     
    Hi, 
    Welcome to the WD Community.
    As a recommendation, please contact WD Support for direct assistance on this case.
    WD contact info: 
    http://support.wdc.com/country/index.asp?lang=en%22
     

  • Disk utility says "RAID STATUS: Disk missing"?

    When I boot up OS 10.3.9 there is error message "detected volume OS X cannot read - initialize, ignore or eject" - then it dissappears and the system drive behaves normally. But in disk utitlity it says "RAID STATUS: Disk missing".
    The disk is not partitioned and it will not let me repair permissions. Any ideas?

    Disk Utility monitors the status of the disks in a RAID set. If you see a message indicating that a disk is missing or has failed, try these troubleshooting steps:
    If you are using a striped RAID set, delete the damaged RAID set. Your data may be lost. Be sure to back up your RAID sets and other data regularly.
    If you are using a mirrored RAID set, there may have been an error writing data to the disk. Click Rebuild in the RAID pane of Disk Utility.
    If a problem persists, replace the damaged disk and click Rebuild to rebuild the RAID set.
    Use the First Aid pane to repair the RAID disk, then check the RAID set to see if it still reports an error. If the problem is still present, quit and reopen Disk Utility, select the RAID disk, and click RAID. Check the RAID set to see if it still reports an error. You may need to restart your computer.
    iBook G4   Mac OS X (10.3.9)  

  • [Solved] RAID 5 degraded after 3.13 upgrade

    Hi there,
    after upgrading my home server to the latest kernel 3.13 I've found out, that my RAID 5 got degraded. One of the drives has been kicked out, but I don't know why. The drive seems okay, I've also done a SMART short test, completed without any errors. The only suspicious looking error message, when upgrading to Linux 3.13 was:
    ERROR: Module 'hci_vhci' has devname (vhci) but lacks major and minor information. Ignoring.
    This is mdstat output:
    [tolga@Ragnarok ~]$ cat /proc/mdstat
    Personalities : [raid6] [raid5] [raid4]
    md127 : active raid5 sda1[0] sdc1[3] sdb1[1]
    5860145664 blocks super 1.2 level 5, 512k chunk, algorithm 2 [4/3] [UUU_]
    unused devices: <none>
    smartctl:
    [tolga@Ragnarok ~]$ sudo smartctl -a /dev/sdd
    smartctl 6.2 2013-07-26 r3841 [x86_64-linux-3.13.4-1-ARCH] (local build)
    Copyright (C) 2002-13, Bruce Allen, Christian Franke, www.smartmontools.org
    === START OF INFORMATION SECTION ===
    Model Family: Western Digital Red (AF)
    Device Model: WDC WD20EFRX-68AX9N0
    Serial Number: [removed]
    LU WWN Device Id: 5 0014ee 2b2cd537a
    Firmware Version: 80.00A80
    User Capacity: 2,000,398,934,016 bytes [2.00 TB]
    Sector Sizes: 512 bytes logical, 4096 bytes physical
    Device is: In smartctl database [for details use: -P show]
    ATA Version is: ACS-2 (minor revision not indicated)
    SATA Version is: SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
    Local Time is: Fri Feb 21 22:26:30 2014 CET
    SMART support is: Available - device has SMART capability.
    SMART support is: Enabled
    === START OF READ SMART DATA SECTION ===
    SMART overall-health self-assessment test result: PASSED
    General SMART Values:
    Offline data collection status: (0x00) Offline data collection activity
    was never started.
    Auto Offline Data Collection: Disabled.
    Self-test execution status: ( 0) The previous self-test routine completed
    without error or no self-test has ever
    been run.
    Total time to complete Offline
    data collection: (26580) seconds.
    Offline data collection
    capabilities: (0x7b) SMART execute Offline immediate.
    Auto Offline data collection on/off support.
    Suspend Offline collection upon new
    command.
    Offline surface scan supported.
    Self-test supported.
    Conveyance Self-test supported.
    Selective Self-test supported.
    SMART capabilities: (0x0003) Saves SMART data before entering
    power-saving mode.
    Supports SMART auto save timer.
    Error logging capability: (0x01) Error logging supported.
    General Purpose Logging supported.
    Short self-test routine
    recommended polling time: ( 2) minutes.
    Extended self-test routine
    recommended polling time: ( 268) minutes.
    Conveyance self-test routine
    recommended polling time: ( 5) minutes.
    SCT capabilities: (0x70bd) SCT Status supported.
    SCT Error Recovery Control supported.
    SCT Feature Control supported.
    SCT Data Table supported.
    SMART Attributes Data Structure revision number: 16
    Vendor Specific SMART Attributes with Thresholds:
    ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
    1 Raw_Read_Error_Rate 0x002f 200 200 051 Pre-fail Always - 0
    3 Spin_Up_Time 0x0027 164 163 021 Pre-fail Always - 6766
    4 Start_Stop_Count 0x0032 100 100 000 Old_age Always - 273
    5 Reallocated_Sector_Ct 0x0033 200 200 140 Pre-fail Always - 0
    7 Seek_Error_Rate 0x002e 200 200 000 Old_age Always - 0
    9 Power_On_Hours 0x0032 098 098 000 Old_age Always - 1954
    10 Spin_Retry_Count 0x0032 100 100 000 Old_age Always - 0
    11 Calibration_Retry_Count 0x0032 100 100 000 Old_age Always - 0
    12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 273
    192 Power-Off_Retract_Count 0x0032 200 200 000 Old_age Always - 6
    193 Load_Cycle_Count 0x0032 200 200 000 Old_age Always - 266
    194 Temperature_Celsius 0x0022 115 104 000 Old_age Always - 35
    196 Reallocated_Event_Count 0x0032 200 200 000 Old_age Always - 0
    197 Current_Pending_Sector 0x0032 200 200 000 Old_age Always - 0
    198 Offline_Uncorrectable 0x0030 100 253 000 Old_age Offline - 0
    199 UDMA_CRC_Error_Count 0x0032 200 200 000 Old_age Always - 0
    200 Multi_Zone_Error_Rate 0x0008 100 253 000 Old_age Offline - 0
    SMART Error Log Version: 1
    ATA Error Count: 306 (device log contains only the most recent five errors)
    CR = Command Register [HEX]
    FR = Features Register [HEX]
    SC = Sector Count Register [HEX]
    SN = Sector Number Register [HEX]
    CL = Cylinder Low Register [HEX]
    CH = Cylinder High Register [HEX]
    DH = Device/Head Register [HEX]
    DC = Device Command Register [HEX]
    ER = Error register [HEX]
    ST = Status register [HEX]
    Powered_Up_Time is measured from power on, and printed as
    DDd+hh:mm:SS.sss where DD=days, hh=hours, mm=minutes,
    SS=sec, and sss=millisec. It "wraps" after 49.710 days.
    Error 306 occurred at disk power-on lifetime: 1706 hours (71 days + 2 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    04 61 02 00 00 00 a0 Device Fault; Error: ABRT
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    ef 10 02 00 00 00 a0 08 22:17:38.065 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.065 IDENTIFY DEVICE
    ef 03 46 00 00 00 a0 08 22:17:38.064 SET FEATURES [Set transfer mode]
    ef 10 02 00 00 00 a0 08 22:17:38.064 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.064 IDENTIFY DEVICE
    Error 305 occurred at disk power-on lifetime: 1706 hours (71 days + 2 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    04 61 46 00 00 00 a0 Device Fault; Error: ABRT
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    ef 03 46 00 00 00 a0 08 22:17:38.064 SET FEATURES [Set transfer mode]
    ef 10 02 00 00 00 a0 08 22:17:38.064 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.064 IDENTIFY DEVICE
    ef 10 02 00 00 00 a0 08 22:17:38.063 SET FEATURES [Enable SATA feature]
    Error 304 occurred at disk power-on lifetime: 1706 hours (71 days + 2 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    04 61 02 00 00 00 a0 Device Fault; Error: ABRT
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    ef 10 02 00 00 00 a0 08 22:17:38.064 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.064 IDENTIFY DEVICE
    ef 10 02 00 00 00 a0 08 22:17:38.063 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.063 IDENTIFY DEVICE
    Error 303 occurred at disk power-on lifetime: 1706 hours (71 days + 2 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    04 61 02 00 00 00 a0 Device Fault; Error: ABRT
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    ef 10 02 00 00 00 a0 08 22:17:38.063 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.063 IDENTIFY DEVICE
    ef 03 46 00 00 00 a0 08 22:17:38.063 SET FEATURES [Set transfer mode]
    ef 10 02 00 00 00 a0 08 22:17:38.063 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.062 IDENTIFY DEVICE
    Error 302 occurred at disk power-on lifetime: 1706 hours (71 days + 2 hours)
    When the command that caused the error occurred, the device was active or idle.
    After command completion occurred, registers were:
    ER ST SC SN CL CH DH
    04 61 46 00 00 00 a0 Device Fault; Error: ABRT
    Commands leading to the command that caused the error were:
    CR FR SC SN CL CH DH DC Powered_Up_Time Command/Feature_Name
    ef 03 46 00 00 00 a0 08 22:17:38.063 SET FEATURES [Set transfer mode]
    ef 10 02 00 00 00 a0 08 22:17:38.063 SET FEATURES [Enable SATA feature]
    ec 00 00 00 00 00 a0 08 22:17:38.062 IDENTIFY DEVICE
    ef 10 02 00 00 00 a0 08 22:17:38.062 SET FEATURES [Enable SATA feature]
    SMART Self-test log structure revision number 1
    Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
    # 1 Short offline Completed without error 00% 1954 -
    # 2 Short offline Completed without error 00% 0 -
    # 3 Conveyance offline Completed without error 00% 0 -
    SMART Selective self-test log data structure revision number 1
    SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
    1 0 0 Not_testing
    2 0 0 Not_testing
    3 0 0 Not_testing
    4 0 0 Not_testing
    5 0 0 Not_testing
    Selective self-test flags (0x0):
    After scanning selected spans, do NOT read-scan remainder of disk.
    If Selective self-test is pending on power-up, resume after 0 minute delay.
    This is my mdadm configuration:
    [tolga@Ragnarok ~]$ cat /etc/mdadm.conf
    ARRAY /dev/md/Asura metadata=1.2 UUID=34bab60a:4d640b50:6228c429:0679bb34 name=Ragnarok:Asura
    I've checked all partition tables, everything seems ok. "Error 30[x] occurred at disk power-on lifetime: 1706 hours (71 days + 2 hours)" seems to be a one-time event, which happened on 1706 hours (I don't know why; no power loss or something similar). Other than those smartctl errors, everything seems fine. I've also inspected the drive; no suspicious noises or anything else, works like the other 3 drives. Am I safe to simply re-add the drive using "sudo mdadm --manage --re-add /dev/md127 /dev/sdd1" and let it re-sync or should I flag it as failed and then re-add it to the RAID?
    I am using 4x 2TB Western Digital Red drives in a RAID 5, which are about 1 year old and they ran perfectly fine until now. The server is currently shut down until this problem is fixed. I currently got a partial backup of my data (most important ones) and will make a full backup, before attempting a repair. At the moment, I'm still able to access all my data, so nothing's wrong there.
    So, what do you guys think, what should I do?
    Last edited by tolga9009 (2014-09-13 12:48:13)

    Thank you brian for the fast reply. I've backed up all my important data and tried the command. It's not working ...
    [tolga@Ragnarok ~]$ sudo mdadm --manage --re-add /dev/md127 /dev/sdd1
    mdadm: --re-add for /dev/sdd1 to /dev/md127 is not possible
    [tolga@Ragnarok ~]$ lsblk
    NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
    sda 8:0 0 1.8T 0 disk
    ââsda1 8:1 0 1.8T 0 part
    ââmd127 9:127 0 5.5T 0 raid5 /media/Asura
    sdb 8:16 0 1.8T 0 disk
    ââsdb1 8:17 0 1.8T 0 part
    ââmd127 9:127 0 5.5T 0 raid5 /media/Asura
    sdc 8:32 0 1.8T 0 disk
    ââsdc1 8:33 0 1.8T 0 part
    ââmd127 9:127 0 5.5T 0 raid5 /media/Asura
    sdd 8:48 0 1.8T 0 disk
    ââsdd1 8:49 0 1.8T 0 part
    sde 8:64 0 59.6G 0 disk
    ââsde1 8:65 0 512M 0 part /boot/efi
    ââsde2 8:66 0 4G 0 part [SWAP]
    ââsde3 8:67 0 54.6G 0 part /
    ââsde4 8:68 0 512M 0 part /boot
    Out of curiosity, I've compared "mdadm -E" of the corrupted and a healthy drive. Here's the output:
    [tolga@Ragnarok ~]$ diff -u sdc sdd
    --- sdc 2014-02-21 23:28:51.051674496 +0100
    +++ sdd 2014-02-21 23:28:55.911816816 +0100
    @@ -1,4 +1,4 @@
    -/dev/sdc1:
    +/dev/sdd1:
    Magic : a92b4efc
    Version : 1.2
    Feature Map : 0x0
    @@ -14,15 +14,15 @@
    Data Offset : 262144 sectors
    Super Offset : 8 sectors
    Unused Space : before=262064 sectors, after=1167 sectors
    - State : clean
    - Device UUID : 4ce2ba99:645b1cc6:60c23336:c4428e2f
    + State : active
    + Device UUID : 4aeef598:64ff6631:826f445e:dbf77ab5
    - Update Time : Fri Feb 21 23:18:20 2014
    - Checksum : a6c42392 - correct
    - Events : 16736
    + Update Time : Sun Jan 12 06:40:56 2014
    + Checksum : bf106b2a - correct
    + Events : 7295
    Layout : left-symmetric
    Chunk Size : 512K
    - Device Role : Active device 2
    - Array State : AAA. ('A' == active, '.' == missing, 'R' == replacing)
    + Device Role : Active device 3
    + Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing)
    So, I guess my only way to fix this is remove the faulty drive from the RAID, zero out the superblock and then re-add it as a new drive. Or is there any other way to fix this?
    //Edit: I've used "mdadm --detail /dev/md127" and found out, that the faulty drive wasn't even listed anymore. So instead of using "re-add", I simply added it as a new drive and it's resyncing now. In about 220mins, I'll know more ! Is there a way to check for corruption after syncing the drives?
    //Edit: ^ this worked. My drive probably didn't got kicked after the 3.13 upgrade, but I've simply noticed it then. The drive seems to be kicked after ~1700 hours for some unknown reason - I've now disconnected and reconnected all drives to opt out any wiring issues. Since the drive was out of sync, simply re-adding it didn't work. I had to manually add it to the array again and this caused a resync, which took around 3,5 hours. I think that's okay for a 4x 2TB RAID 5 array. Everything is working fine again, no data corruption, nothing. I'll mark it as solved.
    Last edited by tolga9009 (2014-09-13 12:59:22)

  • RAID RS1 Degraded - How do I repair it?

    Please note screen shots below.
    RS1 (Bay 3, Bay 4) is showing Viable (Degraded), Bay 3 Disk is showing Degraded but has a green status light next to it in Drives List.
    I can still access data (I assume from Bay 4 Disk if Bay 3 Disk is not working).
    How can I repair the Raid Set using the RAID Utility if the drive is ok; or is it a case of replacing the Bay 3 Drive and allowing the Utility to rebuild the RAID Set?
    Thanks.

    Since this is a Mirrored RAID the data are still available on the remaining drive.
    To rebuild the set you need to introduce a blank good drive as an available spare. You can use a new drive, or you can erase the drive that dropped out and re-introduce it as the "new" drive, ONCE. Beyond that once, you are fooling yourself and reducing the robustness of the set -- a drive that gets kicked out more than once needs to be replaced in the set and re-purposed to something less demanding.

  • [SOLVED] RAID status

    Last night I created a raid5 array using mdadm and 5 x 3TB hdd's.  I let the sync happen overnight and on returning home this evening
    watch -d cat /proc/mdstat
    returned:
    Personalities : [raid6] [raid5] [raid4]
    md0 : active raid5 sda1[0] sdc1[2] sde1[5] sdb1[1] sdd1[3]
    11720536064 blocks super 1.2 level 5, 512k chunk, algorithm 2 [5/5] [UUUUU]
    bitmap: 2/22 pages [8KB], 65536KB chunk
    unused devices: <none>
    which to me pretty much looks like the array sync is completed.
    I then updated the config file, assembled the array and formatted it using:
    mdadm --detail --scan >> /etc/mdadm.conf
    mdadm --assemble --scan
    mkfs.ext4 -v -L offsitestorage -b 4096 -E stride=128,stripe-width=512 /dev/md0
    Running
    mdadm --detail /dev/md0
    returns:
    /dev/md0:
    Version : 1.2
    Creation Time : Thu Apr 17 01:13:52 2014
    Raid Level : raid5
    Array Size : 11720536064 (11177.57 GiB 12001.83 GB)
    Used Dev Size : 2930134016 (2794.39 GiB 3000.46 GB)
    Raid Devices : 5
    Total Devices : 5
    Persistence : Superblock is persistent
    Intent Bitmap : Internal
    Update Time : Thu Apr 17 18:55:01 2014
    State : active
    Active Devices : 5
    Working Devices : 5
    Failed Devices : 0
    Spare Devices : 0
    Layout : left-symmetric
    Chunk Size : 512K
    Name : audioliboffsite:0 (local to host audioliboffsite)
    UUID : aba348c6:8dc7b4a7:4e282ab5:40431aff
    Events : 11306
    Number Major Minor RaidDevice State
    0 8 1 0 active sync /dev/sda1
    1 8 17 1 active sync /dev/sdb1
    2 8 33 2 active sync /dev/sdc1
    3 8 49 3 active sync /dev/sdd1
    5 8 65 4 active sync /dev/sde1
    So, I'm now left wondering why the state of the array isn't "clean"?  Is it normal for arrays to show a state of "active" instead of clean under Arch?
    Last edited by audiomuze (2014-07-02 20:10:33)

    Contrasting two RAID5 arrays - the first created in Arch, the second in Ubuntu Server.  Both were created using the same command set (the only differences being number of drives and stride optimised for number of drives).  Why the difference in status out of the starting blocks?  I've not been able to find anything in the documentation or reading mdadm's manpage to explain this.  If there's additional info required in order to assist please let me know and I'll provide same.
    Thx in advance for considering and assistance.
    /dev/md0:
    Version : 1.2
    Creation Time : Thu Apr 17 01:13:52 2014
    Raid Level : raid5
    Array Size : 11720536064 (11177.57 GiB 12001.83 GB)
    Used Dev Size : 2930134016 (2794.39 GiB 3000.46 GB)
    Raid Devices : 5
    Total Devices : 5
    Persistence : Superblock is persistent
    Intent Bitmap : Internal
    Update Time : Mon May 5 05:35:28 2014
    State : active
    Active Devices : 5
    Working Devices : 5
    Failed Devices : 0
    Spare Devices : 0
    Layout : left-symmetric
    Chunk Size : 512K
    Name : audioliboffsite:0
    UUID : aba348c6:8dc7b4a7:4e282ab5:40431aff
    Events : 11307
    Number Major Minor RaidDevice State
    0 8 1 0 active sync /dev/sda1
    1 8 17 1 active sync /dev/sdb1
    2 8 33 2 active sync /dev/sdc1
    3 8 49 3 active sync /dev/sdd1
    5 8 65 4 active sync /dev/sde1
    /dev/md0:
    Version : 1.2
    Creation Time : Sun Feb 2 21:40:15 2014
    Raid Level : raid5
    Array Size : 8790400512 (8383.18 GiB 9001.37 GB)
    Used Dev Size : 2930133504 (2794.39 GiB 3000.46 GB)
    Raid Devices : 4
    Total Devices : 4
    Persistence : Superblock is persistent
    Update Time : Mon May 5 06:45:45 2014
    State : clean
    Active Devices : 4
    Working Devices : 4
    Failed Devices : 0
    Spare Devices : 0
    Layout : left-symmetric
    Chunk Size : 512K
    Name : fileserver:0 (local to host fileserver)
    UUID : 8389cd99:a86f705a:15c33960:9f1d7cbe
    Events : 208
    Number Major Minor RaidDevice State
    0 8 1 0 active sync /dev/sda1
    1 8 17 1 active sync /dev/sdb1
    2 8 33 2 active sync /dev/sdc1
    4 8 49 3 active sync /dev/sdd1

  • RAID 5 degrades after restart

    Since I configured our Xserve (3x500 GB + Hardware Raid Card) it ran perfectly stable and without any problems.
    The only problem is each time I install new software and restart the RAID 5 is degraded and one of three 500 GB modules has to be rebuilt which takes a long time and decreases performance. Is this normal behaviour or might this be a hardware failure?
    I'm not really sure what to do next. Suggestions welcome!
    Xserve G5 2.3 DP   Mac OS X (10.4.7)   Hardware RAID Level 5

    rgb-digital:
    I am pretty sure that will do it.
    The problem with non-authorized drives in the ADM's is that they are not supported by Apple. That doesn't only mean that they won't support you from a technical standpoint if you have issues, but the other software and hardware as designed will not "support" the drives.
    I am not sure technically why it is working the way it is for you, but if you cannot eliminate one particular drive, sled, or slot in the XServe then I would say that this is definitely the problem.
    Luck-
    -DaddyPaycheck

  • TS130 RAID Status

    I am running Ubuntu 12.04 LTS Server on a headless TS130. I've got my operating system on a single SSD and my data on a RAID Array consisting of two 2TB drives in a RAID 1 configuration.
    Everything is running perfectly, however I am wondering how I would know if there is a problem with the RAID Array... for instance if one of the drives failed and needed to be swapped out, how do I know? Ubuntu recognizes the drives as a RAID array, but there is not a "driver" or a utiility for checking the status of the hardware RAID Array.
    Thanks in advance for your help!

    I believe upon boot, it will flash a screen looking similar to your BIOS that gives the status of your raid configuration, It only holds for a few seconds but the text is either Green, Yellow, or I assume red based off of the status.

  • RAID status in Conky

    I wrote a tiny script to monitor the status of RAID arrays. It is written in awk and is meant to be used with conky. I have this in my conky config file:
    RAID: ${execi 60 gawk -f path-to-the-script /proc/mdstat}
    #!/bin/gawk -f
    # This script checks the status of RAID arrays from /proc/mdstat
    /_/ { state = "Warning!"; exit }
    /blocks/ { state = $NF }
    END { print state }
    If everything is working correctly the ouput will be: "RAID: [UU]". It means that both drives are Up and running.
    If there is something wrong with the drive it will give an error message: "RAID: [WARNING!]".
    Maybe someone will find this usefull.

    Thanks!
    I have multiple RAIDs so I used the following command to check the status of a specific RAID:
    ${execi 60 cat /proc/mdstat | grep md126 -A 1 | gawk -f path-to-the-script --}
    (replace md126 with the array you want to check)

  • Monitoring Raid Status on SRE-910 module in 3945 router

    I'm at my wits end here.  We just recently purchased a 3945 ISR G2 router and have a SRE-910 module (with two hard drives) configured in a Raid 1.  We are running a stand-alone version of ESXi on the service module and I'm trying to figure out how to monitor the status of the Raid on the drives (along with other health issues).  SNMP has revealed nothing so far and even opening a support case for which MIB's to use has proved fruitless.  All the documents I find on monitoring the modules say to use LMS which is now Cisco Prime.  I've downloaded the trial copy, put in the SNMP settings and scanned the router.  I get device results and it shows that I have the SRE-910 module installed, but I get no other configuration / device informaiton from the module itself.
    I tried to create a new Monitoring template using the NAM health as the base template (which I'm assuming this is the correct template).  Unfortunately, when I acutally try to deploy the template against the discovered router, I get an 'Unexpected end of list' error which makes me assume I'm still doing something wrong.  I ANYONE out there monitoring the device health of their service modules in a 3945 router?  What am I missing????

    Oh, and by the way, I tried to monitor this through the ESXi host / vCenter, but even after pulling one of the hard drives from the module, neither software detected that there was an issue.  That is why I'm assuming that this will have to be monitored on the router side somehow.

  • Xserve - Mirrored RAID Failed? Degraded? Confused...

    Hi,
    Can anyone shed some light on this situation please? Disk utility tells me one thing and Server Monitor tells me another.
    I have an X-Serve with 3 drives and a Mirrored RAID Set running 10.5.8.
    Server Monitor is showing a yellow status for disk 1 and disk 2. Raid Status: Degraded (Mirror) Pre-failure Warning: No Warnings
    Disk Utility on the other hand shows that the Mirrored RAID sets are Degraded because disk2s4 = Failed and disk2s2 = Failed.
    Running disk utility commands in terminal also show disk2s4 and disk2s2 as Failed.
    I have not yet tried to Rebuild or eject / reconnect the disk. The only thing I've really done is some research and rebooted the server.
    It sounds like disk 2 has failed. Is there any way to confirm it's dead? Are there any serious risks of trying to rebuild?
    Thanks.
    - F4st

    I'm not sure why you think you're seeing conflicting information…
    Server Monitor is showing a yellow status for disk 1 and disk 2. Raid Status: Degraded (Mirror) Pre-failure Warning: No Warnings
    OK, So Server Monitor says the RAID is in a degraded state…
    Disk Utility […] shows that the Mirrored RAID sets are Degraded because disk2s4 = Failed and disk2s2 = Failed
    Running disk utility commands in terminal also show disk2s4 and disk2s2 as Failed
    I don't see this is anything different. All three methods are telling you disk2 has failed and your mirror is relying on disk1 for all activity.
    Is there any way to confirm it's dead?
    Umm.. you mean other than Server Monitor, Disk Utility.app and diskutil?
    Face it, it's dead. It's had it.
    Are there any serious risks of trying to rebuild?
    Sure. If the disk is dead, it's entirely possible that rebuilding the array on that disk will fail. You run the risk of a problem on disk1 and then all your data is gone.
    My advice: Replace disk2 as soon as possible. If not sooner.

  • RAID degraded - mirrored striped volumes - how to see if mirroring is working

    I've set up a mirrored raid drive. It keeps showing it as 'degraded', even though I've tested each disk separately and they come up OK.
    What does it mean in this report by 'rebuilding'?
    How do I sort out this 'degraded' status?
    How can I tell if mirroring is working correctly?
    The set up is:
    ===============================================================================
    Name:                 Saturn
    Unique ID:            CF319F64-987B-42FC-8C14-751766B37A49
    Type:                 Mirror
    Status:               Degraded
    Size:                 8.0 TB (8000084869120 Bytes)
    Rebuild:              manual
    Device Node:          disk15
    #  DevNode   UUID                                  Status     Size
    0  -none-    384CF206-BCD5-4804-BDB8-FF1C956EEF64  Online     8000084869120
    1  -none-    F886018B-EEE1-4875-853E-BCD4298683F8  0% (Rebuilding)8000084869120
    ===============================================================================
    Name:                 RAID0.1
    Unique ID:            384CF206-BCD5-4804-BDB8-FF1C956EEF64
    Type:                 Stripe
    Status:               Online
    Size:                 8.0 TB (8000219643904 Bytes)
    Rebuild:              manual
    Device Node:          -
    #  DevNode   UUID                                  Status     Size
    0  disk14s2  78CEBE17-8AFA-4849-A5B2-B73D9906FFE2  Online     2000054910976
    1  disk10s2  5D06DB08-6A3E-4041-A3F8-7E195D5B80DD  Online     2000054910976
    2  disk9s2   E430EC4F-AE2C-4B78-B2CC-04ED88315D3A  Online     2000054910976
    3  disk8s2   46EED2A6-BA53-4D54-8B69-FF5D650B97A0  Online     2000054910976
    ===============================================================================
    Name:                 RAID0.0
    Unique ID:            F886018B-EEE1-4875-853E-BCD4298683F8
    Type:                 Stripe
    Status:               Online
    Size:                 8.0 TB (8000084901888 Bytes)
    Rebuild:              manual
    Device Node:          -
    #  DevNode   UUID                                  Status     Size
    0  disk16s2  0B5223E9-2750-493B-A08E-01DD30E65065  Online     2000021225472
    1  disk5s2   F473AEC5-34A0-444E-AE62-DE755ECCE8A5  Online     2000021225472
    2  disk13s2  BDCBFE64-5771-4F3A-AECA-959B20844CD6  Online     2000021225472
    3  disk11s2  63AF3296-C427-4805-9FB5-B496205F49E8  Online     2000021225472
    ===============================================================================

    Hi,
    As I known, RAID1, or mirroring, is the technique of writing the same data to more than one disk drive. Mirrors are usually used to guard against data loss due to drive failure. Each drive in a mirror contains
    an identical copy of the data. No need to compare.
    If the drive is bootable, You should have got a warning when you booted the machine saying that either drive 0 or drive 1 was corrupt/non-bootable/missing and then be given a choice to boot from the second drive.
    Kate Li
    TechNet Community Support

Maybe you are looking for

  • Windows Vista not locating iPod Touch

    Hello all! I recently got an iPod Touch and have run into a problem. I tried syncing it on my PC (Windows Vista Home Premium 32Bit) and was only able to charge it but not sync it. Later today, I tried syncing the iPod Touch on the other PC which has

  • Problem With typecasting at JSONParsing?

    hi.......Every one.... 1.....i have another problem with JSON Parsing 2...i read  .json file , store in Dictionary ,Array  & then  pirnt on Console. 3..... Now i try to print this text on Label  by using..... (UILabel*)[self.view viewWithTag:1].text

  • Ant question about copying dirs

    HI all, I want do copy a folder into another folder (incl. files!) i have for example : /app/foldera and /app/destination/ and i want to copy folder a into destination so that it looks like this in the end: /app/destination/foldera I'm using ANT-151

  • Cannot edit Lightroom 2 photo in Elements 9

    I am using OS 10.6.5. I have  Lightroom 2.7 installed and I just purchased Photoshop Elements 9. I am  trying to edit my Lightroom photos in Elements by going to the Menu,  selecting Photos and then the selecting Edit In, but I am not able to select

  • IPod connected & now constant Do not disconnect on Display

    I recently connected my iPod with Color Display to my Mac, iTunes did not immediately boot up nor did the iPod icon appear on the desktop. I manually started iTunes and ITunes began its synching routine and the "Do not Disconnect" message appeared on