RAID 10 problems, two disks marked as spare

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,

I'm having a bit of trouble with my NAS and I was hoping some of you lads would be able to help my out.

First of all my setup:

The NAS itself is an Iomega StorCenter ix4-200d. It has 4x 1TiB drives, configured to use RAID10. I have already replaced one drive once. The NAS itself doesn't come with shell access but it's fairly easy to 'root' it anyway (which I did).

Several days ago the NAS send me an email that a certain drive was degraded:

The Iomega StorCenter device is degraded and data protection is at risk. A drive may have either failed or been removed from your Iomega StorCenter device. Visit the Dashboard on the management interface for details. To prevent possible data loss, this issue should be repaired as soon as possible.

I decided to first to try and reboot the device, maybe it was a simple error. After reboot I received the following:

Data protection is being reconstructed on your Iomega StorCenter device

So I was happy until several hours later I received the following messages (all at the same time)

The Iomega StorCenter device has completed data protection reconstruction.

The Iomega StorCenter device has failed and some data loss may have occurred. Multiple drives may have either failed or been removed from your storage system. Visit the Dashboard on the management interface for details.

Drive number 4 encountered a recoverable error.

No data was accessible anymore. After that I opened a shell to the device and try to trouble shout it. But I didn’t manage to get it working. The only solution I currently see it to try and rebuild the RAID array but as I have hardly any experience with the mdadm tool I decided to ask the opinions of the people here.

Here is some information regarding the setup:

 root@BauwensNAS:/# mdadm -D /dev/md1
/dev/md1:
        Version : 01.00
  Creation Time : Mon Jan 24 20:57:43 2011
     Raid Level : raid10
  Used Dev Size : 974722176 (929.57 GiB 998.12 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Wed Feb 26 02:44:57 2014
          State : active, degraded, Not Started
 Active Devices : 2
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 2

         Layout : near=2, far=1
     Chunk Size : 64K

           Name : bwns:1
           UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
         Events : 133470

    Number   Major   Minor   RaidDevice State
       0       8        2        0      active sync   /dev/sda2
       4       8       18        1      active sync   /dev/sdb2
       2       0        0        2      removed
       3       0        0        3      removed

       2       8       34        -      spare   /dev/sdc2
       3       8       50        -      spare   /dev/sdd2

As you can see, the two last drives are marked as spare. My multiple attempts to get the drive running with all the disks have only been failures (but I assume that's also due to me not having experience with the tools).

Also, the disks themself appear to be fine (also because the md0 device that hosts /boot works properly)

Some more info:

root@BauwensNAS:/# mdadm --examine /dev/sd[abcd]2
/dev/sda2:
          Magic : a92b4efc
        Version : 1.0
    Feature Map : 0x0
     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
           Name : bwns:1
  Creation Time : Mon Jan 24 20:57:43 2011
     Raid Level : raid10
   Raid Devices : 4

 Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
   Super Offset : 1949444640 sectors
          State : clean
    Device UUID : b05f4c40:819ddbef:76872d9f:abacf3c9

    Update Time : Wed Feb 26 02:44:57 2014
       Checksum : a94e1ae6 - correct
         Events : 133470

         Layout : near=2, far=1
     Chunk Size : 64K

    Array Slot : 0 (0, failed, empty, empty, 1)
   Array State : Uu__ 1 failed
/dev/sdb2:
          Magic : a92b4efc
        Version : 1.0
    Feature Map : 0x0
     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
           Name : bwns:1
  Creation Time : Mon Jan 24 20:57:43 2011
     Raid Level : raid10
   Raid Devices : 4

 Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
   Super Offset : 1949444640 sectors
          State : clean
    Device UUID : 6c34331a:0fda7f73:a1f76d41:a826ac1f

    Update Time : Wed Feb 26 02:44:57 2014
       Checksum : fed2165a - correct
         Events : 133470

         Layout : near=2, far=1
     Chunk Size : 64K

    Array Slot : 4 (0, failed, empty, empty, 1)
   Array State : uU__ 1 failed
/dev/sdc2:
          Magic : a92b4efc
        Version : 1.0
    Feature Map : 0x0
     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
           Name : bwns:1
  Creation Time : Mon Jan 24 20:57:43 2011
     Raid Level : raid10
   Raid Devices : 4

 Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
   Super Offset : 1949444640 sectors
          State : clean
    Device UUID : 773dbfec:07467e62:de7be59b:5c680df5

    Update Time : Wed Feb 26 02:44:57 2014
       Checksum : f035517e - correct
         Events : 133470

         Layout : near=2, far=1
     Chunk Size : 64K

    Array Slot : 2 (0, failed, empty, empty, 1)
   Array State : uu__ 1 failed
/dev/sdd2:
          Magic : a92b4efc
        Version : 1.0
    Feature Map : 0x0
     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
           Name : bwns:1
  Creation Time : Mon Jan 24 20:57:43 2011
     Raid Level : raid10
   Raid Devices : 4

 Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
   Super Offset : 1949444640 sectors
          State : clean
    Device UUID : dbd27546:6b623b53:8f887960:b7cbf424

    Update Time : Wed Feb 26 02:44:57 2014
       Checksum : 2f247322 - correct
         Events : 133470

         Layout : near=2, far=1
     Chunk Size : 64K

    Array Slot : 3 (0, failed, empty, empty, 1)
   Array State : uu__ 1 failed


Taking a look at the event count they seem to be synchronized, so I'm not really sure what's going on here.

root@BauwensNAS:/# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md1 : inactive sda2[0] sdd2[3](S) sdc2[2](S) sdb2[4]
      3898888704 blocks super 1.0

md0 : active raid1 sda1[0] sdd1[3] sdc1[2] sdb1[1]
      2040128 blocks [4/4] [UUUU]

unused devices: <none>

Anyone have an idea how I could resolve this problem (hoping that I don't have any data loss...)? Any help is greatly appreciated. I sure regret rebooting the device without taking some extra backups.

TIA!
Jim
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux