Re: RAID6 grow failed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]


On Sun, 1 Apr 2012 20:02:57 -0400 Bryan Bush <bbushvt@xxxxxxxxx> wrote:

> root@diamond:/# cat /proc/mdstat
> Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
> [raid4] [raid10]
> md1 : inactive sdk1[13](S) sdj1[11](S) sdh1[9](S) sdo1[3](S)
> sdd1[6](S) sde1[7](S) sdn1[2](S) sdl1[0](S) sdb1[4](S) sdc1[8](S)
>       19535134315 blocks super 1.2
> 
> md0 : active raid5 sdg1[3] sda1[0] sdf1[1] sdq1[2]
>       2929686528 blocks level 5, 256k chunk, algorithm 2 [4/4] [UUUU]
> 
> unused devices: <none>
> root@diamond:/# mdadm -S /dev/md1
> mdadm: stopped /dev/md1
> root@diamond:/# cat /proc/mdstat
> Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
> [raid4] [raid10]
> md0 : active raid5 sdg1[3] sda1[0] sdf1[1] sdq1[2]
>       2929686528 blocks level 5, 256k chunk, algorithm 2 [4/4] [UUUU]
> 
> unused devices: <none>
> root@diamond:/# mdadm -A --verbose /dev/md1 /dev/sd[onjlkuhedcb]1
> mdadm: looking for devices for /dev/md1
> mdadm: /dev/sdb1 is identified as a member of /dev/md1, slot 4.
> mdadm: /dev/sdc1 is identified as a member of /dev/md1, slot 5.
> mdadm: /dev/sdd1 is identified as a member of /dev/md1, slot 6.
> mdadm: /dev/sde1 is identified as a member of /dev/md1, slot 7.
> mdadm: /dev/sdh1 is identified as a member of /dev/md1, slot 12.
> mdadm: /dev/sdj1 is identified as a member of /dev/md1, slot 10.
> mdadm: /dev/sdk1 is identified as a member of /dev/md1, slot 8.
> mdadm: /dev/sdl1 is identified as a member of /dev/md1, slot 0.
> mdadm: /dev/sdn1 is identified as a member of /dev/md1, slot 2.
> mdadm: /dev/sdo1 is identified as a member of /dev/md1, slot 3.
> mdadm: device 8 in /dev/md1 has wrong state in superblock, but
> /dev/sdk1 seems ok
> mdadm: device 10 in /dev/md1 has wrong state in superblock, but
> /dev/sdj1 seems ok
> mdadm: device 12 in /dev/md1 has wrong state in superblock, but
> /dev/sdh1 seems ok
> mdadm: SET_ARRAY_INFO failed for /dev/md1: Device or resource busy
> root@diamond:/# cat /proc/mdstat
> Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
> [raid4] [raid10]
> md1 : inactive sdk1[13](S) sdj1[11](S) sdd1[6](S) sdh1[9](S)
> sdo1[3](S) sde1[7](S) sdn1[2](S) sdl1[0](S) sdc1[8](S) sdb1[4](S)
>       19535134315 blocks super 1.2
> 
> md0 : active raid5 sdg1[3] sda1[0] sdf1[1] sdq1[2]
>       2929686528 blocks level 5, 256k chunk, algorithm 2 [4/4] [UUUU]
> 
> unused devices: <none>
> 
> 
> Output from /var/log/messages for the mdadm -A
> Apr  1 20:00:57 diamond kernel: [106978.432900] md: md1 stopped.
> Apr  1 20:00:57 diamond kernel: [106978.493151] md: bind<sdc1>
> Apr  1 20:00:57 diamond kernel: [106978.494551] md: bind<sdb1>
> Apr  1 20:00:57 diamond kernel: [106978.496256] md: bind<sdd1>
> Apr  1 20:00:57 diamond kernel: [106978.516939] md: array md1 already has disks!

That is where SET_ARRAY_INFO is failing ... but why does md1 already have
disks I wonder...

either mdadm has some weird bug - what version are you running???

or something else is messing with md1.

Maybe udev is noticing those devices again for some reason and trying to add
them to the array independently.
You could run
   udevadm monitor

at the same time and see what happens.
Also look in /lib/udev/rules.d or /etc/udev/rules.d to find an entry that 
run "mdadm -I" or "mdadm --incremental" and  try commenting that entry out.

NeilBrown

Attachment: signature.asc
Description: PGP signature


[ATA RAID]     [Linux SCSI Target Infrastructure]     [Managing RAID on Linux]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device-Mapper]     [Kernel]     [Linux Books]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Photos]     [Yosemite Photos]     [Yosemite News]     [AMD 64]     [Linux Networking]

Add to Google Powered by Linux