Re: Removing a failing drive from multiple arrays
|[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Well, it does occur, has on the first bunch of partitions, is now doing the big ~TB one. And because I'm nervous about power cycling sick disks (been there, done that) I am doing the whole rebuild onto drives attached by USB and eSATA connections. On the last one now.On Thu, 19 Apr 2012 14:54:30 -0400 Bill Davidsen<davidsen@xxxxxxx> wrote:I have a failing drive, and partitions are in multiple arrays. I'm looking for the least painful and most reliable way to replace it. It's internal, I have a twin in an external box, and can create all the parts now and then swap the drive physically. The layout is complex, here's what blkdevtra tells me about this device, the full trace is attached. Block device sdd, logical device 8:48 Model Family: Seagate Barracuda 7200.10 Device Model: ST3750640AS Serial Number: 5QD330ZW Device size 732.575 GB sdd1 0.201 GB sdd2 3.912 GB sdd3 24.419 GB sdd4 0.000 GB sdd5 48.838 GB [md123] /mnt/workspace sdd6 0.498 GB sdd7 19.543 GB [md125] sdd8 29.303 GB [md126] sdd9 605.859 GB [md127] /exports/common Unpartitioned 0.003 GB I think what I want to do is to partition the new drive, then one array at a time fail and remove the partition on the bad drive, and add a partition on the new good drive. Then repeat for each array until all are complete and on a new drive. Then I should be able to power off, remove the failed drive, put the good drive in the case, and the arrays should reassemble by UUID. Does that sound right? Is there an easier way?I would add the new partition before failing the old but that isn't a big issues. If you were running a really new kernel, used 1.x metadata, and were happy to try out code that that hasn't had a lot of real-life testing you could (after adding the new partition) do echo want_replacement> /sys/block/md123/md/dev-sdd5/state (for example). Then it would build the spare before failing the original. You need linux 3.3 for this to have any chance of working.
I did them all live and running, although I did "swapoff" the one for swap, it isn't really needed and just seems like a bad thing to be diddling while the system is using it.
Good news, it has worked perfectly, bad news it doesn't do what I thought it did. For RAID- it does what I expected and pulls data off the partition marked for replacement, but with RAID-10 2f layout the "take the best copy" logic seems to take over and data comes from all active drives. I would have expected it to come from the failing drive first and be taken elsewhere only if the failing drive didn't provide the data. I have seen cases where migration failed due to a bad sector on another drive, so that's unexpected. I don't claim it wrong, just "not what I expected."
I think in a perfect world (where you have infinite time to diddle stuff), it would be useful to have three options:
- favor the failing drive, recover what you must - reconstruct all data possible, don't use the failing drive - build the new copy fastest way possible, get it where it's available.In any case this feature worked just fine, and I put my thoughts on the method out for comment. By morning the last rebuild will be done, and I can actually pull the bad drives by serial number, hope the UUID means the new drive can go anywhere, add another eSATA card and Blu-Ray burner, and be up solid.
-- Bill Davidsen <davidsen@xxxxxxx> "We have more to fear from the bungling of the incompetent than from the machinations of the wicked." - from Slashdot -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html