----- Original Message ----- > From: "Thiago Ramon" <thiagoramon@xxxxxxxxx> > On Sun, Feb 10, 2019 at 5:07 AM STEVE LEUNG <sjleung@xxxxxxx> wrote: >> >> Hi all, >> >> I decided to try something a bit crazy, and try multi-device raid1 btrfs on >> top of dm-crypt and bcache. That is: >> >> btrfs -> dm-crypt -> bcache -> physical disks >> >> I have a single cache device in front of 4 disks. Maybe this wasn't >> that good of an idea, because the filesystem went read-only a few >> days after setting it up, and now it won't mount. I'd been running >> btrfs on top of 4 dm-crypt-ed disks for some time without any >> problems, and only added bcache (taking one device out at a time, >> converting it over, adding it back) recently. >> >> This is complete speculation, but I do wonder if having the single >> cache device for multiple btrfs disks triggered the problem. > > But before you try to restore anything, can you go back in your kernel > logs and check for errors? Either one of your devices is failing, you > might have physical link issues or bad memory. Even with a complex > setup like this you shouldn't be getting random corruption like this. Indeed, it looks like plugging in the 5th device for caching may have destabilized things (maybe I'm drawing too much power from the power supply or something), as I've observed some spurious ATA errors trying to boot from rescue media. Things seem to go back to normal if I take the cache device out. This hardware is old, but has seemed reliable enough. Although that said, this is my second btrfs corruption I've run into (fortunately no data lost), so maybe the hardware is not as solid as I'd thought. I guess I should have given it more of a shakedown before rolling out bcache everywhere. :) Thanks for the insight. Steve
