Re: Decoding "unable to fixup (regular)" errors

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Nov 08, 2019 at 11:21:56PM +0100, Richard Weinberger wrote:
> ----- Ursprüngliche Mail -----
> > btrfs found corrupted data on md1.  You appear to be using btrfs
> > -dsingle on a single mdadm raid1 device, so no recovery is possible
> > ("unable to fixup").
> > 
> >> The system has ECC memory with md1 being a RAID1 which passes all health checks.
> > 
> > mdadm doesn't have any way to repair data corruption--it can find
> > differences, but it cannot identify which version of the data is correct.
> > If one of your drives is corrupting data without reporting IO errors,
> > mdadm will simply copy the corruption to the other drive.  If one
> > drive is failing by intermittently injecting corrupted bits into reads
> > (e.g. because of a failure in the RAM on the drive control board),
> > this behavior may not show up in mdadm health checks.
> 
> Well, this is not cheap hardware...
> Possible, but not very likely IMHO

Even the disks?  We see RAM failures in disk drive embedded boards from
time to time.

> >> I tried to find the inodes behind the erroneous addresses without success.
> >> e.g.
> >> $ btrfs inspect-internal logical-resolve -v -P 593483341824 /
> >> ioctl ret=0, total_size=4096, bytes_left=4080, bytes_missing=0, cnt=0, missed=0
> >> $ echo $?
> >> 1
> > 
> > That usually means the file is deleted, or the specific blocks referenced
> > have been overwritten (i.e. there are no references to the given block in
> > any existing file, but a reference to the extent containing the block
> > still exists).  Although it's not possible to reach those blocks by
> > reading a file, a scrub or balance will still hit the corrupted blocks.
> > 
> > You can try adding or subtracting multiples of 4096 to the block number
> > to see if you get a hint about which inodes reference this extent.
> > The first block found in either direction should be a reference to the
> > same extent, though there's no easy way (other than dumping the extent
> > tree with 'btrfs ins dump-tree -t 2' and searching for the extent record
> > containing the block number) to figure out which.  Extents can be up to
> > 128MB long, i.e. 32768 blocks.
> 
> Thanks for the hint!
> 
> > Or modify 'btrfs ins log' to use LOGICAL_INO_V2 and the IGNORE_OFFSETS
> > flag.
> > 
> >> My kernel is 4.12.14-lp150.12.64-default (OpenSUSE 15.0), so not super recent
> >> but AFAICT btrfs should be sane
> >> there. :-)
> > 
> > I don't know of specific problems with csums in 4.12, but I'd upgrade that
> > for a dozen other reasons anyway.  One of those is that LOGICAL_INO_V2
> > was merged in 4.15.
> > 
> >> What could cause the errors and how to dig further?
> > 
> > Probably a silent data corruption on one of the underlying disks.
> > If you convert this mdadm raid1 to btrfs raid1, btrfs will tell you
> > which disk the errors are coming from while also correcting them.
> 
> Hmm, I don't really buy this reasoning. Like I said, this is not
> cheap/consumer hardware.
> 
> Thanks,
> //richard

Attachment: signature.asc
Description: PGP signature


[Index of Archives]     [Linux Filesystem Development]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux