Re: corrupt leaf; invalid root item size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

any more ideas to investigate this?

On Thu, Jun 4, 2020 at 7:57 PM Thorsten Rehm <thorsten.rehm@xxxxxxxxx> wrote:
>
> Hmm, ok wait a minute:
>
> "But still, if you're using metadata without copy (aka, SINGLE, RAID0)
> then it would be a completely different story."
>
> It's a single disk (SSD):
>
> root@grml ~ # btrfs filesystem usage /mnt
> Overall:
>     Device size:         115.23GiB
>     Device allocated:          26.08GiB
>     Device unallocated:          89.15GiB
>     Device missing:             0.00B
>     Used:               7.44GiB
>     Free (estimated):         104.04GiB    (min: 59.47GiB)
>     Data ratio:                  1.00
>     Metadata ratio:              2.00
>     Global reserve:          25.25MiB    (used: 0.00B)
>
> Data,single: Size:22.01GiB, Used:7.11GiB (32.33%)
>    /dev/mapper/foo      22.01GiB
>
> Metadata,single: Size:8.00MiB, Used:0.00B (0.00%)
>    /dev/mapper/foo       8.00MiB
>
> Metadata,DUP: Size:2.00GiB, Used:167.81MiB (8.19%)
>    /dev/mapper/foo       4.00GiB
>
> System,single: Size:4.00MiB, Used:0.00B (0.00%)
>    /dev/mapper/foo       4.00MiB
>
> System,DUP: Size:32.00MiB, Used:4.00KiB (0.01%)
>    /dev/mapper/foo      64.00MiB
>
> Unallocated:
>    /dev/mapper/foo      89.15GiB
>
>
> root@grml ~ # btrfs filesystem df /mnt
> Data, single: total=22.01GiB, used=7.11GiB
> System, DUP: total=32.00MiB, used=4.00KiB
> System, single: total=4.00MiB, used=0.00B
> Metadata, DUP: total=2.00GiB, used=167.81MiB
> Metadata, single: total=8.00MiB, used=0.00B
> GlobalReserve, single: total=25.25MiB, used=0.00B
>
> I did also a fstrim:
>
> root@grml ~ # cryptsetup --allow-discards open /dev/sda5 foo
> Enter passphrase for /dev/sda5:
> root@grml ~ # mount -o discard /dev/mapper/foo /mnt
> root@grml ~ # fstrim -v /mnt/
> /mnt/: 105.8 GiB (113600049152 bytes) trimmed
> fstrim -v /mnt/  0.00s user 5.34s system 0% cpu 10:28.70 total
>
> The kern.log in the runtime of fstrim:
> --- snip ---
> Jun 04 12:32:02 grml kernel: BTRFS critical (device dm-0): corrupt
> leaf: root=1 block=32505856 slot=32, invalid root item size, have 239
> expect 439
> Jun 04 12:32:02 grml kernel: BTRFS critical (device dm-0): corrupt
> leaf: root=1 block=32813056 slot=32, invalid root item size, have 239
> expect 439
> Jun 04 12:32:37 grml kernel: BTRFS info (device dm-0): turning on sync discard
> Jun 04 12:32:37 grml kernel: BTRFS info (device dm-0): disk space
> caching is enabled
> Jun 04 12:32:37 grml kernel: BTRFS critical (device dm-0): corrupt
> leaf: root=1 block=32813056 slot=32, invalid root item size, have 239
> expect 439
> Jun 04 12:32:37 grml kernel: BTRFS info (device dm-0): enabling ssd
> optimizations
> Jun 04 12:34:35 grml kernel: BTRFS critical (device dm-0): corrupt
> leaf: root=1 block=32382976 slot=32, invalid root item size, have 239
> expect 439
> Jun 04 12:36:50 grml kernel: BTRFS info (device dm-0): turning on sync discard
> Jun 04 12:36:50 grml kernel: BTRFS info (device dm-0): disk space
> caching is enabled
> Jun 04 12:36:50 grml kernel: BTRFS critical (device dm-0): corrupt
> leaf: root=1 block=32382976 slot=32, invalid root item size, have 239
> expect 439
> Jun 04 12:36:50 grml kernel: BTRFS info (device dm-0): enabling ssd
> optimizations
> --- snap ---
>
> Furthermore the system runs for years now. I can't remember exactly,
> but think for 4-5 years. I've started with Debian Testing and just
> upgraded my system on a regular basis. And and I started with btrfs of
> course, but I can't remember with which version...
>
> The problem is still there after the fstrim. Any further suggestions?
>
> And isn't it a little bit strange, that someone had a very similiar problem?
> https://lore.kernel.org/linux-btrfs/19acbd39-475f-bd72-e280-5f6c6496035c@xxxxxx/
>
> root=1, slot=32, and "invalid root item size, have 239 expect 439" are
> identical to my errors.
>
> Thx so far!
>
>
>
> On Thu, Jun 4, 2020 at 2:06 PM Qu Wenruo <quwenruo.btrfs@xxxxxxx> wrote:
> >
> >
> >
> > On 2020/6/4 下午6:52, Thorsten Rehm wrote:
> > > The disk in question is my root (/) partition. If the filesystem is
> > > that highly damaged, I have to reinstall my system. We will see, if
> > > it's come to that. Maybe we find something interesting on the way...
> > > I've downloaded the latest grml daily image and started my system from
> > > a usb stick. Here we go:
> > >
> > > root@grml ~ # uname -r
> > > 5.6.0-2-amd64
> > >
> > > root@grml ~ # cryptsetup open /dev/sda5 foo
> > >
> > >                                                                   :(
> > > Enter passphrase for /dev/sda5:
> > >
> > > root@grml ~ # file -L -s /dev/mapper/foo
> > > /dev/mapper/foo: BTRFS Filesystem label "slash", sectorsize 4096,
> > > nodesize 4096, leafsize 4096,
> > > UUID=65005d0f-f8ea-4f77-8372-eb8b53198685, 7815716864/123731968000
> > > bytes used, 1 devices
> > >
> > > root@grml ~ # btrfs check /dev/mapper/foo
> > > Opening filesystem to check...
> > > Checking filesystem on /dev/mapper/foo
> > > UUID: 65005d0f-f8ea-4f77-8372-eb8b53198685
> > > [1/7] checking root items
> > > [2/7] checking extents
> > > [3/7] checking free space cache
> > > [4/7] checking fs roots
> > > [5/7] checking only csums items (without verifying data)
> > > [6/7] checking root refs
> > > [7/7] checking quota groups skipped (not enabled on this FS)
> > > found 7815716864 bytes used, no error found
> > > total csum bytes: 6428260
> > > total tree bytes: 175968256
> > > total fs tree bytes: 149475328
> > > total extent tree bytes: 16052224
> > > btree space waste bytes: 43268911
> > > file data blocks allocated: 10453221376
> > >  referenced 8746053632
> >
> > Errr, this is a super good news, all your fs metadata is completely fine
> > (at least for the first copy).
> > Which is completely different from the kernel dmesg.
> >
> > >
> > > root@grml ~ # lsblk /dev/sda5 --fs
> > > NAME  FSTYPE      FSVER LABEL UUID
> > > FSAVAIL FSUSE% MOUNTPOINT
> > > sda5  crypto_LUKS 1           d2b4fa40-8afd-4e16-b207-4d106096fd22
> > > └─foo btrfs             slash 65005d0f-f8ea-4f77-8372-eb8b53198685
> > >
> > > root@grml ~ # mount /dev/mapper/foo /mnt
> > > root@grml ~ # btrfs scrub start /mnt
> > >
> > > root@grml ~ # journalctl -k --no-pager | grep BTRFS
> > > Jun 04 10:33:04 grml kernel: BTRFS: device label slash devid 1 transid
> > > 24750795 /dev/dm-0 scanned by systemd-udevd (3233)
> > > Jun 04 10:45:17 grml kernel: BTRFS info (device dm-0): disk space
> > > caching is enabled
> > > Jun 04 10:45:17 grml kernel: BTRFS critical (device dm-0): corrupt
> > > leaf: root=1 block=54222848 slot=32, invalid root item size, have 239
> > > expect 439
> >
> > One error line without "read time corruption" line means btrfs kernel
> > indeed skipped to next copy.
> > In this case, there is one copy (aka the first copy) corrupted.
> > Strangely, if it's the first copy in kernel, it should also be the first
> > copy in btrfs check.
> >
> > And no problem reported from btrfs check, that's already super strange.
> >
> > > Jun 04 10:45:17 grml kernel: BTRFS info (device dm-0): enabling ssd
> > > optimizations
> > > Jun 04 10:45:17 grml kernel: BTRFS info (device dm-0): checking UUID tree
> > > Jun 04 10:45:38 grml kernel: BTRFS info (device dm-0): scrub: started on devid 1
> > > Jun 04 10:45:49 grml kernel: BTRFS critical (device dm-0): corrupt
> > > leaf: root=1 block=29552640 slot=32, invalid root item size, have 239
> > > expect 439
> > > Jun 04 10:46:25 grml kernel: BTRFS critical (device dm-0): corrupt
> > > leaf: root=1 block=29741056 slot=32, invalid root item size, have 239
> > > expect 439
> > > Jun 04 10:46:31 grml kernel: BTRFS info (device dm-0): scrub: finished
> > > on devid 1 with status: 0
> > > Jun 04 10:46:56 grml kernel: BTRFS critical (device dm-0): corrupt
> > > leaf: root=1 block=29974528 slot=32, invalid root item size, have 239
> > > expect 439
> >
> > This means the corrupted copy are also there for several (and I guess
> > unrelated) tree blocks.
> > For scrub I guess it just try to read the good copy without bothering
> > the bad one it found, so no error reported in scrub.
> >
> > But still, if you're using metadata without copy (aka, SINGLE, RAID0)
> > then it would be a completely different story.
> >
> >
> > >
> > > root@grml ~ # btrfs scrub status /mnt
> > > UUID:             65005d0f-f8ea-4f77-8372-eb8b53198685
> > > Scrub started:    Thu Jun  4 10:45:38 2020
> > > Status:           finished
> > > Duration:         0:00:53
> > > Total to scrub:   7.44GiB
> > > Rate:             143.80MiB/s
> > > Error summary:    no errors found
> > >
> > >
> > > root@grml ~ # for block in 54222848 29552640 29741056 29974528; do
> > > btrfs ins dump-tree -b $block /dev/dm-0; done
> > > btrfs-progs v5.6
> > > leaf 54222848 items 33 free space 1095 generation 24750795 owner ROOT_TREE
> > > leaf 54222848 flags 0x1(WRITTEN) backref revision 1
> > > fs uuid 65005d0f-f8ea-4f77-8372-eb8b53198685
> > > chunk uuid 137764f6-c8e6-45b3-b275-82d8558c1ff9
> > >     item 0 key (289 INODE_ITEM 0) itemoff 3835 itemsize 160
> > >         generation 24703953 transid 24703953 size 262144 nbytes 8595701760
> > ...
> > >         cache generation 24750791 entries 139 bitmaps 8
> > >     item 32 key (DATA_RELOC_TREE ROOT_ITEM 0) itemoff 1920 itemsize 239
> >
> > So it's still there. The first copy is corrupted. Just btrfs-progs can't
> > detect it.
> >
> > >         generation 4 root_dirid 256 bytenr 29380608 level 0 refs 1
> > >         lastsnap 0 byte_limit 0 bytes_used 4096 flags 0x0(none)
> > >         drop key (0 UNKNOWN.0 0) level 0
> > > btrfs-progs v5.6
> > > leaf 29552640 items 33 free space 1095 generation 24750796 owner ROOT_TREE
> > > leaf 29552640 flags 0x1(WRITTEN) backref revision 1
> > > fs uuid 65005d0f-f8ea-4f77-8372-eb8b53198685
> > > chunk uuid 137764f6-c8e6-45b3-b275-82d8558c1ff9
> > ...
> > >     item 32 key (DATA_RELOC_TREE ROOT_ITEM 0) itemoff 1920 itemsize 239
> > >         generation 4 root_dirid 256 bytenr 29380608 level 0 refs 1
> > >         lastsnap 0 byte_limit 0 bytes_used 4096 flags 0x0(none)
> > >         drop key (0 UNKNOWN.0 0) level 0
> >
> > This is different from previous copy, which means it should be an CoWed
> > tree blocks.
> >
> > > btrfs-progs v5.6
> > > leaf 29741056 items 33 free space 1095 generation 24750797 owner ROOT_TREE
> >
> > Even newer one.
> >
> > ...
> > > btrfs-progs v5.6
> > > leaf 29974528 items 33 free space 1095 generation 24750798 owner ROOT_TREE
> >
> > Newer.
> >
> > So It looks the bad copy exists for a while, but at the same time we
> > still have one good copy to let everything float.
> >
> > To kill all the old corrupted copies, if it supports TRIM/DISCARD, I
> > recommend to run scrub first, then fstrim on the fs.
> >
> > If it's HDD, I recommend to run a btrfs balance -m to relocate all
> > metadata blocks, to get rid the bad copies.
> >
> > Of course, all using v5.3+ kernels.
> >
> > Thanks,
> > Qu
> > >
> > > On Thu, Jun 4, 2020 at 12:00 PM Qu Wenruo <quwenruo.btrfs@xxxxxxx> wrote:
> > >>
> > >>
> > >>
> > >> On 2020/6/4 下午5:45, Thorsten Rehm wrote:
> > >>> Thank you for you answer.
> > >>> I've just updated my system, did a reboot and it's running with a
> > >>> 5.6.0-2-amd64 now.
> > >>> So, this is how my kern.log looks like, just right after the start:
> > >>>
> > >>
> > >>>
> > >>> There are too many blocks. I just picked three randomly:
> > >>
> > >> Looks like we need more result, especially some result doesn't match at all.
> > >>
> > >>>
> > >>> === Block 33017856 ===
> > >>> $ btrfs ins dump-tree -b 33017856 /dev/dm-0
> > >>> btrfs-progs v5.6
> > >>> leaf 33017856 items 51 free space 17 generation 24749502 owner FS_TREE
> > >>> leaf 33017856 flags 0x1(WRITTEN) backref revision 1
> > >>> fs uuid 65005d0f-f8ea-4f77-8372-eb8b53198685
> > >>> chunk uuid 137764f6-c8e6-45b3-b275-82d8558c1ff9
> > >> ...
> > >>>         item 31 key (4000670 EXTENT_DATA 1933312) itemoff 2299 itemsize 53
> > >>>                 generation 24749502 type 1 (regular)
> > >>>                 extent data disk byte 1126502400 nr 4096
> > >>>                 extent data offset 0 nr 8192 ram 8192
> > >>>                 extent compression 2 (lzo)
> > >>>         item 32 key (4000670 EXTENT_DATA 1941504) itemoff 2246 itemsize 53
> > >>>                 generation 24749502 type 1 (regular)
> > >>>                 extent data disk byte 0 nr 0
> > >>>                 extent data offset 1937408 nr 4096 ram 4194304
> > >>>                 extent compression 0 (none)
> > >> Not root item at all.
> > >> At least for this copy, it looks like kernel got one completely bad
> > >> copy, then discarded it and found a good copy.
> > >>
> > >> That's very strange, especially when all the other involved ones seems
> > >> random and all at slot 32 is not a coincident.
> > >>
> > >>
> > >>> === Block 44900352  ===
> > >>> btrfs ins dump-tree -b 44900352 /dev/dm-0
> > >>> btrfs-progs v5.6
> > >>> leaf 44900352 items 19 free space 591 generation 24749527 owner FS_TREE
> > >>> leaf 44900352 flags 0x1(WRITTEN) backref revision 1
> > >>
> > >> This block doesn't even have slot 32... It only have 19 items, thus slot
> > >> 0 ~ slot 18.
> > >> And its owner, FS_TREE shouldn't have ROOT_ITEM.
> > >>
> > >>>
> > >>>
> > >>> === Block 55352561664 ===
> > >>> $ btrfs ins dump-tree -b 55352561664 /dev/dm-0
> > >>> btrfs-progs v5.6
> > >>> leaf 55352561664 items 33 free space 1095 generation 24749497 owner ROOT_TREE
> > >>> leaf 55352561664 flags 0x1(WRITTEN) backref revision 1
> > >>> fs uuid 65005d0f-f8ea-4f77-8372-eb8b53198685
> > >>> chunk uuid 137764f6-c8e6-45b3-b275-82d8558c1ff9
> > >> ...
> > >>>         item 32 key (DATA_RELOC_TREE ROOT_ITEM 0) itemoff 1920 itemsize 239
> > >>>                 generation 4 root_dirid 256 bytenr 29380608 level 0 refs 1
> > >>>                 lastsnap 0 byte_limit 0 bytes_used 4096 flags 0x0(none)
> > >>>                 drop key (0 UNKNOWN.0 0) level 0
> > >>
> > >> This looks like the offending tree block.
> > >> Slot 32, item size 239, which is ROOT_ITEM, but in valid size.
> > >>
> > >> Since you're here, I guess a btrfs check without --repair on the
> > >> unmounted fs would help to identify the real damage.
> > >>
> > >> And again, the fs looks very damaged, it's highly recommended to backup
> > >> your data asap.
> > >>
> > >> Thanks,
> > >> Qu
> > >>
> > >>> --- snap ---
> > >>>
> > >>>
> > >>>
> > >>> On Thu, Jun 4, 2020 at 3:31 AM Qu Wenruo <quwenruo.btrfs@xxxxxxx> wrote:
> > >>>>
> > >>>>
> > >>>>
> > >>>> On 2020/6/3 下午9:37, Thorsten Rehm wrote:
> > >>>>> Hi,
> > >>>>>
> > >>>>> I've updated my system (Debian testing) [1] several months ago (~
> > >>>>> December) and I noticed a lot of corrupt leaf messages flooding my
> > >>>>> kern.log [2]. Furthermore my system had some trouble, e.g.
> > >>>>> applications were terminated after some uptime, due to the btrfs
> > >>>>> filesystem errors. This was with kernel 5.3.
> > >>>>> The last time I tried was with Kernel 5.6.0-1-amd64 and the problem persists.
> > >>>>>
> > >>>>> I've downgraded my kernel to 4.19.0-8-amd64 from the Debian Stable
> > >>>>> release and with this kernel there aren't any corrupt leaf messages
> > >>>>> and the problem is gone. IMHO, it must be something coming with kernel
> > >>>>> 5.3 (or 5.x).
> > >>>>
> > >>>> V5.3 introduced a lot of enhanced metadata sanity checks, and they catch
> > >>>> such *obviously* wrong metadata.
> > >>>>>
> > >>>>> My harddisk is a SSD which is responsible for the root partition. I've
> > >>>>> encrypted my filesystem with LUKS and just right after I entered my
> > >>>>> password at the boot, the first corrupt leaf errors appear.
> > >>>>>
> > >>>>> An error message looks like this:
> > >>>>> May  7 14:39:34 foo kernel: [  100.162145] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=35799040 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>
> > >>>> Btrfs root items have fixed size. This is already something very bad.
> > >>>>
> > >>>> Furthermore, the item size is smaller than expected, which means we can
> > >>>> easily get garbage. I'm a little surprised that older kernel can even
> > >>>> work without crashing the whole kernel.
> > >>>>
> > >>>> Some extra info could help us to find out how badly the fs is corrupted.
> > >>>> # btrfs ins dump-tree -b 35799040 /dev/dm-0
> > >>>>
> > >>>>>
> > >>>>> "root=1", "slot=32", "have 239 expect 439" is always the same at every
> > >>>>> error line. Only the block number changes.
> > >>>>
> > >>>> And dumps for the other block numbers too.
> > >>>>
> > >>>>>
> > >>>>> Interestingly it's the very same as reported to the ML here [3]. I've
> > >>>>> contacted the reporter, but he didn't have a solution for me, because
> > >>>>> he changed to a different filesystem.
> > >>>>>
> > >>>>> I've already tried "btrfs scrub" and "btrfs check --readonly /" in
> > >>>>> rescue mode, but w/o any errors. I've also checked the S.M.A.R.T.
> > >>>>> values of the SSD, which are fine. Furthermore I've tested my RAM, but
> > >>>>> again, w/o any errors.
> > >>>>
> > >>>> This doesn't look like a bit flip, so not RAM problems.
> > >>>>
> > >>>> Don't have any better advice until we got the dumps, but I'd recommend
> > >>>> to backup your data since it's still possible.
> > >>>>
> > >>>> Thanks,
> > >>>> Qu
> > >>>>
> > >>>>>
> > >>>>> So, I have no more ideas what I can do. Could you please help me to
> > >>>>> investigate this further? Could it be a bug?
> > >>>>>
> > >>>>> Thank you very much.
> > >>>>>
> > >>>>> Best regards,
> > >>>>> Thorsten
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>> 1:
> > >>>>> $ cat /etc/debian_version
> > >>>>> bullseye/sid
> > >>>>>
> > >>>>> $ uname -a
> > >>>>> [no problem with this kernel]
> > >>>>> Linux foo 4.19.0-8-amd64 #1 SMP Debian 4.19.98-1 (2020-01-26) x86_64 GNU/Linux
> > >>>>>
> > >>>>> $ btrfs --version
> > >>>>> btrfs-progs v5.6
> > >>>>>
> > >>>>> $ sudo btrfs fi show
> > >>>>> Label: 'slash'  uuid: 65005d0f-f8ea-4f77-8372-eb8b53198685
> > >>>>>         Total devices 1 FS bytes used 7.33GiB
> > >>>>>         devid    1 size 115.23GiB used 26.08GiB path /dev/mapper/sda5_crypt
> > >>>>>
> > >>>>> $ btrfs fi df /
> > >>>>> Data, single: total=22.01GiB, used=7.16GiB
> > >>>>> System, DUP: total=32.00MiB, used=4.00KiB
> > >>>>> System, single: total=4.00MiB, used=0.00B
> > >>>>> Metadata, DUP: total=2.00GiB, used=168.19MiB
> > >>>>> Metadata, single: total=8.00MiB, used=0.00B
> > >>>>> GlobalReserve, single: total=25.42MiB, used=0.00B
> > >>>>>
> > >>>>>
> > >>>>> 2:
> > >>>>> [several messages per second]
> > >>>>> May  7 14:39:34 foo kernel: [  100.162145] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=35799040 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:35 foo kernel: [  100.998530] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=35885056 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:35 foo kernel: [  101.348650] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=35926016 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:36 foo kernel: [  101.619437] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=35995648 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:36 foo kernel: [  101.874069] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=36184064 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:36 foo kernel: [  102.339087] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=36319232 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:37 foo kernel: [  102.629429] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=36380672 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:37 foo kernel: [  102.839669] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=36487168 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:37 foo kernel: [  103.109183] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=36597760 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>> May  7 14:39:37 foo kernel: [  103.299101] BTRFS critical (device
> > >>>>> dm-0): corrupt leaf: root=1 block=36626432 slot=32, invalid root item
> > >>>>> size, have 239 expect 439
> > >>>>>
> > >>>>> 3:
> > >>>>> https://lore.kernel.org/linux-btrfs/19acbd39-475f-bd72-e280-5f6c6496035c@xxxxxx/
> > >>>>>
> > >>>>
> > >>
> >




[Index of Archives]     [Linux Filesystem Development]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux