Hi, This email is actually several questions clubbed as one... I have a btrfs filesystem mounted at /btrfs_vol/ Every N minutes, I run bedup for deduplication of data in /btrfs_vol Inside /btrfs_vol, I have several subvolumes (consider this as home directories of several users) I have set individual qgroup limits for each of these subvolumes. Additionally, I'm using btrfs-send/receive to backup each subvolume to an identical btrfs filesystem on a remote server. I hope I've explained the setup clearly. Now, here are my questions: 1. I understand that bedup does deduplication on whole-file basis. I also read that duperemove works at extent level. How do the two compare in terms of performance? How stable is duperemove? I read a few posts suggesting bugs related to extent-same. duperemove seems ideal for my workload. However, have stayed away from it, fearing instability. 2. Is it possible to have nested subvolumes (like in my tree structure), and run btrfs-send and expect the data whole filesystem (including all subvolumes) to be sent? My problem is that, since I send btrfs subvolumes, if there is any dedplicated data between these subvolumes, it needs to send 2 copies of the same data to the remote host. You might ask why not have just a single subvolume and backup the whole thing. But there are many cool features, like quotas and btrfs send/receive based backup for selected user home directories, which works only at the subvolume level. Please help me with your customized solutions to my problem. Thanks in advance for your help. -- -Shyam -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html
