On Thu, Jun 18, 2015 at 02:10:03PM -0700, Mark Fasheh wrote:
> So I split btrfs_cmp_data() into 3 parts with a small context structure that
> is passed between them. The first, btrfs_cmp_data_prepare() gathers up the
> pages needed (taking page lock as required) and puts them on our context
> structure. At this point, we are safe to lock the extent range. Afterwards,
> we use btrfs_cmp_data() to do the data compare as usual and btrfs_cmp_data_free()
> to clean up our context.
Sounds good.
I see some inconsitencies in the double locking.
> @@ -2808,52 +2819,120 @@ static inline void lock_extent_range(struct inode *inode, u64 off, u64 len)
> }
> }
>
> -static void btrfs_double_unlock(struct inode *inode1, u64 loff1,
> - struct inode *inode2, u64 loff2, u64 len)
> +static void btrfs_double_inode_unlock(struct inode *inode1, struct inode *inode2)
> {
> - unlock_extent(&BTRFS_I(inode1)->io_tree, loff1, loff1 + len - 1);
> - unlock_extent(&BTRFS_I(inode2)->io_tree, loff2, loff2 + len - 1);
> -
> mutex_unlock(&inode1->i_mutex);
> mutex_unlock(&inode2->i_mutex);
unlock parent lock first, child lock second -- should it be reversed?
> }
>
> -static void btrfs_double_lock(struct inode *inode1, u64 loff1,
> - struct inode *inode2, u64 loff2, u64 len)
> +static void btrfs_double_inode_lock(struct inode *inode1, struct inode *inode2)
> +{
> + if (inode1 < inode2)
> + swap(inode1, inode2);
after that, inode1 > inode2
> +
> + mutex_lock_nested(&inode1->i_mutex, I_MUTEX_PARENT);
higher address, locked first, parent lock
> + if (inode1 != inode2)
> + mutex_lock_nested(&inode2->i_mutex, I_MUTEX_CHILD);
lower address, locked second, child lock
> +}
> +
> +static void btrfs_double_extent_unlock(struct inode *inode1, u64 loff1,
> + struct inode *inode2, u64 loff2, u64 len)
> +{
> + unlock_extent(&BTRFS_I(inode1)->io_tree, loff1, loff1 + len - 1);
> + unlock_extent(&BTRFS_I(inode2)->io_tree, loff2, loff2 + len - 1);
reversed?
> +}
> +
> +static void btrfs_double_extent_lock(struct inode *inode1, u64 loff1,
> + struct inode *inode2, u64 loff2, u64 len)
> {
> if (inode1 < inode2) {
> swap(inode1, inode2);
> swap(loff1, loff2);
> }
> -
> - mutex_lock_nested(&inode1->i_mutex, I_MUTEX_PARENT);
> lock_extent_range(inode1, loff1, len);
> - if (inode1 != inode2) {
> - mutex_lock_nested(&inode2->i_mutex, I_MUTEX_CHILD);
> + if (inode1 != inode2)
> lock_extent_range(inode2, loff2, len);
> +}
higher address, locked first
If the locking sequence is always the same, it's not a problem
deadlock-wise, but see btrfs_ioctl_clone:
3639 if (inode < src) {
3640 mutex_lock_nested(&inode->i_mutex, I_MUTEX_PARENT);
3641 mutex_lock_nested(&src->i_mutex, I_MUTEX_CHILD);
3642 } else {
3643 mutex_lock_nested(&src->i_mutex, I_MUTEX_PARENT);
3644 mutex_lock_nested(&inode->i_mutex, I_MUTEX_CHILD);
3645 }
lower address, locked first, parent lock
different from the order in btrfs_double_inode_lock. What happens if we
get the locks interleaved when extent same and clone are called in
parallel?
lock(i1)
lock(i2)
lock(i2) <-- lockup?
lock(i1)
I haven't looked further whether the locking classes (parent, child)
could prevent that, but the code should be clear enough so that I don't
have to dig into the locking code to see if it's ok.
To fix it, the clone ioctl should use the same locking helper and we're
set.
Besides that,
Reviewed-by: David Sterba <dsterba@xxxxxxx>
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in