On 12.02.2018 16:17, Liu Bo wrote:
> On Tue, Feb 06, 2018 at 11:11:55AM +0200, Nikolay Borisov wrote:
>>
>>
>> On 6.02.2018 01:15, Liu Bo wrote:
>>> Btrfs tries its best to tolerate write errors, but kind of silently
>>> (except some messages in kernel log).
>>>
>>> For raid1 and raid10, this is usually not a problem because there is a
>>> copy as backup, while for parity based raid setup, i.e. raid5 and
>>> raid6, the problem is that, if a write error occurs due to some bad
>>> sectors, one horizonal stripe becomes degraded and the number of write
>>> errors it can tolerate gets reduced by one, now if two disk fails,
>>> data may be lost forever.
>>>
>>> One way to mitigate the data loss pain is to expose 'bad chunks',
>>> i.e. degraded chunks, to users, so that they can use 'btrfs balance'
>>> to relocate the whole chunk and get the full raid6 protection again
>>> (if the relocation works).
>>>
>>> This introduces 'bad_chunks' in btrfs's per-fs sysfs directory. Once
>>> a chunk of raid5 or raid6 becomes degraded, it will appear in
>>> 'bad_chunks'.
>>>
>>> Signed-off-by: Liu Bo <bo.li.liu@xxxxxxxxxx>
>>> ---
>>> - In this patch, 'bad chunks' is not persistent on disk, but it can be
>>> added if it's thought to be a good idea.
>>> - This is lightly tested, comments are very welcome.
>>>
>>> fs/btrfs/ctree.h | 8 +++++++
>>> fs/btrfs/disk-io.c | 2 ++
>>> fs/btrfs/extent-tree.c | 13 +++++++++++
>>> fs/btrfs/raid56.c | 59 ++++++++++++++++++++++++++++++++++++++++++++++++--
>>> fs/btrfs/sysfs.c | 26 ++++++++++++++++++++++
>>> fs/btrfs/volumes.c | 15 +++++++++++--
>>> fs/btrfs/volumes.h | 2 ++
>>> 7 files changed, 121 insertions(+), 4 deletions(-)
>>>
>>> diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h
>>> index 13c260b..08aad65 100644
>>> --- a/fs/btrfs/ctree.h
>>> +++ b/fs/btrfs/ctree.h
>>> @@ -1101,6 +1101,9 @@ struct btrfs_fs_info {
>>> spinlock_t ref_verify_lock;
>>> struct rb_root block_tree;
>>> #endif
>>> +
>>> + struct list_head bad_chunks;
>>> + seqlock_t bc_lock;
>>> };
>>>
>>> static inline struct btrfs_fs_info *btrfs_sb(struct super_block *sb)
>>> @@ -2568,6 +2571,11 @@ static inline gfp_t btrfs_alloc_write_mask(struct address_space *mapping)
>>>
>>> /* extent-tree.c */
>>>
>>> +struct btrfs_bad_chunk {
>>> + u64 chunk_offset;
>>> + struct list_head list;
>>> +};
>>> +
>>> enum btrfs_inline_ref_type {
>>> BTRFS_REF_TYPE_INVALID = 0,
>>> BTRFS_REF_TYPE_BLOCK = 1,
>>> diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c
>>> index a8ecccf..061e7f94 100644
>>> --- a/fs/btrfs/disk-io.c
>>> +++ b/fs/btrfs/disk-io.c
>>> @@ -2568,6 +2568,8 @@ int open_ctree(struct super_block *sb,
>>> init_waitqueue_head(&fs_info->async_submit_wait);
>>>
>>> INIT_LIST_HEAD(&fs_info->pinned_chunks);
>>> + INIT_LIST_HEAD(&fs_info->bad_chunks);
>>> + seqlock_init(&fs_info->bc_lock);
>>>
>>> /* Usable values until the real ones are cached from the superblock */
>>> fs_info->nodesize = 4096;
>>> diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c
>>> index 2f43285..3ca7cb4 100644
>>> --- a/fs/btrfs/extent-tree.c
>>> +++ b/fs/btrfs/extent-tree.c
>>> @@ -9903,6 +9903,19 @@ int btrfs_free_block_groups(struct btrfs_fs_info *info)
>>> kobject_del(&space_info->kobj);
>>> kobject_put(&space_info->kobj);
>>> }
>>> +
>>> + /* Clean up bad chunks. */
>>> + write_seqlock_irq(&info->bc_lock);
>>> + while (!list_empty(&info->bad_chunks)) {
>>
>> Why not the idiomatic list_for_each_entry_safe, that way you remove the
>> list_first_entry invocation altogether and still get a well-formed
>> btrfs_bad_chunk object.
>>
>>> + struct btrfs_bad_chunk *bc;
>>> +
>>> + bc = list_first_entry(&info->bad_chunks,
>>> + struct btrfs_bad_chunk, list);
>>> + list_del_init(&bc->list);
>>
>> nit: no need to use the _init variant, you are directly freeing the
>> entry, less code to execute :)
>>
>>> + kfree(bc);
>>> + }
>>> + write_sequnlock_irq(&info->bc_lock);
>>> +
>>> return 0;
>>> }
>>>
>>> diff --git a/fs/btrfs/raid56.c b/fs/btrfs/raid56.c
>>> index a7f7925..e960247 100644
>>> --- a/fs/btrfs/raid56.c
>>> +++ b/fs/btrfs/raid56.c
>>> @@ -888,14 +888,19 @@ static void rbio_orig_end_io(struct btrfs_raid_bio *rbio, blk_status_t err)
>>> }
>>>
>>> /*
>>> - * end io function used by finish_rmw. When we finally
>>> - * get here, we've written a full stripe
>>> + * end io function used by finish_rmw. When we finally get here, we've written
>>> + * a full stripe.
>>> + *
>>> + * Note that this is not under interrupt context as we queued endio to workers.
>>> */
>>> static void raid_write_end_io(struct bio *bio)
>>> {
>>> struct btrfs_raid_bio *rbio = bio->bi_private;
>>> blk_status_t err = bio->bi_status;
>>> int max_errors;
>>> + u64 stripe_start = rbio->bbio->raid_map[0];
>>> + struct btrfs_fs_info *fs_info = rbio->fs_info;
>>> + int err_cnt;
>>>
>>> if (err)
>>> fail_bio_stripe(rbio, bio);
>>> @@ -908,12 +913,58 @@ static void raid_write_end_io(struct bio *bio)
>>> err = BLK_STS_OK;
>>>
>>> /* OK, we have read all the stripes we need to. */
>>> + err_cnt = atomic_read(&rbio->error);
>>> max_errors = (rbio->operation == BTRFS_RBIO_PARITY_SCRUB) ?
>>> 0 : rbio->bbio->max_errors;
>>> if (atomic_read(&rbio->error) > max_errors)
>>> err = BLK_STS_IOERR;
>>>
>>> rbio_orig_end_io(rbio, err);
>>> +
>>> + /*
>>> + * If there is any error, this stripe is a degraded one, so is the whole
>>> + * chunk, expose this chunk info to sysfs.
>>> + */
>>> + if (unlikely(err_cnt)) {
>>> + struct btrfs_bad_chunk *bc;
>>> + struct btrfs_bad_chunk *tmp;
>>> + struct extent_map *em;
>>> + unsigned long flags;
>>> +
>>> + em = get_chunk_map(fs_info, stripe_start, 1);
>>> + if (IS_ERR(em))
>>> + return;
>>> +
>>> + bc = kzalloc(sizeof(*bc), GFP_NOFS);
>>> + /* If allocation fails, it's OK. */
>>> + if (!bc) {
>>> + free_extent_map(em);
>>> + return;
>>> + }
>>> +
>>> + write_seqlock_irqsave(&fs_info->bc_lock, flags);
>>
>> Why do you disable interrupts here and the comment at the beginning of
>> the function claims this code can't be executed in irq context? Given
>> the comment I'd expect if you put the following assert at the beginning
>> of the function it should never trigger:
>>
>> ASSERT(in_irq())
>
> I think you're right, no one is processing the object in irq context.
>
>>
>>> + list_for_each_entry(tmp, &fs_info->bad_chunks, list) {
>>> + if (tmp->chunk_offset != em->start)
>>> + continue;
>>> +
>>> + /*
>>> + * Don't bother if this chunk has already been recorded.
>>> + */
>>> + write_sequnlock_irqrestore(&fs_info->bc_lock, flags);
>>> + kfree(bc);
>>> + free_extent_map(em);
>>> + return;
>>> + }
>>> +
>>> + /* Add new bad chunk to list. */
>>> + bc->chunk_offset = em->start;
>>> + free_extent_map(em);
>>> +
>>> + INIT_LIST_HEAD(&bc->list);
>>
>> nit: There is no need to initialize the list head of the entry itself.
>>
>>> + list_add(&bc->list, &fs_info->bad_chunks);
>>> +
>>> + write_sequnlock_irqrestore(&fs_info->bc_lock, flags);
>>> + }
>>> }
>>>
>>> /*
>>> @@ -1320,6 +1371,8 @@ static noinline void finish_rmw(struct btrfs_raid_bio *rbio)
>>> bio->bi_end_io = raid_write_end_io;
>>> bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>>
>>> + btrfs_bio_wq_end_io(rbio->fs_info, bio, BTRFS_WQ_ENDIO_RAID56);
>>> +
>>> submit_bio(bio);
>>> }
>>> return;
>>> @@ -2465,6 +2518,8 @@ static noinline void finish_parity_scrub(struct btrfs_raid_bio *rbio,
>>> bio->bi_end_io = raid_write_end_io;
>>> bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>>
>>> + btrfs_bio_wq_end_io(rbio->fs_info, bio, BTRFS_WQ_ENDIO_RAID56);
>>> +
>>> submit_bio(bio);
>>> }
>>> return;
>>> diff --git a/fs/btrfs/sysfs.c b/fs/btrfs/sysfs.c
>>> index a28bba8..0baaa33 100644
>>> --- a/fs/btrfs/sysfs.c
>>> +++ b/fs/btrfs/sysfs.c
>>> @@ -490,12 +490,38 @@ static ssize_t quota_override_store(struct kobject *kobj,
>>>
>>> BTRFS_ATTR_RW(, quota_override, quota_override_show, quota_override_store);
>>>
>>> +static ssize_t btrfs_bad_chunks_show(struct kobject *kobj,
>>> + struct kobj_attribute *a, char *buf)
>>> +{
>>> + struct btrfs_fs_info *fs_info = to_fs_info(kobj);
>>> + struct btrfs_bad_chunk *bc;
>>> + int len = 0;
>>> + unsigned int seq;
>>> +
>>> + /* read lock please */
>>> + do {
>>> + seq = read_seqbegin(&fs_info->bc_lock);
>>> + list_for_each_entry(bc, &fs_info->bad_chunks, list) {
>>> + len += snprintf(buf + len, PAGE_SIZE - len, "%llu\n",
>>> + bc->chunk_offset);
>>> + /* chunk offset is u64 */
>>> + if (len >= PAGE_SIZE)
>>> + break;
>>> + }
>>> + } while (read_seqretry(&fs_info->bc_lock, seq));
>>> +
>>> + return len;
>>> +}
>>> +
>>> +BTRFS_ATTR(, bad_chunks, btrfs_bad_chunks_show);
>>> +
>>> static const struct attribute *btrfs_attrs[] = {
>>> BTRFS_ATTR_PTR(, label),
>>> BTRFS_ATTR_PTR(, nodesize),
>>> BTRFS_ATTR_PTR(, sectorsize),
>>> BTRFS_ATTR_PTR(, clone_alignment),
>>> BTRFS_ATTR_PTR(, quota_override),
>>> + BTRFS_ATTR_PTR(, bad_chunks),
>>> NULL,
>>> };
>>>
>>> diff --git a/fs/btrfs/volumes.c b/fs/btrfs/volumes.c
>>> index a256842..d71f11a 100644
>>> --- a/fs/btrfs/volumes.c
>>> +++ b/fs/btrfs/volumes.c
>>> @@ -2803,8 +2803,8 @@ static int btrfs_del_sys_chunk(struct btrfs_fs_info *fs_info, u64 chunk_offset)
>>> return ret;
>>> }
>>>
>>> -static struct extent_map *get_chunk_map(struct btrfs_fs_info *fs_info,
>>> - u64 logical, u64 length)
>>> +struct extent_map *get_chunk_map(struct btrfs_fs_info *fs_info,
>>> + u64 logical, u64 length)
>>
>> nit: Since you are exposing the function as an API I think this is a
>> good opportunity to add proper kernel doc for it.
>>
>
> It has nothing to do with the patch's purpose, lets leave it to a
> seperate one.
>
>>> {
>>> struct extent_map_tree *em_tree;
>>> struct extent_map *em;
>>> @@ -2840,6 +2840,7 @@ int btrfs_remove_chunk(struct btrfs_trans_handle *trans,
>>> u64 dev_extent_len = 0;
>>> int i, ret = 0;
>>> struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
>>> + struct btrfs_bad_chunk *bc;
>>>
>>> em = get_chunk_map(fs_info, chunk_offset, 1);
>>> if (IS_ERR(em)) {
>>> @@ -2916,6 +2917,16 @@ int btrfs_remove_chunk(struct btrfs_trans_handle *trans,
>>> }
>>>
>>> out:
>>> + write_seqlock_irq(&fs_info->bc_lock);
>>> + list_for_each_entry(bc, &fs_info->bad_chunks, list) {
>>
>> Use list_for_each_entry_safe to make it more apparent you are going to
>> be removing from the list. The code as-is works since you are doing a
>> break after deleting element from the list but this is somewhat subtle.
>
> To be honest, I don't see much difference.
>
> I think the _safe version is to protect us from some race when others
> remove objects from list, and write lock is held so we're safe.
No, the _safe version uses the second argument (n) as the list iterator.
The non-safe version just uses 'pos', and in case you remove 'pos' from
the list AND continue iterating you will deref an invalid pointer. So
_safe is actually really necessary for correctness when you intend to
remove an entry from a list you are iterating, irrespective of any locks
you might have.
>
>> Also it's not necessary to re-init the deleted entry since you are
>> directly freeing it.
>>
>
> OK.
>
> Thanks for the comments.
>
> Thanks,
>
> -liubo
>>> + if (bc->chunk_offset == chunk_offset) {
>>> + list_del_init(&bc->list);
>>> + kfree(bc);
>>> + break;
>>> + }
>>> + }
>>> + write_sequnlock_irq(&fs_info->bc_lock);
>>> +
>>> /* once for us */
>>> free_extent_map(em);
>>> return ret;
>>> diff --git a/fs/btrfs/volumes.h b/fs/btrfs/volumes.h
>>> index ff15208..4e846ba 100644
>>> --- a/fs/btrfs/volumes.h
>>> +++ b/fs/btrfs/volumes.h
>>> @@ -396,6 +396,8 @@ static inline enum btrfs_map_op btrfs_op(struct bio *bio)
>>> }
>>> }
>>>
>>> +struct extent_map *get_chunk_map(struct btrfs_fs_info *fs_info,
>>> + u64 logical, u64 length);
>>> int btrfs_account_dev_extents_size(struct btrfs_device *device, u64 start,
>>> u64 end, u64 *length);
>>> void btrfs_get_bbio(struct btrfs_bio *bbio);
>>>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html