On 12.02.2018 16:17, Liu Bo wrote:
> On Tue, Feb 06, 2018 at 11:11:55AM +0200, Nikolay Borisov wrote:
>>
>>
>> On  6.02.2018 01:15, Liu Bo wrote:
>>> Btrfs tries its best to tolerate write errors, but kind of silently
>>> (except some messages in kernel log).
>>>
>>> For raid1 and raid10, this is usually not a problem because there is a
>>> copy as backup, while for parity based raid setup, i.e. raid5 and
>>> raid6, the problem is that, if a write error occurs due to some bad
>>> sectors, one horizonal stripe becomes degraded and the number of write
>>> errors it can tolerate gets reduced by one, now if two disk fails,
>>> data may be lost forever.
>>>
>>> One way to mitigate the data loss pain is to expose 'bad chunks',
>>> i.e. degraded chunks, to users, so that they can use 'btrfs balance'
>>> to relocate the whole chunk and get the full raid6 protection again
>>> (if the relocation works).
>>>
>>> This introduces 'bad_chunks' in btrfs's per-fs sysfs directory.  Once
>>> a chunk of raid5 or raid6 becomes degraded, it will appear in
>>> 'bad_chunks'.
>>>
>>> Signed-off-by: Liu Bo <bo.li....@oracle.com>
>>> ---
>>> - In this patch, 'bad chunks' is not persistent on disk, but it can be
>>>   added if it's thought to be a good idea.
>>> - This is lightly tested, comments are very welcome.
>>>
>>>  fs/btrfs/ctree.h       |  8 +++++++
>>>  fs/btrfs/disk-io.c     |  2 ++
>>>  fs/btrfs/extent-tree.c | 13 +++++++++++
>>>  fs/btrfs/raid56.c      | 59 
>>> ++++++++++++++++++++++++++++++++++++++++++++++++--
>>>  fs/btrfs/sysfs.c       | 26 ++++++++++++++++++++++
>>>  fs/btrfs/volumes.c     | 15 +++++++++++--
>>>  fs/btrfs/volumes.h     |  2 ++
>>>  7 files changed, 121 insertions(+), 4 deletions(-)
>>>
>>> diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h
>>> index 13c260b..08aad65 100644
>>> --- a/fs/btrfs/ctree.h
>>> +++ b/fs/btrfs/ctree.h
>>> @@ -1101,6 +1101,9 @@ struct btrfs_fs_info {
>>>     spinlock_t ref_verify_lock;
>>>     struct rb_root block_tree;
>>>  #endif
>>> +
>>> +   struct list_head bad_chunks;
>>> +   seqlock_t bc_lock;
>>>  };
>>>  
>>>  static inline struct btrfs_fs_info *btrfs_sb(struct super_block *sb)
>>> @@ -2568,6 +2571,11 @@ static inline gfp_t btrfs_alloc_write_mask(struct 
>>> address_space *mapping)
>>>  
>>>  /* extent-tree.c */
>>>  
>>> +struct btrfs_bad_chunk {
>>> +   u64 chunk_offset;
>>> +   struct list_head list;
>>> +};
>>> +
>>>  enum btrfs_inline_ref_type {
>>>     BTRFS_REF_TYPE_INVALID =         0,
>>>     BTRFS_REF_TYPE_BLOCK =           1,
>>> diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c
>>> index a8ecccf..061e7f94 100644
>>> --- a/fs/btrfs/disk-io.c
>>> +++ b/fs/btrfs/disk-io.c
>>> @@ -2568,6 +2568,8 @@ int open_ctree(struct super_block *sb,
>>>     init_waitqueue_head(&fs_info->async_submit_wait);
>>>  
>>>     INIT_LIST_HEAD(&fs_info->pinned_chunks);
>>> +   INIT_LIST_HEAD(&fs_info->bad_chunks);
>>> +   seqlock_init(&fs_info->bc_lock);
>>>  
>>>     /* Usable values until the real ones are cached from the superblock */
>>>     fs_info->nodesize = 4096;
>>> diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c
>>> index 2f43285..3ca7cb4 100644
>>> --- a/fs/btrfs/extent-tree.c
>>> +++ b/fs/btrfs/extent-tree.c
>>> @@ -9903,6 +9903,19 @@ int btrfs_free_block_groups(struct btrfs_fs_info 
>>> *info)
>>>             kobject_del(&space_info->kobj);
>>>             kobject_put(&space_info->kobj);
>>>     }
>>> +
>>> +   /* Clean up bad chunks. */
>>> +   write_seqlock_irq(&info->bc_lock);
>>> +   while (!list_empty(&info->bad_chunks)) {
>>
>> Why not the idiomatic list_for_each_entry_safe, that way you remove the
>> list_first_entry invocation altogether and still get a well-formed
>> btrfs_bad_chunk object.
>>
>>> +           struct btrfs_bad_chunk *bc;
>>> +
>>> +           bc = list_first_entry(&info->bad_chunks,
>>> +                                 struct btrfs_bad_chunk, list);
>>> +           list_del_init(&bc->list);
>>
>> nit: no need to use the _init variant, you are directly freeing the
>> entry, less code to execute :)
>>
>>> +           kfree(bc);
>>> +   }
>>> +   write_sequnlock_irq(&info->bc_lock);
>>> +
>>>     return 0;
>>>  }
>>>  
>>> diff --git a/fs/btrfs/raid56.c b/fs/btrfs/raid56.c
>>> index a7f7925..e960247 100644
>>> --- a/fs/btrfs/raid56.c
>>> +++ b/fs/btrfs/raid56.c
>>> @@ -888,14 +888,19 @@ static void rbio_orig_end_io(struct btrfs_raid_bio 
>>> *rbio, blk_status_t err)
>>>  }
>>>  
>>>  /*
>>> - * end io function used by finish_rmw.  When we finally
>>> - * get here, we've written a full stripe
>>> + * end io function used by finish_rmw.  When we finally get here, we've 
>>> written
>>> + * a full stripe.
>>> + *
>>> + * Note that this is not under interrupt context as we queued endio to 
>>> workers.
>>>   */
>>>  static void raid_write_end_io(struct bio *bio)
>>>  {
>>>     struct btrfs_raid_bio *rbio = bio->bi_private;
>>>     blk_status_t err = bio->bi_status;
>>>     int max_errors;
>>> +   u64 stripe_start = rbio->bbio->raid_map[0];
>>> +   struct btrfs_fs_info *fs_info = rbio->fs_info;
>>> +   int err_cnt;
>>>  
>>>     if (err)
>>>             fail_bio_stripe(rbio, bio);
>>> @@ -908,12 +913,58 @@ static void raid_write_end_io(struct bio *bio)
>>>     err = BLK_STS_OK;
>>>  
>>>     /* OK, we have read all the stripes we need to. */
>>> +   err_cnt = atomic_read(&rbio->error);
>>>     max_errors = (rbio->operation == BTRFS_RBIO_PARITY_SCRUB) ?
>>>                  0 : rbio->bbio->max_errors;
>>>     if (atomic_read(&rbio->error) > max_errors)
>>>             err = BLK_STS_IOERR;
>>>  
>>>     rbio_orig_end_io(rbio, err);
>>> +
>>> +   /*
>>> +    * If there is any error, this stripe is a degraded one, so is the whole
>>> +    * chunk, expose this chunk info to sysfs.
>>> +    */
>>> +   if (unlikely(err_cnt)) {
>>> +           struct btrfs_bad_chunk *bc;
>>> +           struct btrfs_bad_chunk *tmp;
>>> +           struct extent_map *em;
>>> +           unsigned long flags;
>>> +
>>> +           em = get_chunk_map(fs_info, stripe_start, 1);
>>> +           if (IS_ERR(em))
>>> +                   return;
>>> +
>>> +           bc = kzalloc(sizeof(*bc), GFP_NOFS);
>>> +           /* If allocation fails, it's OK. */
>>> +           if (!bc) {
>>> +                   free_extent_map(em);
>>> +                   return;
>>> +           }
>>> +
>>> +           write_seqlock_irqsave(&fs_info->bc_lock, flags);
>>
>> Why do you disable interrupts here and the comment at the beginning of
>> the function claims this code can't be executed in irq context? Given
>> the comment I'd expect if you put the following assert at the beginning
>> of the function it should never trigger:
>>
>> ASSERT(in_irq())
> 
> I think you're right, no one is processing the object in irq context.
> 
>>
>>> +           list_for_each_entry(tmp, &fs_info->bad_chunks, list) {
>>> +                   if (tmp->chunk_offset != em->start)
>>> +                           continue;
>>> +
>>> +                   /*
>>> +                    * Don't bother if this chunk has already been recorded.
>>> +                    */
>>> +                   write_sequnlock_irqrestore(&fs_info->bc_lock, flags);
>>> +                   kfree(bc);
>>> +                   free_extent_map(em);
>>> +                   return;
>>> +           }
>>> +
>>> +           /* Add new bad chunk to list. */
>>> +           bc->chunk_offset = em->start;
>>> +           free_extent_map(em);
>>> +
>>> +           INIT_LIST_HEAD(&bc->list);
>>
>> nit: There is no need to initialize the list head of the entry itself.
>>
>>> +           list_add(&bc->list, &fs_info->bad_chunks);
>>> +
>>> +           write_sequnlock_irqrestore(&fs_info->bc_lock, flags);
>>> +   }
>>>  }
>>>  
>>>  /*
>>> @@ -1320,6 +1371,8 @@ static noinline void finish_rmw(struct btrfs_raid_bio 
>>> *rbio)
>>>             bio->bi_end_io = raid_write_end_io;
>>>             bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>>  
>>> +           btrfs_bio_wq_end_io(rbio->fs_info, bio, BTRFS_WQ_ENDIO_RAID56);
>>> +
>>>             submit_bio(bio);
>>>     }
>>>     return;
>>> @@ -2465,6 +2518,8 @@ static noinline void finish_parity_scrub(struct 
>>> btrfs_raid_bio *rbio,
>>>             bio->bi_end_io = raid_write_end_io;
>>>             bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>>  
>>> +           btrfs_bio_wq_end_io(rbio->fs_info, bio, BTRFS_WQ_ENDIO_RAID56);
>>> +
>>>             submit_bio(bio);
>>>     }
>>>     return;
>>> diff --git a/fs/btrfs/sysfs.c b/fs/btrfs/sysfs.c
>>> index a28bba8..0baaa33 100644
>>> --- a/fs/btrfs/sysfs.c
>>> +++ b/fs/btrfs/sysfs.c
>>> @@ -490,12 +490,38 @@ static ssize_t quota_override_store(struct kobject 
>>> *kobj,
>>>  
>>>  BTRFS_ATTR_RW(, quota_override, quota_override_show, quota_override_store);
>>>  
>>> +static ssize_t btrfs_bad_chunks_show(struct kobject *kobj,
>>> +                                struct kobj_attribute *a, char *buf)
>>> +{
>>> +   struct btrfs_fs_info *fs_info = to_fs_info(kobj);
>>> +   struct btrfs_bad_chunk *bc;
>>> +   int len = 0;
>>> +   unsigned int seq;
>>> +
>>> +   /* read lock please */
>>> +   do {
>>> +           seq = read_seqbegin(&fs_info->bc_lock);
>>> +           list_for_each_entry(bc, &fs_info->bad_chunks, list) {
>>> +                   len += snprintf(buf + len, PAGE_SIZE - len, "%llu\n",
>>> +                                   bc->chunk_offset);
>>> +                   /* chunk offset is u64 */
>>> +                   if (len >= PAGE_SIZE)
>>> +                           break;
>>> +           }
>>> +   } while (read_seqretry(&fs_info->bc_lock, seq));
>>> +
>>> +   return len;
>>> +}
>>> +
>>> +BTRFS_ATTR(, bad_chunks, btrfs_bad_chunks_show);
>>> +
>>>  static const struct attribute *btrfs_attrs[] = {
>>>     BTRFS_ATTR_PTR(, label),
>>>     BTRFS_ATTR_PTR(, nodesize),
>>>     BTRFS_ATTR_PTR(, sectorsize),
>>>     BTRFS_ATTR_PTR(, clone_alignment),
>>>     BTRFS_ATTR_PTR(, quota_override),
>>> +   BTRFS_ATTR_PTR(, bad_chunks),
>>>     NULL,
>>>  };
>>>  
>>> diff --git a/fs/btrfs/volumes.c b/fs/btrfs/volumes.c
>>> index a256842..d71f11a 100644
>>> --- a/fs/btrfs/volumes.c
>>> +++ b/fs/btrfs/volumes.c
>>> @@ -2803,8 +2803,8 @@ static int btrfs_del_sys_chunk(struct btrfs_fs_info 
>>> *fs_info, u64 chunk_offset)
>>>     return ret;
>>>  }
>>>  
>>> -static struct extent_map *get_chunk_map(struct btrfs_fs_info *fs_info,
>>> -                                   u64 logical, u64 length)
>>> +struct extent_map *get_chunk_map(struct btrfs_fs_info *fs_info,
>>> +                            u64 logical, u64 length)
>>
>> nit: Since you are exposing the function as an API I think this is a
>> good opportunity to add proper kernel doc for it.
>>
> 
> It has nothing to do with the patch's purpose, lets leave it to a
> seperate one.
> 
>>>  {
>>>     struct extent_map_tree *em_tree;
>>>     struct extent_map *em;
>>> @@ -2840,6 +2840,7 @@ int btrfs_remove_chunk(struct btrfs_trans_handle 
>>> *trans,
>>>     u64 dev_extent_len = 0;
>>>     int i, ret = 0;
>>>     struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
>>> +   struct btrfs_bad_chunk *bc;
>>>  
>>>     em = get_chunk_map(fs_info, chunk_offset, 1);
>>>     if (IS_ERR(em)) {
>>> @@ -2916,6 +2917,16 @@ int btrfs_remove_chunk(struct btrfs_trans_handle 
>>> *trans,
>>>     }
>>>  
>>>  out:
>>> +   write_seqlock_irq(&fs_info->bc_lock);
>>> +   list_for_each_entry(bc, &fs_info->bad_chunks, list) {
>>
>> Use list_for_each_entry_safe to make it more apparent you are going to
>> be removing from the list. The code as-is works since you are doing a
>> break after deleting element from the list but this is somewhat subtle.
> 
> To be honest, I don't see much difference.
> 
> I think the _safe version is to protect us from some race when others
> remove objects from list, and write lock is held so we're safe.

No, the _safe version uses the second argument (n) as the list iterator.
The non-safe version just uses 'pos', and in case you remove 'pos' from
the list AND continue iterating you will deref an invalid pointer. So
_safe is actually really necessary for correctness when you intend to
remove an entry from a list you are iterating, irrespective of any locks
you might have.
> 
>> Also it's not necessary to re-init the deleted entry since you are
>> directly freeing it.
>>
> 
> OK.
> 
> Thanks for the comments.
> 
> Thanks,
> 
> -liubo
>>> +           if (bc->chunk_offset == chunk_offset) {
>>> +                   list_del_init(&bc->list);
>>> +                   kfree(bc);
>>> +                   break;
>>> +           }
>>> +   }
>>> +   write_sequnlock_irq(&fs_info->bc_lock);
>>> +
>>>     /* once for us */
>>>     free_extent_map(em);
>>>     return ret;
>>> diff --git a/fs/btrfs/volumes.h b/fs/btrfs/volumes.h
>>> index ff15208..4e846ba 100644
>>> --- a/fs/btrfs/volumes.h
>>> +++ b/fs/btrfs/volumes.h
>>> @@ -396,6 +396,8 @@ static inline enum btrfs_map_op btrfs_op(struct bio 
>>> *bio)
>>>     }
>>>  }
>>>  
>>> +struct extent_map *get_chunk_map(struct btrfs_fs_info *fs_info,
>>> +                            u64 logical, u64 length);
>>>  int btrfs_account_dev_extents_size(struct btrfs_device *device, u64 start,
>>>                                u64 end, u64 *length);
>>>  void btrfs_get_bbio(struct btrfs_bio *bbio);
>>>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
> the body of a message to majord...@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to