On Mon, Jul 10, 2017 at 03:25:41PM +0800, Ming Lei wrote:
> On Mon, Jul 10, 2017 at 02:38:19PM +1000, NeilBrown wrote:
> > On Mon, Jul 10 2017, Ming Lei wrote:
> >
> > > On Mon, Jul 10, 2017 at 11:35:12AM +0800, Ming Lei wrote:
> > >> On Mon, Jul 10, 2017 at 7:09 AM, NeilBrown <[email protected]> wrote:
> > ...
> > >> >> +
> > >> >> + rp->idx = 0;
> > >> >
> > >> > This is the only place the ->idx is initialized, in r1buf_pool_alloc().
> > >> > The mempool alloc function is suppose to allocate memory, not
> > >> > initialize
> > >> > it.
> > >> >
> > >> > If the mempool_alloc() call cannot allocate memory it will use memory
> > >> > from the pool. If this memory has already been used, then it will no
> > >> > longer have the initialized value.
> > >> >
> > >> > In short: you need to initialise memory *after* calling
> > >> > mempool_alloc(), unless you ensure it is reset to the init values
> > >> > before
> > >> > calling mempool_free().
> > >> >
> > >> > https://bugzilla.kernel.org/show_bug.cgi?id=196307
> > >>
> > >> OK, thanks for posting it out.
> > >>
> > >> Another fix might be to reinitialize the variable(rp->idx = 0) in
> > >> r1buf_pool_free().
> > >> Or just set it as zero every time when it is used.
> > >>
> > >> But I don't understand why mempool_free() calls pool->free() at the end
> > >> of
> > >> this function, which may cause to run pool->free() on a new allocated
> > >> buf,
> > >> seems a bug in mempool?
> > >
> > > Looks I missed the 'return' in mempool_free(), so it is fine.
> > >
> > > How about the following fix?
> >
> > It looks like it would probably work, but it is rather unusual to
> > initialise something just before freeing it.
> >
> > Couldn't you just move the initialization to shortly after the
> > mempool_alloc() call. There looks like a good place that already loops
> > over all the bios....
>
> OK, follows the revised patch according to your suggestion.
> ---
>
> From 68f9936635b3dda13c87a6b6125ac543145bb940 Mon Sep 17 00:00:00 2001
> From: Ming Lei <[email protected]>
> Date: Mon, 10 Jul 2017 15:16:16 +0800
> Subject: [PATCH] MD: move initialization of resync pages' index out of mempool
> allocator
>
> mempool_alloc() is only responsible for allocation, not for initialization,
> so we need to move the initialization of resync pages's index out of the
> allocator function.
>
> Reported-by: NeilBrown <[email protected]>
> Fixes: f0250618361d(md: raid10: don't use bio's vec table to manage resync
> pages)
> Fixes: 98d30c5812c3(md: raid1: don't use bio's vec table to manage resync
> pages)
> Signed-off-by: Ming Lei <[email protected]>
> ---
> drivers/md/raid1.c | 4 +++-
> drivers/md/raid10.c | 6 +++++-
> 2 files changed, 8 insertions(+), 2 deletions(-)
>
> diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c
> index e1a7e3d4c5e4..26f5efba0504 100644
> --- a/drivers/md/raid1.c
> +++ b/drivers/md/raid1.c
> @@ -170,7 +170,6 @@ static void * r1buf_pool_alloc(gfp_t gfp_flags, void
> *data)
> resync_get_all_pages(rp);
> }
>
> - rp->idx = 0;
> rp->raid_bio = r1_bio;
> bio->bi_private = rp;
> }
> @@ -2698,6 +2697,9 @@ static sector_t raid1_sync_request(struct mddev *mddev,
> sector_t sector_nr,
> struct md_rdev *rdev;
> bio = r1_bio->bios[i];
>
> + /* This initialization should follow mempool_alloc() */
> + get_resync_pages(bio)->idx = 0;
> +
This is fragile and hard to maintain. Can we add a wrap for the
allocation/init?
Thanks,
Shaohua
> rdev = rcu_dereference(conf->mirrors[i].rdev);
> if (rdev == NULL ||
> test_bit(Faulty, &rdev->flags)) {
> diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c
> index 797ed60abd5e..5ebcb7487284 100644
> --- a/drivers/md/raid10.c
> +++ b/drivers/md/raid10.c
> @@ -221,7 +221,6 @@ static void * r10buf_pool_alloc(gfp_t gfp_flags, void
> *data)
> resync_get_all_pages(rp);
> }
>
> - rp->idx = 0;
> rp->raid_bio = r10_bio;
> bio->bi_private = rp;
> if (rbio) {
> @@ -3095,6 +3094,7 @@ static sector_t raid10_sync_request(struct mddev
> *mddev, sector_t sector_nr,
> bio = r10_bio->devs[0].bio;
> bio->bi_next = biolist;
> biolist = bio;
> + get_resync_pages(bio)->idx = 0;
> bio->bi_end_io = end_sync_read;
> bio_set_op_attrs(bio, REQ_OP_READ, 0);
> if (test_bit(FailFast, &rdev->flags))
> @@ -3120,6 +3120,7 @@ static sector_t raid10_sync_request(struct mddev
> *mddev, sector_t sector_nr,
> bio = r10_bio->devs[1].bio;
> bio->bi_next = biolist;
> biolist = bio;
> + get_resync_pages(bio)->idx = 0;
> bio->bi_end_io = end_sync_write;
> bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
> bio->bi_iter.bi_sector = to_addr
> @@ -3146,6 +3147,7 @@ static sector_t raid10_sync_request(struct mddev
> *mddev, sector_t sector_nr,
> break;
> bio->bi_next = biolist;
> biolist = bio;
> + get_resync_pages(bio)->idx = 0;
> bio->bi_end_io = end_sync_write;
> bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
> bio->bi_iter.bi_sector = to_addr +
> @@ -3291,6 +3293,7 @@ static sector_t raid10_sync_request(struct mddev
> *mddev, sector_t sector_nr,
> atomic_inc(&r10_bio->remaining);
> bio->bi_next = biolist;
> biolist = bio;
> + get_resync_pages(bio)->idx = 0;
> bio->bi_end_io = end_sync_read;
> bio_set_op_attrs(bio, REQ_OP_READ, 0);
> if (test_bit(FailFast, &conf->mirrors[d].rdev->flags))
> @@ -3314,6 +3317,7 @@ static sector_t raid10_sync_request(struct mddev
> *mddev, sector_t sector_nr,
> sector = r10_bio->devs[i].addr;
> bio->bi_next = biolist;
> biolist = bio;
> + get_resync_pages(bio)->idx = 0;
> bio->bi_end_io = end_sync_write;
> bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
> if (test_bit(FailFast, &conf->mirrors[d].rdev->flags))
> --
> 2.9.4
>
>
>
> --
> Ming