On Fri, Jun 19, 2015 at 03:15:01PM +0530, Chandan Rajendra wrote:
> On Friday 19 Jun 2015 12:45:37 Liu Bo wrote:
> > On Mon, Jun 01, 2015 at 08:52:36PM +0530, Chandan Rajendra wrote:
> > > For the subpagesize-blocksize scenario, a page can contain multiple
> > > blocks. In such cases, this patch handles reading data from files.
> > > 
> > > To track the status of individual blocks of a page, this patch makes use
> > > of a bitmap pointed to by page->private.
> > 
> > Start going through the patchset, it's not easy though.
> > 
> > Several comments are following.
> 
> Thanks for the review comments Liu.
> 
> > > +static int modify_page_blks_state(struct page *page,
> > > +                         unsigned long blk_states,
> > > +                         u64 start, u64 end, int set)
> > > +{
> > > + struct inode *inode = page->mapping->host;
> > > + unsigned long *bitmap;
> > > + unsigned long state;
> > > + u64 nr_blks;
> > > + u64 blk;
> > > +
> > > + BUG_ON(!PagePrivate(page));
> > > +
> > > + bitmap = ((struct btrfs_page_private *)page->private)->bstate;
> > > +
> > > + blk = (start & (PAGE_CACHE_SIZE - 1)) >> inode->i_blkbits;
> > > + nr_blks = (end - start + 1) >> inode->i_blkbits;
> > > +
> > > + while (nr_blks--) {
> > > +         state = find_next_bit(&blk_states, BLK_NR_STATE, 0);
> > 
> > Looks like we don't need to do find_next_bit for every block.
> 
> Yes, I agree. The find_next_bit() invocation in the outer loop can be moved
> outside the loop.
> > 
> > > +
> > > +         while (state < BLK_NR_STATE) {
> > > +                 if (set)
> > > +                         set_bit((blk * BLK_NR_STATE) + state, bitmap);
> > > +                 else
> > > +                         clear_bit((blk * BLK_NR_STATE) + state, 
> bitmap);
> > > +
> > > +                 state = find_next_bit(&blk_states, BLK_NR_STATE,
> > > +                                 state + 1);
> > > +         }
> > > +
> > > +         ++blk;
> > > + }
> > > +
> > > + return 0;
> > > +}
> > > +
> > > 
> > >  /*
> > >  
> > >   * after a readpage IO is done, we need to:
> > >   * clear the uptodate bits on error
> > > 
> > > @@ -2548,14 +2628,16 @@ static void end_bio_extent_readpage(struct bio
> > > *bio, int err)> 
> > >   struct bio_vec *bvec;
> > >   int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
> > >   struct btrfs_io_bio *io_bio = btrfs_io_bio(bio);
> > > 
> > > + struct extent_state *cached = NULL;
> > > + struct btrfs_page_private *pg_private;
> > > 
> > >   struct extent_io_tree *tree;
> > > 
> > > + unsigned long flags;
> > > 
> > >   u64 offset = 0;
> > >   u64 start;
> > >   u64 end;
> > > 
> > > - u64 len;
> > > - u64 extent_start = 0;
> > > - u64 extent_len = 0;
> > > + int nr_sectors;
> > > 
> > >   int mirror;
> > > 
> > > + int unlock;
> > > 
> > >   int ret;
> > >   int i;
> > > 
> > > @@ -2565,54 +2647,31 @@ static void end_bio_extent_readpage(struct bio
> > > *bio, int err)> 
> > >   bio_for_each_segment_all(bvec, bio, i) {
> > >   
> > >           struct page *page = bvec->bv_page;
> > >           struct inode *inode = page->mapping->host;
> > > 
> > > +         struct btrfs_root *root = BTRFS_I(inode)->root;
> > > 
> > >           pr_debug("end_bio_extent_readpage: bi_sector=%llu, err=%d, "
> > >           
> > >                    "mirror=%u\n", (u64)bio->bi_iter.bi_sector, err,
> > >                    io_bio->mirror_num);
> > >           
> > >           tree = &BTRFS_I(inode)->io_tree;
> > > 
> > > -         /* We always issue full-page reads, but if some block
> > > -          * in a page fails to read, blk_update_request() will
> > > -          * advance bv_offset and adjust bv_len to compensate.
> > > -          * Print a warning for nonzero offsets, and an error
> > > -          * if they don't add up to a full page.  */
> > > -         if (bvec->bv_offset || bvec->bv_len != PAGE_CACHE_SIZE) {
> > > -                 if (bvec->bv_offset + bvec->bv_len != PAGE_CACHE_SIZE)
> > > -                         btrfs_err(BTRFS_I(page->mapping->host)->root-
> >fs_info,
> > > -                            "partial page read in btrfs with offset %u 
> and length %u",
> > > -                                 bvec->bv_offset, bvec->bv_len);
> > > -                 else
> > > -                         btrfs_info(BTRFS_I(page->mapping->host)->root-
> >fs_info,
> > > -                            "incomplete page read in btrfs with offset 
> %u and "
> > > -                            "length %u",
> > > -                                 bvec->bv_offset, bvec->bv_len);
> > > -         }
> > > -
> > > -         start = page_offset(page);
> > > -         end = start + bvec->bv_offset + bvec->bv_len - 1;
> > > -         len = bvec->bv_len;
> > > -
> > > +         start = page_offset(page) + bvec->bv_offset;
> > > +         end = start + bvec->bv_len - 1;
> > > +         nr_sectors = bvec->bv_len >> inode->i_sb->s_blocksize_bits;
> > > 
> > >           mirror = io_bio->mirror_num;
> > > 
> > > -         if (likely(uptodate && tree->ops &&
> > > -                    tree->ops->readpage_end_io_hook)) {
> > > +
> > > +next_block:
> > > +         if (likely(uptodate)) {
> > 
> > Any reason of killing (tree->ops && tree->ops->readpage_end_io_hook)?
> 
> In subpagesize-blocksize scenario, For extent buffers we need the ability to
> read just a single extent buffer rather than reading the complete contents of
> the page containing the extent buffer. Similarly in the corresponding endio
> function we need to verify a single extent buffer rather than the contents of
> the full page.  Hence I ended up removing btree_readpage_end_io_hook() and
> btree_io_failed_hook() functions and had verfication functions being
> invoked directly by the endio function.
> 
> So since data "read page code" was the only one left to have
> extent_io_tree->ops->readpage_end_io_hook set, I removed the code to check for
> its existance. Now i realize that it is not the right thing to do. I will
> restore back the condition check to its original state.
> 
> > 
> > >                   ret = tree->ops->readpage_end_io_hook(io_bio, offset,
> > > 
> > > -                                                       page, start, 
> end,
> > > -                                                       mirror);
> > > +                                                 page, start,
> > > +                                                 start + root-
> >sectorsize - 1,
> > > +                                                 mirror);
> > > 
> > >                   if (ret)
> > >                   
> > >                           uptodate = 0;
> > >                   
> > >                   else
> > >                   
> > >                           clean_io_failure(inode, start, page, 0);
> > >           
> > >           }
> > > 
> > > -         if (likely(uptodate))
> > > -                 goto readpage_ok;
> > > -
> > > -         if (tree->ops && tree->ops->readpage_io_failed_hook) {
> > > -                 ret = tree->ops->readpage_io_failed_hook(page, 
> mirror);
> > > -                 if (!ret && !err &&
> > > -                     test_bit(BIO_UPTODATE, &bio->bi_flags))
> > > -                         uptodate = 1;
> > > -         } else {
> > > +         if (!uptodate) {
> > > 
> > >                   /*
> > >                   
> > >                    * The generic bio_readpage_error handles errors the
> > >                    * following way: If possible, new read requests are
> > > 
> > > @@ -2623,61 +2682,63 @@ static void end_bio_extent_readpage(struct bio
> > > *bio, int err)> 
> > >                    * can't handle the error it will return -EIO and we
> > >                    * remain responsible for that page.
> > >                    */
> > > 
> > > -                 ret = bio_readpage_error(bio, offset, page, start, 
> end,
> > > -                                          mirror);
> > > +                 ret = bio_readpage_error(bio, offset, page,
> > > +                                         start, start + root-
> >sectorsize - 1,
> > > +                                         mirror);
> > > 
> > >                   if (ret == 0) {
> > > 
> > > -                         uptodate =
> > > -                                 test_bit(BIO_UPTODATE, &bio-
> >bi_flags);
> > > +                         uptodate = test_bit(BIO_UPTODATE, &bio-
> >bi_flags);
> > > 
> > >                           if (err)
> > >                           
> > >                                   uptodate = 0;
> > > 
> > > -                         offset += len;
> > > -                         continue;
> > > +                         offset += root->sectorsize;
> > > +                         if (--nr_sectors) {
> > > +                                 start += root->sectorsize;
> > > +                                 goto next_block;
> > > +                         } else {
> > > +                                 continue;
> > > +                         }
> > > 
> > >                   }
> > >           
> > >           }
> > > 
> > > -readpage_ok:
> > > -         if (likely(uptodate)) {
> > > -                 loff_t i_size = i_size_read(inode);
> > > -                 pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT;
> > > -                 unsigned off;
> > > -
> > > -                 /* Zero out the end if this page straddles i_size */
> > > -                 off = i_size & (PAGE_CACHE_SIZE-1);
> > > -                 if (page->index == end_index && off)
> > > -                         zero_user_segment(page, off, PAGE_CACHE_SIZE);
> > > -                 SetPageUptodate(page);
> > > +
> > > +         if (uptodate) {
> > > +                 set_page_blks_state(page, 1 << BLK_STATE_UPTODATE, 
> start,
> > > +                                 start + root->sectorsize - 1);
> > > +                 check_page_uptodate(page);
> > > 
> > >           } else {
> > >           
> > >                   ClearPageUptodate(page);
> > >                   SetPageError(page);
> > >           
> > >           }
> > > 
> > > -         unlock_page(page);
> > > -         offset += len;
> > > -
> > > -         if (unlikely(!uptodate)) {
> > > -                 if (extent_len) {
> > > -                         endio_readpage_release_extent(tree,
> > > -                                                       extent_start,
> > > -                                                       extent_len, 1);
> > > -                         extent_start = 0;
> > > -                         extent_len = 0;
> > > -                 }
> > > -                 endio_readpage_release_extent(tree, start,
> > > -                                               end - start + 1, 0);
> > > -         } else if (!extent_len) {
> > > -                 extent_start = start;
> > > -                 extent_len = end + 1 - start;
> > > -         } else if (extent_start + extent_len == start) {
> > > -                 extent_len += end + 1 - start;
> > > -         } else {
> > > -                 endio_readpage_release_extent(tree, extent_start,
> > > -                                               extent_len, uptodate);
> > > -                 extent_start = start;
> > > -                 extent_len = end + 1 - start;
> > > +
> > > +         offset += root->sectorsize;
> > > +
> > > +         if (--nr_sectors) {
> > > +                 clear_page_blks_state(page, 1 << BLK_STATE_IO,
> > > +                                 start, start + root->sectorsize - 1);
> > 
> > private->io_lock is not acquired here but not in below.
> > 
> > IIUC, this can be protected by EXTENT_LOCKED.
> >
> 
> private->io_lock plays the same role as BH_Uptodate_Lock (see
> end_buffer_async_read()) i.e. without the io_lock we may end up in the
> following situation,
> 
> NOTE: Assume 64k page size and 4k block size. Also assume that the first 12
> blocks of the page are contiguous while the next 4 blocks are contiguous. When
> reading the page we end up submitting two "logical address space" bios. So
> end_bio_extent_readpage function is invoked twice (once for each bio).
> 
> |-------------------------+-------------------------+-------------|
> | Task A                  | Task B                  | Task C      |
> |-------------------------+-------------------------+-------------|
> | end_bio_extent_readpage |                         |             |
> | process block 0         |                         |             |
> | - clear BLK_STATE_IO    |                         |             |
> | - page_read_complete    |                         |             |
> | process block 1         |                         |             |
> | ...                     |                         |             |
> | ...                     |                         |             |
> | ...                     | end_bio_extent_readpage |             |
> | ...                     | process block 0         |             |
> | ...                     | - clear BLK_STATE_IO    |             |
> | ...                     | - page_read_complete    |             |
> | ...                     | process block 1         |             |
> | ...                     | ...                     |             |
> | process block 11        | process block 3         |             |
> | - clear BLK_STATE_IO    | - clear BLK_STATE_IO    |             |
> | - page_read_complete    | - page_read_complete    |             |
> |   - returns true        |   - returns true        |             |
> |   - unlock_page()       |                         |             |
> |                         |                         | lock_page() |
> |                         |   - unlock_page()       |             |
> |-------------------------+-------------------------+-------------|
> 
> So we end up incorrectly unlocking the page twice and "Task C" ends up working
> on an unlocked page. So private->io_lock makes sure that only one of the tasks
> gets "true" as the return value when page_read_complete() is invoked. As an
> optimization the patch gets the io_lock only when nr_sectors counter reaches
> the value 0 (i.e. when the last block of the bio_vec is being processed).
> Please let me know if my analysis was incorrect.

Thanks for the nice explanation, it looks reasonable to me.

Thanks,

-liubo

> 
> Also, I noticed that page_read_complete() and page_write_complete() can be
> replaced by just one function i.e. page_io_complete().
> 
> 
> -- 
> chandan
> 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
> the body of a message to [email protected]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to