On Fri, Jun 19, 2015 at 03:15:01PM +0530, Chandan Rajendra wrote:
> On Friday 19 Jun 2015 12:45:37 Liu Bo wrote:
> > On Mon, Jun 01, 2015 at 08:52:36PM +0530, Chandan Rajendra wrote:
> > > For the subpagesize-blocksize scenario, a page can contain multiple
> > > blocks. In such cases, this patch handles reading data from files.
> > >
> > > To track the status of individual blocks of a page, this patch makes use
> > > of a bitmap pointed to by page->private.
> >
> > Start going through the patchset, it's not easy though.
> >
> > Several comments are following.
>
> Thanks for the review comments Liu.
>
> > > +static int modify_page_blks_state(struct page *page,
> > > + unsigned long blk_states,
> > > + u64 start, u64 end, int set)
> > > +{
> > > + struct inode *inode = page->mapping->host;
> > > + unsigned long *bitmap;
> > > + unsigned long state;
> > > + u64 nr_blks;
> > > + u64 blk;
> > > +
> > > + BUG_ON(!PagePrivate(page));
> > > +
> > > + bitmap = ((struct btrfs_page_private *)page->private)->bstate;
> > > +
> > > + blk = (start & (PAGE_CACHE_SIZE - 1)) >> inode->i_blkbits;
> > > + nr_blks = (end - start + 1) >> inode->i_blkbits;
> > > +
> > > + while (nr_blks--) {
> > > + state = find_next_bit(&blk_states, BLK_NR_STATE, 0);
> >
> > Looks like we don't need to do find_next_bit for every block.
>
> Yes, I agree. The find_next_bit() invocation in the outer loop can be moved
> outside the loop.
> >
> > > +
> > > + while (state < BLK_NR_STATE) {
> > > + if (set)
> > > + set_bit((blk * BLK_NR_STATE) + state, bitmap);
> > > + else
> > > + clear_bit((blk * BLK_NR_STATE) + state,
> bitmap);
> > > +
> > > + state = find_next_bit(&blk_states, BLK_NR_STATE,
> > > + state + 1);
> > > + }
> > > +
> > > + ++blk;
> > > + }
> > > +
> > > + return 0;
> > > +}
> > > +
> > >
> > > /*
> > >
> > > * after a readpage IO is done, we need to:
> > > * clear the uptodate bits on error
> > >
> > > @@ -2548,14 +2628,16 @@ static void end_bio_extent_readpage(struct bio
> > > *bio, int err)>
> > > struct bio_vec *bvec;
> > > int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags);
> > > struct btrfs_io_bio *io_bio = btrfs_io_bio(bio);
> > >
> > > + struct extent_state *cached = NULL;
> > > + struct btrfs_page_private *pg_private;
> > >
> > > struct extent_io_tree *tree;
> > >
> > > + unsigned long flags;
> > >
> > > u64 offset = 0;
> > > u64 start;
> > > u64 end;
> > >
> > > - u64 len;
> > > - u64 extent_start = 0;
> > > - u64 extent_len = 0;
> > > + int nr_sectors;
> > >
> > > int mirror;
> > >
> > > + int unlock;
> > >
> > > int ret;
> > > int i;
> > >
> > > @@ -2565,54 +2647,31 @@ static void end_bio_extent_readpage(struct bio
> > > *bio, int err)>
> > > bio_for_each_segment_all(bvec, bio, i) {
> > >
> > > struct page *page = bvec->bv_page;
> > > struct inode *inode = page->mapping->host;
> > >
> > > + struct btrfs_root *root = BTRFS_I(inode)->root;
> > >
> > > pr_debug("end_bio_extent_readpage: bi_sector=%llu, err=%d, "
> > >
> > > "mirror=%u\n", (u64)bio->bi_iter.bi_sector, err,
> > > io_bio->mirror_num);
> > >
> > > tree = &BTRFS_I(inode)->io_tree;
> > >
> > > - /* We always issue full-page reads, but if some block
> > > - * in a page fails to read, blk_update_request() will
> > > - * advance bv_offset and adjust bv_len to compensate.
> > > - * Print a warning for nonzero offsets, and an error
> > > - * if they don't add up to a full page. */
> > > - if (bvec->bv_offset || bvec->bv_len != PAGE_CACHE_SIZE) {
> > > - if (bvec->bv_offset + bvec->bv_len != PAGE_CACHE_SIZE)
> > > - btrfs_err(BTRFS_I(page->mapping->host)->root-
> >fs_info,
> > > - "partial page read in btrfs with offset %u
> and length %u",
> > > - bvec->bv_offset, bvec->bv_len);
> > > - else
> > > - btrfs_info(BTRFS_I(page->mapping->host)->root-
> >fs_info,
> > > - "incomplete page read in btrfs with offset
> %u and "
> > > - "length %u",
> > > - bvec->bv_offset, bvec->bv_len);
> > > - }
> > > -
> > > - start = page_offset(page);
> > > - end = start + bvec->bv_offset + bvec->bv_len - 1;
> > > - len = bvec->bv_len;
> > > -
> > > + start = page_offset(page) + bvec->bv_offset;
> > > + end = start + bvec->bv_len - 1;
> > > + nr_sectors = bvec->bv_len >> inode->i_sb->s_blocksize_bits;
> > >
> > > mirror = io_bio->mirror_num;
> > >
> > > - if (likely(uptodate && tree->ops &&
> > > - tree->ops->readpage_end_io_hook)) {
> > > +
> > > +next_block:
> > > + if (likely(uptodate)) {
> >
> > Any reason of killing (tree->ops && tree->ops->readpage_end_io_hook)?
>
> In subpagesize-blocksize scenario, For extent buffers we need the ability to
> read just a single extent buffer rather than reading the complete contents of
> the page containing the extent buffer. Similarly in the corresponding endio
> function we need to verify a single extent buffer rather than the contents of
> the full page. Hence I ended up removing btree_readpage_end_io_hook() and
> btree_io_failed_hook() functions and had verfication functions being
> invoked directly by the endio function.
>
> So since data "read page code" was the only one left to have
> extent_io_tree->ops->readpage_end_io_hook set, I removed the code to check for
> its existance. Now i realize that it is not the right thing to do. I will
> restore back the condition check to its original state.
>
> >
> > > ret = tree->ops->readpage_end_io_hook(io_bio, offset,
> > >
> > > - page, start,
> end,
> > > - mirror);
> > > + page, start,
> > > + start + root-
> >sectorsize - 1,
> > > + mirror);
> > >
> > > if (ret)
> > >
> > > uptodate = 0;
> > >
> > > else
> > >
> > > clean_io_failure(inode, start, page, 0);
> > >
> > > }
> > >
> > > - if (likely(uptodate))
> > > - goto readpage_ok;
> > > -
> > > - if (tree->ops && tree->ops->readpage_io_failed_hook) {
> > > - ret = tree->ops->readpage_io_failed_hook(page,
> mirror);
> > > - if (!ret && !err &&
> > > - test_bit(BIO_UPTODATE, &bio->bi_flags))
> > > - uptodate = 1;
> > > - } else {
> > > + if (!uptodate) {
> > >
> > > /*
> > >
> > > * The generic bio_readpage_error handles errors the
> > > * following way: If possible, new read requests are
> > >
> > > @@ -2623,61 +2682,63 @@ static void end_bio_extent_readpage(struct bio
> > > *bio, int err)>
> > > * can't handle the error it will return -EIO and we
> > > * remain responsible for that page.
> > > */
> > >
> > > - ret = bio_readpage_error(bio, offset, page, start,
> end,
> > > - mirror);
> > > + ret = bio_readpage_error(bio, offset, page,
> > > + start, start + root-
> >sectorsize - 1,
> > > + mirror);
> > >
> > > if (ret == 0) {
> > >
> > > - uptodate =
> > > - test_bit(BIO_UPTODATE, &bio-
> >bi_flags);
> > > + uptodate = test_bit(BIO_UPTODATE, &bio-
> >bi_flags);
> > >
> > > if (err)
> > >
> > > uptodate = 0;
> > >
> > > - offset += len;
> > > - continue;
> > > + offset += root->sectorsize;
> > > + if (--nr_sectors) {
> > > + start += root->sectorsize;
> > > + goto next_block;
> > > + } else {
> > > + continue;
> > > + }
> > >
> > > }
> > >
> > > }
> > >
> > > -readpage_ok:
> > > - if (likely(uptodate)) {
> > > - loff_t i_size = i_size_read(inode);
> > > - pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT;
> > > - unsigned off;
> > > -
> > > - /* Zero out the end if this page straddles i_size */
> > > - off = i_size & (PAGE_CACHE_SIZE-1);
> > > - if (page->index == end_index && off)
> > > - zero_user_segment(page, off, PAGE_CACHE_SIZE);
> > > - SetPageUptodate(page);
> > > +
> > > + if (uptodate) {
> > > + set_page_blks_state(page, 1 << BLK_STATE_UPTODATE,
> start,
> > > + start + root->sectorsize - 1);
> > > + check_page_uptodate(page);
> > >
> > > } else {
> > >
> > > ClearPageUptodate(page);
> > > SetPageError(page);
> > >
> > > }
> > >
> > > - unlock_page(page);
> > > - offset += len;
> > > -
> > > - if (unlikely(!uptodate)) {
> > > - if (extent_len) {
> > > - endio_readpage_release_extent(tree,
> > > - extent_start,
> > > - extent_len, 1);
> > > - extent_start = 0;
> > > - extent_len = 0;
> > > - }
> > > - endio_readpage_release_extent(tree, start,
> > > - end - start + 1, 0);
> > > - } else if (!extent_len) {
> > > - extent_start = start;
> > > - extent_len = end + 1 - start;
> > > - } else if (extent_start + extent_len == start) {
> > > - extent_len += end + 1 - start;
> > > - } else {
> > > - endio_readpage_release_extent(tree, extent_start,
> > > - extent_len, uptodate);
> > > - extent_start = start;
> > > - extent_len = end + 1 - start;
> > > +
> > > + offset += root->sectorsize;
> > > +
> > > + if (--nr_sectors) {
> > > + clear_page_blks_state(page, 1 << BLK_STATE_IO,
> > > + start, start + root->sectorsize - 1);
> >
> > private->io_lock is not acquired here but not in below.
> >
> > IIUC, this can be protected by EXTENT_LOCKED.
> >
>
> private->io_lock plays the same role as BH_Uptodate_Lock (see
> end_buffer_async_read()) i.e. without the io_lock we may end up in the
> following situation,
>
> NOTE: Assume 64k page size and 4k block size. Also assume that the first 12
> blocks of the page are contiguous while the next 4 blocks are contiguous. When
> reading the page we end up submitting two "logical address space" bios. So
> end_bio_extent_readpage function is invoked twice (once for each bio).
>
> |-------------------------+-------------------------+-------------|
> | Task A | Task B | Task C |
> |-------------------------+-------------------------+-------------|
> | end_bio_extent_readpage | | |
> | process block 0 | | |
> | - clear BLK_STATE_IO | | |
> | - page_read_complete | | |
> | process block 1 | | |
> | ... | | |
> | ... | | |
> | ... | end_bio_extent_readpage | |
> | ... | process block 0 | |
> | ... | - clear BLK_STATE_IO | |
> | ... | - page_read_complete | |
> | ... | process block 1 | |
> | ... | ... | |
> | process block 11 | process block 3 | |
> | - clear BLK_STATE_IO | - clear BLK_STATE_IO | |
> | - page_read_complete | - page_read_complete | |
> | - returns true | - returns true | |
> | - unlock_page() | | |
> | | | lock_page() |
> | | - unlock_page() | |
> |-------------------------+-------------------------+-------------|
>
> So we end up incorrectly unlocking the page twice and "Task C" ends up working
> on an unlocked page. So private->io_lock makes sure that only one of the tasks
> gets "true" as the return value when page_read_complete() is invoked. As an
> optimization the patch gets the io_lock only when nr_sectors counter reaches
> the value 0 (i.e. when the last block of the bio_vec is being processed).
> Please let me know if my analysis was incorrect.
Thanks for the nice explanation, it looks reasonable to me.
Thanks,
-liubo
>
> Also, I noticed that page_read_complete() and page_write_complete() can be
> replaced by just one function i.e. page_io_complete().
>
>
> --
> chandan
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
> the body of a message to [email protected]
> More majordomo info at http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html