Message ID | 20210126083402.142577-18-wqu@suse.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | btrfs: add read-only support for subpage sector size | expand |
On 1/26/21 3:34 AM, Qu Wenruo wrote: > In btrfs data page read path, the page status update are handled in two > different locations: > > btrfs_do_read_page() > { > while (cur <= end) { > /* No need to read from disk */ > if (HOLE/PREALLOC/INLINE){ > memset(); > set_extent_uptodate(); > continue; > } > /* Read from disk */ > ret = submit_extent_page(end_bio_extent_readpage); > } > > end_bio_extent_readpage() > { > endio_readpage_uptodate_page_status(); > } > > This is fine for sectorsize == PAGE_SIZE case, as for above loop we > should only hit one branch and then exit. > > But for subpage, there are more works to be done in page status update: > - Page Unlock condition > Unlike regular page size == sectorsize case, we can no longer just > unlock a page without a brain. > Only the last reader of the page can unlock the page. > This means, we can unlock the page either in the while() loop, or in > the endio function. > > - Page uptodate condition > Since we have multiple sectors to read for a page, we can only mark > the full page uptodate if all sectors are uptodate. > > To handle both subpage and regular cases, introduce a pair of functions > to help handling page status update: > > - being_page_read() > For regular case, it does nothing. > For subpage case, it update the reader counters so that later > end_page_read() can know who is the last one to unlock the page. > > - end_page_read() > This is just endio_readpage_uptodate_page_status() renamed. > The original name is a little too long and too specific for endio. > > The only new trick added is the condition for page unlock. > Now for subage data, we unlock the page if we're the last reader. > > This does not only provide the basis for subpage data read, but also > hide the special handling of page read from the main read loop. > > Signed-off-by: Qu Wenruo <wqu@suse.com> > Signed-off-by: David Sterba <dsterba@suse.com> > --- > fs/btrfs/extent_io.c | 38 ++++++++++++++++++++---------- > fs/btrfs/subpage.c | 56 ++++++++++++++++++++++++++++++++++---------- > fs/btrfs/subpage.h | 8 +++++++ > 3 files changed, 78 insertions(+), 24 deletions(-) > > diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c > index eeee3213daaa..7fc2c62d4eb9 100644 > --- a/fs/btrfs/extent_io.c > +++ b/fs/btrfs/extent_io.c > @@ -2839,8 +2839,17 @@ static void endio_readpage_release_extent(struct processed_extent *processed, > processed->uptodate = uptodate; > } > > -static void endio_readpage_update_page_status(struct page *page, bool uptodate, > - u64 start, u32 len) > +static void begin_data_page_read(struct btrfs_fs_info *fs_info, struct page *page) > +{ > + ASSERT(PageLocked(page)); > + if (fs_info->sectorsize == PAGE_SIZE) > + return; > + > + ASSERT(PagePrivate(page)); > + btrfs_subpage_start_reader(fs_info, page, page_offset(page), PAGE_SIZE); > +} > + > +static void end_page_read(struct page *page, bool uptodate, u64 start, u32 len) > { > struct btrfs_fs_info *fs_info = btrfs_sb(page->mapping->host->i_sb); > > @@ -2856,7 +2865,12 @@ static void endio_readpage_update_page_status(struct page *page, bool uptodate, > > if (fs_info->sectorsize == PAGE_SIZE) > unlock_page(page); > - /* Subpage locking will be handled in later patches */ > + else if (is_data_inode(page->mapping->host)) > + /* > + * For subpage data, unlock the page if we're the last reader. > + * For subpage metadata, page lock is not utilized for read. > + */ > + btrfs_subpage_end_reader(fs_info, page, start, len); > } > > /* > @@ -2993,7 +3007,7 @@ static void end_bio_extent_readpage(struct bio *bio) > bio_offset += len; > > /* Update page status and unlock */ > - endio_readpage_update_page_status(page, uptodate, start, len); > + end_page_read(page, uptodate, start, len); > endio_readpage_release_extent(&processed, BTRFS_I(inode), > start, end, uptodate); > } > @@ -3263,6 +3277,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > unsigned int read_flags, u64 *prev_em_start) > { > struct inode *inode = page->mapping->host; > + struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb); > u64 start = page_offset(page); > const u64 end = start + PAGE_SIZE - 1; > u64 cur = start; > @@ -3306,6 +3321,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > kunmap_atomic(userpage); > } > } You have two error cases above this ret = set_page_extent_mapped(page); if (ret < 0) { unlock_extent(tree, start, end); SetPageError(page); goto out; } and if (!PageUptodate(page)) { if (cleancache_get_page(page) == 0) { BUG_ON(blocksize != PAGE_SIZE); unlock_extent(tree, start, end); goto out; } } which will now leave the page locked when it errors out. Not to mention I'm pretty sure you want to use btrfs_page_set_error() instead of SetPageError() in that first case. > + begin_data_page_read(fs_info, page); > while (cur <= end) { > bool force_bio_submit = false; > u64 disk_bytenr; > @@ -3323,13 +3339,14 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > &cached, GFP_NOFS); > unlock_extent_cached(tree, cur, > cur + iosize - 1, &cached); > + end_page_read(page, true, cur, iosize); > break; > } > em = __get_extent_map(inode, page, pg_offset, cur, > end - cur + 1, em_cached); > if (IS_ERR_OR_NULL(em)) { > - SetPageError(page); > unlock_extent(tree, cur, end); > + end_page_read(page, false, cur, end + 1 - cur); > break; > } > extent_offset = cur - em->start; > @@ -3412,6 +3429,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > &cached, GFP_NOFS); > unlock_extent_cached(tree, cur, > cur + iosize - 1, &cached); > + end_page_read(page, true, cur, iosize); > cur = cur + iosize; > pg_offset += iosize; > continue; > @@ -3421,6 +3439,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > EXTENT_UPTODATE, 1, NULL)) { > check_page_uptodate(tree, page); > unlock_extent(tree, cur, cur + iosize - 1); > + end_page_read(page, true, cur, iosize); > cur = cur + iosize; > pg_offset += iosize; > continue; > @@ -3429,8 +3448,8 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > * to date. Error out > */ > if (block_start == EXTENT_MAP_INLINE) { > - SetPageError(page); > unlock_extent(tree, cur, cur + iosize - 1); > + end_page_read(page, false, cur, iosize); > cur = cur + iosize; > pg_offset += iosize; > continue; > @@ -3447,19 +3466,14 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > nr++; > *bio_flags = this_bio_flag; > } else { > - SetPageError(page); > unlock_extent(tree, cur, cur + iosize - 1); > + end_page_read(page, false, cur, iosize); > goto out; > } > cur = cur + iosize; > pg_offset += iosize; > } > out: > - if (!nr) { > - if (!PageError(page)) > - SetPageUptodate(page); > - unlock_page(page); > - } We can just delete out: here and either return on error or break from the main loop. Thanks, Josef
On Wed, Jan 27, 2021 at 12:13:27PM -0500, Josef Bacik wrote: > On 1/26/21 3:34 AM, Qu Wenruo wrote: > > @@ -3263,6 +3277,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > > unsigned int read_flags, u64 *prev_em_start) > > { > > struct inode *inode = page->mapping->host; > > + struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb); > > u64 start = page_offset(page); > > const u64 end = start + PAGE_SIZE - 1; > > u64 cur = start; > > @@ -3306,6 +3321,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, > > kunmap_atomic(userpage); > > } > > } > > You have two error cases above this > > ret = set_page_extent_mapped(page); > if (ret < 0) { > unlock_extent(tree, start, end); > SetPageError(page); > goto out; > } > > and > > if (!PageUptodate(page)) { > if (cleancache_get_page(page) == 0) { > BUG_ON(blocksize != PAGE_SIZE); > unlock_extent(tree, start, end); > goto out; > } > } > > which will now leave the page locked when it errors out. Not to mention I'm > pretty sure you want to use btrfs_page_set_error() instead of SetPageError() in > that first case. Qu, please send a fixed version, just this patch, thanks.
diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c index eeee3213daaa..7fc2c62d4eb9 100644 --- a/fs/btrfs/extent_io.c +++ b/fs/btrfs/extent_io.c @@ -2839,8 +2839,17 @@ static void endio_readpage_release_extent(struct processed_extent *processed, processed->uptodate = uptodate; } -static void endio_readpage_update_page_status(struct page *page, bool uptodate, - u64 start, u32 len) +static void begin_data_page_read(struct btrfs_fs_info *fs_info, struct page *page) +{ + ASSERT(PageLocked(page)); + if (fs_info->sectorsize == PAGE_SIZE) + return; + + ASSERT(PagePrivate(page)); + btrfs_subpage_start_reader(fs_info, page, page_offset(page), PAGE_SIZE); +} + +static void end_page_read(struct page *page, bool uptodate, u64 start, u32 len) { struct btrfs_fs_info *fs_info = btrfs_sb(page->mapping->host->i_sb); @@ -2856,7 +2865,12 @@ static void endio_readpage_update_page_status(struct page *page, bool uptodate, if (fs_info->sectorsize == PAGE_SIZE) unlock_page(page); - /* Subpage locking will be handled in later patches */ + else if (is_data_inode(page->mapping->host)) + /* + * For subpage data, unlock the page if we're the last reader. + * For subpage metadata, page lock is not utilized for read. + */ + btrfs_subpage_end_reader(fs_info, page, start, len); } /* @@ -2993,7 +3007,7 @@ static void end_bio_extent_readpage(struct bio *bio) bio_offset += len; /* Update page status and unlock */ - endio_readpage_update_page_status(page, uptodate, start, len); + end_page_read(page, uptodate, start, len); endio_readpage_release_extent(&processed, BTRFS_I(inode), start, end, uptodate); } @@ -3263,6 +3277,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, unsigned int read_flags, u64 *prev_em_start) { struct inode *inode = page->mapping->host; + struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb); u64 start = page_offset(page); const u64 end = start + PAGE_SIZE - 1; u64 cur = start; @@ -3306,6 +3321,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, kunmap_atomic(userpage); } } + begin_data_page_read(fs_info, page); while (cur <= end) { bool force_bio_submit = false; u64 disk_bytenr; @@ -3323,13 +3339,14 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, &cached, GFP_NOFS); unlock_extent_cached(tree, cur, cur + iosize - 1, &cached); + end_page_read(page, true, cur, iosize); break; } em = __get_extent_map(inode, page, pg_offset, cur, end - cur + 1, em_cached); if (IS_ERR_OR_NULL(em)) { - SetPageError(page); unlock_extent(tree, cur, end); + end_page_read(page, false, cur, end + 1 - cur); break; } extent_offset = cur - em->start; @@ -3412,6 +3429,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, &cached, GFP_NOFS); unlock_extent_cached(tree, cur, cur + iosize - 1, &cached); + end_page_read(page, true, cur, iosize); cur = cur + iosize; pg_offset += iosize; continue; @@ -3421,6 +3439,7 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, EXTENT_UPTODATE, 1, NULL)) { check_page_uptodate(tree, page); unlock_extent(tree, cur, cur + iosize - 1); + end_page_read(page, true, cur, iosize); cur = cur + iosize; pg_offset += iosize; continue; @@ -3429,8 +3448,8 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, * to date. Error out */ if (block_start == EXTENT_MAP_INLINE) { - SetPageError(page); unlock_extent(tree, cur, cur + iosize - 1); + end_page_read(page, false, cur, iosize); cur = cur + iosize; pg_offset += iosize; continue; @@ -3447,19 +3466,14 @@ int btrfs_do_readpage(struct page *page, struct extent_map **em_cached, nr++; *bio_flags = this_bio_flag; } else { - SetPageError(page); unlock_extent(tree, cur, cur + iosize - 1); + end_page_read(page, false, cur, iosize); goto out; } cur = cur + iosize; pg_offset += iosize; } out: - if (!nr) { - if (!PageError(page)) - SetPageUptodate(page); - unlock_page(page); - } return ret; } diff --git a/fs/btrfs/subpage.c b/fs/btrfs/subpage.c index 2fe55a712557..c85f0f1c7441 100644 --- a/fs/btrfs/subpage.c +++ b/fs/btrfs/subpage.c @@ -54,6 +54,8 @@ int btrfs_alloc_subpage(const struct btrfs_fs_info *fs_info, spin_lock_init(&(*ret)->lock); if (type == BTRFS_SUBPAGE_METADATA) atomic_set(&(*ret)->eb_refs, 0); + else + atomic_set(&(*ret)->readers, 0); return 0; } @@ -102,23 +104,13 @@ void btrfs_page_dec_eb_refs(const struct btrfs_fs_info *fs_info, atomic_dec(&subpage->eb_refs); } -/* - * Convert the [start, start + len) range into a u16 bitmap - * - * For example: if start == page_offset() + 16K, len = 16K, we get 0x00f0. - */ -static inline u16 btrfs_subpage_calc_bitmap( - const struct btrfs_fs_info *fs_info, struct page *page, - u64 start, u32 len) +static void btrfs_subpage_assert(const struct btrfs_fs_info *fs_info, + struct page *page, u64 start, u32 len) { - const int bit_start = offset_in_page(start) >> fs_info->sectorsize_bits; - const int nbits = len >> fs_info->sectorsize_bits; - /* Basic checks */ ASSERT(PagePrivate(page) && page->private); ASSERT(IS_ALIGNED(start, fs_info->sectorsize) && IS_ALIGNED(len, fs_info->sectorsize)); - /* * The range check only works for mapped page, we can still have * unampped page like dummy extent buffer pages. @@ -126,6 +118,46 @@ static inline u16 btrfs_subpage_calc_bitmap( if (page->mapping) ASSERT(page_offset(page) <= start && start + len <= page_offset(page) + PAGE_SIZE); +} + +void btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, + struct page *page, u64 start, u32 len) +{ + struct btrfs_subpage *subpage = (struct btrfs_subpage *)page->private; + const int nbits = len >> fs_info->sectorsize_bits; + int ret; + + btrfs_subpage_assert(fs_info, page, start, len); + + ret = atomic_add_return(nbits, &subpage->readers); + ASSERT(ret == nbits); +} + +void btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, + struct page *page, u64 start, u32 len) +{ + struct btrfs_subpage *subpage = (struct btrfs_subpage *)page->private; + const int nbits = len >> fs_info->sectorsize_bits; + + btrfs_subpage_assert(fs_info, page, start, len); + ASSERT(atomic_read(&subpage->readers) >= nbits); + if (atomic_sub_and_test(nbits, &subpage->readers)) + unlock_page(page); +} + +/* + * Convert the [start, start + len) range into a u16 bitmap + * + * For example: if start == page_offset() + 16K, len = 16K, we get 0x00f0. + */ +static u16 btrfs_subpage_calc_bitmap(const struct btrfs_fs_info *fs_info, + struct page *page, u64 start, u32 len) +{ + const int bit_start = offset_in_page(start) >> fs_info->sectorsize_bits; + const int nbits = len >> fs_info->sectorsize_bits; + + btrfs_subpage_assert(fs_info, page, start, len); + /* * Here nbits can be 16, thus can go beyond u16 range. We make the * first left shift to be calculate in unsigned long (at least u32), diff --git a/fs/btrfs/subpage.h b/fs/btrfs/subpage.h index 68cbfc4f6765..bf5f565c8d1d 100644 --- a/fs/btrfs/subpage.h +++ b/fs/btrfs/subpage.h @@ -30,6 +30,9 @@ struct btrfs_subpage { */ atomic_t eb_refs; /* Structures only used by data */ + struct { + atomic_t readers; + }; }; }; @@ -54,6 +57,11 @@ void btrfs_page_inc_eb_refs(const struct btrfs_fs_info *fs_info, void btrfs_page_dec_eb_refs(const struct btrfs_fs_info *fs_info, struct page *page); +void btrfs_subpage_start_reader(const struct btrfs_fs_info *fs_info, + struct page *page, u64 start, u32 len); +void btrfs_subpage_end_reader(const struct btrfs_fs_info *fs_info, + struct page *page, u64 start, u32 len); + /* * Template for subpage related operations. *