Message ID | 20170221173410.GF5846@birch.djwong.org (mailing list archive) |
---|---|
State | Accepted |
Headers | show |
On Tue, Feb 21, 2017 at 09:34:10AM -0800, Darrick J. Wong wrote: > Plumb in the pieces (init_high_key, diff_two_keys) necessary to call > query_range on the free space btrees. Remove the debugging asserts > so that we can make queries starting from block 0. > > While we're at it, merge the redundant "if (btnum ==" hunks. > > Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> > --- Reviewed-by: Brian Foster <bfoster@redhat.com> > v2: Add forgotten cntbt high key function. > --- > fs/xfs/libxfs/xfs_alloc_btree.c | 172 +++++++++++++++++++++++++++++---------- > 1 file changed, 129 insertions(+), 43 deletions(-) > > diff --git a/fs/xfs/libxfs/xfs_alloc_btree.c b/fs/xfs/libxfs/xfs_alloc_btree.c > index efb467b..e1fcfe7 100644 > --- a/fs/xfs/libxfs/xfs_alloc_btree.c > +++ b/fs/xfs/libxfs/xfs_alloc_btree.c > @@ -205,19 +205,37 @@ xfs_allocbt_init_key_from_rec( > union xfs_btree_key *key, > union xfs_btree_rec *rec) > { > - ASSERT(rec->alloc.ar_startblock != 0); > - > key->alloc.ar_startblock = rec->alloc.ar_startblock; > key->alloc.ar_blockcount = rec->alloc.ar_blockcount; > } > > STATIC void > +xfs_bnobt_init_high_key_from_rec( > + union xfs_btree_key *key, > + union xfs_btree_rec *rec) > +{ > + __u32 x; > + > + x = be32_to_cpu(rec->alloc.ar_startblock); > + x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; > + key->alloc.ar_startblock = cpu_to_be32(x); > + key->alloc.ar_blockcount = 0; > +} > + > +STATIC void > +xfs_cntbt_init_high_key_from_rec( > + union xfs_btree_key *key, > + union xfs_btree_rec *rec) > +{ > + key->alloc.ar_blockcount = rec->alloc.ar_blockcount; > + key->alloc.ar_startblock = 0; > +} > + > +STATIC void > xfs_allocbt_init_rec_from_cur( > struct xfs_btree_cur *cur, > union xfs_btree_rec *rec) > { > - ASSERT(cur->bc_rec.a.ar_startblock != 0); > - > rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock); > rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount); > } > @@ -236,18 +254,24 @@ xfs_allocbt_init_ptr_from_cur( > } > > STATIC __int64_t > -xfs_allocbt_key_diff( > +xfs_bnobt_key_diff( > struct xfs_btree_cur *cur, > union xfs_btree_key *key) > { > xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a; > xfs_alloc_key_t *kp = &key->alloc; > - __int64_t diff; > > - if (cur->bc_btnum == XFS_BTNUM_BNO) { > - return (__int64_t)be32_to_cpu(kp->ar_startblock) - > - rec->ar_startblock; > - } > + return (__int64_t)be32_to_cpu(kp->ar_startblock) - rec->ar_startblock; > +} > + > +STATIC __int64_t > +xfs_cntbt_key_diff( > + struct xfs_btree_cur *cur, > + union xfs_btree_key *key) > +{ > + xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a; > + xfs_alloc_key_t *kp = &key->alloc; > + __int64_t diff; > > diff = (__int64_t)be32_to_cpu(kp->ar_blockcount) - rec->ar_blockcount; > if (diff) > @@ -256,6 +280,33 @@ xfs_allocbt_key_diff( > return (__int64_t)be32_to_cpu(kp->ar_startblock) - rec->ar_startblock; > } > > +STATIC __int64_t > +xfs_bnobt_diff_two_keys( > + struct xfs_btree_cur *cur, > + union xfs_btree_key *k1, > + union xfs_btree_key *k2) > +{ > + return (__int64_t)be32_to_cpu(k1->alloc.ar_startblock) - > + be32_to_cpu(k2->alloc.ar_startblock); > +} > + > +STATIC __int64_t > +xfs_cntbt_diff_two_keys( > + struct xfs_btree_cur *cur, > + union xfs_btree_key *k1, > + union xfs_btree_key *k2) > +{ > + __int64_t diff; > + > + diff = be32_to_cpu(k1->alloc.ar_blockcount) - > + be32_to_cpu(k2->alloc.ar_blockcount); > + if (diff) > + return diff; > + > + return be32_to_cpu(k1->alloc.ar_startblock) - > + be32_to_cpu(k2->alloc.ar_startblock); > +} > + > static bool > xfs_allocbt_verify( > struct xfs_buf *bp) > @@ -346,44 +397,54 @@ const struct xfs_buf_ops xfs_allocbt_buf_ops = { > > #if defined(DEBUG) || defined(XFS_WARN) > STATIC int > -xfs_allocbt_keys_inorder( > +xfs_bnobt_keys_inorder( > struct xfs_btree_cur *cur, > union xfs_btree_key *k1, > union xfs_btree_key *k2) > { > - if (cur->bc_btnum == XFS_BTNUM_BNO) { > - return be32_to_cpu(k1->alloc.ar_startblock) < > - be32_to_cpu(k2->alloc.ar_startblock); > - } else { > - return be32_to_cpu(k1->alloc.ar_blockcount) < > - be32_to_cpu(k2->alloc.ar_blockcount) || > - (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && > - be32_to_cpu(k1->alloc.ar_startblock) < > - be32_to_cpu(k2->alloc.ar_startblock)); > - } > + return be32_to_cpu(k1->alloc.ar_startblock) < > + be32_to_cpu(k2->alloc.ar_startblock); > } > > STATIC int > -xfs_allocbt_recs_inorder( > +xfs_bnobt_recs_inorder( > struct xfs_btree_cur *cur, > union xfs_btree_rec *r1, > union xfs_btree_rec *r2) > { > - if (cur->bc_btnum == XFS_BTNUM_BNO) { > - return be32_to_cpu(r1->alloc.ar_startblock) + > - be32_to_cpu(r1->alloc.ar_blockcount) <= > - be32_to_cpu(r2->alloc.ar_startblock); > - } else { > - return be32_to_cpu(r1->alloc.ar_blockcount) < > - be32_to_cpu(r2->alloc.ar_blockcount) || > - (r1->alloc.ar_blockcount == r2->alloc.ar_blockcount && > - be32_to_cpu(r1->alloc.ar_startblock) < > - be32_to_cpu(r2->alloc.ar_startblock)); > - } > + return be32_to_cpu(r1->alloc.ar_startblock) + > + be32_to_cpu(r1->alloc.ar_blockcount) <= > + be32_to_cpu(r2->alloc.ar_startblock); > +} > + > +STATIC int > +xfs_cntbt_keys_inorder( > + struct xfs_btree_cur *cur, > + union xfs_btree_key *k1, > + union xfs_btree_key *k2) > +{ > + return be32_to_cpu(k1->alloc.ar_blockcount) < > + be32_to_cpu(k2->alloc.ar_blockcount) || > + (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && > + be32_to_cpu(k1->alloc.ar_startblock) < > + be32_to_cpu(k2->alloc.ar_startblock)); > } > -#endif /* DEBUG */ > > -static const struct xfs_btree_ops xfs_allocbt_ops = { > +STATIC int > +xfs_cntbt_recs_inorder( > + struct xfs_btree_cur *cur, > + union xfs_btree_rec *r1, > + union xfs_btree_rec *r2) > +{ > + return be32_to_cpu(r1->alloc.ar_blockcount) < > + be32_to_cpu(r2->alloc.ar_blockcount) || > + (r1->alloc.ar_blockcount == r2->alloc.ar_blockcount && > + be32_to_cpu(r1->alloc.ar_startblock) < > + be32_to_cpu(r2->alloc.ar_startblock)); > +} > +#endif /* DEBUG */ > + > +static const struct xfs_btree_ops xfs_bnobt_ops = { > .rec_len = sizeof(xfs_alloc_rec_t), > .key_len = sizeof(xfs_alloc_key_t), > > @@ -395,13 +456,39 @@ static const struct xfs_btree_ops xfs_allocbt_ops = { > .get_minrecs = xfs_allocbt_get_minrecs, > .get_maxrecs = xfs_allocbt_get_maxrecs, > .init_key_from_rec = xfs_allocbt_init_key_from_rec, > + .init_high_key_from_rec = xfs_bnobt_init_high_key_from_rec, > .init_rec_from_cur = xfs_allocbt_init_rec_from_cur, > .init_ptr_from_cur = xfs_allocbt_init_ptr_from_cur, > - .key_diff = xfs_allocbt_key_diff, > + .key_diff = xfs_bnobt_key_diff, > .buf_ops = &xfs_allocbt_buf_ops, > + .diff_two_keys = xfs_bnobt_diff_two_keys, > #if defined(DEBUG) || defined(XFS_WARN) > - .keys_inorder = xfs_allocbt_keys_inorder, > - .recs_inorder = xfs_allocbt_recs_inorder, > + .keys_inorder = xfs_bnobt_keys_inorder, > + .recs_inorder = xfs_bnobt_recs_inorder, > +#endif > +}; > + > +static const struct xfs_btree_ops xfs_cntbt_ops = { > + .rec_len = sizeof(xfs_alloc_rec_t), > + .key_len = sizeof(xfs_alloc_key_t), > + > + .dup_cursor = xfs_allocbt_dup_cursor, > + .set_root = xfs_allocbt_set_root, > + .alloc_block = xfs_allocbt_alloc_block, > + .free_block = xfs_allocbt_free_block, > + .update_lastrec = xfs_allocbt_update_lastrec, > + .get_minrecs = xfs_allocbt_get_minrecs, > + .get_maxrecs = xfs_allocbt_get_maxrecs, > + .init_key_from_rec = xfs_allocbt_init_key_from_rec, > + .init_high_key_from_rec = xfs_cntbt_init_high_key_from_rec, > + .init_rec_from_cur = xfs_allocbt_init_rec_from_cur, > + .init_ptr_from_cur = xfs_allocbt_init_ptr_from_cur, > + .key_diff = xfs_cntbt_key_diff, > + .buf_ops = &xfs_allocbt_buf_ops, > + .diff_two_keys = xfs_cntbt_diff_two_keys, > +#if defined(DEBUG) || defined(XFS_WARN) > + .keys_inorder = xfs_cntbt_keys_inorder, > + .recs_inorder = xfs_cntbt_recs_inorder, > #endif > }; > > @@ -427,16 +514,15 @@ xfs_allocbt_init_cursor( > cur->bc_mp = mp; > cur->bc_btnum = btnum; > cur->bc_blocklog = mp->m_sb.sb_blocklog; > - cur->bc_ops = &xfs_allocbt_ops; > - if (btnum == XFS_BTNUM_BNO) > - cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2); > - else > - cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2); > > if (btnum == XFS_BTNUM_CNT) { > + cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2); > + cur->bc_ops = &xfs_cntbt_ops; > cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]); > cur->bc_flags = XFS_BTREE_LASTREC_UPDATE; > } else { > + cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2); > + cur->bc_ops = &xfs_bnobt_ops; > cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]); > } > > -- > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/fs/xfs/libxfs/xfs_alloc_btree.c b/fs/xfs/libxfs/xfs_alloc_btree.c index efb467b..e1fcfe7 100644 --- a/fs/xfs/libxfs/xfs_alloc_btree.c +++ b/fs/xfs/libxfs/xfs_alloc_btree.c @@ -205,19 +205,37 @@ xfs_allocbt_init_key_from_rec( union xfs_btree_key *key, union xfs_btree_rec *rec) { - ASSERT(rec->alloc.ar_startblock != 0); - key->alloc.ar_startblock = rec->alloc.ar_startblock; key->alloc.ar_blockcount = rec->alloc.ar_blockcount; } STATIC void +xfs_bnobt_init_high_key_from_rec( + union xfs_btree_key *key, + union xfs_btree_rec *rec) +{ + __u32 x; + + x = be32_to_cpu(rec->alloc.ar_startblock); + x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; + key->alloc.ar_startblock = cpu_to_be32(x); + key->alloc.ar_blockcount = 0; +} + +STATIC void +xfs_cntbt_init_high_key_from_rec( + union xfs_btree_key *key, + union xfs_btree_rec *rec) +{ + key->alloc.ar_blockcount = rec->alloc.ar_blockcount; + key->alloc.ar_startblock = 0; +} + +STATIC void xfs_allocbt_init_rec_from_cur( struct xfs_btree_cur *cur, union xfs_btree_rec *rec) { - ASSERT(cur->bc_rec.a.ar_startblock != 0); - rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock); rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount); } @@ -236,18 +254,24 @@ xfs_allocbt_init_ptr_from_cur( } STATIC __int64_t -xfs_allocbt_key_diff( +xfs_bnobt_key_diff( struct xfs_btree_cur *cur, union xfs_btree_key *key) { xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a; xfs_alloc_key_t *kp = &key->alloc; - __int64_t diff; - if (cur->bc_btnum == XFS_BTNUM_BNO) { - return (__int64_t)be32_to_cpu(kp->ar_startblock) - - rec->ar_startblock; - } + return (__int64_t)be32_to_cpu(kp->ar_startblock) - rec->ar_startblock; +} + +STATIC __int64_t +xfs_cntbt_key_diff( + struct xfs_btree_cur *cur, + union xfs_btree_key *key) +{ + xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a; + xfs_alloc_key_t *kp = &key->alloc; + __int64_t diff; diff = (__int64_t)be32_to_cpu(kp->ar_blockcount) - rec->ar_blockcount; if (diff) @@ -256,6 +280,33 @@ xfs_allocbt_key_diff( return (__int64_t)be32_to_cpu(kp->ar_startblock) - rec->ar_startblock; } +STATIC __int64_t +xfs_bnobt_diff_two_keys( + struct xfs_btree_cur *cur, + union xfs_btree_key *k1, + union xfs_btree_key *k2) +{ + return (__int64_t)be32_to_cpu(k1->alloc.ar_startblock) - + be32_to_cpu(k2->alloc.ar_startblock); +} + +STATIC __int64_t +xfs_cntbt_diff_two_keys( + struct xfs_btree_cur *cur, + union xfs_btree_key *k1, + union xfs_btree_key *k2) +{ + __int64_t diff; + + diff = be32_to_cpu(k1->alloc.ar_blockcount) - + be32_to_cpu(k2->alloc.ar_blockcount); + if (diff) + return diff; + + return be32_to_cpu(k1->alloc.ar_startblock) - + be32_to_cpu(k2->alloc.ar_startblock); +} + static bool xfs_allocbt_verify( struct xfs_buf *bp) @@ -346,44 +397,54 @@ const struct xfs_buf_ops xfs_allocbt_buf_ops = { #if defined(DEBUG) || defined(XFS_WARN) STATIC int -xfs_allocbt_keys_inorder( +xfs_bnobt_keys_inorder( struct xfs_btree_cur *cur, union xfs_btree_key *k1, union xfs_btree_key *k2) { - if (cur->bc_btnum == XFS_BTNUM_BNO) { - return be32_to_cpu(k1->alloc.ar_startblock) < - be32_to_cpu(k2->alloc.ar_startblock); - } else { - return be32_to_cpu(k1->alloc.ar_blockcount) < - be32_to_cpu(k2->alloc.ar_blockcount) || - (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && - be32_to_cpu(k1->alloc.ar_startblock) < - be32_to_cpu(k2->alloc.ar_startblock)); - } + return be32_to_cpu(k1->alloc.ar_startblock) < + be32_to_cpu(k2->alloc.ar_startblock); } STATIC int -xfs_allocbt_recs_inorder( +xfs_bnobt_recs_inorder( struct xfs_btree_cur *cur, union xfs_btree_rec *r1, union xfs_btree_rec *r2) { - if (cur->bc_btnum == XFS_BTNUM_BNO) { - return be32_to_cpu(r1->alloc.ar_startblock) + - be32_to_cpu(r1->alloc.ar_blockcount) <= - be32_to_cpu(r2->alloc.ar_startblock); - } else { - return be32_to_cpu(r1->alloc.ar_blockcount) < - be32_to_cpu(r2->alloc.ar_blockcount) || - (r1->alloc.ar_blockcount == r2->alloc.ar_blockcount && - be32_to_cpu(r1->alloc.ar_startblock) < - be32_to_cpu(r2->alloc.ar_startblock)); - } + return be32_to_cpu(r1->alloc.ar_startblock) + + be32_to_cpu(r1->alloc.ar_blockcount) <= + be32_to_cpu(r2->alloc.ar_startblock); +} + +STATIC int +xfs_cntbt_keys_inorder( + struct xfs_btree_cur *cur, + union xfs_btree_key *k1, + union xfs_btree_key *k2) +{ + return be32_to_cpu(k1->alloc.ar_blockcount) < + be32_to_cpu(k2->alloc.ar_blockcount) || + (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && + be32_to_cpu(k1->alloc.ar_startblock) < + be32_to_cpu(k2->alloc.ar_startblock)); } -#endif /* DEBUG */ -static const struct xfs_btree_ops xfs_allocbt_ops = { +STATIC int +xfs_cntbt_recs_inorder( + struct xfs_btree_cur *cur, + union xfs_btree_rec *r1, + union xfs_btree_rec *r2) +{ + return be32_to_cpu(r1->alloc.ar_blockcount) < + be32_to_cpu(r2->alloc.ar_blockcount) || + (r1->alloc.ar_blockcount == r2->alloc.ar_blockcount && + be32_to_cpu(r1->alloc.ar_startblock) < + be32_to_cpu(r2->alloc.ar_startblock)); +} +#endif /* DEBUG */ + +static const struct xfs_btree_ops xfs_bnobt_ops = { .rec_len = sizeof(xfs_alloc_rec_t), .key_len = sizeof(xfs_alloc_key_t), @@ -395,13 +456,39 @@ static const struct xfs_btree_ops xfs_allocbt_ops = { .get_minrecs = xfs_allocbt_get_minrecs, .get_maxrecs = xfs_allocbt_get_maxrecs, .init_key_from_rec = xfs_allocbt_init_key_from_rec, + .init_high_key_from_rec = xfs_bnobt_init_high_key_from_rec, .init_rec_from_cur = xfs_allocbt_init_rec_from_cur, .init_ptr_from_cur = xfs_allocbt_init_ptr_from_cur, - .key_diff = xfs_allocbt_key_diff, + .key_diff = xfs_bnobt_key_diff, .buf_ops = &xfs_allocbt_buf_ops, + .diff_two_keys = xfs_bnobt_diff_two_keys, #if defined(DEBUG) || defined(XFS_WARN) - .keys_inorder = xfs_allocbt_keys_inorder, - .recs_inorder = xfs_allocbt_recs_inorder, + .keys_inorder = xfs_bnobt_keys_inorder, + .recs_inorder = xfs_bnobt_recs_inorder, +#endif +}; + +static const struct xfs_btree_ops xfs_cntbt_ops = { + .rec_len = sizeof(xfs_alloc_rec_t), + .key_len = sizeof(xfs_alloc_key_t), + + .dup_cursor = xfs_allocbt_dup_cursor, + .set_root = xfs_allocbt_set_root, + .alloc_block = xfs_allocbt_alloc_block, + .free_block = xfs_allocbt_free_block, + .update_lastrec = xfs_allocbt_update_lastrec, + .get_minrecs = xfs_allocbt_get_minrecs, + .get_maxrecs = xfs_allocbt_get_maxrecs, + .init_key_from_rec = xfs_allocbt_init_key_from_rec, + .init_high_key_from_rec = xfs_cntbt_init_high_key_from_rec, + .init_rec_from_cur = xfs_allocbt_init_rec_from_cur, + .init_ptr_from_cur = xfs_allocbt_init_ptr_from_cur, + .key_diff = xfs_cntbt_key_diff, + .buf_ops = &xfs_allocbt_buf_ops, + .diff_two_keys = xfs_cntbt_diff_two_keys, +#if defined(DEBUG) || defined(XFS_WARN) + .keys_inorder = xfs_cntbt_keys_inorder, + .recs_inorder = xfs_cntbt_recs_inorder, #endif }; @@ -427,16 +514,15 @@ xfs_allocbt_init_cursor( cur->bc_mp = mp; cur->bc_btnum = btnum; cur->bc_blocklog = mp->m_sb.sb_blocklog; - cur->bc_ops = &xfs_allocbt_ops; - if (btnum == XFS_BTNUM_BNO) - cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2); - else - cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2); if (btnum == XFS_BTNUM_CNT) { + cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2); + cur->bc_ops = &xfs_cntbt_ops; cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]); cur->bc_flags = XFS_BTREE_LASTREC_UPDATE; } else { + cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2); + cur->bc_ops = &xfs_bnobt_ops; cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]); }
Plumb in the pieces (init_high_key, diff_two_keys) necessary to call query_range on the free space btrees. Remove the debugging asserts so that we can make queries starting from block 0. While we're at it, merge the redundant "if (btnum ==" hunks. Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> --- v2: Add forgotten cntbt high key function. --- fs/xfs/libxfs/xfs_alloc_btree.c | 172 +++++++++++++++++++++++++++++---------- 1 file changed, 129 insertions(+), 43 deletions(-) -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html