From patchwork Sat Dec 3 01:40:32 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Darrick J. Wong" X-Patchwork-Id: 9486615 X-Mozilla-Keys: nonjunk Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on sandeen.net X-Spam-Level: X-Spam-Status: No, score=-7.0 required=5.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,RCVD_IN_DNSWL_HI,RP_MATCHES_RCVD, UNPARSEABLE_RELAY, URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 X-Spam-HP: BAYES_00=-1.9,HEADER_FROM_DIFFERENT_DOMAINS=0.001, RCVD_IN_DNSWL_HI=-5,RP_MATCHES_RCVD=-0.1,UNPARSEABLE_RELAY=0.001, URIBL_BLOCKED=0.001 X-Original-To: sandeen@sandeen.net Delivered-To: sandeen@sandeen.net Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by sandeen.net (Postfix) with ESMTP id EFBFD325413 for ; Fri, 2 Dec 2016 19:39:42 -0600 (CST) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1759489AbcLCBkg (ORCPT ); Fri, 2 Dec 2016 20:40:36 -0500 Received: from userp1040.oracle.com ([156.151.31.81]:23828 "EHLO userp1040.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1759149AbcLCBkg (ORCPT ); Fri, 2 Dec 2016 20:40:36 -0500 Received: from aserv0022.oracle.com (aserv0022.oracle.com [141.146.126.234]) by userp1040.oracle.com (Sentrion-MTA-4.3.2/Sentrion-MTA-4.3.2) with ESMTP id uB31eYkC021722 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Sat, 3 Dec 2016 01:40:34 GMT Received: from userv0122.oracle.com (userv0122.oracle.com [156.151.31.75]) by aserv0022.oracle.com (8.14.4/8.14.4) with ESMTP id uB31eX12011090 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Sat, 3 Dec 2016 01:40:33 GMT Received: from abhmp0018.oracle.com (abhmp0018.oracle.com [141.146.116.24]) by userv0122.oracle.com (8.14.4/8.14.4) with ESMTP id uB31eXJS023242; Sat, 3 Dec 2016 01:40:33 GMT Received: from localhost (/24.21.211.40) by default (Oracle Beehive Gateway v4.0) with ESMTP ; Fri, 02 Dec 2016 17:40:33 -0800 Subject: [PATCH 48/55] xfs: repair inode btrees From: "Darrick J. Wong" To: david@fromorbit.com, darrick.wong@oracle.com Cc: linux-xfs@vger.kernel.org Date: Fri, 02 Dec 2016 17:40:32 -0800 Message-ID: <148072923208.12995.2302009439133685364.stgit@birch.djwong.org> In-Reply-To: <148072891404.12995.15510849192837089093.stgit@birch.djwong.org> References: <148072891404.12995.15510849192837089093.stgit@birch.djwong.org> User-Agent: StGit/0.17.1-dirty MIME-Version: 1.0 X-Source-IP: aserv0022.oracle.com [141.146.126.234] Sender: linux-xfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-xfs@vger.kernel.org Use the rmapbt to find inode chunks, query the chunks to compute hole and free masks, and with that information rebuild the inobt and finobt. Signed-off-by: Darrick J. Wong --- fs/xfs/libxfs/xfs_ialloc.c | 2 fs/xfs/libxfs/xfs_ialloc.h | 3 fs/xfs/repair/common.c | 4 - fs/xfs/repair/common.h | 1 fs/xfs/repair/ialloc.c | 343 ++++++++++++++++++++++++++++++++++++++++++++ 5 files changed, 350 insertions(+), 3 deletions(-) -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html diff --git a/fs/xfs/libxfs/xfs_ialloc.c b/fs/xfs/libxfs/xfs_ialloc.c index c725eb9..282e3e6e 100644 --- a/fs/xfs/libxfs/xfs_ialloc.c +++ b/fs/xfs/libxfs/xfs_ialloc.c @@ -147,7 +147,7 @@ xfs_inobt_get_rec( /* * Insert a single inobt record. Cursor must already point to desired location. */ -STATIC int +int xfs_inobt_insert_rec( struct xfs_btree_cur *cur, __uint16_t holemask, diff --git a/fs/xfs/libxfs/xfs_ialloc.h b/fs/xfs/libxfs/xfs_ialloc.h index 17f0f1b..95216bb 100644 --- a/fs/xfs/libxfs/xfs_ialloc.h +++ b/fs/xfs/libxfs/xfs_ialloc.h @@ -177,5 +177,8 @@ int xfs_ialloc_has_inode_record(struct xfs_btree_cur *cur, xfs_agino_t low, xfs_agino_t high, bool *exists); int xfs_ialloc_count_inodes(struct xfs_btree_cur *cur, xfs_agino_t *count, xfs_agino_t *freecount); +int xfs_inobt_insert_rec(struct xfs_btree_cur *cur, __uint16_t holemask, + __uint8_t count, __int32_t freecount, xfs_inofree_t free, + int *stat); #endif /* __XFS_IALLOC_H__ */ diff --git a/fs/xfs/repair/common.c b/fs/xfs/repair/common.c index 07134c9..630ec8f 100644 --- a/fs/xfs/repair/common.c +++ b/fs/xfs/repair/common.c @@ -1008,8 +1008,8 @@ static const struct xfs_scrub_meta_fns meta_scrub_fns[] = { {xfs_scrub_setup_ag, xfs_scrub_agi, xfs_repair_agi, NULL}, {xfs_scrub_setup_ag_header, xfs_scrub_bnobt, xfs_repair_allocbt, NULL}, {xfs_scrub_setup_ag_header, xfs_scrub_cntbt, xfs_repair_allocbt, NULL}, - {xfs_scrub_setup_ag_header, xfs_scrub_inobt, NULL, NULL}, - {xfs_scrub_setup_ag_header, xfs_scrub_finobt, NULL, xfs_sb_version_hasfinobt}, + {xfs_scrub_setup_ag_header, xfs_scrub_inobt, xfs_repair_iallocbt, NULL}, + {xfs_scrub_setup_ag_header, xfs_scrub_finobt, xfs_repair_iallocbt, xfs_sb_version_hasfinobt}, {xfs_scrub_setup_ag_header, xfs_scrub_rmapbt, NULL, xfs_sb_version_hasrmapbt}, {xfs_scrub_setup_ag_header, xfs_scrub_refcountbt, NULL, xfs_sb_version_hasreflink}, {xfs_scrub_setup_inode_raw, xfs_scrub_inode, NULL, NULL}, diff --git a/fs/xfs/repair/common.h b/fs/xfs/repair/common.h index 158a002..ec83c3e 100644 --- a/fs/xfs/repair/common.h +++ b/fs/xfs/repair/common.h @@ -285,5 +285,6 @@ int xfs_repair_agf(struct xfs_scrub_context *sc); int xfs_repair_agfl(struct xfs_scrub_context *sc); int xfs_repair_agi(struct xfs_scrub_context *sc); int xfs_repair_allocbt(struct xfs_scrub_context *sc); +int xfs_repair_iallocbt(struct xfs_scrub_context *sc); #endif /* __XFS_REPAIR_COMMON_H__ */ diff --git a/fs/xfs/repair/ialloc.c b/fs/xfs/repair/ialloc.c index 4981c65..05af2145 100644 --- a/fs/xfs/repair/ialloc.c +++ b/fs/xfs/repair/ialloc.c @@ -36,7 +36,9 @@ #include "xfs_icache.h" #include "xfs_rmap.h" #include "xfs_alloc.h" +#include "xfs_rmap_btree.h" #include "xfs_refcount.h" +#include "xfs_error.h" #include "repair/common.h" #include "repair/btree.h" @@ -363,3 +365,344 @@ xfs_scrub_finobt( { return xfs_scrub_iallocbt(sc, XFS_BTNUM_FINO); } + +/* Inode btree repair. */ + +struct xfs_repair_ialloc_extent { + struct list_head list; + xfs_inofree_t freemask; + xfs_agino_t startino; + unsigned int count; + unsigned int usedcount; + __uint16_t holemask; +}; + +struct xfs_repair_ialloc { + struct list_head extlist; + struct list_head btlist; + uint64_t nr_records; +}; + +/* Record extents that belong to inode btrees. */ +STATIC int +xfs_repair_ialloc_extent_fn( + struct xfs_btree_cur *cur, + struct xfs_rmap_irec *rec, + void *priv) +{ + struct xfs_imap imap; + struct xfs_repair_ialloc *ri = priv; + struct xfs_repair_ialloc_extent *rie; + struct xfs_dinode *dip; + struct xfs_buf *bp; + struct xfs_mount *mp = cur->bc_mp; + xfs_ino_t fsino; + xfs_inofree_t usedmask; + xfs_fsblock_t fsbno; + xfs_agnumber_t agno; + xfs_agblock_t agbno; + xfs_agino_t agino; + xfs_agino_t startino; + xfs_agino_t chunkino; + xfs_agino_t nr_inodes; + xfs_agino_t i; + __uint16_t fillmask; + int blks_per_cluster; + int usedcount; + int error = 0; + + if (xfs_scrub_should_terminate(&error)) + return error; + + /* Fragment of the old btrees; dispose of them later. */ + if (rec->rm_owner == XFS_RMAP_OWN_INOBT) { + fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, + rec->rm_startblock); + return xfs_repair_collect_btree_extent(mp, &ri->btlist, + fsbno, rec->rm_blockcount); + } + + /* Skip extents which are not owned by this inode and fork. */ + if (rec->rm_owner != XFS_RMAP_OWN_INODES) + return 0; + + agno = cur->bc_private.a.agno; + blks_per_cluster = xfs_icluster_size_fsb(mp); + nr_inodes = XFS_OFFBNO_TO_AGINO(mp, blks_per_cluster, 0); + + ASSERT(rec->rm_startblock % blks_per_cluster == 0); + + trace_xfs_repair_ialloc_extent_fn(mp, cur->bc_private.a.agno, + rec->rm_startblock, rec->rm_blockcount, rec->rm_owner, + rec->rm_offset, rec->rm_flags); + + for (agbno = rec->rm_startblock; + agbno < rec->rm_startblock + rec->rm_blockcount; + agbno += blks_per_cluster) { + agino = XFS_OFFBNO_TO_AGINO(mp, agbno, 0); + fsino = XFS_AGINO_TO_INO(mp, agno, agino); + chunkino = agino & (XFS_INODES_PER_CHUNK - 1); + startino = agino & ~(XFS_INODES_PER_CHUNK - 1); + + /* Which inodes are not holes? */ + fillmask = xfs_inobt_maskn( + chunkino / XFS_INODES_PER_HOLEMASK_BIT, + nr_inodes / XFS_INODES_PER_HOLEMASK_BIT); + + /* Grab the inode cluster buffer. */ + imap.im_blkno = XFS_AGB_TO_DADDR(mp, agno, agbno); + imap.im_len = XFS_FSB_TO_BB(mp, blks_per_cluster); + imap.im_boffset = 0; + + error = xfs_imap_to_bp(mp, cur->bc_tp, &imap, + &dip, &bp, 0, XFS_IGET_UNTRUSTED); + if (error) + return error; + + /* Which inodes are free? */ + for (usedmask = 0, usedcount = 0, i = 0; i < nr_inodes; i++) { + dip = xfs_buf_offset(bp, i * mp->m_sb.sb_inodesize); + if (be16_to_cpu(dip->di_magic) != XFS_DINODE_MAGIC) { + xfs_trans_brelse(cur->bc_tp, bp); + return -EFSCORRUPTED; + } + if (dip->di_version >= 3 && + be64_to_cpu(dip->di_ino) != fsino + i) { + xfs_trans_brelse(cur->bc_tp, bp); + return -EFSCORRUPTED; + } + + if (dip->di_nlink || dip->di_onlink) { + usedmask |= 1ULL << (chunkino + i); + usedcount++; + } + } + xfs_trans_brelse(cur->bc_tp, bp); + + /* + * If the last item in the list is our chunk record, + * update that. + */ + if (!list_empty(&ri->extlist)) { + rie = list_last_entry(&ri->extlist, + struct xfs_repair_ialloc_extent, list); + if (rie->startino == startino) { + rie->freemask &= ~usedmask; + rie->holemask &= ~fillmask; + rie->count += nr_inodes; + rie->usedcount += usedcount; + continue; + } + } + + /* New inode chunk; add to the list. */ + rie = kmem_alloc(sizeof(*rie), KM_NOFS); + if (!rie) + return -ENOMEM; + + INIT_LIST_HEAD(&rie->list); + rie->startino = startino; + rie->freemask = XFS_INOBT_ALL_FREE & ~usedmask; + rie->holemask = XFS_INOBT_ALL_FREE & ~fillmask; + rie->count = nr_inodes; + rie->usedcount = usedcount; + list_add_tail(&rie->list, &ri->extlist); + ri->nr_records++; + } + + return 0; +} + +/* Compare two ialloc extents. */ +static int +xfs_repair_ialloc_extent_cmp( + void *priv, + struct list_head *a, + struct list_head *b) +{ + struct xfs_repair_ialloc_extent *ap; + struct xfs_repair_ialloc_extent *bp; + + ap = container_of(a, struct xfs_repair_ialloc_extent, list); + bp = container_of(b, struct xfs_repair_ialloc_extent, list); + + if (ap->startino > bp->startino) + return 1; + else if (ap->startino < bp->startino) + return -1; + return 0; +} + +/* Repair both inode btrees. */ +int +xfs_repair_iallocbt( + struct xfs_scrub_context *sc) +{ + struct xfs_repair_ialloc ri; + struct xfs_owner_info oinfo; + struct xfs_mount *mp = sc->tp->t_mountp; + struct xfs_buf *bp; + struct xfs_repair_ialloc_extent *rie; + struct xfs_repair_ialloc_extent *n; + struct xfs_agi *agi; + struct xfs_btree_cur *cur = NULL; + struct xfs_perag *pag; + xfs_fsblock_t inofsb; + xfs_fsblock_t finofsb; + xfs_extlen_t nr_blocks; + unsigned int count; + unsigned int usedcount; + int stat; + int logflags; + int error = 0; + + /* We require the rmapbt to rebuild anything. */ + if (!xfs_sb_version_hasrmapbt(&mp->m_sb)) + return -EOPNOTSUPP; + + /* Collect all reverse mappings for inode blocks. */ + xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_INOBT); + INIT_LIST_HEAD(&ri.extlist); + INIT_LIST_HEAD(&ri.btlist); + ri.nr_records = 0; + cur = xfs_rmapbt_init_cursor(mp, sc->tp, sc->sa.agf_bp, sc->sa.agno); + error = xfs_rmap_query_all(cur, xfs_repair_ialloc_extent_fn, &ri); + if (error) + goto out; + xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); + cur = NULL; + + /* Do we actually have enough space to do this? */ + pag = xfs_perag_get(mp, sc->sa.agno); + nr_blocks = xfs_iallocbt_calc_size(mp, ri.nr_records); + if (xfs_sb_version_hasfinobt(&mp->m_sb)) + nr_blocks *= 2; + if (!xfs_repair_ag_has_space(pag, nr_blocks, XFS_AG_RESV_NONE)) { + xfs_perag_put(pag); + error = -ENOSPC; + goto out; + } + xfs_perag_put(pag); + + agi = XFS_BUF_TO_AGI(sc->sa.agi_bp); + /* Initialize new btree roots. */ + error = xfs_repair_alloc_ag_block(sc, &oinfo, &inofsb, + XFS_AG_RESV_NONE); + if (error) + goto out; + error = xfs_repair_init_btblock(sc, inofsb, &bp, XFS_IBT_CRC_MAGIC, + &xfs_inobt_buf_ops); + if (error) + goto out; + agi->agi_root = cpu_to_be32(XFS_FSB_TO_AGBNO(mp, inofsb)); + agi->agi_level = cpu_to_be32(1); + logflags = XFS_AGI_ROOT | XFS_AGI_LEVEL; + + if (xfs_sb_version_hasfinobt(&mp->m_sb)) { + error = xfs_repair_alloc_ag_block(sc, &oinfo, &finofsb, + XFS_AG_RESV_NONE); + if (error) + goto out; + error = xfs_repair_init_btblock(sc, finofsb, &bp, + XFS_FIBT_CRC_MAGIC, &xfs_inobt_buf_ops); + if (error) + goto out; + agi->agi_free_root = cpu_to_be32(XFS_FSB_TO_AGBNO(mp, finofsb)); + agi->agi_free_level = cpu_to_be32(1); + logflags |= XFS_AGI_FREE_ROOT | XFS_AGI_FREE_LEVEL; + } + + xfs_ialloc_log_agi(sc->tp, sc->sa.agi_bp, logflags); + error = xfs_repair_roll_ag_trans(sc); + if (error) + goto out; + + /* Insert records into the new btrees. */ + count = 0; + usedcount = 0; + list_sort(NULL, &ri.extlist, xfs_repair_ialloc_extent_cmp); + list_for_each_entry_safe(rie, n, &ri.extlist, list) { + count += rie->count; + usedcount += rie->usedcount; + + trace_xfs_repair_ialloc_insert(mp, sc->sa.agno, rie->startino, + rie->holemask, rie->count, + rie->count - rie->usedcount, rie->freemask); + + /* Insert into the inobt. */ + cur = xfs_inobt_init_cursor(mp, sc->tp, sc->sa.agi_bp, + sc->sa.agno, XFS_BTNUM_INO); + error = xfs_inobt_lookup(cur, rie->startino, XFS_LOOKUP_EQ, + &stat); + if (error) + goto out; + XFS_WANT_CORRUPTED_GOTO(mp, stat == 0, out); + error = xfs_inobt_insert_rec(cur, rie->holemask, rie->count, + rie->count - rie->usedcount, rie->freemask, + &stat); + if (error) + goto out; + XFS_WANT_CORRUPTED_GOTO(mp, stat == 1, out); + xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); + cur = NULL; + + /* Insert into the finobt. */ + if (rie->count != rie->usedcount && + xfs_sb_version_hasfinobt(&mp->m_sb)) { + cur = xfs_inobt_init_cursor(mp, sc->tp, sc->sa.agi_bp, + sc->sa.agno, XFS_BTNUM_FINO); + error = xfs_inobt_lookup(cur, rie->startino, + XFS_LOOKUP_EQ, &stat); + if (error) + goto out; + XFS_WANT_CORRUPTED_GOTO(mp, stat == 0, out); + error = xfs_inobt_insert_rec(cur, rie->holemask, + rie->count, rie->count - rie->usedcount, + rie->freemask, &stat); + if (error) + goto out; + XFS_WANT_CORRUPTED_GOTO(mp, stat == 1, out); + xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); + cur = NULL; + } + + error = xfs_repair_roll_ag_trans(sc); + if (error) + goto out; + + list_del(&rie->list); + kmem_free(rie); + } + + /* Update the AGI counters. */ + agi = XFS_BUF_TO_AGI(sc->sa.agi_bp); + if (be32_to_cpu(agi->agi_count) != count || + be32_to_cpu(agi->agi_freecount) != count - usedcount) { + pag = xfs_perag_get(mp, sc->sa.agno); + pag->pagi_init = 0; + xfs_perag_put(pag); + + agi->agi_count = cpu_to_be32(count); + agi->agi_freecount = cpu_to_be32(count - usedcount); + xfs_ialloc_log_agi(sc->tp, sc->sa.agi_bp, + XFS_AGI_COUNT | XFS_AGI_FREECOUNT); + sc->reset_counters = true; + } + + /* Free the old inode btree blocks if they're not in use. */ + error = xfs_repair_reap_btree_extents(sc, &ri.btlist, &oinfo, + XFS_AG_RESV_NONE); + if (error) + goto out; + + return error; +out: + if (cur) + xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); + xfs_repair_cancel_btree_extents(sc, &ri.btlist); + list_for_each_entry_safe(rie, n, &ri.extlist, list) { + list_del(&rie->list); + kmem_free(rie); + } + return error; +}