1051 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1051 lines
		
	
	
		
			28 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-or-later
 | |
| /*
 | |
|  * Copyright (C) 2022-2023 Oracle.  All Rights Reserved.
 | |
|  * Author: Darrick J. Wong <djwong@kernel.org>
 | |
|  */
 | |
| #include "xfs.h"
 | |
| #include "xfs_fs.h"
 | |
| #include "xfs_shared.h"
 | |
| #include "xfs_format.h"
 | |
| #include "xfs_trans_resv.h"
 | |
| #include "xfs_mount.h"
 | |
| #include "xfs_btree.h"
 | |
| #include "xfs_log_format.h"
 | |
| #include "xfs_trans.h"
 | |
| #include "xfs_sb.h"
 | |
| #include "xfs_inode.h"
 | |
| #include "xfs_alloc.h"
 | |
| #include "xfs_alloc_btree.h"
 | |
| #include "xfs_ialloc.h"
 | |
| #include "xfs_ialloc_btree.h"
 | |
| #include "xfs_rmap.h"
 | |
| #include "xfs_rmap_btree.h"
 | |
| #include "xfs_refcount.h"
 | |
| #include "xfs_refcount_btree.h"
 | |
| #include "xfs_extent_busy.h"
 | |
| #include "xfs_ag.h"
 | |
| #include "xfs_ag_resv.h"
 | |
| #include "xfs_quota.h"
 | |
| #include "xfs_qm.h"
 | |
| #include "xfs_bmap.h"
 | |
| #include "xfs_da_format.h"
 | |
| #include "xfs_da_btree.h"
 | |
| #include "xfs_attr.h"
 | |
| #include "xfs_attr_remote.h"
 | |
| #include "xfs_defer.h"
 | |
| #include "scrub/scrub.h"
 | |
| #include "scrub/common.h"
 | |
| #include "scrub/trace.h"
 | |
| #include "scrub/repair.h"
 | |
| #include "scrub/bitmap.h"
 | |
| #include "scrub/agb_bitmap.h"
 | |
| #include "scrub/fsb_bitmap.h"
 | |
| #include "scrub/reap.h"
 | |
| 
 | |
| /*
 | |
|  * Disposal of Blocks from Old Metadata
 | |
|  *
 | |
|  * Now that we've constructed a new btree to replace the damaged one, we want
 | |
|  * to dispose of the blocks that (we think) the old btree was using.
 | |
|  * Previously, we used the rmapbt to collect the extents (bitmap) with the
 | |
|  * rmap owner corresponding to the tree we rebuilt, collected extents for any
 | |
|  * blocks with the same rmap owner that are owned by another data structure
 | |
|  * (sublist), and subtracted sublist from bitmap.  In theory the extents
 | |
|  * remaining in bitmap are the old btree's blocks.
 | |
|  *
 | |
|  * Unfortunately, it's possible that the btree was crosslinked with other
 | |
|  * blocks on disk.  The rmap data can tell us if there are multiple owners, so
 | |
|  * if the rmapbt says there is an owner of this block other than @oinfo, then
 | |
|  * the block is crosslinked.  Remove the reverse mapping and continue.
 | |
|  *
 | |
|  * If there is one rmap record, we can free the block, which removes the
 | |
|  * reverse mapping but doesn't add the block to the free space.  Our repair
 | |
|  * strategy is to hope the other metadata objects crosslinked on this block
 | |
|  * will be rebuilt (atop different blocks), thereby removing all the cross
 | |
|  * links.
 | |
|  *
 | |
|  * If there are no rmap records at all, we also free the block.  If the btree
 | |
|  * being rebuilt lives in the free space (bnobt/cntbt/rmapbt) then there isn't
 | |
|  * supposed to be a rmap record and everything is ok.  For other btrees there
 | |
|  * had to have been an rmap entry for the block to have ended up on @bitmap,
 | |
|  * so if it's gone now there's something wrong and the fs will shut down.
 | |
|  *
 | |
|  * Note: If there are multiple rmap records with only the same rmap owner as
 | |
|  * the btree we're trying to rebuild and the block is indeed owned by another
 | |
|  * data structure with the same rmap owner, then the block will be in sublist
 | |
|  * and therefore doesn't need disposal.  If there are multiple rmap records
 | |
|  * with only the same rmap owner but the block is not owned by something with
 | |
|  * the same rmap owner, the block will be freed.
 | |
|  *
 | |
|  * The caller is responsible for locking the AG headers/inode for the entire
 | |
|  * rebuild operation so that nothing else can sneak in and change the incore
 | |
|  * state while we're not looking.  We must also invalidate any buffers
 | |
|  * associated with @bitmap.
 | |
|  */
 | |
| 
 | |
| /* Information about reaping extents after a repair. */
 | |
| struct xreap_state {
 | |
| 	struct xfs_scrub		*sc;
 | |
| 
 | |
| 	/* Reverse mapping owner and metadata reservation type. */
 | |
| 	const struct xfs_owner_info	*oinfo;
 | |
| 	enum xfs_ag_resv_type		resv;
 | |
| 
 | |
| 	/* If true, roll the transaction before reaping the next extent. */
 | |
| 	bool				force_roll;
 | |
| 
 | |
| 	/* Number of deferred reaps attached to the current transaction. */
 | |
| 	unsigned int			deferred;
 | |
| 
 | |
| 	/* Number of invalidated buffers logged to the current transaction. */
 | |
| 	unsigned int			invalidated;
 | |
| 
 | |
| 	/* Number of deferred reaps queued during the whole reap sequence. */
 | |
| 	unsigned long long		total_deferred;
 | |
| };
 | |
| 
 | |
| /* Put a block back on the AGFL. */
 | |
| STATIC int
 | |
| xreap_put_freelist(
 | |
| 	struct xfs_scrub	*sc,
 | |
| 	xfs_agblock_t		agbno)
 | |
| {
 | |
| 	struct xfs_buf		*agfl_bp;
 | |
| 	int			error;
 | |
| 
 | |
| 	/* Make sure there's space on the freelist. */
 | |
| 	error = xrep_fix_freelist(sc, 0);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 
 | |
| 	/*
 | |
| 	 * Since we're "freeing" a lost block onto the AGFL, we have to
 | |
| 	 * create an rmap for the block prior to merging it or else other
 | |
| 	 * parts will break.
 | |
| 	 */
 | |
| 	error = xfs_rmap_alloc(sc->tp, sc->sa.agf_bp, sc->sa.pag, agbno, 1,
 | |
| 			&XFS_RMAP_OINFO_AG);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 
 | |
| 	/* Put the block on the AGFL. */
 | |
| 	error = xfs_alloc_read_agfl(sc->sa.pag, sc->tp, &agfl_bp);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 
 | |
| 	error = xfs_alloc_put_freelist(sc->sa.pag, sc->tp, sc->sa.agf_bp,
 | |
| 			agfl_bp, agbno, 0);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 	xfs_extent_busy_insert(sc->tp, sc->sa.pag, agbno, 1,
 | |
| 			XFS_EXTENT_BUSY_SKIP_DISCARD);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Are there any uncommitted reap operations? */
 | |
| static inline bool xreap_dirty(const struct xreap_state *rs)
 | |
| {
 | |
| 	if (rs->force_roll)
 | |
| 		return true;
 | |
| 	if (rs->deferred)
 | |
| 		return true;
 | |
| 	if (rs->invalidated)
 | |
| 		return true;
 | |
| 	if (rs->total_deferred)
 | |
| 		return true;
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| #define XREAP_MAX_BINVAL	(2048)
 | |
| 
 | |
| /*
 | |
|  * Decide if we want to roll the transaction after reaping an extent.  We don't
 | |
|  * want to overrun the transaction reservation, so we prohibit more than
 | |
|  * 128 EFIs per transaction.  For the same reason, we limit the number
 | |
|  * of buffer invalidations to 2048.
 | |
|  */
 | |
| static inline bool xreap_want_roll(const struct xreap_state *rs)
 | |
| {
 | |
| 	if (rs->force_roll)
 | |
| 		return true;
 | |
| 	if (rs->deferred > XREP_MAX_ITRUNCATE_EFIS)
 | |
| 		return true;
 | |
| 	if (rs->invalidated > XREAP_MAX_BINVAL)
 | |
| 		return true;
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| static inline void xreap_reset(struct xreap_state *rs)
 | |
| {
 | |
| 	rs->total_deferred += rs->deferred;
 | |
| 	rs->deferred = 0;
 | |
| 	rs->invalidated = 0;
 | |
| 	rs->force_roll = false;
 | |
| }
 | |
| 
 | |
| #define XREAP_MAX_DEFER_CHAIN		(2048)
 | |
| 
 | |
| /*
 | |
|  * Decide if we want to finish the deferred ops that are attached to the scrub
 | |
|  * transaction.  We don't want to queue huge chains of deferred ops because
 | |
|  * that can consume a lot of log space and kernel memory.  Hence we trigger a
 | |
|  * xfs_defer_finish if there are more than 2048 deferred reap operations or the
 | |
|  * caller did some real work.
 | |
|  */
 | |
| static inline bool
 | |
| xreap_want_defer_finish(const struct xreap_state *rs)
 | |
| {
 | |
| 	if (rs->force_roll)
 | |
| 		return true;
 | |
| 	if (rs->total_deferred > XREAP_MAX_DEFER_CHAIN)
 | |
| 		return true;
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| static inline void xreap_defer_finish_reset(struct xreap_state *rs)
 | |
| {
 | |
| 	rs->total_deferred = 0;
 | |
| 	rs->deferred = 0;
 | |
| 	rs->invalidated = 0;
 | |
| 	rs->force_roll = false;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Compute the maximum length of a buffer cache scan (in units of sectors),
 | |
|  * given a quantity of fs blocks.
 | |
|  */
 | |
| xfs_daddr_t
 | |
| xrep_bufscan_max_sectors(
 | |
| 	struct xfs_mount	*mp,
 | |
| 	xfs_extlen_t		fsblocks)
 | |
| {
 | |
| 	int			max_fsbs;
 | |
| 
 | |
| 	/* Remote xattr values are the largest buffers that we support. */
 | |
| 	max_fsbs = xfs_attr3_max_rmt_blocks(mp);
 | |
| 
 | |
| 	return XFS_FSB_TO_BB(mp, min_t(xfs_extlen_t, fsblocks, max_fsbs));
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Return an incore buffer from a sector scan, or NULL if there are no buffers
 | |
|  * left to return.
 | |
|  */
 | |
| struct xfs_buf *
 | |
| xrep_bufscan_advance(
 | |
| 	struct xfs_mount	*mp,
 | |
| 	struct xrep_bufscan	*scan)
 | |
| {
 | |
| 	scan->__sector_count += scan->daddr_step;
 | |
| 	while (scan->__sector_count <= scan->max_sectors) {
 | |
| 		struct xfs_buf	*bp = NULL;
 | |
| 		int		error;
 | |
| 
 | |
| 		error = xfs_buf_incore(mp->m_ddev_targp, scan->daddr,
 | |
| 				scan->__sector_count, XBF_LIVESCAN, &bp);
 | |
| 		if (!error)
 | |
| 			return bp;
 | |
| 
 | |
| 		scan->__sector_count += scan->daddr_step;
 | |
| 	}
 | |
| 
 | |
| 	return NULL;
 | |
| }
 | |
| 
 | |
| /* Try to invalidate the incore buffers for an extent that we're freeing. */
 | |
| STATIC void
 | |
| xreap_agextent_binval(
 | |
| 	struct xreap_state	*rs,
 | |
| 	xfs_agblock_t		agbno,
 | |
| 	xfs_extlen_t		*aglenp)
 | |
| {
 | |
| 	struct xfs_scrub	*sc = rs->sc;
 | |
| 	struct xfs_perag	*pag = sc->sa.pag;
 | |
| 	struct xfs_mount	*mp = sc->mp;
 | |
| 	xfs_agnumber_t		agno = sc->sa.pag->pag_agno;
 | |
| 	xfs_agblock_t		agbno_next = agbno + *aglenp;
 | |
| 	xfs_agblock_t		bno = agbno;
 | |
| 
 | |
| 	/*
 | |
| 	 * Avoid invalidating AG headers and post-EOFS blocks because we never
 | |
| 	 * own those.
 | |
| 	 */
 | |
| 	if (!xfs_verify_agbno(pag, agbno) ||
 | |
| 	    !xfs_verify_agbno(pag, agbno_next - 1))
 | |
| 		return;
 | |
| 
 | |
| 	/*
 | |
| 	 * If there are incore buffers for these blocks, invalidate them.  We
 | |
| 	 * assume that the lack of any other known owners means that the buffer
 | |
| 	 * can be locked without risk of deadlocking.  The buffer cache cannot
 | |
| 	 * detect aliasing, so employ nested loops to scan for incore buffers
 | |
| 	 * of any plausible size.
 | |
| 	 */
 | |
| 	while (bno < agbno_next) {
 | |
| 		struct xrep_bufscan	scan = {
 | |
| 			.daddr		= XFS_AGB_TO_DADDR(mp, agno, bno),
 | |
| 			.max_sectors	= xrep_bufscan_max_sectors(mp,
 | |
| 							agbno_next - bno),
 | |
| 			.daddr_step	= XFS_FSB_TO_BB(mp, 1),
 | |
| 		};
 | |
| 		struct xfs_buf	*bp;
 | |
| 
 | |
| 		while ((bp = xrep_bufscan_advance(mp, &scan)) != NULL) {
 | |
| 			xfs_trans_bjoin(sc->tp, bp);
 | |
| 			xfs_trans_binval(sc->tp, bp);
 | |
| 			rs->invalidated++;
 | |
| 
 | |
| 			/*
 | |
| 			 * Stop invalidating if we've hit the limit; we should
 | |
| 			 * still have enough reservation left to free however
 | |
| 			 * far we've gotten.
 | |
| 			 */
 | |
| 			if (rs->invalidated > XREAP_MAX_BINVAL) {
 | |
| 				*aglenp -= agbno_next - bno;
 | |
| 				goto out;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		bno++;
 | |
| 	}
 | |
| 
 | |
| out:
 | |
| 	trace_xreap_agextent_binval(sc->sa.pag, agbno, *aglenp);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Figure out the longest run of blocks that we can dispose of with a single
 | |
|  * call.  Cross-linked blocks should have their reverse mappings removed, but
 | |
|  * single-owner extents can be freed.  AGFL blocks can only be put back one at
 | |
|  * a time.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_agextent_select(
 | |
| 	struct xreap_state	*rs,
 | |
| 	xfs_agblock_t		agbno,
 | |
| 	xfs_agblock_t		agbno_next,
 | |
| 	bool			*crosslinked,
 | |
| 	xfs_extlen_t		*aglenp)
 | |
| {
 | |
| 	struct xfs_scrub	*sc = rs->sc;
 | |
| 	struct xfs_btree_cur	*cur;
 | |
| 	xfs_agblock_t		bno = agbno + 1;
 | |
| 	xfs_extlen_t		len = 1;
 | |
| 	int			error;
 | |
| 
 | |
| 	/*
 | |
| 	 * Determine if there are any other rmap records covering the first
 | |
| 	 * block of this extent.  If so, the block is crosslinked.
 | |
| 	 */
 | |
| 	cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
 | |
| 			sc->sa.pag);
 | |
| 	error = xfs_rmap_has_other_keys(cur, agbno, 1, rs->oinfo,
 | |
| 			crosslinked);
 | |
| 	if (error)
 | |
| 		goto out_cur;
 | |
| 
 | |
| 	/* AGFL blocks can only be deal with one at a time. */
 | |
| 	if (rs->resv == XFS_AG_RESV_AGFL)
 | |
| 		goto out_found;
 | |
| 
 | |
| 	/*
 | |
| 	 * Figure out how many of the subsequent blocks have the same crosslink
 | |
| 	 * status.
 | |
| 	 */
 | |
| 	while (bno < agbno_next) {
 | |
| 		bool		also_crosslinked;
 | |
| 
 | |
| 		error = xfs_rmap_has_other_keys(cur, bno, 1, rs->oinfo,
 | |
| 				&also_crosslinked);
 | |
| 		if (error)
 | |
| 			goto out_cur;
 | |
| 
 | |
| 		if (*crosslinked != also_crosslinked)
 | |
| 			break;
 | |
| 
 | |
| 		len++;
 | |
| 		bno++;
 | |
| 	}
 | |
| 
 | |
| out_found:
 | |
| 	*aglenp = len;
 | |
| 	trace_xreap_agextent_select(sc->sa.pag, agbno, len, *crosslinked);
 | |
| out_cur:
 | |
| 	xfs_btree_del_cursor(cur, error);
 | |
| 	return error;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Dispose of as much of the beginning of this AG extent as possible.  The
 | |
|  * number of blocks disposed of will be returned in @aglenp.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_agextent_iter(
 | |
| 	struct xreap_state	*rs,
 | |
| 	xfs_agblock_t		agbno,
 | |
| 	xfs_extlen_t		*aglenp,
 | |
| 	bool			crosslinked)
 | |
| {
 | |
| 	struct xfs_scrub	*sc = rs->sc;
 | |
| 	xfs_fsblock_t		fsbno;
 | |
| 	int			error = 0;
 | |
| 
 | |
| 	fsbno = XFS_AGB_TO_FSB(sc->mp, sc->sa.pag->pag_agno, agbno);
 | |
| 
 | |
| 	/*
 | |
| 	 * If there are other rmappings, this block is cross linked and must
 | |
| 	 * not be freed.  Remove the reverse mapping and move on.  Otherwise,
 | |
| 	 * we were the only owner of the block, so free the extent, which will
 | |
| 	 * also remove the rmap.
 | |
| 	 *
 | |
| 	 * XXX: XFS doesn't support detecting the case where a single block
 | |
| 	 * metadata structure is crosslinked with a multi-block structure
 | |
| 	 * because the buffer cache doesn't detect aliasing problems, so we
 | |
| 	 * can't fix 100% of crosslinking problems (yet).  The verifiers will
 | |
| 	 * blow on writeout, the filesystem will shut down, and the admin gets
 | |
| 	 * to run xfs_repair.
 | |
| 	 */
 | |
| 	if (crosslinked) {
 | |
| 		trace_xreap_dispose_unmap_extent(sc->sa.pag, agbno, *aglenp);
 | |
| 
 | |
| 		rs->force_roll = true;
 | |
| 
 | |
| 		if (rs->oinfo == &XFS_RMAP_OINFO_COW) {
 | |
| 			/*
 | |
| 			 * If we're unmapping CoW staging extents, remove the
 | |
| 			 * records from the refcountbt, which will remove the
 | |
| 			 * rmap record as well.
 | |
| 			 */
 | |
| 			xfs_refcount_free_cow_extent(sc->tp, fsbno, *aglenp);
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		return xfs_rmap_free(sc->tp, sc->sa.agf_bp, sc->sa.pag, agbno,
 | |
| 				*aglenp, rs->oinfo);
 | |
| 	}
 | |
| 
 | |
| 	trace_xreap_dispose_free_extent(sc->sa.pag, agbno, *aglenp);
 | |
| 
 | |
| 	/*
 | |
| 	 * Invalidate as many buffers as we can, starting at agbno.  If this
 | |
| 	 * function sets *aglenp to zero, the transaction is full of logged
 | |
| 	 * buffer invalidations, so we need to return early so that we can
 | |
| 	 * roll and retry.
 | |
| 	 */
 | |
| 	xreap_agextent_binval(rs, agbno, aglenp);
 | |
| 	if (*aglenp == 0) {
 | |
| 		ASSERT(xreap_want_roll(rs));
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * If we're getting rid of CoW staging extents, use deferred work items
 | |
| 	 * to remove the refcountbt records (which removes the rmap records)
 | |
| 	 * and free the extent.  We're not worried about the system going down
 | |
| 	 * here because log recovery walks the refcount btree to clean out the
 | |
| 	 * CoW staging extents.
 | |
| 	 */
 | |
| 	if (rs->oinfo == &XFS_RMAP_OINFO_COW) {
 | |
| 		ASSERT(rs->resv == XFS_AG_RESV_NONE);
 | |
| 
 | |
| 		xfs_refcount_free_cow_extent(sc->tp, fsbno, *aglenp);
 | |
| 		error = xfs_free_extent_later(sc->tp, fsbno, *aglenp, NULL,
 | |
| 				rs->resv, XFS_FREE_EXTENT_SKIP_DISCARD);
 | |
| 		if (error)
 | |
| 			return error;
 | |
| 
 | |
| 		rs->force_roll = true;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/* Put blocks back on the AGFL one at a time. */
 | |
| 	if (rs->resv == XFS_AG_RESV_AGFL) {
 | |
| 		ASSERT(*aglenp == 1);
 | |
| 		error = xreap_put_freelist(sc, agbno);
 | |
| 		if (error)
 | |
| 			return error;
 | |
| 
 | |
| 		rs->force_roll = true;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Use deferred frees to get rid of the old btree blocks to try to
 | |
| 	 * minimize the window in which we could crash and lose the old blocks.
 | |
| 	 * Add a defer ops barrier every other extent to avoid stressing the
 | |
| 	 * system with large EFIs.
 | |
| 	 */
 | |
| 	error = xfs_free_extent_later(sc->tp, fsbno, *aglenp, rs->oinfo,
 | |
| 			rs->resv, XFS_FREE_EXTENT_SKIP_DISCARD);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 
 | |
| 	rs->deferred++;
 | |
| 	if (rs->deferred % 2 == 0)
 | |
| 		xfs_defer_add_barrier(sc->tp);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Break an AG metadata extent into sub-extents by fate (crosslinked, not
 | |
|  * crosslinked), and dispose of each sub-extent separately.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_agmeta_extent(
 | |
| 	uint32_t		agbno,
 | |
| 	uint32_t		len,
 | |
| 	void			*priv)
 | |
| {
 | |
| 	struct xreap_state	*rs = priv;
 | |
| 	struct xfs_scrub	*sc = rs->sc;
 | |
| 	xfs_agblock_t		agbno_next = agbno + len;
 | |
| 	int			error = 0;
 | |
| 
 | |
| 	ASSERT(len <= XFS_MAX_BMBT_EXTLEN);
 | |
| 	ASSERT(sc->ip == NULL);
 | |
| 
 | |
| 	while (agbno < agbno_next) {
 | |
| 		xfs_extlen_t	aglen;
 | |
| 		bool		crosslinked;
 | |
| 
 | |
| 		error = xreap_agextent_select(rs, agbno, agbno_next,
 | |
| 				&crosslinked, &aglen);
 | |
| 		if (error)
 | |
| 			return error;
 | |
| 
 | |
| 		error = xreap_agextent_iter(rs, agbno, &aglen, crosslinked);
 | |
| 		if (error)
 | |
| 			return error;
 | |
| 
 | |
| 		if (xreap_want_defer_finish(rs)) {
 | |
| 			error = xrep_defer_finish(sc);
 | |
| 			if (error)
 | |
| 				return error;
 | |
| 			xreap_defer_finish_reset(rs);
 | |
| 		} else if (xreap_want_roll(rs)) {
 | |
| 			error = xrep_roll_ag_trans(sc);
 | |
| 			if (error)
 | |
| 				return error;
 | |
| 			xreap_reset(rs);
 | |
| 		}
 | |
| 
 | |
| 		agbno += aglen;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Dispose of every block of every AG metadata extent in the bitmap. */
 | |
| int
 | |
| xrep_reap_agblocks(
 | |
| 	struct xfs_scrub		*sc,
 | |
| 	struct xagb_bitmap		*bitmap,
 | |
| 	const struct xfs_owner_info	*oinfo,
 | |
| 	enum xfs_ag_resv_type		type)
 | |
| {
 | |
| 	struct xreap_state		rs = {
 | |
| 		.sc			= sc,
 | |
| 		.oinfo			= oinfo,
 | |
| 		.resv			= type,
 | |
| 	};
 | |
| 	int				error;
 | |
| 
 | |
| 	ASSERT(xfs_has_rmapbt(sc->mp));
 | |
| 	ASSERT(sc->ip == NULL);
 | |
| 
 | |
| 	error = xagb_bitmap_walk(bitmap, xreap_agmeta_extent, &rs);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 
 | |
| 	if (xreap_dirty(&rs))
 | |
| 		return xrep_defer_finish(sc);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Break a file metadata extent into sub-extents by fate (crosslinked, not
 | |
|  * crosslinked), and dispose of each sub-extent separately.  The extent must
 | |
|  * not cross an AG boundary.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_fsmeta_extent(
 | |
| 	uint64_t		fsbno,
 | |
| 	uint64_t		len,
 | |
| 	void			*priv)
 | |
| {
 | |
| 	struct xreap_state	*rs = priv;
 | |
| 	struct xfs_scrub	*sc = rs->sc;
 | |
| 	xfs_agnumber_t		agno = XFS_FSB_TO_AGNO(sc->mp, fsbno);
 | |
| 	xfs_agblock_t		agbno = XFS_FSB_TO_AGBNO(sc->mp, fsbno);
 | |
| 	xfs_agblock_t		agbno_next = agbno + len;
 | |
| 	int			error = 0;
 | |
| 
 | |
| 	ASSERT(len <= XFS_MAX_BMBT_EXTLEN);
 | |
| 	ASSERT(sc->ip != NULL);
 | |
| 	ASSERT(!sc->sa.pag);
 | |
| 
 | |
| 	/*
 | |
| 	 * We're reaping blocks after repairing file metadata, which means that
 | |
| 	 * we have to init the xchk_ag structure ourselves.
 | |
| 	 */
 | |
| 	sc->sa.pag = xfs_perag_get(sc->mp, agno);
 | |
| 	if (!sc->sa.pag)
 | |
| 		return -EFSCORRUPTED;
 | |
| 
 | |
| 	error = xfs_alloc_read_agf(sc->sa.pag, sc->tp, 0, &sc->sa.agf_bp);
 | |
| 	if (error)
 | |
| 		goto out_pag;
 | |
| 
 | |
| 	while (agbno < agbno_next) {
 | |
| 		xfs_extlen_t	aglen;
 | |
| 		bool		crosslinked;
 | |
| 
 | |
| 		error = xreap_agextent_select(rs, agbno, agbno_next,
 | |
| 				&crosslinked, &aglen);
 | |
| 		if (error)
 | |
| 			goto out_agf;
 | |
| 
 | |
| 		error = xreap_agextent_iter(rs, agbno, &aglen, crosslinked);
 | |
| 		if (error)
 | |
| 			goto out_agf;
 | |
| 
 | |
| 		if (xreap_want_defer_finish(rs)) {
 | |
| 			/*
 | |
| 			 * Holds the AGF buffer across the deferred chain
 | |
| 			 * processing.
 | |
| 			 */
 | |
| 			error = xrep_defer_finish(sc);
 | |
| 			if (error)
 | |
| 				goto out_agf;
 | |
| 			xreap_defer_finish_reset(rs);
 | |
| 		} else if (xreap_want_roll(rs)) {
 | |
| 			/*
 | |
| 			 * Hold the AGF buffer across the transaction roll so
 | |
| 			 * that we don't have to reattach it to the scrub
 | |
| 			 * context.
 | |
| 			 */
 | |
| 			xfs_trans_bhold(sc->tp, sc->sa.agf_bp);
 | |
| 			error = xfs_trans_roll_inode(&sc->tp, sc->ip);
 | |
| 			xfs_trans_bjoin(sc->tp, sc->sa.agf_bp);
 | |
| 			if (error)
 | |
| 				goto out_agf;
 | |
| 			xreap_reset(rs);
 | |
| 		}
 | |
| 
 | |
| 		agbno += aglen;
 | |
| 	}
 | |
| 
 | |
| out_agf:
 | |
| 	xfs_trans_brelse(sc->tp, sc->sa.agf_bp);
 | |
| 	sc->sa.agf_bp = NULL;
 | |
| out_pag:
 | |
| 	xfs_perag_put(sc->sa.pag);
 | |
| 	sc->sa.pag = NULL;
 | |
| 	return error;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Dispose of every block of every fs metadata extent in the bitmap.
 | |
|  * Do not use this to dispose of the mappings in an ondisk inode fork.
 | |
|  */
 | |
| int
 | |
| xrep_reap_fsblocks(
 | |
| 	struct xfs_scrub		*sc,
 | |
| 	struct xfsb_bitmap		*bitmap,
 | |
| 	const struct xfs_owner_info	*oinfo)
 | |
| {
 | |
| 	struct xreap_state		rs = {
 | |
| 		.sc			= sc,
 | |
| 		.oinfo			= oinfo,
 | |
| 		.resv			= XFS_AG_RESV_NONE,
 | |
| 	};
 | |
| 	int				error;
 | |
| 
 | |
| 	ASSERT(xfs_has_rmapbt(sc->mp));
 | |
| 	ASSERT(sc->ip != NULL);
 | |
| 
 | |
| 	error = xfsb_bitmap_walk(bitmap, xreap_fsmeta_extent, &rs);
 | |
| 	if (error)
 | |
| 		return error;
 | |
| 
 | |
| 	if (xreap_dirty(&rs))
 | |
| 		return xrep_defer_finish(sc);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Metadata files are not supposed to share blocks with anything else.
 | |
|  * If blocks are shared, we remove the reverse mapping (thus reducing the
 | |
|  * crosslink factor); if blocks are not shared, we also need to free them.
 | |
|  *
 | |
|  * This first step determines the longest subset of the passed-in imap
 | |
|  * (starting at its beginning) that is either crosslinked or not crosslinked.
 | |
|  * The blockcount will be adjust down as needed.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_bmapi_select(
 | |
| 	struct xfs_scrub	*sc,
 | |
| 	struct xfs_inode	*ip,
 | |
| 	int			whichfork,
 | |
| 	struct xfs_bmbt_irec	*imap,
 | |
| 	bool			*crosslinked)
 | |
| {
 | |
| 	struct xfs_owner_info	oinfo;
 | |
| 	struct xfs_btree_cur	*cur;
 | |
| 	xfs_filblks_t		len = 1;
 | |
| 	xfs_agblock_t		bno;
 | |
| 	xfs_agblock_t		agbno;
 | |
| 	xfs_agblock_t		agbno_next;
 | |
| 	int			error;
 | |
| 
 | |
| 	agbno = XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock);
 | |
| 	agbno_next = agbno + imap->br_blockcount;
 | |
| 
 | |
| 	cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
 | |
| 			sc->sa.pag);
 | |
| 
 | |
| 	xfs_rmap_ino_owner(&oinfo, ip->i_ino, whichfork, imap->br_startoff);
 | |
| 	error = xfs_rmap_has_other_keys(cur, agbno, 1, &oinfo, crosslinked);
 | |
| 	if (error)
 | |
| 		goto out_cur;
 | |
| 
 | |
| 	bno = agbno + 1;
 | |
| 	while (bno < agbno_next) {
 | |
| 		bool		also_crosslinked;
 | |
| 
 | |
| 		oinfo.oi_offset++;
 | |
| 		error = xfs_rmap_has_other_keys(cur, bno, 1, &oinfo,
 | |
| 				&also_crosslinked);
 | |
| 		if (error)
 | |
| 			goto out_cur;
 | |
| 
 | |
| 		if (also_crosslinked != *crosslinked)
 | |
| 			break;
 | |
| 
 | |
| 		len++;
 | |
| 		bno++;
 | |
| 	}
 | |
| 
 | |
| 	imap->br_blockcount = len;
 | |
| 	trace_xreap_bmapi_select(sc->sa.pag, agbno, len, *crosslinked);
 | |
| out_cur:
 | |
| 	xfs_btree_del_cursor(cur, error);
 | |
| 	return error;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Decide if this buffer can be joined to a transaction.  This is true for most
 | |
|  * buffers, but there are two cases that we want to catch: large remote xattr
 | |
|  * value buffers are not logged and can overflow the buffer log item dirty
 | |
|  * bitmap size; and oversized cached buffers if things have really gone
 | |
|  * haywire.
 | |
|  */
 | |
| static inline bool
 | |
| xreap_buf_loggable(
 | |
| 	const struct xfs_buf	*bp)
 | |
| {
 | |
| 	int			i;
 | |
| 
 | |
| 	for (i = 0; i < bp->b_map_count; i++) {
 | |
| 		int		chunks;
 | |
| 		int		map_size;
 | |
| 
 | |
| 		chunks = DIV_ROUND_UP(BBTOB(bp->b_maps[i].bm_len),
 | |
| 				XFS_BLF_CHUNK);
 | |
| 		map_size = DIV_ROUND_UP(chunks, NBWORD);
 | |
| 		if (map_size > XFS_BLF_DATAMAP_SIZE)
 | |
| 			return false;
 | |
| 	}
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Invalidate any buffers for this file mapping.  The @imap blockcount may be
 | |
|  * adjusted downward if we need to roll the transaction.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_bmapi_binval(
 | |
| 	struct xfs_scrub	*sc,
 | |
| 	struct xfs_inode	*ip,
 | |
| 	int			whichfork,
 | |
| 	struct xfs_bmbt_irec	*imap)
 | |
| {
 | |
| 	struct xfs_mount	*mp = sc->mp;
 | |
| 	struct xfs_perag	*pag = sc->sa.pag;
 | |
| 	int			bmap_flags = xfs_bmapi_aflag(whichfork);
 | |
| 	xfs_fileoff_t		off;
 | |
| 	xfs_fileoff_t		max_off;
 | |
| 	xfs_extlen_t		scan_blocks;
 | |
| 	xfs_agnumber_t		agno = sc->sa.pag->pag_agno;
 | |
| 	xfs_agblock_t		bno;
 | |
| 	xfs_agblock_t		agbno;
 | |
| 	xfs_agblock_t		agbno_next;
 | |
| 	unsigned int		invalidated = 0;
 | |
| 	int			error;
 | |
| 
 | |
| 	/*
 | |
| 	 * Avoid invalidating AG headers and post-EOFS blocks because we never
 | |
| 	 * own those.
 | |
| 	 */
 | |
| 	agbno = bno = XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock);
 | |
| 	agbno_next = agbno + imap->br_blockcount;
 | |
| 	if (!xfs_verify_agbno(pag, agbno) ||
 | |
| 	    !xfs_verify_agbno(pag, agbno_next - 1))
 | |
| 		return 0;
 | |
| 
 | |
| 	/*
 | |
| 	 * Buffers for file blocks can span multiple contiguous mappings.  This
 | |
| 	 * means that for each block in the mapping, there could exist an
 | |
| 	 * xfs_buf indexed by that block with any length up to the maximum
 | |
| 	 * buffer size (remote xattr values) or to the next hole in the fork.
 | |
| 	 * To set up our binval scan, first we need to figure out the location
 | |
| 	 * of the next hole.
 | |
| 	 */
 | |
| 	off = imap->br_startoff + imap->br_blockcount;
 | |
| 	max_off = off + xfs_attr3_max_rmt_blocks(mp);
 | |
| 	while (off < max_off) {
 | |
| 		struct xfs_bmbt_irec	hmap;
 | |
| 		int			nhmaps = 1;
 | |
| 
 | |
| 		error = xfs_bmapi_read(ip, off, max_off - off, &hmap,
 | |
| 				&nhmaps, bmap_flags);
 | |
| 		if (error)
 | |
| 			return error;
 | |
| 		if (nhmaps != 1 || hmap.br_startblock == DELAYSTARTBLOCK) {
 | |
| 			ASSERT(0);
 | |
| 			return -EFSCORRUPTED;
 | |
| 		}
 | |
| 
 | |
| 		if (!xfs_bmap_is_real_extent(&hmap))
 | |
| 			break;
 | |
| 
 | |
| 		off = hmap.br_startoff + hmap.br_blockcount;
 | |
| 	}
 | |
| 	scan_blocks = off - imap->br_startoff;
 | |
| 
 | |
| 	trace_xreap_bmapi_binval_scan(sc, imap, scan_blocks);
 | |
| 
 | |
| 	/*
 | |
| 	 * If there are incore buffers for these blocks, invalidate them.  If
 | |
| 	 * we can't (try)lock the buffer we assume it's owned by someone else
 | |
| 	 * and leave it alone.  The buffer cache cannot detect aliasing, so
 | |
| 	 * employ nested loops to detect incore buffers of any plausible size.
 | |
| 	 */
 | |
| 	while (bno < agbno_next) {
 | |
| 		struct xrep_bufscan	scan = {
 | |
| 			.daddr		= XFS_AGB_TO_DADDR(mp, agno, bno),
 | |
| 			.max_sectors	= xrep_bufscan_max_sectors(mp,
 | |
| 								scan_blocks),
 | |
| 			.daddr_step	= XFS_FSB_TO_BB(mp, 1),
 | |
| 		};
 | |
| 		struct xfs_buf		*bp;
 | |
| 
 | |
| 		while ((bp = xrep_bufscan_advance(mp, &scan)) != NULL) {
 | |
| 			if (xreap_buf_loggable(bp)) {
 | |
| 				xfs_trans_bjoin(sc->tp, bp);
 | |
| 				xfs_trans_binval(sc->tp, bp);
 | |
| 			} else {
 | |
| 				xfs_buf_stale(bp);
 | |
| 				xfs_buf_relse(bp);
 | |
| 			}
 | |
| 			invalidated++;
 | |
| 
 | |
| 			/*
 | |
| 			 * Stop invalidating if we've hit the limit; we should
 | |
| 			 * still have enough reservation left to free however
 | |
| 			 * much of the mapping we've seen so far.
 | |
| 			 */
 | |
| 			if (invalidated > XREAP_MAX_BINVAL) {
 | |
| 				imap->br_blockcount = agbno_next - bno;
 | |
| 				goto out;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		bno++;
 | |
| 		scan_blocks--;
 | |
| 	}
 | |
| 
 | |
| out:
 | |
| 	trace_xreap_bmapi_binval(sc->sa.pag, agbno, imap->br_blockcount);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Dispose of as much of the beginning of this file fork mapping as possible.
 | |
|  * The number of blocks disposed of is returned in @imap->br_blockcount.
 | |
|  */
 | |
| STATIC int
 | |
| xrep_reap_bmapi_iter(
 | |
| 	struct xfs_scrub		*sc,
 | |
| 	struct xfs_inode		*ip,
 | |
| 	int				whichfork,
 | |
| 	struct xfs_bmbt_irec		*imap,
 | |
| 	bool				crosslinked)
 | |
| {
 | |
| 	int				error;
 | |
| 
 | |
| 	if (crosslinked) {
 | |
| 		/*
 | |
| 		 * If there are other rmappings, this block is cross linked and
 | |
| 		 * must not be freed.  Remove the reverse mapping, leave the
 | |
| 		 * buffer cache in its possibly confused state, and move on.
 | |
| 		 * We don't want to risk discarding valid data buffers from
 | |
| 		 * anybody else who thinks they own the block, even though that
 | |
| 		 * runs the risk of stale buffer warnings in the future.
 | |
| 		 */
 | |
| 		trace_xreap_dispose_unmap_extent(sc->sa.pag,
 | |
| 				XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock),
 | |
| 				imap->br_blockcount);
 | |
| 
 | |
| 		/*
 | |
| 		 * Schedule removal of the mapping from the fork.  We use
 | |
| 		 * deferred log intents in this function to control the exact
 | |
| 		 * sequence of metadata updates.
 | |
| 		 */
 | |
| 		xfs_bmap_unmap_extent(sc->tp, ip, whichfork, imap);
 | |
| 		xfs_trans_mod_dquot_byino(sc->tp, ip, XFS_TRANS_DQ_BCOUNT,
 | |
| 				-(int64_t)imap->br_blockcount);
 | |
| 		xfs_rmap_unmap_extent(sc->tp, ip, whichfork, imap);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * If the block is not crosslinked, we can invalidate all the incore
 | |
| 	 * buffers for the extent, and then free the extent.  This is a bit of
 | |
| 	 * a mess since we don't detect discontiguous buffers that are indexed
 | |
| 	 * by a block starting before the first block of the extent but overlap
 | |
| 	 * anyway.
 | |
| 	 */
 | |
| 	trace_xreap_dispose_free_extent(sc->sa.pag,
 | |
| 			XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock),
 | |
| 			imap->br_blockcount);
 | |
| 
 | |
| 	/*
 | |
| 	 * Invalidate as many buffers as we can, starting at the beginning of
 | |
| 	 * this mapping.  If this function sets blockcount to zero, the
 | |
| 	 * transaction is full of logged buffer invalidations, so we need to
 | |
| 	 * return early so that we can roll and retry.
 | |
| 	 */
 | |
| 	error = xreap_bmapi_binval(sc, ip, whichfork, imap);
 | |
| 	if (error || imap->br_blockcount == 0)
 | |
| 		return error;
 | |
| 
 | |
| 	/*
 | |
| 	 * Schedule removal of the mapping from the fork.  We use deferred log
 | |
| 	 * intents in this function to control the exact sequence of metadata
 | |
| 	 * updates.
 | |
| 	 */
 | |
| 	xfs_bmap_unmap_extent(sc->tp, ip, whichfork, imap);
 | |
| 	xfs_trans_mod_dquot_byino(sc->tp, ip, XFS_TRANS_DQ_BCOUNT,
 | |
| 			-(int64_t)imap->br_blockcount);
 | |
| 	return xfs_free_extent_later(sc->tp, imap->br_startblock,
 | |
| 			imap->br_blockcount, NULL, XFS_AG_RESV_NONE,
 | |
| 			XFS_FREE_EXTENT_SKIP_DISCARD);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Dispose of as much of this file extent as we can.  Upon successful return,
 | |
|  * the imap will reflect the mapping that was removed from the fork.
 | |
|  */
 | |
| STATIC int
 | |
| xreap_ifork_extent(
 | |
| 	struct xfs_scrub		*sc,
 | |
| 	struct xfs_inode		*ip,
 | |
| 	int				whichfork,
 | |
| 	struct xfs_bmbt_irec		*imap)
 | |
| {
 | |
| 	xfs_agnumber_t			agno;
 | |
| 	bool				crosslinked;
 | |
| 	int				error;
 | |
| 
 | |
| 	ASSERT(sc->sa.pag == NULL);
 | |
| 
 | |
| 	trace_xreap_ifork_extent(sc, ip, whichfork, imap);
 | |
| 
 | |
| 	agno = XFS_FSB_TO_AGNO(sc->mp, imap->br_startblock);
 | |
| 	sc->sa.pag = xfs_perag_get(sc->mp, agno);
 | |
| 	if (!sc->sa.pag)
 | |
| 		return -EFSCORRUPTED;
 | |
| 
 | |
| 	error = xfs_alloc_read_agf(sc->sa.pag, sc->tp, 0, &sc->sa.agf_bp);
 | |
| 	if (error)
 | |
| 		goto out_pag;
 | |
| 
 | |
| 	/*
 | |
| 	 * Decide the fate of the blocks at the beginning of the mapping, then
 | |
| 	 * update the mapping to use it with the unmap calls.
 | |
| 	 */
 | |
| 	error = xreap_bmapi_select(sc, ip, whichfork, imap, &crosslinked);
 | |
| 	if (error)
 | |
| 		goto out_agf;
 | |
| 
 | |
| 	error = xrep_reap_bmapi_iter(sc, ip, whichfork, imap, crosslinked);
 | |
| 	if (error)
 | |
| 		goto out_agf;
 | |
| 
 | |
| out_agf:
 | |
| 	xfs_trans_brelse(sc->tp, sc->sa.agf_bp);
 | |
| 	sc->sa.agf_bp = NULL;
 | |
| out_pag:
 | |
| 	xfs_perag_put(sc->sa.pag);
 | |
| 	sc->sa.pag = NULL;
 | |
| 	return error;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Dispose of each block mapped to the given fork of the given file.  Callers
 | |
|  * must hold ILOCK_EXCL, and ip can only be sc->ip or sc->tempip.  The fork
 | |
|  * must not have any delalloc reservations.
 | |
|  */
 | |
| int
 | |
| xrep_reap_ifork(
 | |
| 	struct xfs_scrub	*sc,
 | |
| 	struct xfs_inode	*ip,
 | |
| 	int			whichfork)
 | |
| {
 | |
| 	xfs_fileoff_t		off = 0;
 | |
| 	int			bmap_flags = xfs_bmapi_aflag(whichfork);
 | |
| 	int			error;
 | |
| 
 | |
| 	ASSERT(xfs_has_rmapbt(sc->mp));
 | |
| 	ASSERT(ip == sc->ip || ip == sc->tempip);
 | |
| 	ASSERT(whichfork == XFS_ATTR_FORK || !XFS_IS_REALTIME_INODE(ip));
 | |
| 
 | |
| 	while (off < XFS_MAX_FILEOFF) {
 | |
| 		struct xfs_bmbt_irec	imap;
 | |
| 		int			nimaps = 1;
 | |
| 
 | |
| 		/* Read the next extent, skip past holes and delalloc. */
 | |
| 		error = xfs_bmapi_read(ip, off, XFS_MAX_FILEOFF - off, &imap,
 | |
| 				&nimaps, bmap_flags);
 | |
| 		if (error)
 | |
| 			return error;
 | |
| 		if (nimaps != 1 || imap.br_startblock == DELAYSTARTBLOCK) {
 | |
| 			ASSERT(0);
 | |
| 			return -EFSCORRUPTED;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * If this is a real space mapping, reap as much of it as we
 | |
| 		 * can in a single transaction.
 | |
| 		 */
 | |
| 		if (xfs_bmap_is_real_extent(&imap)) {
 | |
| 			error = xreap_ifork_extent(sc, ip, whichfork, &imap);
 | |
| 			if (error)
 | |
| 				return error;
 | |
| 
 | |
| 			error = xfs_defer_finish(&sc->tp);
 | |
| 			if (error)
 | |
| 				return error;
 | |
| 		}
 | |
| 
 | |
| 		off = imap.br_startoff + imap.br_blockcount;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 |