0dc63c8a1c
As we've noted in various places, all current users of in-memory btrees are online fsck. Online fsck only stages a btree long enough to rebuild an ondisk data structure, which means that the in-memory btree is ephemeral. Furthermore, if we encounter /any/ errors while updating an in-memory btree, all we do is tear down all the staged data and return an errno to userspace. In-memory btrees need not be transactional, so their buffers should not be committed to the ondisk log, nor should they be checkpointed by the AIL. That's just as well since the ephemeral nature of the btree means that the buftarg and the buffers may disappear quickly anyway. Therefore, we need a way to launder the btree buffers that get attached to the transaction by the generic btree code. Because the buffers are directly mapped to backing file pages, there's no need to bwrite them back to the tmpfs file. All we need to do is clean enough of the buffer log item state so that the bli can be detached from the buffer, remove the bli from the transaction's log item list, and reset the transaction dirty state as if the laundered items had never been there. For simplicity, create xfbtree transaction commit and cancel helpers that launder the in-memory btree buffers for callers. Once laundered, call the write verifier on non-stale buffers to avoid integrity issues, or punch a hole in the backing file for stale buffers. Signed-off-by: Darrick J. Wong <djwong@kernel.org> Reviewed-by: Christoph Hellwig <hch@lst.de>
76 lines
2.2 KiB
C
76 lines
2.2 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (c) 2021-2024 Oracle. All Rights Reserved.
|
|
* Author: Darrick J. Wong <djwong@kernel.org>
|
|
*/
|
|
#ifndef __XFS_BTREE_MEM_H__
|
|
#define __XFS_BTREE_MEM_H__
|
|
|
|
typedef uint64_t xfbno_t;
|
|
|
|
#define XFBNO_BLOCKSIZE (XMBUF_BLOCKSIZE)
|
|
#define XFBNO_BBSHIFT (XMBUF_BLOCKSHIFT - BBSHIFT)
|
|
#define XFBNO_BBSIZE (XFBNO_BLOCKSIZE >> BBSHIFT)
|
|
|
|
static inline xfs_daddr_t xfbno_to_daddr(xfbno_t blkno)
|
|
{
|
|
return blkno << XFBNO_BBSHIFT;
|
|
}
|
|
|
|
static inline xfbno_t xfs_daddr_to_xfbno(xfs_daddr_t daddr)
|
|
{
|
|
return daddr >> XFBNO_BBSHIFT;
|
|
}
|
|
|
|
struct xfbtree {
|
|
/* buffer cache target for this in-memory btree */
|
|
struct xfs_buftarg *target;
|
|
|
|
/* Highest block number that has been written to. */
|
|
xfbno_t highest_bno;
|
|
|
|
/* Owner of this btree. */
|
|
unsigned long long owner;
|
|
|
|
/* Btree header */
|
|
union xfs_btree_ptr root;
|
|
unsigned int nlevels;
|
|
|
|
/* Minimum and maximum records per block. */
|
|
unsigned int maxrecs[2];
|
|
unsigned int minrecs[2];
|
|
};
|
|
|
|
#ifdef CONFIG_XFS_BTREE_IN_MEM
|
|
static inline bool xfbtree_verify_bno(struct xfbtree *xfbt, xfbno_t bno)
|
|
{
|
|
return xmbuf_verify_daddr(xfbt->target, xfbno_to_daddr(bno));
|
|
}
|
|
|
|
void xfbtree_set_root(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *ptr, int inc);
|
|
void xfbtree_init_ptr_from_cur(struct xfs_btree_cur *cur,
|
|
union xfs_btree_ptr *ptr);
|
|
struct xfs_btree_cur *xfbtree_dup_cursor(struct xfs_btree_cur *cur);
|
|
|
|
int xfbtree_get_minrecs(struct xfs_btree_cur *cur, int level);
|
|
int xfbtree_get_maxrecs(struct xfs_btree_cur *cur, int level);
|
|
|
|
int xfbtree_alloc_block(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *start, union xfs_btree_ptr *ptr,
|
|
int *stat);
|
|
int xfbtree_free_block(struct xfs_btree_cur *cur, struct xfs_buf *bp);
|
|
|
|
/* Callers must set xfbt->target and xfbt->owner before calling this */
|
|
int xfbtree_init(struct xfs_mount *mp, struct xfbtree *xfbt,
|
|
struct xfs_buftarg *btp, const struct xfs_btree_ops *ops);
|
|
void xfbtree_destroy(struct xfbtree *xfbt);
|
|
|
|
int xfbtree_trans_commit(struct xfbtree *xfbt, struct xfs_trans *tp);
|
|
void xfbtree_trans_cancel(struct xfbtree *xfbt, struct xfs_trans *tp);
|
|
#else
|
|
# define xfbtree_verify_bno(...) (false)
|
|
#endif /* CONFIG_XFS_BTREE_IN_MEM */
|
|
|
|
#endif /* __XFS_BTREE_MEM_H__ */
|