xfs: convert kmem_alloc() to kmalloc()
kmem_alloc() is just a thin wrapper around kmalloc() these days. Convert everything to use kmalloc() so we can get rid of the wrapper. Note: the transaction region allocation in xlog_add_to_transaction() can be a high order allocation. Converting it to use kmalloc(__GFP_NOFAIL) results in warnings in the page allocation code being triggered because the mm subsystem does not want us to use __GFP_NOFAIL with high order allocations like we've been doing with the kmem_alloc() wrapper for a couple of decades. Hence this specific case gets converted to xlog_kvmalloc() rather than kmalloc() to avoid this issue. Signed-off-by: Dave Chinner <dchinner@redhat.com> Reviewed-by: "Darrick J. Wong" <djwong@kernel.org> Signed-off-by: Chandan Babu R <chandanbabu@kernel.org>
This commit is contained in:
parent
10634530f7
commit
f078d4ea82
@ -92,8 +92,7 @@ xfs-y += xfs_aops.o \
|
|||||||
xfs_symlink.o \
|
xfs_symlink.o \
|
||||||
xfs_sysfs.o \
|
xfs_sysfs.o \
|
||||||
xfs_trans.o \
|
xfs_trans.o \
|
||||||
xfs_xattr.o \
|
xfs_xattr.o
|
||||||
kmem.o
|
|
||||||
|
|
||||||
# low-level transaction/log code
|
# low-level transaction/log code
|
||||||
xfs-y += xfs_log.o \
|
xfs-y += xfs_log.o \
|
||||||
|
@ -1,30 +0,0 @@
|
|||||||
// SPDX-License-Identifier: GPL-2.0
|
|
||||||
/*
|
|
||||||
* Copyright (c) 2000-2005 Silicon Graphics, Inc.
|
|
||||||
* All Rights Reserved.
|
|
||||||
*/
|
|
||||||
#include "xfs.h"
|
|
||||||
#include "xfs_message.h"
|
|
||||||
#include "xfs_trace.h"
|
|
||||||
|
|
||||||
void *
|
|
||||||
kmem_alloc(size_t size, xfs_km_flags_t flags)
|
|
||||||
{
|
|
||||||
int retries = 0;
|
|
||||||
gfp_t lflags = kmem_flags_convert(flags);
|
|
||||||
void *ptr;
|
|
||||||
|
|
||||||
trace_kmem_alloc(size, flags, _RET_IP_);
|
|
||||||
|
|
||||||
do {
|
|
||||||
ptr = kmalloc(size, lflags);
|
|
||||||
if (ptr || (flags & KM_MAYFAIL))
|
|
||||||
return ptr;
|
|
||||||
if (!(++retries % 100))
|
|
||||||
xfs_err(NULL,
|
|
||||||
"%s(%u) possible memory allocation deadlock size %u in %s (mode:0x%x)",
|
|
||||||
current->comm, current->pid,
|
|
||||||
(unsigned int)size, __func__, lflags);
|
|
||||||
memalloc_retry_wait(lflags);
|
|
||||||
} while (1);
|
|
||||||
}
|
|
@ -15,48 +15,6 @@
|
|||||||
* General memory allocation interfaces
|
* General memory allocation interfaces
|
||||||
*/
|
*/
|
||||||
|
|
||||||
typedef unsigned __bitwise xfs_km_flags_t;
|
|
||||||
#define KM_NOFS ((__force xfs_km_flags_t)0x0004u)
|
|
||||||
#define KM_MAYFAIL ((__force xfs_km_flags_t)0x0008u)
|
|
||||||
#define KM_ZERO ((__force xfs_km_flags_t)0x0010u)
|
|
||||||
#define KM_NOLOCKDEP ((__force xfs_km_flags_t)0x0020u)
|
|
||||||
|
|
||||||
/*
|
|
||||||
* We use a special process flag to avoid recursive callbacks into
|
|
||||||
* the filesystem during transactions. We will also issue our own
|
|
||||||
* warnings, so we explicitly skip any generic ones (silly of us).
|
|
||||||
*/
|
|
||||||
static inline gfp_t
|
|
||||||
kmem_flags_convert(xfs_km_flags_t flags)
|
|
||||||
{
|
|
||||||
gfp_t lflags;
|
|
||||||
|
|
||||||
BUG_ON(flags & ~(KM_NOFS | KM_MAYFAIL | KM_ZERO | KM_NOLOCKDEP));
|
|
||||||
|
|
||||||
lflags = GFP_KERNEL | __GFP_NOWARN;
|
|
||||||
if (flags & KM_NOFS)
|
|
||||||
lflags &= ~__GFP_FS;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Default page/slab allocator behavior is to retry for ever
|
|
||||||
* for small allocations. We can override this behavior by using
|
|
||||||
* __GFP_RETRY_MAYFAIL which will tell the allocator to retry as long
|
|
||||||
* as it is feasible but rather fail than retry forever for all
|
|
||||||
* request sizes.
|
|
||||||
*/
|
|
||||||
if (flags & KM_MAYFAIL)
|
|
||||||
lflags |= __GFP_RETRY_MAYFAIL;
|
|
||||||
|
|
||||||
if (flags & KM_ZERO)
|
|
||||||
lflags |= __GFP_ZERO;
|
|
||||||
|
|
||||||
if (flags & KM_NOLOCKDEP)
|
|
||||||
lflags |= __GFP_NOLOCKDEP;
|
|
||||||
|
|
||||||
return lflags;
|
|
||||||
}
|
|
||||||
|
|
||||||
extern void *kmem_alloc(size_t, xfs_km_flags_t);
|
|
||||||
static inline void kmem_free(const void *ptr)
|
static inline void kmem_free(const void *ptr)
|
||||||
{
|
{
|
||||||
kvfree(ptr);
|
kvfree(ptr);
|
||||||
|
@ -879,8 +879,7 @@ xfs_attr_shortform_to_leaf(
|
|||||||
|
|
||||||
trace_xfs_attr_sf_to_leaf(args);
|
trace_xfs_attr_sf_to_leaf(args);
|
||||||
|
|
||||||
tmpbuffer = kmem_alloc(size, 0);
|
tmpbuffer = kmalloc(size, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
ASSERT(tmpbuffer != NULL);
|
|
||||||
memcpy(tmpbuffer, ifp->if_data, size);
|
memcpy(tmpbuffer, ifp->if_data, size);
|
||||||
sf = (struct xfs_attr_sf_hdr *)tmpbuffer;
|
sf = (struct xfs_attr_sf_hdr *)tmpbuffer;
|
||||||
|
|
||||||
@ -1059,7 +1058,7 @@ xfs_attr3_leaf_to_shortform(
|
|||||||
|
|
||||||
trace_xfs_attr_leaf_to_sf(args);
|
trace_xfs_attr_leaf_to_sf(args);
|
||||||
|
|
||||||
tmpbuffer = kmem_alloc(args->geo->blksize, 0);
|
tmpbuffer = kmalloc(args->geo->blksize, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
if (!tmpbuffer)
|
if (!tmpbuffer)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
@ -1533,7 +1532,7 @@ xfs_attr3_leaf_compact(
|
|||||||
|
|
||||||
trace_xfs_attr_leaf_compact(args);
|
trace_xfs_attr_leaf_compact(args);
|
||||||
|
|
||||||
tmpbuffer = kmem_alloc(args->geo->blksize, 0);
|
tmpbuffer = kmalloc(args->geo->blksize, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
memcpy(tmpbuffer, bp->b_addr, args->geo->blksize);
|
memcpy(tmpbuffer, bp->b_addr, args->geo->blksize);
|
||||||
memset(bp->b_addr, 0, args->geo->blksize);
|
memset(bp->b_addr, 0, args->geo->blksize);
|
||||||
leaf_src = (xfs_attr_leafblock_t *)tmpbuffer;
|
leaf_src = (xfs_attr_leafblock_t *)tmpbuffer;
|
||||||
|
@ -139,7 +139,7 @@ xfs_btree_stage_afakeroot(
|
|||||||
ASSERT(!(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE));
|
ASSERT(!(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE));
|
||||||
ASSERT(cur->bc_tp == NULL);
|
ASSERT(cur->bc_tp == NULL);
|
||||||
|
|
||||||
nops = kmem_alloc(sizeof(struct xfs_btree_ops), KM_NOFS);
|
nops = kmalloc(sizeof(struct xfs_btree_ops), GFP_NOFS | __GFP_NOFAIL);
|
||||||
memcpy(nops, cur->bc_ops, sizeof(struct xfs_btree_ops));
|
memcpy(nops, cur->bc_ops, sizeof(struct xfs_btree_ops));
|
||||||
nops->alloc_block = xfs_btree_fakeroot_alloc_block;
|
nops->alloc_block = xfs_btree_fakeroot_alloc_block;
|
||||||
nops->free_block = xfs_btree_fakeroot_free_block;
|
nops->free_block = xfs_btree_fakeroot_free_block;
|
||||||
@ -220,7 +220,7 @@ xfs_btree_stage_ifakeroot(
|
|||||||
ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE);
|
ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE);
|
||||||
ASSERT(cur->bc_tp == NULL);
|
ASSERT(cur->bc_tp == NULL);
|
||||||
|
|
||||||
nops = kmem_alloc(sizeof(struct xfs_btree_ops), KM_NOFS);
|
nops = kmalloc(sizeof(struct xfs_btree_ops), GFP_NOFS | __GFP_NOFAIL);
|
||||||
memcpy(nops, cur->bc_ops, sizeof(struct xfs_btree_ops));
|
memcpy(nops, cur->bc_ops, sizeof(struct xfs_btree_ops));
|
||||||
nops->alloc_block = xfs_btree_fakeroot_alloc_block;
|
nops->alloc_block = xfs_btree_fakeroot_alloc_block;
|
||||||
nops->free_block = xfs_btree_fakeroot_free_block;
|
nops->free_block = xfs_btree_fakeroot_free_block;
|
||||||
|
@ -2182,7 +2182,8 @@ xfs_da_grow_inode_int(
|
|||||||
* If we didn't get it and the block might work if fragmented,
|
* If we didn't get it and the block might work if fragmented,
|
||||||
* try without the CONTIG flag. Loop until we get it all.
|
* try without the CONTIG flag. Loop until we get it all.
|
||||||
*/
|
*/
|
||||||
mapp = kmem_alloc(sizeof(*mapp) * count, 0);
|
mapp = kmalloc(sizeof(*mapp) * count,
|
||||||
|
GFP_KERNEL | __GFP_NOFAIL);
|
||||||
for (b = *bno, mapi = 0; b < *bno + count; ) {
|
for (b = *bno, mapi = 0; b < *bno + count; ) {
|
||||||
c = (int)(*bno + count - b);
|
c = (int)(*bno + count - b);
|
||||||
nmap = min(XFS_BMAP_MAX_NMAP, c);
|
nmap = min(XFS_BMAP_MAX_NMAP, c);
|
||||||
|
@ -333,7 +333,7 @@ xfs_dir_cilookup_result(
|
|||||||
!(args->op_flags & XFS_DA_OP_CILOOKUP))
|
!(args->op_flags & XFS_DA_OP_CILOOKUP))
|
||||||
return -EEXIST;
|
return -EEXIST;
|
||||||
|
|
||||||
args->value = kmem_alloc(len, KM_NOFS | KM_MAYFAIL);
|
args->value = kmalloc(len, GFP_NOFS | __GFP_RETRY_MAYFAIL);
|
||||||
if (!args->value)
|
if (!args->value)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -1108,7 +1108,7 @@ xfs_dir2_sf_to_block(
|
|||||||
* Copy the directory into a temporary buffer.
|
* Copy the directory into a temporary buffer.
|
||||||
* Then pitch the incore inode data so we can make extents.
|
* Then pitch the incore inode data so we can make extents.
|
||||||
*/
|
*/
|
||||||
sfp = kmem_alloc(ifp->if_bytes, 0);
|
sfp = kmalloc(ifp->if_bytes, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
memcpy(sfp, oldsfp, ifp->if_bytes);
|
memcpy(sfp, oldsfp, ifp->if_bytes);
|
||||||
|
|
||||||
xfs_idata_realloc(dp, -ifp->if_bytes, XFS_DATA_FORK);
|
xfs_idata_realloc(dp, -ifp->if_bytes, XFS_DATA_FORK);
|
||||||
|
@ -276,7 +276,7 @@ xfs_dir2_block_to_sf(
|
|||||||
* format the data into. Once we have formatted the data, we can free
|
* format the data into. Once we have formatted the data, we can free
|
||||||
* the block and copy the formatted data into the inode literal area.
|
* the block and copy the formatted data into the inode literal area.
|
||||||
*/
|
*/
|
||||||
sfp = kmem_alloc(mp->m_sb.sb_inodesize, 0);
|
sfp = kmalloc(mp->m_sb.sb_inodesize, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
memcpy(sfp, sfhp, xfs_dir2_sf_hdr_size(sfhp->i8count));
|
memcpy(sfp, sfhp, xfs_dir2_sf_hdr_size(sfhp->i8count));
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -524,7 +524,7 @@ xfs_dir2_sf_addname_hard(
|
|||||||
* Copy the old directory to the stack buffer.
|
* Copy the old directory to the stack buffer.
|
||||||
*/
|
*/
|
||||||
old_isize = (int)dp->i_disk_size;
|
old_isize = (int)dp->i_disk_size;
|
||||||
buf = kmem_alloc(old_isize, 0);
|
buf = kmalloc(old_isize, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
oldsfp = (xfs_dir2_sf_hdr_t *)buf;
|
oldsfp = (xfs_dir2_sf_hdr_t *)buf;
|
||||||
memcpy(oldsfp, dp->i_df.if_data, old_isize);
|
memcpy(oldsfp, dp->i_df.if_data, old_isize);
|
||||||
/*
|
/*
|
||||||
@ -1151,7 +1151,7 @@ xfs_dir2_sf_toino4(
|
|||||||
* Don't want xfs_idata_realloc copying the data here.
|
* Don't want xfs_idata_realloc copying the data here.
|
||||||
*/
|
*/
|
||||||
oldsize = dp->i_df.if_bytes;
|
oldsize = dp->i_df.if_bytes;
|
||||||
buf = kmem_alloc(oldsize, 0);
|
buf = kmalloc(oldsize, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
ASSERT(oldsfp->i8count == 1);
|
ASSERT(oldsfp->i8count == 1);
|
||||||
memcpy(buf, oldsfp, oldsize);
|
memcpy(buf, oldsfp, oldsize);
|
||||||
/*
|
/*
|
||||||
@ -1223,7 +1223,7 @@ xfs_dir2_sf_toino8(
|
|||||||
* Don't want xfs_idata_realloc copying the data here.
|
* Don't want xfs_idata_realloc copying the data here.
|
||||||
*/
|
*/
|
||||||
oldsize = dp->i_df.if_bytes;
|
oldsize = dp->i_df.if_bytes;
|
||||||
buf = kmem_alloc(oldsize, 0);
|
buf = kmalloc(oldsize, GFP_KERNEL | __GFP_NOFAIL);
|
||||||
ASSERT(oldsfp->i8count == 0);
|
ASSERT(oldsfp->i8count == 0);
|
||||||
memcpy(buf, oldsfp, oldsize);
|
memcpy(buf, oldsfp, oldsize);
|
||||||
/*
|
/*
|
||||||
|
@ -50,7 +50,7 @@ xfs_init_local_fork(
|
|||||||
mem_size++;
|
mem_size++;
|
||||||
|
|
||||||
if (size) {
|
if (size) {
|
||||||
char *new_data = kmem_alloc(mem_size, KM_NOFS);
|
char *new_data = kmalloc(mem_size, GFP_NOFS | __GFP_NOFAIL);
|
||||||
|
|
||||||
memcpy(new_data, data, size);
|
memcpy(new_data, data, size);
|
||||||
if (zero_terminate)
|
if (zero_terminate)
|
||||||
@ -77,7 +77,7 @@ xfs_iformat_local(
|
|||||||
/*
|
/*
|
||||||
* If the size is unreasonable, then something
|
* If the size is unreasonable, then something
|
||||||
* is wrong and we just bail out rather than crash in
|
* is wrong and we just bail out rather than crash in
|
||||||
* kmem_alloc() or memcpy() below.
|
* kmalloc() or memcpy() below.
|
||||||
*/
|
*/
|
||||||
if (unlikely(size > XFS_DFORK_SIZE(dip, ip->i_mount, whichfork))) {
|
if (unlikely(size > XFS_DFORK_SIZE(dip, ip->i_mount, whichfork))) {
|
||||||
xfs_warn(ip->i_mount,
|
xfs_warn(ip->i_mount,
|
||||||
@ -116,7 +116,7 @@ xfs_iformat_extents(
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* If the number of extents is unreasonable, then something is wrong and
|
* If the number of extents is unreasonable, then something is wrong and
|
||||||
* we just bail out rather than crash in kmem_alloc() or memcpy() below.
|
* we just bail out rather than crash in kmalloc() or memcpy() below.
|
||||||
*/
|
*/
|
||||||
if (unlikely(size < 0 || size > XFS_DFORK_SIZE(dip, mp, whichfork))) {
|
if (unlikely(size < 0 || size > XFS_DFORK_SIZE(dip, mp, whichfork))) {
|
||||||
xfs_warn(ip->i_mount, "corrupt inode %llu ((a)extents = %llu).",
|
xfs_warn(ip->i_mount, "corrupt inode %llu ((a)extents = %llu).",
|
||||||
@ -205,7 +205,7 @@ xfs_iformat_btree(
|
|||||||
}
|
}
|
||||||
|
|
||||||
ifp->if_broot_bytes = size;
|
ifp->if_broot_bytes = size;
|
||||||
ifp->if_broot = kmem_alloc(size, KM_NOFS);
|
ifp->if_broot = kmalloc(size, GFP_NOFS | __GFP_NOFAIL);
|
||||||
ASSERT(ifp->if_broot != NULL);
|
ASSERT(ifp->if_broot != NULL);
|
||||||
/*
|
/*
|
||||||
* Copy and convert from the on-disk structure
|
* Copy and convert from the on-disk structure
|
||||||
@ -399,7 +399,8 @@ xfs_iroot_realloc(
|
|||||||
*/
|
*/
|
||||||
if (ifp->if_broot_bytes == 0) {
|
if (ifp->if_broot_bytes == 0) {
|
||||||
new_size = XFS_BMAP_BROOT_SPACE_CALC(mp, rec_diff);
|
new_size = XFS_BMAP_BROOT_SPACE_CALC(mp, rec_diff);
|
||||||
ifp->if_broot = kmem_alloc(new_size, KM_NOFS);
|
ifp->if_broot = kmalloc(new_size,
|
||||||
|
GFP_NOFS | __GFP_NOFAIL);
|
||||||
ifp->if_broot_bytes = (int)new_size;
|
ifp->if_broot_bytes = (int)new_size;
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@ -440,7 +441,7 @@ xfs_iroot_realloc(
|
|||||||
else
|
else
|
||||||
new_size = 0;
|
new_size = 0;
|
||||||
if (new_size > 0) {
|
if (new_size > 0) {
|
||||||
new_broot = kmem_alloc(new_size, KM_NOFS);
|
new_broot = kmalloc(new_size, GFP_NOFS | __GFP_NOFAIL);
|
||||||
/*
|
/*
|
||||||
* First copy over the btree block header.
|
* First copy over the btree block header.
|
||||||
*/
|
*/
|
||||||
@ -488,7 +489,7 @@ xfs_iroot_realloc(
|
|||||||
*
|
*
|
||||||
* If the amount of space needed has decreased below the size of the
|
* If the amount of space needed has decreased below the size of the
|
||||||
* inline buffer, then switch to using the inline buffer. Otherwise,
|
* inline buffer, then switch to using the inline buffer. Otherwise,
|
||||||
* use kmem_realloc() or kmem_alloc() to adjust the size of the buffer
|
* use krealloc() or kmalloc() to adjust the size of the buffer
|
||||||
* to what is needed.
|
* to what is needed.
|
||||||
*
|
*
|
||||||
* ip -- the inode whose if_data area is changing
|
* ip -- the inode whose if_data area is changing
|
||||||
|
@ -109,7 +109,7 @@ xfs_attr_shortform_list(
|
|||||||
* It didn't all fit, so we have to sort everything on hashval.
|
* It didn't all fit, so we have to sort everything on hashval.
|
||||||
*/
|
*/
|
||||||
sbsize = sf->count * sizeof(*sbuf);
|
sbsize = sf->count * sizeof(*sbuf);
|
||||||
sbp = sbuf = kmem_alloc(sbsize, KM_NOFS);
|
sbp = sbuf = kmalloc(sbsize, GFP_NOFS | __GFP_NOFAIL);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Scan the attribute list for the rest of the entries, storing
|
* Scan the attribute list for the rest of the entries, storing
|
||||||
|
@ -325,14 +325,14 @@ xfs_buf_alloc_kmem(
|
|||||||
struct xfs_buf *bp,
|
struct xfs_buf *bp,
|
||||||
xfs_buf_flags_t flags)
|
xfs_buf_flags_t flags)
|
||||||
{
|
{
|
||||||
xfs_km_flags_t kmflag_mask = KM_NOFS;
|
gfp_t gfp_mask = GFP_NOFS | __GFP_NOFAIL;
|
||||||
size_t size = BBTOB(bp->b_length);
|
size_t size = BBTOB(bp->b_length);
|
||||||
|
|
||||||
/* Assure zeroed buffer for non-read cases. */
|
/* Assure zeroed buffer for non-read cases. */
|
||||||
if (!(flags & XBF_READ))
|
if (!(flags & XBF_READ))
|
||||||
kmflag_mask |= KM_ZERO;
|
gfp_mask |= __GFP_ZERO;
|
||||||
|
|
||||||
bp->b_addr = kmem_alloc(size, kmflag_mask);
|
bp->b_addr = kmalloc(size, gfp_mask);
|
||||||
if (!bp->b_addr)
|
if (!bp->b_addr)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -85,7 +85,7 @@ xlog_add_buffer_cancelled(
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
bcp = kmem_alloc(sizeof(struct xfs_buf_cancel), 0);
|
bcp = kmalloc(sizeof(struct xfs_buf_cancel), GFP_KERNEL | __GFP_NOFAIL);
|
||||||
bcp->bc_blkno = blkno;
|
bcp->bc_blkno = blkno;
|
||||||
bcp->bc_len = len;
|
bcp->bc_len = len;
|
||||||
bcp->bc_refcount = 1;
|
bcp->bc_refcount = 1;
|
||||||
|
@ -313,7 +313,7 @@ xfs_filestream_create_association(
|
|||||||
* we return a referenced AG, the allocation can still go ahead just
|
* we return a referenced AG, the allocation can still go ahead just
|
||||||
* fine.
|
* fine.
|
||||||
*/
|
*/
|
||||||
item = kmem_alloc(sizeof(*item), KM_MAYFAIL);
|
item = kmalloc(sizeof(*item), GFP_KERNEL | __GFP_RETRY_MAYFAIL);
|
||||||
if (!item)
|
if (!item)
|
||||||
goto out_put_fstrms;
|
goto out_put_fstrms;
|
||||||
|
|
||||||
|
@ -291,7 +291,8 @@ xlog_recover_inode_commit_pass2(
|
|||||||
if (item->ri_buf[0].i_len == sizeof(struct xfs_inode_log_format)) {
|
if (item->ri_buf[0].i_len == sizeof(struct xfs_inode_log_format)) {
|
||||||
in_f = item->ri_buf[0].i_addr;
|
in_f = item->ri_buf[0].i_addr;
|
||||||
} else {
|
} else {
|
||||||
in_f = kmem_alloc(sizeof(struct xfs_inode_log_format), 0);
|
in_f = kmalloc(sizeof(struct xfs_inode_log_format),
|
||||||
|
GFP_KERNEL | __GFP_NOFAIL);
|
||||||
need_free = 1;
|
need_free = 1;
|
||||||
error = xfs_inode_item_format_convert(&item->ri_buf[0], in_f);
|
error = xfs_inode_item_format_convert(&item->ri_buf[0], in_f);
|
||||||
if (error)
|
if (error)
|
||||||
|
@ -160,7 +160,7 @@ xfs_iwalk_alloc(
|
|||||||
|
|
||||||
/* Allocate a prefetch buffer for inobt records. */
|
/* Allocate a prefetch buffer for inobt records. */
|
||||||
size = iwag->sz_recs * sizeof(struct xfs_inobt_rec_incore);
|
size = iwag->sz_recs * sizeof(struct xfs_inobt_rec_incore);
|
||||||
iwag->recs = kmem_alloc(size, KM_MAYFAIL);
|
iwag->recs = kmalloc(size, GFP_KERNEL | __GFP_RETRY_MAYFAIL);
|
||||||
if (iwag->recs == NULL)
|
if (iwag->recs == NULL)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -2161,7 +2161,7 @@ xlog_recover_add_to_trans(
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
ptr = kmem_alloc(len, 0);
|
ptr = xlog_kvmalloc(len);
|
||||||
memcpy(ptr, dp, len);
|
memcpy(ptr, dp, len);
|
||||||
in_f = (struct xfs_inode_log_format *)ptr;
|
in_f = (struct xfs_inode_log_format *)ptr;
|
||||||
|
|
||||||
|
@ -997,7 +997,8 @@ xfs_qm_reset_dqcounts_buf(
|
|||||||
if (qip->i_nblocks == 0)
|
if (qip->i_nblocks == 0)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
map = kmem_alloc(XFS_DQITER_MAP_SIZE * sizeof(*map), 0);
|
map = kmalloc(XFS_DQITER_MAP_SIZE * sizeof(*map),
|
||||||
|
GFP_KERNEL | __GFP_NOFAIL);
|
||||||
|
|
||||||
lblkno = 0;
|
lblkno = 0;
|
||||||
maxlblkcnt = XFS_B_TO_FSB(mp, mp->m_super->s_maxbytes);
|
maxlblkcnt = XFS_B_TO_FSB(mp, mp->m_super->s_maxbytes);
|
||||||
|
@ -903,7 +903,7 @@ xfs_growfs_rt(
|
|||||||
/*
|
/*
|
||||||
* Allocate a new (fake) mount/sb.
|
* Allocate a new (fake) mount/sb.
|
||||||
*/
|
*/
|
||||||
nmp = kmem_alloc(sizeof(*nmp), 0);
|
nmp = kmalloc(sizeof(*nmp), GFP_KERNEL | __GFP_NOFAIL);
|
||||||
/*
|
/*
|
||||||
* Loop over the bitmap blocks.
|
* Loop over the bitmap blocks.
|
||||||
* We will do everything one bitmap block at a time.
|
* We will do everything one bitmap block at a time.
|
||||||
|
@ -1987,7 +1987,7 @@ static int xfs_init_fs_context(
|
|||||||
{
|
{
|
||||||
struct xfs_mount *mp;
|
struct xfs_mount *mp;
|
||||||
|
|
||||||
mp = kmem_alloc(sizeof(struct xfs_mount), KM_ZERO);
|
mp = kzalloc(sizeof(struct xfs_mount), GFP_KERNEL | __GFP_NOFAIL);
|
||||||
if (!mp)
|
if (!mp)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
@ -4040,31 +4040,6 @@ TRACE_EVENT(xfs_pwork_init,
|
|||||||
__entry->nr_threads, __entry->pid)
|
__entry->nr_threads, __entry->pid)
|
||||||
)
|
)
|
||||||
|
|
||||||
DECLARE_EVENT_CLASS(xfs_kmem_class,
|
|
||||||
TP_PROTO(ssize_t size, int flags, unsigned long caller_ip),
|
|
||||||
TP_ARGS(size, flags, caller_ip),
|
|
||||||
TP_STRUCT__entry(
|
|
||||||
__field(ssize_t, size)
|
|
||||||
__field(int, flags)
|
|
||||||
__field(unsigned long, caller_ip)
|
|
||||||
),
|
|
||||||
TP_fast_assign(
|
|
||||||
__entry->size = size;
|
|
||||||
__entry->flags = flags;
|
|
||||||
__entry->caller_ip = caller_ip;
|
|
||||||
),
|
|
||||||
TP_printk("size %zd flags 0x%x caller %pS",
|
|
||||||
__entry->size,
|
|
||||||
__entry->flags,
|
|
||||||
(char *)__entry->caller_ip)
|
|
||||||
)
|
|
||||||
|
|
||||||
#define DEFINE_KMEM_EVENT(name) \
|
|
||||||
DEFINE_EVENT(xfs_kmem_class, name, \
|
|
||||||
TP_PROTO(ssize_t size, int flags, unsigned long caller_ip), \
|
|
||||||
TP_ARGS(size, flags, caller_ip))
|
|
||||||
DEFINE_KMEM_EVENT(kmem_alloc);
|
|
||||||
|
|
||||||
TRACE_EVENT(xfs_check_new_dalign,
|
TRACE_EVENT(xfs_check_new_dalign,
|
||||||
TP_PROTO(struct xfs_mount *mp, int new_dalign, xfs_ino_t calc_rootino),
|
TP_PROTO(struct xfs_mount *mp, int new_dalign, xfs_ino_t calc_rootino),
|
||||||
TP_ARGS(mp, new_dalign, calc_rootino),
|
TP_ARGS(mp, new_dalign, calc_rootino),
|
||||||
|
Loading…
Reference in New Issue
Block a user