mm/vma: add expand-only VMA merge mode and optimise do_brk_flags()
Patch series "introduce VMA merge mode to improve brk() performance".
A ~5% performance regression was discovered on the
aim9.brk_test.ops_per_sec by the linux kernel test bot [0].
In the past to satisfy brk() performance we duplicated VMA expansion code
and special-cased do_brk_flags(). This is however horrid and undoes work
to abstract this logic, so in resolving the issue I have endeavoured to
avoid this.
Investigating further I was able to observe that the use of a
vma_iter_next_range() and vma_prev() pair, causing an unnecessary maple
tree walk. In addition there is work that we do that is simply
unnecessary for brk().
Therefore, add a special VMA merge mode VMG_FLAG_JUST_EXPAND to avoid
doing any of this - it assumes the VMA iterator is pointing at the
previous VMA and which skips logic that brk() does not require.
This mostly eliminates the performance regression reducing it to ~2% which
is in the realm of noise. In addition, the will-it-scale test brk2,
written to be more representative of real-world brk() usage, shows a
modest performance improvement - which gives me confidence that we are not
meaningfully regressing real workloads here.
This series includes a test asserting that the 'just expand' mode works as
expected.
With many thanks to Oliver Sang for helping with performance testing of
candidate patch sets!
[0]:https://lore.kernel.org/linux-mm/202409301043.629bea78-oliver.sang@intel.com
This patch (of 2):
We know in advance that do_brk_flags() wants only to perform a VMA
expansion (if the prior VMA is compatible), and that we assume no
mergeable VMA follows it.
These are the semantics of this function prior to the recent rewrite of
the VMA merging logic, however we are now doing more work than necessary -
positioning the VMA iterator at the prior VMA and performing tasks that
are not required.
Add a new field to the vmg struct to permit merge flags and add a new
merge flag VMG_FLAG_JUST_EXPAND which implies this behaviour, and have
do_brk_flags() use this.
This fixes a reported performance regression in a brk() benchmarking suite.
Link: https://lkml.kernel.org/r/cover.1729174352.git.lorenzo.stoakes@oracle.com
Link: https://lkml.kernel.org/r/4e65d4395e5841c5acf8470dbcb714016364fd39.1729174352.git.lorenzo.stoakes@oracle.com
Fixes: cacded5e42
("mm: avoid using vma_merge() for new VMAs")
Reported-by: kernel test robot <oliver.sang@intel.com>
Closes: https://lore.kernel.org/linux-mm/202409301043.629bea78-oliver.sang@intel.com
Signed-off-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
Reviewed-by: Liam R. Howlett <Liam.Howlett@Oracle.com>
Cc: Jann Horn <jannh@google.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
b125a0def2
commit
c4d91e225f
@ -1756,7 +1756,8 @@ static int do_brk_flags(struct vma_iterator *vmi, struct vm_area_struct *vma,
|
|||||||
VMG_STATE(vmg, mm, vmi, addr, addr + len, flags, PHYS_PFN(addr));
|
VMG_STATE(vmg, mm, vmi, addr, addr + len, flags, PHYS_PFN(addr));
|
||||||
|
|
||||||
vmg.prev = vma;
|
vmg.prev = vma;
|
||||||
vma_iter_next_range(vmi);
|
/* vmi is positioned at prev, which this mode expects. */
|
||||||
|
vmg.merge_flags = VMG_FLAG_JUST_EXPAND;
|
||||||
|
|
||||||
if (vma_merge_new_range(&vmg))
|
if (vma_merge_new_range(&vmg))
|
||||||
goto out;
|
goto out;
|
||||||
|
23
mm/vma.c
23
mm/vma.c
@ -917,6 +917,7 @@ struct vm_area_struct *vma_merge_new_range(struct vma_merge_struct *vmg)
|
|||||||
pgoff_t pgoff = vmg->pgoff;
|
pgoff_t pgoff = vmg->pgoff;
|
||||||
pgoff_t pglen = PHYS_PFN(end - start);
|
pgoff_t pglen = PHYS_PFN(end - start);
|
||||||
bool can_merge_left, can_merge_right;
|
bool can_merge_left, can_merge_right;
|
||||||
|
bool just_expand = vmg->merge_flags & VMG_FLAG_JUST_EXPAND;
|
||||||
|
|
||||||
mmap_assert_write_locked(vmg->mm);
|
mmap_assert_write_locked(vmg->mm);
|
||||||
VM_WARN_ON(vmg->vma);
|
VM_WARN_ON(vmg->vma);
|
||||||
@ -930,7 +931,7 @@ struct vm_area_struct *vma_merge_new_range(struct vma_merge_struct *vmg)
|
|||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
can_merge_left = can_vma_merge_left(vmg);
|
can_merge_left = can_vma_merge_left(vmg);
|
||||||
can_merge_right = can_vma_merge_right(vmg, can_merge_left);
|
can_merge_right = !just_expand && can_vma_merge_right(vmg, can_merge_left);
|
||||||
|
|
||||||
/* If we can merge with the next VMA, adjust vmg accordingly. */
|
/* If we can merge with the next VMA, adjust vmg accordingly. */
|
||||||
if (can_merge_right) {
|
if (can_merge_right) {
|
||||||
@ -953,7 +954,11 @@ struct vm_area_struct *vma_merge_new_range(struct vma_merge_struct *vmg)
|
|||||||
if (can_merge_right && !can_merge_remove_vma(next))
|
if (can_merge_right && !can_merge_remove_vma(next))
|
||||||
vmg->end = end;
|
vmg->end = end;
|
||||||
|
|
||||||
vma_prev(vmg->vmi); /* Equivalent to going to the previous range */
|
/* In expand-only case we are already positioned at prev. */
|
||||||
|
if (!just_expand) {
|
||||||
|
/* Equivalent to going to the previous range. */
|
||||||
|
vma_prev(vmg->vmi);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -967,12 +972,14 @@ struct vm_area_struct *vma_merge_new_range(struct vma_merge_struct *vmg)
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* If expansion failed, reset state. Allows us to retry merge later. */
|
/* If expansion failed, reset state. Allows us to retry merge later. */
|
||||||
vmg->vma = NULL;
|
if (!just_expand) {
|
||||||
vmg->start = start;
|
vmg->vma = NULL;
|
||||||
vmg->end = end;
|
vmg->start = start;
|
||||||
vmg->pgoff = pgoff;
|
vmg->end = end;
|
||||||
if (vmg->vma == prev)
|
vmg->pgoff = pgoff;
|
||||||
vma_iter_set(vmg->vmi, start);
|
if (vmg->vma == prev)
|
||||||
|
vma_iter_set(vmg->vmi, start);
|
||||||
|
}
|
||||||
|
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
14
mm/vma.h
14
mm/vma.h
@ -59,6 +59,17 @@ enum vma_merge_state {
|
|||||||
VMA_MERGE_SUCCESS,
|
VMA_MERGE_SUCCESS,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enum vma_merge_flags {
|
||||||
|
VMG_FLAG_DEFAULT = 0,
|
||||||
|
/*
|
||||||
|
* If we can expand, simply do so. We know there is nothing to merge to
|
||||||
|
* the right. Does not reset state upon failure to merge. The VMA
|
||||||
|
* iterator is assumed to be positioned at the previous VMA, rather than
|
||||||
|
* at the gap.
|
||||||
|
*/
|
||||||
|
VMG_FLAG_JUST_EXPAND = 1 << 0,
|
||||||
|
};
|
||||||
|
|
||||||
/* Represents a VMA merge operation. */
|
/* Represents a VMA merge operation. */
|
||||||
struct vma_merge_struct {
|
struct vma_merge_struct {
|
||||||
struct mm_struct *mm;
|
struct mm_struct *mm;
|
||||||
@ -75,6 +86,7 @@ struct vma_merge_struct {
|
|||||||
struct mempolicy *policy;
|
struct mempolicy *policy;
|
||||||
struct vm_userfaultfd_ctx uffd_ctx;
|
struct vm_userfaultfd_ctx uffd_ctx;
|
||||||
struct anon_vma_name *anon_name;
|
struct anon_vma_name *anon_name;
|
||||||
|
enum vma_merge_flags merge_flags;
|
||||||
enum vma_merge_state state;
|
enum vma_merge_state state;
|
||||||
};
|
};
|
||||||
|
|
||||||
@ -99,6 +111,7 @@ static inline pgoff_t vma_pgoff_offset(struct vm_area_struct *vma,
|
|||||||
.flags = flags_, \
|
.flags = flags_, \
|
||||||
.pgoff = pgoff_, \
|
.pgoff = pgoff_, \
|
||||||
.state = VMA_MERGE_START, \
|
.state = VMA_MERGE_START, \
|
||||||
|
.merge_flags = VMG_FLAG_DEFAULT, \
|
||||||
}
|
}
|
||||||
|
|
||||||
#define VMG_VMA_STATE(name, vmi_, prev_, vma_, start_, end_) \
|
#define VMG_VMA_STATE(name, vmi_, prev_, vma_, start_, end_) \
|
||||||
@ -118,6 +131,7 @@ static inline pgoff_t vma_pgoff_offset(struct vm_area_struct *vma,
|
|||||||
.uffd_ctx = vma_->vm_userfaultfd_ctx, \
|
.uffd_ctx = vma_->vm_userfaultfd_ctx, \
|
||||||
.anon_name = anon_vma_name(vma_), \
|
.anon_name = anon_vma_name(vma_), \
|
||||||
.state = VMA_MERGE_START, \
|
.state = VMA_MERGE_START, \
|
||||||
|
.merge_flags = VMG_FLAG_DEFAULT, \
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_DEBUG_VM_MAPLE_TREE
|
#ifdef CONFIG_DEBUG_VM_MAPLE_TREE
|
||||||
|
Loading…
Reference in New Issue
Block a user