Skip to content

Commit b9bf6c2

Browse files
Dev Jainakpm00
authored andcommitted
mm: refactor MM_CP_PROT_NUMA skipping case into new function
Patch series "Optimize mprotect() for large folios", v5. Use folio_pte_batch() to optimize change_pte_range(). On arm64, if the ptes are painted with the contig bit, then ptep_get() will iterate through all 16 entries to collect a/d bits. Hence this optimization will result in a 16x reduction in the number of ptep_get() calls. Next, ptep_modify_prot_start() will eventually call contpte_try_unfold() on every contig block, thus flushing the TLB for the complete large folio range. Instead, use get_and_clear_full_ptes() so as to elide TLBIs on each contig block, and only do them on the starting and ending contig block. For split folios, there will be no pte batching; the batch size returned by folio_pte_batch() will be 1. For pagetable split folios, the ptes will still point to the same large folio; for arm64, this results in the optimization described above, and for other arches, a minor improvement is expected due to a reduction in the number of function calls. mm-selftests pass on arm64. I have some failing tests on my x86 VM already; no new tests fail as a result of this patchset. We use the following test cases to measure performance, mprotect()'ing the mapped memory to read-only then read-write 40 times: Test case 1: Mapping 1G of memory, touching it to get PMD-THPs, then pte-mapping those THPs Test case 2: Mapping 1G of memory with 64K mTHPs Test case 3: Mapping 1G of memory with 4K pages Average execution time on arm64, Apple M3: Before the patchset: T1: 2.1 seconds T2: 2 seconds T3: 1 second After the patchset: T1: 0.65 seconds T2: 0.7 seconds T3: 1.1 seconds Observing T1/T2 and T3 before the patchset, we also remove the regression introduced by ptep_get() on a contpte block. And, for large folios we get an almost 74% performance improvement, albeit the trade-off being a slight degradation in the small folio case. For x86: Before the patchset: T1: 3.75 seconds T2: 3.7 seconds T3: 3.85 seconds After the patchset: T1: 3.7 seconds T2: 3.7 seconds T3: 3.9 seconds So there is a minor improvement due to reduction in number of function calls, and a slight degradation in the small folio case due to the overhead of vm_normal_folio() + folio_test_large(). Here is the test program: #define _GNU_SOURCE #include <sys/mman.h> #include <stdlib.h> #include <string.h> #include <stdio.h> #include <unistd.h> #define SIZE (1024*1024*1024) unsigned long pmdsize = (1UL << 21); unsigned long pagesize = (1UL << 12); static void pte_map_thps(char *mem, size_t size) { size_t offs; int ret = 0; /* PTE-map each THP by temporarily splitting the VMAs. */ for (offs = 0; offs < size; offs += pmdsize) { ret |= madvise(mem + offs, pagesize, MADV_DONTFORK); ret |= madvise(mem + offs, pagesize, MADV_DOFORK); } if (ret) { fprintf(stderr, "ERROR: mprotect() failed\n"); exit(1); } } int main(int argc, char *argv[]) { char *p; int ret = 0; p = mmap((1UL << 30), SIZE, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); if (p != (1UL << 30)) { perror("mmap"); return 1; } memset(p, 0, SIZE); if (madvise(p, SIZE, MADV_NOHUGEPAGE)) perror("madvise"); explicit_bzero(p, SIZE); pte_map_thps(p, SIZE); for (int loops = 0; loops < 40; loops++) { if (mprotect(p, SIZE, PROT_READ)) perror("mprotect"), exit(1); if (mprotect(p, SIZE, PROT_READ|PROT_WRITE)) perror("mprotect"), exit(1); explicit_bzero(p, SIZE); } } This patch (of 7): Reduce indentation by refactoring the prot_numa case into a new function. No functional change intended. Link: https://lkml.kernel.org/r/20250718090244.21092-1-dev.jain@arm.com Link: https://lkml.kernel.org/r/20250718090244.21092-2-dev.jain@arm.com Signed-off-by: Dev Jain <dev.jain@arm.com> Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Reviewed-by: Barry Song <baohua@kernel.org> Reviewed-by: Zi Yan <ziy@nvidia.com> Cc: Anshuman Khandual <anshuman.khandual@arm.com> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Christophe Leroy <christophe.leroy@csgroup.eu> Cc: David Hildenbrand <david@redhat.com> Cc: Hugh Dickins <hughd@google.com> Cc: Jann Horn <jannh@google.com> Cc: Joey Gouly <joey.gouly@arm.com> Cc: Kevin Brodsky <kevin.brodsky@arm.com> Cc: Lance Yang <ioworker0@gmail.com> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: Peter Xu <peterx@redhat.com> Cc: Ryan Roberts <ryan.roberts@arm.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Will Deacon <will@kernel.org> Cc: Yang Shi <yang@os.amperecomputing.com> Cc: Yicong Yang <yangyicong@hisilicon.com> Cc: Zhenhua Huang <quic_zhenhuah@quicinc.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
1 parent fde4770 commit b9bf6c2

File tree

1 file changed

+55
-46
lines changed

1 file changed

+55
-46
lines changed

mm/mprotect.c

Lines changed: 55 additions & 46 deletions
Original file line numberDiff line numberDiff line change
@@ -83,6 +83,59 @@ bool can_change_pte_writable(struct vm_area_struct *vma, unsigned long addr,
8383
return pte_dirty(pte);
8484
}
8585

86+
static bool prot_numa_skip(struct vm_area_struct *vma, unsigned long addr,
87+
pte_t oldpte, pte_t *pte, int target_node)
88+
{
89+
struct folio *folio;
90+
bool toptier;
91+
int nid;
92+
93+
/* Avoid TLB flush if possible */
94+
if (pte_protnone(oldpte))
95+
return true;
96+
97+
folio = vm_normal_folio(vma, addr, oldpte);
98+
if (!folio)
99+
return true;
100+
101+
if (folio_is_zone_device(folio) || folio_test_ksm(folio))
102+
return true;
103+
104+
/* Also skip shared copy-on-write pages */
105+
if (is_cow_mapping(vma->vm_flags) &&
106+
(folio_maybe_dma_pinned(folio) || folio_maybe_mapped_shared(folio)))
107+
return true;
108+
109+
/*
110+
* While migration can move some dirty pages,
111+
* it cannot move them all from MIGRATE_ASYNC
112+
* context.
113+
*/
114+
if (folio_is_file_lru(folio) && folio_test_dirty(folio))
115+
return true;
116+
117+
/*
118+
* Don't mess with PTEs if page is already on the node
119+
* a single-threaded process is running on.
120+
*/
121+
nid = folio_nid(folio);
122+
if (target_node == nid)
123+
return true;
124+
125+
toptier = node_is_toptier(nid);
126+
127+
/*
128+
* Skip scanning top tier node if normal numa
129+
* balancing is disabled
130+
*/
131+
if (!(sysctl_numa_balancing_mode & NUMA_BALANCING_NORMAL) && toptier)
132+
return true;
133+
134+
if (folio_use_access_time(folio))
135+
folio_xchg_access_time(folio, jiffies_to_msecs(jiffies));
136+
return false;
137+
}
138+
86139
static long change_pte_range(struct mmu_gather *tlb,
87140
struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr,
88141
unsigned long end, pgprot_t newprot, unsigned long cp_flags)
@@ -117,53 +170,9 @@ static long change_pte_range(struct mmu_gather *tlb,
117170
* pages. See similar comment in change_huge_pmd.
118171
*/
119172
if (prot_numa) {
120-
struct folio *folio;
121-
int nid;
122-
bool toptier;
123-
124-
/* Avoid TLB flush if possible */
125-
if (pte_protnone(oldpte))
126-
continue;
127-
128-
folio = vm_normal_folio(vma, addr, oldpte);
129-
if (!folio || folio_is_zone_device(folio) ||
130-
folio_test_ksm(folio))
131-
continue;
132-
133-
/* Also skip shared copy-on-write pages */
134-
if (is_cow_mapping(vma->vm_flags) &&
135-
(folio_maybe_dma_pinned(folio) ||
136-
folio_maybe_mapped_shared(folio)))
137-
continue;
138-
139-
/*
140-
* While migration can move some dirty pages,
141-
* it cannot move them all from MIGRATE_ASYNC
142-
* context.
143-
*/
144-
if (folio_is_file_lru(folio) &&
145-
folio_test_dirty(folio))
146-
continue;
147-
148-
/*
149-
* Don't mess with PTEs if page is already on the node
150-
* a single-threaded process is running on.
151-
*/
152-
nid = folio_nid(folio);
153-
if (target_node == nid)
154-
continue;
155-
toptier = node_is_toptier(nid);
156-
157-
/*
158-
* Skip scanning top tier node if normal numa
159-
* balancing is disabled
160-
*/
161-
if (!(sysctl_numa_balancing_mode & NUMA_BALANCING_NORMAL) &&
162-
toptier)
173+
if (prot_numa_skip(vma, addr, oldpte, pte,
174+
target_node))
163175
continue;
164-
if (folio_use_access_time(folio))
165-
folio_xchg_access_time(folio,
166-
jiffies_to_msecs(jiffies));
167176
}
168177

169178
oldpte = ptep_modify_prot_start(vma, addr, pte);

0 commit comments

Comments
 (0)