summaryrefslogtreecommitdiff
path: root/mm/huge_memory.c
diff options
context:
space:
mode:
authorWei Yang <richard.weiyang@gmail.com>2025-10-21 21:21:40 +0000
committerAndrew Morton <akpm@linux-foundation.org>2025-11-16 17:28:19 -0800
commit092ef3899707ac2a5ad195b96bcfb3c85d97f25c (patch)
tree25317a82c5591acff4346f5b08a8b624b8785d63 /mm/huge_memory.c
parentef0258857d40dd8e33c52e28ea3a1c82099a6879 (diff)
mm/huge_memory: update folio stat after successful split
The current implementation complicates this process: * It iterates over the resulting new folios. * It uses a flag (@stop_split) to conditionally skip updating the stat for the folio at @split_at during the loop. * It then attempts to update the skipped stat on a subsequent failure path. This logic is unnecessarily hard to follow. This commit refactors the code to update the folio statistics only after a successful split. This makes the logic much cleaner and sets the stage for further simplification of the stat-handling code. Link: https://lkml.kernel.org/r/20251021212142.25766-3-richard.weiyang@gmail.com Signed-off-by: Wei Yang <richard.weiyang@gmail.com> Acked-by: David Hildenbrand <david@redhat.com> Reviewed-by: Zi Yan <ziy@nvidia.com> Reviewed-by: wang lian <lianux.mm@gmail.com> Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Cc: Baolin Wang <baolin.wang@linux.alibaba.com> Cc: Barry Song <baohua@kernel.org> Cc: Dev Jain <dev.jain@arm.com> Cc: Lance Yang <lance.yang@linux.dev> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Mariano Pache <npache@redhat.com> Cc: Ryan Roberts <ryan.roberts@arm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Diffstat (limited to 'mm/huge_memory.c')
-rw-r--r--mm/huge_memory.c44
1 files changed, 11 insertions, 33 deletions
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index 5d95e3462c43..85c472fbcbfa 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -3404,20 +3404,15 @@ static int __split_unmapped_folio(struct folio *folio, int new_order,
const bool is_anon = folio_test_anon(folio);
int order = folio_order(folio);
int start_order = uniform_split ? new_order : order - 1;
- bool stop_split = false;
struct folio *next;
int split_order;
- int ret = 0;
-
- if (is_anon)
- mod_mthp_stat(order, MTHP_STAT_NR_ANON, -1);
/*
* split to new_order one order at a time. For uniform split,
* folio is split to new_order directly.
*/
for (split_order = start_order;
- split_order >= new_order && !stop_split;
+ split_order >= new_order;
split_order--) {
struct folio *end_folio = folio_next(folio);
int old_order = folio_order(folio);
@@ -3440,49 +3435,32 @@ static int __split_unmapped_folio(struct folio *folio, int new_order,
else {
xas_set_order(xas, folio->index, split_order);
xas_try_split(xas, folio, old_order);
- if (xas_error(xas)) {
- ret = xas_error(xas);
- stop_split = true;
- }
+ if (xas_error(xas))
+ return xas_error(xas);
}
}
- if (!stop_split) {
- folio_split_memcg_refs(folio, old_order, split_order);
- split_page_owner(&folio->page, old_order, split_order);
- pgalloc_tag_split(folio, old_order, split_order);
-
- __split_folio_to_order(folio, old_order, split_order);
- }
+ folio_split_memcg_refs(folio, old_order, split_order);
+ split_page_owner(&folio->page, old_order, split_order);
+ pgalloc_tag_split(folio, old_order, split_order);
+ __split_folio_to_order(folio, old_order, split_order);
+ if (is_anon)
+ mod_mthp_stat(old_order, MTHP_STAT_NR_ANON, -1);
/*
* Iterate through after-split folios and update folio stats.
- * But in buddy allocator like split, the folio
- * containing the specified page is skipped until its order
- * is new_order, since the folio will be worked on in next
- * iteration.
*/
for (new_folio = folio; new_folio != end_folio; new_folio = next) {
next = folio_next(new_folio);
- /*
- * for buddy allocator like split, new_folio containing
- * @split_at page could be split again, thus do not
- * change stats yet. Wait until new_folio's order is
- * @new_order or stop_split is set to true by the above
- * xas_split() failure.
- */
- if (new_folio == page_folio(split_at)) {
+ if (new_folio == page_folio(split_at))
folio = new_folio;
- if (split_order != new_order && !stop_split)
- continue;
- }
if (is_anon)
mod_mthp_stat(folio_order(new_folio),
MTHP_STAT_NR_ANON, 1);
}
}
- return ret;
+ return 0;
}
bool non_uniform_split_supported(struct folio *folio, unsigned int new_order,