On Wed, Feb 26, 2025 at 04:00:25PM -0500, Zi Yan wrote:
+static int __split_unmapped_folio(struct folio *folio, int new_order,
struct page *split_at, struct page *lock_at,
struct list_head *list, pgoff_t end,
struct xa_state *xas, struct address_space *mapping,
bool uniform_split)
+{
[...]
/* complete memcg works before add pages to LRU */
split_page_memcg(&folio->page, old_order, split_order);
split_page_owner(&folio->page, old_order, split_order);
pgalloc_tag_split(folio, old_order, split_order);
At least split_page_memcg() needs to become aware of 'uniform_split'.
if (folio_memcg_kmem(folio)) obj_cgroup_get_many(__folio_objcg(folio), old_nr / new_nr - 1);
If we're doing uniform_split, that calculation should be old_order - new_order - 1