[PATCH 01/49] mm/sparse: fix vmemmap accounting imbalance on memory hotplug error
Muchun Song
muchun.song at linux.dev
Mon Apr 13 19:49:17 AEST 2026
> On Apr 13, 2026, at 17:35, Mike Rapoport <rppt at kernel.org> wrote:
>
> On Mon, Apr 13, 2026 at 12:19:50PM +0300, Mike Rapoport wrote:
>> On Sun, Apr 05, 2026 at 08:51:52PM +0800, Muchun Song wrote:
>>> In section_activate(), if populate_section_memmap() fails, the error
>>> handling path calls section_deactivate() to roll back the state. This
>>> approach introduces an accounting imbalance.
>>>
>>> Since the commit c3576889d87b ("mm: fix accounting of memmap pages"),
>>> memmap pages are accounted for only after populate_section_memmap()
>>> succeeds. However, section_deactivate() unconditionally decrements the
>>> vmemmap account. Consequently, a failure in populate_section_memmap()
>>> leads to a negative offset (underflow) in the system's vmemmap tracking.
>>>
>>> We can fix this by ensuring that the vmemmap accounting is incremented
>>> immediately before checking for the success of populate_section_memmap().
>>> If populate_section_memmap() fails, the subsequent call to
>>> section_deactivate() will decrement the accounting, perfectly offsetting
>>> the increment and maintaining balance.
>>>
>>> Fixes: c3576889d87b ("mm: fix accounting of memmap pages")
>>> Signed-off-by: Muchun Song <songmuchun at bytedance.com>
>>> ---
>>> mm/sparse-vmemmap.c | 2 +-
>>> 1 file changed, 1 insertion(+), 1 deletion(-)
>>>
>>> diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c
>>> index 6eadb9d116e4..ee27d0c0efe2 100644
>>> --- a/mm/sparse-vmemmap.c
>>> +++ b/mm/sparse-vmemmap.c
>>> @@ -822,11 +822,11 @@ static struct page * __meminit section_activate(int nid, unsigned long pfn,
>>> return pfn_to_page(pfn);
>>>
>>> memmap = populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);
>>> + memmap_pages_add(DIV_ROUND_UP(nr_pages * sizeof(struct page), PAGE_SIZE));
>>
>> This logically belongs to success path in populate_section_memmap(). If we
>> update the counter there, we won't need to temporarily increase it at all.
>
> Not strictly related to this patchset, but it seems, we can have a single
> memmap_boot_pages_add() in memmap_alloc() rather than to update the counter
> in memmap_alloc() callers.
It will indeed become simpler and is a good cleanup direction, but there
is a slight change in semantics: the page tables used for vmemmap page
mapping will also be counted in memmap_boot_pages_add(). This might not
be an issue (after all, the size of the page tables is very small compared
to struct pages, right?).
Additionally, I still lean toward making no changes to this patch, because
this is a pure bugfix patch — of course, it is meant to facilitate backporting
for those who need it. The cleanup would involve many more changes, so I
prefer to do that in a separate patch. What do you think?
Thanks,
Muchun.
>
>>> if (!memmap) {
>>> section_deactivate(pfn, nr_pages, altmap);
>>> return ERR_PTR(-ENOMEM);
>>> }
>>> - memmap_pages_add(DIV_ROUND_UP(nr_pages * sizeof(struct page), PAGE_SIZE));
>>>
>>> return memmap;
>>> }
>>> --
>>> 2.20.1
>>>
>>
>> --
>> Sincerely yours,
>> Mike.
>
> --
> Sincerely yours,
> Mike.
More information about the Linuxppc-dev
mailing list