[PATCH 01/49] mm/sparse: fix vmemmap accounting imbalance on memory hotplug error
Muchun Song
muchun.song at linux.dev
Tue Apr 14 00:16:20 AEST 2026
> On Apr 13, 2026, at 22:05, Muchun Song <muchun.song at linux.dev> wrote:
>
>
>
>> On Apr 13, 2026, at 21:36, Mike Rapoport <rppt at kernel.org> wrote:
>>
>> Hi Muchun,
>
> Hi,
>
>>
>> On Mon, Apr 13, 2026 at 08:47:45PM +0800, Muchun Song wrote:
>>>>>> On Apr 13, 2026, at 20:05, Mike Rapoport <rppt at kernel.org> wrote:
>>>>>>>>>>
>>>>>>>>>> diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c
>>>>>>>>>> index 6eadb9d116e4..ee27d0c0efe2 100644
>>>>>>>>>> --- a/mm/sparse-vmemmap.c
>>>>>>>>>> +++ b/mm/sparse-vmemmap.c
>>>>>>>>>> @@ -822,11 +822,11 @@ static struct page * __meminit section_activate(int nid, unsigned long pfn,
>>>>>>>>>> return pfn_to_page(pfn);
>>>>>>>>>>
>>>>>>>>>> memmap = populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);
>>>>>>>>>> + memmap_pages_add(DIV_ROUND_UP(nr_pages * sizeof(struct page), PAGE_SIZE));
>>>>>>>>>
>>>>>>>>> This logically belongs to success path in populate_section_memmap(). If we
>>>>>>>>> update the counter there, we won't need to temporarily increase it at all.
>>>>>>>>
>>>>>>>> Not strictly related to this patchset, but it seems, we can have a single
>>>>>>>> memmap_boot_pages_add() in memmap_alloc() rather than to update the counter
>>>>>>>> in memmap_alloc() callers.
>>>>>>>
>>>>>>> It will indeed become simpler and is a good cleanup direction, but there
>>>>>>> is a slight change in semantics: the page tables used for vmemmap page
>>>>>>> mapping will also be counted in memmap_boot_pages_add(). This might not
>>>>>>> be an issue (after all, the size of the page tables is very small compared
>>>>>>> to struct pages, right?).
>>>>>>>
>>>>>>> Additionally, I still lean toward making no changes to this patch, because
>>>>>>> this is a pure bugfix patch — of course, it is meant to facilitate backporting
>>>>>>> for those who need it. The cleanup would involve many more changes, so I
>>>>>>> prefer to do that in a separate patch. What do you think?
>>>>>
>>>>> For this patch and easy backporting I still think that cleaner to have the
>>>>> counter incremented in populate_section_memmap() rather immediately after
>>>>> it.
>>>
>>> Hi Mike,
>>>
>>> Alright, let’s revisit your solution. After we’ve moved the counter into the
>>> populate_section_memmap(), we still need to increase the counter temporarily
>>> (but in populate_section_memmap()) even if we fail to populate. That’s
>>> because section_deactivate() reduces the counter without exception, isn’t it?
>>> Just want to make sure we are on the same page on the meaning of “temporarily
>>> increase”. Maybe you do not mean “temporarily” in this case.
>>
>> I suggest to increase the counter only if we succeeded to populate:
>>
>> diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c
>> index 6eadb9d116e4..247fd54f1003 100644
>> --- a/mm/sparse-vmemmap.c
>> +++ b/mm/sparse-vmemmap.c
>> @@ -656,7 +656,13 @@ static struct page * __meminit populate_section_memmap(unsigned long pfn,
>> unsigned long nr_pages, int nid, struct vmem_altmap *altmap,
>> struct dev_pagemap *pgmap)
>> {
>> - return __populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);
>> + struct page *p = __populate_section_memmap(pfn, nr_pages, nid, altmap,
>> + pgmap);
>> +
>> + if (p)
>> + memmap_pages_add(DIV_ROUND_UP(nr_pages * sizeof(struct page), PAGE_SIZE));
>
> We don’t increase the counter on failure, then
>
>> +
>> + return p;
>> }
>>
>> static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages,
>> @@ -826,7 +832,6 @@ static struct page * __meminit section_activate(int nid, unsigned long pfn,
>> section_deactivate(pfn, nr_pages, altmap);
>
> here section_deactivate() is called, which decrease the counter unconditionally,
> the issue still exists. We didn't fix anything.
>
> Thanks.
>
>> return ERR_PTR(-ENOMEM);
>> }
>> - memmap_pages_add(DIV_ROUND_UP(nr_pages * sizeof(struct page), PAGE_SIZE));
>>
>> return memmap;
>> }
>>
>>
>> Then we'll better follow "all or nothing" principle and won't have
>> exceptional cases in section_deactivate().
To follow "all or nothing" principle here, I think we should not call
section_deactivate() to do the cleanup in section_activate().
After all, if section_activate() didn't succeed, how can we use
section_deactivate() to release the resources? What do you think?
Thanks.
>>
>>> Thanks,
>>> Muchun.
>>
>> --
>> Sincerely yours,
>> Mike.
More information about the Linuxppc-dev
mailing list