[PATCH v3 1/2] mm/cma: provide option to opt out from exposing pages on activation failure
Hari Bathini
hbathini at linux.ibm.com
Mon Jan 24 16:21:22 AEDT 2022
Hi Andrew,
Could you please pick these patches via -mm tree.
On 17/01/22 1:22 pm, Hari Bathini wrote:
> Commit 072355c1cf2d ("mm/cma: expose all pages to the buddy if
> activation of an area fails") started exposing all pages to buddy
> allocator on CMA activation failure. But there can be CMA users that
> want to handle the reserved memory differently on CMA allocation
> failure. Provide an option to opt out from exposing pages to buddy
> for such cases.
>
> Signed-off-by: Hari Bathini <hbathini at linux.ibm.com>
> Reviewed-by: David Hildenbrand <david at redhat.com>
> ---
>
> Changes in v3:
> * Dropped NULL check in cma_reserve_pages_on_error().
> * Dropped explicit initialization of cma->reserve_pages_on_error to
> 'false' in cma_init_reserved_mem().
> * Added Reviewed-by tag from David.
>
> Changes in v2:
> * Changed cma->free_pages_on_error to cma->reserve_pages_on_error and
> cma_dont_free_pages_on_error() to cma_reserve_pages_on_error() to
> avoid confusion.
>
>
> include/linux/cma.h | 2 ++
> mm/cma.c | 11 +++++++++--
> mm/cma.h | 1 +
> 3 files changed, 12 insertions(+), 2 deletions(-)
>
> diff --git a/include/linux/cma.h b/include/linux/cma.h
> index bd801023504b..51d540eee18a 100644
> --- a/include/linux/cma.h
> +++ b/include/linux/cma.h
> @@ -50,4 +50,6 @@ extern bool cma_pages_valid(struct cma *cma, const struct page *pages, unsigned
> extern bool cma_release(struct cma *cma, const struct page *pages, unsigned long count);
>
> extern int cma_for_each_area(int (*it)(struct cma *cma, void *data), void *data);
> +
> +extern void cma_reserve_pages_on_error(struct cma *cma);
> #endif
> diff --git a/mm/cma.c b/mm/cma.c
> index bc9ca8f3c487..766f1b82b532 100644
> --- a/mm/cma.c
> +++ b/mm/cma.c
> @@ -131,8 +131,10 @@ static void __init cma_activate_area(struct cma *cma)
> bitmap_free(cma->bitmap);
> out_error:
> /* Expose all pages to the buddy, they are useless for CMA. */
> - for (pfn = base_pfn; pfn < base_pfn + cma->count; pfn++)
> - free_reserved_page(pfn_to_page(pfn));
> + if (!cma->reserve_pages_on_error) {
> + for (pfn = base_pfn; pfn < base_pfn + cma->count; pfn++)
> + free_reserved_page(pfn_to_page(pfn));
> + }
> totalcma_pages -= cma->count;
> cma->count = 0;
> pr_err("CMA area %s could not be activated\n", cma->name);
> @@ -150,6 +152,11 @@ static int __init cma_init_reserved_areas(void)
> }
> core_initcall(cma_init_reserved_areas);
>
> +void __init cma_reserve_pages_on_error(struct cma *cma)
> +{
> + cma->reserve_pages_on_error = true;
> +}
> +
> /**
> * cma_init_reserved_mem() - create custom contiguous area from reserved memory
> * @base: Base address of the reserved area
> diff --git a/mm/cma.h b/mm/cma.h
> index 2c775877eae2..88a0595670b7 100644
> --- a/mm/cma.h
> +++ b/mm/cma.h
> @@ -30,6 +30,7 @@ struct cma {
> /* kobject requires dynamic object */
> struct cma_kobject *cma_kobj;
> #endif
> + bool reserve_pages_on_error;
> };
>
> extern struct cma cma_areas[MAX_CMA_AREAS];
Thanks
Hari
More information about the Linuxppc-dev
mailing list