[RFC v1 01/10] book3s64/hash: Remove kfence support temporarily

Christophe Leroy christophe.leroy at csgroup.eu
Thu Aug 15 03:07:13 AEST 2024



Le 31/07/2024 à 09:56, Ritesh Harjani (IBM) a écrit :
> [Vous ne recevez pas souvent de courriers de ritesh.list at gmail.com. Découvrez pourquoi ceci est important à https://aka.ms/LearnAboutSenderIdentification ]
> 
> Kfence on book3s Hash on pseries is anyways broken. It fails to boot
> due to RMA size limitation. That is because, kfence with Hash uses
> debug_pagealloc infrastructure. debug_pagealloc allocates linear map
> for entire dram size instead of just kfence relevant objects.
> This means for 16TB of DRAM it will require (16TB >> PAGE_SHIFT)
> which is 256MB which is half of RMA region on P8.
> crash kernel reserves 256MB and we also need 2048 * 16KB * 3 for
> emergency stack and some more for paca allocations.
> That means there is not enough memory for reserving the full linear map
> in the RMA region, if the DRAM size is too big (>=16TB)
> (The issue is seen above 8TB with crash kernel 256 MB reservation).
> 
> Now Kfence does not require linear memory map for entire DRAM.
> It only needs for kfence objects. So this patch temporarily removes the
> kfence functionality since debug_pagealloc code needs some refactoring.
> We will bring in kfence on Hash support in later patches.
> 
> Signed-off-by: Ritesh Harjani (IBM) <ritesh.list at gmail.com>
> ---
>   arch/powerpc/include/asm/kfence.h     |  5 +++++
>   arch/powerpc/mm/book3s64/hash_utils.c | 16 +++++++++++-----
>   2 files changed, 16 insertions(+), 5 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/kfence.h b/arch/powerpc/include/asm/kfence.h
> index fab124ada1c7..f3a9476a71b3 100644
> --- a/arch/powerpc/include/asm/kfence.h
> +++ b/arch/powerpc/include/asm/kfence.h
> @@ -10,6 +10,7 @@
> 
>   #include <linux/mm.h>
>   #include <asm/pgtable.h>
> +#include <asm/mmu.h>
> 
>   #ifdef CONFIG_PPC64_ELF_ABI_V1
>   #define ARCH_FUNC_PREFIX "."
> @@ -25,6 +26,10 @@ static inline void disable_kfence(void)
> 
>   static inline bool arch_kfence_init_pool(void)
>   {
> +#ifdef CONFIG_PPC64
> +       if (!radix_enabled())
> +               return false;
> +#endif

Avoid #ifdefs whenever possible. Here you can do:

	if (IS_ENABLED(CONFIG_PPC64) && !radix_enabled())
		return false;

>          return !kfence_disabled;
>   }
>   #endif
> diff --git a/arch/powerpc/mm/book3s64/hash_utils.c b/arch/powerpc/mm/book3s64/hash_utils.c
> index 01c3b4b65241..1a1b50735fa0 100644
> --- a/arch/powerpc/mm/book3s64/hash_utils.c
> +++ b/arch/powerpc/mm/book3s64/hash_utils.c
> @@ -431,7 +431,7 @@ int htab_bolt_mapping(unsigned long vstart, unsigned long vend,
>                          break;
> 
>                  cond_resched();
> -               if (debug_pagealloc_enabled_or_kfence() &&
> +               if (debug_pagealloc_enabled() &&
>                          (paddr >> PAGE_SHIFT) < linear_map_hash_count)
>                          linear_map_hash_slots[paddr >> PAGE_SHIFT] = ret | 0x80;
>          }
> @@ -814,7 +814,7 @@ static void __init htab_init_page_sizes(void)
>          bool aligned = true;
>          init_hpte_page_sizes();
> 
> -       if (!debug_pagealloc_enabled_or_kfence()) {
> +       if (!debug_pagealloc_enabled()) {
>                  /*
>                   * Pick a size for the linear mapping. Currently, we only
>                   * support 16M, 1M and 4K which is the default
> @@ -1134,7 +1134,7 @@ static void __init htab_initialize(void)
> 
>          prot = pgprot_val(PAGE_KERNEL);
> 
> -       if (debug_pagealloc_enabled_or_kfence()) {
> +       if (debug_pagealloc_enabled()) {
>                  linear_map_hash_count = memblock_end_of_DRAM() >> PAGE_SHIFT;
>                  linear_map_hash_slots = memblock_alloc_try_nid(
>                                  linear_map_hash_count, 1, MEMBLOCK_LOW_LIMIT,
> @@ -2117,7 +2117,7 @@ void hpt_do_stress(unsigned long ea, unsigned long hpte_group)
>          }
>   }
> 
> -#if defined(CONFIG_DEBUG_PAGEALLOC) || defined(CONFIG_KFENCE)
> +#if defined(CONFIG_DEBUG_PAGEALLOC)

Use #ifdef

>   static DEFINE_RAW_SPINLOCK(linear_map_hash_lock);
> 
>   static void kernel_map_linear_page(unsigned long vaddr, unsigned long lmi)
> @@ -2191,7 +2191,13 @@ int hash__kernel_map_pages(struct page *page, int numpages, int enable)
>          local_irq_restore(flags);
>          return 0;
>   }
> -#endif /* CONFIG_DEBUG_PAGEALLOC || CONFIG_KFENCE */
> +#else /* CONFIG_DEBUG_PAGEALLOC */
> +int hash__kernel_map_pages(struct page *page, int numpages,
> +                                        int enable)
> +{
> +       return 0;
> +}
> +#endif /* CONFIG_DEBUG_PAGEALLOC */
> 
>   void hash__setup_initial_memory_limit(phys_addr_t first_memblock_base,
>                                  phys_addr_t first_memblock_size)
> --
> 2.45.2
> 


More information about the Linuxppc-dev mailing list