[RFC 7/8] powerpc/book3s64/slb: Add no_slb_preload early cmdline param
Christophe Leroy
christophe.leroy at csgroup.eu
Sat Aug 30 16:42:42 AEST 2025
Le 30/08/2025 à 05:51, Ritesh Harjani (IBM) a écrit :
> no_slb_preload cmdline can come useful in quickly disabling and/or
> testing the performance impact of userspace slb preloads. Recently there
> was a slb multi-hit issue due to slb preload cache which was very
> difficult to triage. This cmdline option allows to quickly disable
> preloads and verify if the issue exists in preload cache or somewhere
> else. This can also be a useful option to see the effect of slb preloads
> for any application workload e.g. number of slb faults with or w/o slb
> preloads.
>
> For e.g. with the next patch where we added slb_faults counter to /proc/vmstat:
>
> with slb_preload:
> slb_faults (minimal initrd boot): 15
> slb_faults (full systemd boot): 300
>
> with no_slb_preload:
> slb_faults (minimal initrd boot): 33
> slb_faults (full systemd boot): 138180
>
> Cc: Madhavan Srinivasan <maddy at linux.ibm.com>
> Cc: Michael Ellerman <mpe at ellerman.id.au>
> Cc: Nicholas Piggin <npiggin at gmail.com>
> Cc: Christophe Leroy <christophe.leroy at csgroup.eu>
> Cc: Paul Mackerras <paulus at ozlabs.org>
> Cc: "Aneesh Kumar K.V" <aneesh.kumar at kernel.org>
> Cc: Donet Tom <donettom at linux.ibm.com>
> Cc: linuxppc-dev at lists.ozlabs.org
> Signed-off-by: Ritesh Harjani (IBM) <ritesh.list at gmail.com>
> ---
> Documentation/admin-guide/kernel-parameters.txt | 3 +++
> arch/powerpc/mm/book3s64/hash_utils.c | 3 +++
> arch/powerpc/mm/book3s64/internal.h | 7 +++++++
> arch/powerpc/mm/book3s64/slb.c | 15 +++++++++++++++
> 4 files changed, 28 insertions(+)
>
> diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt
> index 747a55abf494..9a66f255b659 100644
> --- a/Documentation/admin-guide/kernel-parameters.txt
> +++ b/Documentation/admin-guide/kernel-parameters.txt
> @@ -7135,6 +7135,9 @@
> them frequently to increase the rate of SLB faults
> on kernel addresses.
>
> + no_slb_preload [PPC,EARLY]
> + Disables slb preloading for userspace.
> +
> sunrpc.min_resvport=
> sunrpc.max_resvport=
> [NFS,SUNRPC]
> diff --git a/arch/powerpc/mm/book3s64/hash_utils.c b/arch/powerpc/mm/book3s64/hash_utils.c
> index 495b6da6f5d4..abf703563ea3 100644
> --- a/arch/powerpc/mm/book3s64/hash_utils.c
> +++ b/arch/powerpc/mm/book3s64/hash_utils.c
> @@ -1319,6 +1319,9 @@ static void __init htab_initialize(void)
> if (stress_slb_enabled)
> static_branch_enable(&stress_slb_key);
>
> + if (no_slb_preload)
> + static_branch_enable(&no_slb_preload_key);
> +
> if (stress_hpt_enabled) {
> unsigned long tmp;
> static_branch_enable(&stress_hpt_key);
> diff --git a/arch/powerpc/mm/book3s64/internal.h b/arch/powerpc/mm/book3s64/internal.h
> index c26a6f0c90fc..cad08d83369c 100644
> --- a/arch/powerpc/mm/book3s64/internal.h
> +++ b/arch/powerpc/mm/book3s64/internal.h
> @@ -22,6 +22,13 @@ static inline bool stress_hpt(void)
> return static_branch_unlikely(&stress_hpt_key);
> }
>
> +extern bool no_slb_preload;
> +DECLARE_STATIC_KEY_FALSE(no_slb_preload_key);
> +static inline bool slb_preload_disabled(void)
> +{
> + return static_branch_unlikely(&no_slb_preload_key);
> +}
> +
> void hpt_do_stress(unsigned long ea, unsigned long hpte_group);
>
> void exit_lazy_flush_tlb(struct mm_struct *mm, bool always_flush);
> diff --git a/arch/powerpc/mm/book3s64/slb.c b/arch/powerpc/mm/book3s64/slb.c
> index 780792b9a1e5..297ab0e93c1e 100644
> --- a/arch/powerpc/mm/book3s64/slb.c
> +++ b/arch/powerpc/mm/book3s64/slb.c
> @@ -42,6 +42,15 @@ early_param("stress_slb", parse_stress_slb);
>
> __ro_after_init DEFINE_STATIC_KEY_FALSE(stress_slb_key);
>
> +bool no_slb_preload __initdata;
> +static int __init parse_no_slb_preload(char *p)
> +{
> + no_slb_preload = true;
> + return 0;
Can't you call static_branch_disable() directly from here and avoid
doing it in hash_utils.c ?
> +}
> +early_param("no_slb_preload", parse_no_slb_preload);
> +__ro_after_init DEFINE_STATIC_KEY_FALSE(no_slb_preload_key);
> +
> static void assert_slb_presence(bool present, unsigned long ea)
> {
> #ifdef CONFIG_DEBUG_VM
> @@ -299,6 +308,9 @@ static void preload_add(struct thread_info *ti, unsigned long ea)
> unsigned char idx;
> unsigned long esid;
>
> + if (slb_preload_disabled())
> + return;
> +
> if (mmu_has_feature(MMU_FTR_1T_SEGMENT)) {
> /* EAs are stored >> 28 so 256MB segments don't need clearing */
> if (ea & ESID_MASK_1T)
> @@ -414,6 +426,9 @@ void switch_slb(struct task_struct *tsk, struct mm_struct *mm)
>
> copy_mm_to_paca(mm);
>
> + if (slb_preload_disabled())
> + return;
> +
> /*
> * We gradually age out SLBs after a number of context switches to
> * reduce reload overhead of unused entries (like we do with FP/VEC
> --
> 2.50.1
>
More information about the Linuxppc-dev
mailing list