[PATCH 05/12] powerpc/64s/hash: Use POWER6 SLBIA IH=1 variant in switch_slb
Nicholas Piggin
npiggin at gmail.com
Mon Sep 17 18:42:04 AEST 2018
On Mon, 17 Sep 2018 11:38:35 +0530
"Aneesh Kumar K.V" <aneesh.kumar at linux.ibm.com> wrote:
> Nicholas Piggin <npiggin at gmail.com> writes:
>
> > The SLBIA IH=1 hint will remove all non-zero SLBEs, but only
> > invalidate ERAT entries associated with a class value of 1, for
> > processors that support the hint (e.g., POWER6 and newer), which
> > Linux assigns to user addresses.
> >
> > This prevents kernel ERAT entries from being invalidated when
> > context switchig (if the thread faulted in more than 8 user SLBEs).
>
>
> how about renaming stuff to indicate kernel ERAT entries are kept?
> something like slb_flush_and_rebolt_user()?
User mappings aren't bolted though. I consider rebolt to mean update
the bolted kernel mappings when something has changed (like vmalloc
segment update). That doesn't need to be done here, so I think this
is okay. I can add a comment though.
Thanks,
Nick
>
> >
> > Signed-off-by: Nicholas Piggin <npiggin at gmail.com>
> > ---
> > arch/powerpc/mm/slb.c | 38 +++++++++++++++++++++++---------------
> > 1 file changed, 23 insertions(+), 15 deletions(-)
> >
> > diff --git a/arch/powerpc/mm/slb.c b/arch/powerpc/mm/slb.c
> > index a5e58f11d676..03fa1c663ccf 100644
> > --- a/arch/powerpc/mm/slb.c
> > +++ b/arch/powerpc/mm/slb.c
> > @@ -128,13 +128,21 @@ void slb_flush_all_realmode(void)
> > asm volatile("slbmte %0,%0; slbia" : : "r" (0));
> > }
> >
> > -static void __slb_flush_and_rebolt(void)
> > +void slb_flush_and_rebolt(void)
> > {
> > /* If you change this make sure you change SLB_NUM_BOLTED
> > * and PR KVM appropriately too. */
> > unsigned long linear_llp, lflags;
> > unsigned long ksp_esid_data, ksp_vsid_data;
> >
> > + WARN_ON(!irqs_disabled());
> > +
> > + /*
> > + * We can't take a PMU exception in the following code, so hard
> > + * disable interrupts.
> > + */
> > + hard_irq_disable();
> > +
> > linear_llp = mmu_psize_defs[mmu_linear_psize].sllp;
> > lflags = SLB_VSID_KERNEL | linear_llp;
> >
> > @@ -160,20 +168,7 @@ static void __slb_flush_and_rebolt(void)
> > :: "r"(ksp_vsid_data),
> > "r"(ksp_esid_data)
> > : "memory");
> > -}
> >
> > -void slb_flush_and_rebolt(void)
> > -{
> > -
> > - WARN_ON(!irqs_disabled());
> > -
> > - /*
> > - * We can't take a PMU exception in the following code, so hard
> > - * disable interrupts.
> > - */
> > - hard_irq_disable();
> > -
> > - __slb_flush_and_rebolt();
> > get_paca()->slb_cache_ptr = 0;
> > }
> >
> > @@ -248,7 +243,20 @@ void switch_slb(struct task_struct *tsk, struct mm_struct *mm)
> >
> > asm volatile("isync" : : : "memory");
> > } else {
> > - __slb_flush_and_rebolt();
> > + struct slb_shadow *p = get_slb_shadow();
> > + unsigned long ksp_esid_data =
> > + be64_to_cpu(p->save_area[KSTACK_INDEX].esid);
> > + unsigned long ksp_vsid_data =
> > + be64_to_cpu(p->save_area[KSTACK_INDEX].vsid);
> > +
> > + asm volatile("isync\n"
> > + PPC_SLBIA(1) "\n"
> > + "slbmte %0,%1\n"
> > + "isync"
> > + :: "r"(ksp_vsid_data),
> > + "r"(ksp_esid_data));
> > +
> > + asm volatile("isync" : : : "memory");
> > }
> >
> > get_paca()->slb_cache_ptr = 0;
> > --
> > 2.18.0
>
More information about the Linuxppc-dev
mailing list