[PATCH 4/4] powerpc/mm/radix: Workaround prefetch issue with KVM
Benjamin Herrenschmidt
benh at kernel.crashing.org
Wed Jul 19 13:54:13 AEST 2017
On Wed, 2017-07-19 at 12:29 +1000, Balbir Singh wrote:
>
> > diff --git a/arch/powerpc/kvm/book3s_hv_rmhandlers.S b/arch/powerpc/kvm/book3s_hv_rmhandlers.S
> > index 6ea4b53..4fb3581b 100644
> > --- a/arch/powerpc/kvm/book3s_hv_rmhandlers.S
> > +++ b/arch/powerpc/kvm/book3s_hv_rmhandlers.S
> > @@ -1522,6 +1522,18 @@ END_FTR_SECTION_IFCLR(CPU_FTR_ARCH_207S)
> > std r6, VCPU_BESCR(r9)
> > stw r7, VCPU_GUEST_PID(r9)
> > std r8, VCPU_WORT(r9)
> > +
> > + /* Handle the case where the guest used an illegal PID */
> > + LOAD_REG_ADDR(r4, mmu_base_pid)
> > + lwz r3, 0(r4)
> > + cmpw cr0,r7,r3
> > + blt 1f
>
> So the boundary is [1..(1<<mmu_pid_bits-1)]? Do we flush the tlb
> for pid 0 always?
What do you mean ? mmu_base_pid is going to be 1 << (mmu_pid_bits-1),
we flush the TLB if the guest PID r7 is not lower than the boundary.
Ntoe: this is patch v1, look at v2.
> > +
> > + /* Illegal PID, flush the TLB */
> > + bl radix_flush_pid
> > + ld r9, HSTATE_KVM_VCPU(r13)
> > +1:
> > +
> > BEGIN_FTR_SECTION
> > mfspr r5, SPRN_TCSCR
> > mfspr r6, SPRN_ACOP
> > diff --git a/arch/powerpc/mm/mmu_context_book3s64.c b/arch/powerpc/mm/mmu_context_book3s64.c
> > index abed1fe..183a67b 100644
> > --- a/arch/powerpc/mm/mmu_context_book3s64.c
> > +++ b/arch/powerpc/mm/mmu_context_book3s64.c
> > @@ -25,6 +25,10 @@
> > #include <asm/mmu_context.h>
> > #include <asm/pgalloc.h>
> >
> > +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE
> > +#include <asm/kvm_book3s_asm.h>
> > +#endif
> > +
> > #include "icswx.h"
> >
> > static DEFINE_SPINLOCK(mmu_context_lock);
> > @@ -126,9 +130,10 @@ static int hash__init_new_context(struct mm_struct *mm)
> > static int radix__init_new_context(struct mm_struct *mm)
> > {
> > unsigned long rts_field;
> > - int index;
> > + int index, max_id;
> >
> > - index = alloc_context_id(1, PRTB_ENTRIES - 1);
> > + max_id = (1 << mmu_pid_bits) - 1;
> > + index = alloc_context_id(mmu_base_pid, max_id);
> > if (index < 0)
> > return index;
> >
> > @@ -247,8 +252,40 @@ void destroy_context(struct mm_struct *mm)
> > }
> >
> > #ifdef CONFIG_PPC_RADIX_MMU
> > -void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next)
> > +void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
> > + bool new_on_cpu)
> > {
> > + /*
> > + * If this context hasn't run on that CPU before and KVM is
> > + * around, there's a slim chance that the guest on another
> > + * CPU just brought in obsolete translation into the TLB of
> > + * this CPU due to a bad prefetch using the guest PID on
> > + * the way into the hypervisor.
> > + *
> > + * We work around this here. If KVM is possible, we check if
> > + * any sibling thread is in KVM. If it is, the window may exist
> > + * and thus we flush that PID from the core.
> > + *
> > + * A potential future improvement would be to mark which PIDs
> > + * have never been used on the system and avoid it if the PID
> > + * is new and the process has no other cpumask bit set.
>
> Also due to the pid split, the chances of the context bringing in
> TLB entries is low, but a bad guest could bring in stale entries
>
> > + */
> > +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE
> > + if (cpu_has_feature(CPU_FTR_HVMODE) && new_on_cpu) {
> > + int cpu = smp_processor_id();
> > + int sib = cpu_first_thread_sibling(cpu);
> > + bool flush = false;
> > +
> > + for (; sib <= cpu_last_thread_sibling(cpu) && !flush; sib++) {
> > + if (sib == cpu)
> > + continue;
> > + if (paca[sib].kvm_hstate.kvm_vcpu)
> > + flush = true;
> > + }
> > + if (flush)
> > + radix__local_flush_all_mm(next);
> > + }
> > +#endif /* CONFIG_KVM_BOOK3S_HV_POSSIBLE */
> >
More information about the Linuxppc-dev
mailing list