[PATCH v3 24/33] KVM: PPC: Book3S HV: Emulate Privileged TLBIE for guest hypervisors
David Gibson
david at gibson.dropbear.id.au
Wed Oct 3 15:56:58 AEST 2018
On Tue, Oct 02, 2018 at 09:31:23PM +1000, Paul Mackerras wrote:
> From: Suraj Jitindar Singh <sjitindarsingh at gmail.com>
>
> When running a nested (L2) guest the guest (L1) hypervisor will use
> hypervisor privileged tlb invalidation instructions (to manage the
> partition scoped page tables) which will result in hypervisor
> emulation assistance interrupts. We emulate these instructions on behalf
> of the L1 guest.
>
> The tlbie instruction can invalidate different scopes:
>
> Invalidate TLB for a given target address:
> - This invalidates a single L2 -> L1 pte
> - We need to invalidate any L2 -> L0 shadow_pgtable ptes which map the L2
> address space which is being invalidated. This is because a single
> L2 -> L1 pte may have been mapped with more than one pte in the
> L2 -> L0 page tables.
>
> Invalidate the entire TLB for a given LPID or for all LPIDs:
> - Invalidate the entire shadow_pgtable for a given nested guest, or
> for all nested guests.
>
> Invalidate the PWC (page walk cache) for a given LPID or for all LPIDs:
> - We don't cache the PWC, so nothing to do
>
> Invalidate the entire TLB, PWC and partition table for a given/all LPIDs:
> - Here we re-read the partition table entry and remove the nested state
> for any nested guest for which the first doubleword of the partition
> table entry is now zero.
>
> This also implements the H_TLB_INVALIDATE hcall. It takes as parameters
> the tlbie instruction word (of which the RIC, PRS and R fields are used),
> the rS value (giving the lpid, where required) and the rB value (giving
> the IS, AP and EPN values).
>
> [paulus at ozlabs.org - adapted to having the partition table in guest
> memory, added the H_TLB_INVALIDATE implementation.]
>
> Signed-off-by: Suraj Jitindar Singh <sjitindarsingh at gmail.com>
> Signed-off-by: Paul Mackerras <paulus at ozlabs.org>
Again, do we need this if we're moving to a paravirt tlbie?
> ---
> arch/powerpc/include/asm/book3s/64/mmu-hash.h | 12 ++
> arch/powerpc/include/asm/kvm_book3s.h | 1 +
> arch/powerpc/include/asm/ppc-opcode.h | 1 +
> arch/powerpc/kvm/book3s_emulate.c | 1 -
> arch/powerpc/kvm/book3s_hv.c | 3 +
> arch/powerpc/kvm/book3s_hv_nested.c | 210 +++++++++++++++++++++++++-
> 6 files changed, 225 insertions(+), 3 deletions(-)
>
> diff --git a/arch/powerpc/include/asm/book3s/64/mmu-hash.h b/arch/powerpc/include/asm/book3s/64/mmu-hash.h
> index b3520b5..66db23e 100644
> --- a/arch/powerpc/include/asm/book3s/64/mmu-hash.h
> +++ b/arch/powerpc/include/asm/book3s/64/mmu-hash.h
> @@ -203,6 +203,18 @@ static inline unsigned int mmu_psize_to_shift(unsigned int mmu_psize)
> BUG();
> }
>
> +static inline unsigned int ap_to_shift(unsigned long ap)
> +{
> + int psize;
> +
> + for (psize = 0; psize < MMU_PAGE_COUNT; psize++) {
> + if (mmu_psize_defs[psize].ap == ap)
> + return mmu_psize_defs[psize].shift;
> + }
> +
> + return -1;
> +}
> +
> static inline unsigned long get_sllp_encoding(int psize)
> {
> unsigned long sllp;
> diff --git a/arch/powerpc/include/asm/kvm_book3s.h b/arch/powerpc/include/asm/kvm_book3s.h
> index 1d2286d..210e550 100644
> --- a/arch/powerpc/include/asm/kvm_book3s.h
> +++ b/arch/powerpc/include/asm/kvm_book3s.h
> @@ -301,6 +301,7 @@ long kvmhv_set_partition_table(struct kvm_vcpu *vcpu);
> void kvmhv_set_ptbl_entry(unsigned int lpid, u64 dw0, u64 dw1);
> void kvmhv_release_all_nested(struct kvm *kvm);
> long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu);
> +long kvmhv_do_nested_tlbie(struct kvm_vcpu *vcpu);
> int kvmhv_run_single_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu,
> u64 time_limit, unsigned long lpcr);
> void kvmhv_save_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr);
> diff --git a/arch/powerpc/include/asm/ppc-opcode.h b/arch/powerpc/include/asm/ppc-opcode.h
> index 665af14..6093bc8 100644
> --- a/arch/powerpc/include/asm/ppc-opcode.h
> +++ b/arch/powerpc/include/asm/ppc-opcode.h
> @@ -104,6 +104,7 @@
> #define OP_31_XOP_LHZUX 311
> #define OP_31_XOP_MSGSNDP 142
> #define OP_31_XOP_MSGCLRP 174
> +#define OP_31_XOP_TLBIE 306
> #define OP_31_XOP_MFSPR 339
> #define OP_31_XOP_LWAX 341
> #define OP_31_XOP_LHAX 343
> diff --git a/arch/powerpc/kvm/book3s_emulate.c b/arch/powerpc/kvm/book3s_emulate.c
> index 2654df2..8c7e933 100644
> --- a/arch/powerpc/kvm/book3s_emulate.c
> +++ b/arch/powerpc/kvm/book3s_emulate.c
> @@ -36,7 +36,6 @@
> #define OP_31_XOP_MTSR 210
> #define OP_31_XOP_MTSRIN 242
> #define OP_31_XOP_TLBIEL 274
> -#define OP_31_XOP_TLBIE 306
> /* Opcode is officially reserved, reuse it as sc 1 when sc 1 doesn't trap */
> #define OP_31_XOP_FAKE_SC1 308
> #define OP_31_XOP_SLBMTE 402
> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
> index 6629df4..3aa5d11e 100644
> --- a/arch/powerpc/kvm/book3s_hv.c
> +++ b/arch/powerpc/kvm/book3s_hv.c
> @@ -974,6 +974,9 @@ int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu)
> break;
> case H_TLB_INVALIDATE:
> ret = H_FUNCTION;
> + if (!vcpu->kvm->arch.nested_enable)
> + break;
> + ret = kvmhv_do_nested_tlbie(vcpu);
> break;
>
> default:
> diff --git a/arch/powerpc/kvm/book3s_hv_nested.c b/arch/powerpc/kvm/book3s_hv_nested.c
> index 4efa1f7..7abcc2b 100644
> --- a/arch/powerpc/kvm/book3s_hv_nested.c
> +++ b/arch/powerpc/kvm/book3s_hv_nested.c
> @@ -465,7 +465,7 @@ void kvmhv_release_all_nested(struct kvm *kvm)
> }
>
> /* caller must hold gp->tlb_lock */
> -void kvmhv_flush_nested(struct kvm_nested_guest *gp)
> +static void kvmhv_flush_nested(struct kvm_nested_guest *gp)
> {
> struct kvm *kvm = gp->l1_host;
>
> @@ -685,10 +685,216 @@ static int kvmhv_emulate_priv_mfspr(struct kvm_run *run, struct kvm_vcpu *vcpu,
> return EMULATE_FAIL;
> }
>
> +static inline int get_ric(unsigned int instr)
> +{
> + return (instr >> 18) & 0x3;
> +}
> +
> +static inline int get_prs(unsigned int instr)
> +{
> + return (instr >> 17) & 0x1;
> +}
> +
> +static inline int get_r(unsigned int instr)
> +{
> + return (instr >> 16) & 0x1;
> +}
> +
> +static inline int get_lpid(unsigned long r_val)
> +{
> + return r_val & 0xffffffff;
> +}
> +
> +static inline int get_is(unsigned long r_val)
> +{
> + return (r_val >> 10) & 0x3;
> +}
> +
> +static inline int get_ap(unsigned long r_val)
> +{
> + return (r_val >> 5) & 0x7;
> +}
> +
> +static inline long get_epn(unsigned long r_val)
> +{
> + return r_val >> 12;
> +}
> +
> +static int kvmhv_emulate_tlbie_tlb_addr(struct kvm_vcpu *vcpu, int lpid,
> + int ap, long epn)
> +{
> + struct kvm *kvm = vcpu->kvm;
> + struct kvm_nested_guest *gp;
> + long npages;
> + int shift;
> + unsigned long addr;
> +
> + shift = ap_to_shift(ap);
> + addr = epn << 12;
> + if (shift < 0)
> + /* Invalid ap encoding */
> + return -EINVAL;
> +
> + addr &= ~((1UL << shift) - 1);
> + npages = 1UL << (shift - PAGE_SHIFT);
> +
> + gp = kvmhv_get_nested(kvm, lpid, false);
> + if (!gp) /* No such guest -> nothing to do */
> + return 0;
> + mutex_lock(&gp->tlb_lock);
> +
> + /* There may be more than one host page backing this single guest pte */
> + do {
> + kvmhv_invalidate_shadow_pte(vcpu, gp, addr, &shift);
> +
> + npages -= 1UL << (shift - PAGE_SHIFT);
> + addr += 1UL << shift;
> + } while (npages > 0);
> +
> + mutex_unlock(&gp->tlb_lock);
> + kvmhv_put_nested(gp);
> + return 0;
> +}
> +
> +static void kvmhv_emulate_tlbie_lpid(struct kvm_vcpu *vcpu,
> + struct kvm_nested_guest *gp, int ric)
> +{
> + struct kvm *kvm = vcpu->kvm;
> +
> + mutex_lock(&gp->tlb_lock);
> + switch (ric) {
> + case 0:
> + /* Invalidate TLB */
> + spin_lock(&kvm->mmu_lock);
> + kvmppc_free_pgtable_radix(kvm, gp->shadow_pgtable,
> + gp->shadow_lpid);
> + radix__flush_tlb_lpid(gp->shadow_lpid);
> + spin_unlock(&kvm->mmu_lock);
> + break;
> + case 1:
> + /*
> + * Invalidate PWC
> + * We don't cache this -> nothing to do
> + */
> + break;
> + case 2:
> + /* Invalidate TLB, PWC and caching of partition table entries */
> + kvmhv_flush_nested(gp);
> + break;
> + default:
> + break;
> + }
> + mutex_unlock(&gp->tlb_lock);
> +}
> +
> +static void kvmhv_emulate_tlbie_all_lpid(struct kvm_vcpu *vcpu, int ric)
> +{
> + struct kvm *kvm = vcpu->kvm;
> + struct kvm_nested_guest *gp;
> + int i;
> +
> + spin_lock(&kvm->mmu_lock);
> + for (i = 0; i <= kvm->arch.max_nested_lpid; i++) {
> + gp = kvm->arch.nested_guests[i];
> + if (gp) {
> + spin_unlock(&kvm->mmu_lock);
> + kvmhv_emulate_tlbie_lpid(vcpu, gp, ric);
> + spin_lock(&kvm->mmu_lock);
> + }
> + }
> + spin_unlock(&kvm->mmu_lock);
> +}
> +
> +static int kvmhv_emulate_priv_tlbie(struct kvm_vcpu *vcpu, unsigned int instr,
> + int rs, int rb)
> +{
> + struct kvm *kvm = vcpu->kvm;
> + struct kvm_nested_guest *gp;
> + int r, ric, prs, is, ap;
> + int lpid;
> + long epn;
> + int ret = 0;
> +
> + ric = get_ric(instr);
> + prs = get_prs(instr);
> + r = get_r(instr);
> + lpid = get_lpid(kvmppc_get_gpr(vcpu, rs));
> + is = get_is(kvmppc_get_gpr(vcpu, rb));
> +
> + /*
> + * These cases are invalid and __should__ have caused a machine check
> + * r != 1 -> Only radix supported
> + * prs == 1 -> Not HV privileged
> + * ric == 3 -> No clusted bombs for radix
> + * is == 1 -> Partition scoped translations not associated with pid
> + * (!is) && (ric == 1 || ric == 2) -> Not supported by ISA
> + */
> + if ((!r) || (prs) || (ric == 3) || (is == 1) ||
> + ((!is) && (ric == 1 || ric == 2)))
> + return -EINVAL;
> +
> + switch (is) {
> + case 0:
> + /*
> + * We know ric == 0
> + * Invalidate TLB for a given target address
> + */
> + epn = get_epn(kvmppc_get_gpr(vcpu, rb));
> + ap = get_ap(kvmppc_get_gpr(vcpu, rb));
> + ret = kvmhv_emulate_tlbie_tlb_addr(vcpu, lpid, ap, epn);
> + break;
> + case 2:
> + /* Invalidate matching LPID */
> + gp = kvmhv_get_nested(kvm, lpid, false);
> + if (gp) {
> + kvmhv_emulate_tlbie_lpid(vcpu, gp, ric);
> + kvmhv_put_nested(gp);
> + }
> + break;
> + case 3:
> + /* Invalidate ALL LPIDs */
> + kvmhv_emulate_tlbie_all_lpid(vcpu, ric);
> + break;
> + default:
> + ret = -EINVAL;
> + break;
> + }
> +
> + return ret;
> +}
> +
> static int kvmhv_emulate_priv_op_31(struct kvm_run *run, struct kvm_vcpu *vcpu,
> unsigned int instr)
> {
> - return EMULATE_FAIL;
> + int ret;
> +
> + switch (get_xop(instr)) {
> + case OP_31_XOP_TLBIE:
> + ret = kvmhv_emulate_priv_tlbie(vcpu, instr, get_rs(instr),
> + get_rb(instr));
> + if (ret) {
> + kvmppc_core_queue_program(vcpu, SRR1_PROGILL);
> + return EMULATE_AGAIN;
> + }
> + return EMULATE_DONE;
> + default:
> + return EMULATE_FAIL;
> + }
> +}
> +
> +/*
> + * This handles the H_TLB_INVALIDATE hcall.
> + * Parameters are (r4) tlbie instruction code, (r5) rS contents,
> + * (r6) rB contents.
> + */
> +long kvmhv_do_nested_tlbie(struct kvm_vcpu *vcpu)
> +{
> + int ret;
> +
> + ret = kvmhv_emulate_priv_tlbie(vcpu, kvmppc_get_gpr(vcpu, 4), 5, 6);
> + if (ret)
> + return H_PARAMETER;
> + return H_SUCCESS;
> }
>
> static int kvmhv_emulate_priv_op(struct kvm_run *run, struct kvm_vcpu *vcpu,
--
David Gibson | I'll have my music baroque, and my code
david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_ _other_
| _way_ _around_!
http://www.ozlabs.org/~dgibson
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 833 bytes
Desc: not available
URL: <http://lists.ozlabs.org/pipermail/linuxppc-dev/attachments/20181003/ca7f43df/attachment.sig>
More information about the Linuxppc-dev
mailing list