[PATCH] powerpc/64: Simplify adaptation to new ISA v3.00 HPTE format
Aneesh Kumar K.V
aneesh.kumar at linux.vnet.ibm.com
Fri Nov 11 23:29:33 AEDT 2016
Paul Mackerras <paulus at ozlabs.org> writes:
> This changes the way that we support the new ISA v3.00 HPTE format.
> Instead of adapting everything that uses HPTE values to handle either
> the old format or the new format, depending on which CPU we are on,
> we now convert explicitly between old and new formats if necessary
> in the low-level routines that actually access HPTEs in memory.
> This limits the amount of code that needs to know about the new
> format and makes the conversions explicit. This is OK because the
> old format contains all the information that is in the new format.
>
> This also fixes operation under a hypervisor, because the H_ENTER
> hypercall (and other hypercalls that deal with HPTEs) will continue
> to require the HPTE value to be supplied in the old format. At
> present the kernel will not boot in HPT mode on POWER9 under a
> hypervisor.
>
> This fixes and partially reverts commit 50de596de8be
> ("powerpc/mm/hash: Add support for Power9 Hash", 2016-04-29).
This is better than the patch I sent. Do you think we need
s/new_v/3_0_v/ ?
Reviewed-by: Aneesh Kumar K.V <aneesh.kumar at linux.vnet.ibm.com>
>
> Fixes: 50de596de8be
> Signed-off-by: Paul Mackerras <paulus at ozlabs.org>
> ---
> arch/powerpc/include/asm/book3s/64/mmu-hash.h | 47 ++++++++++++++++++++++-----
> arch/powerpc/mm/hash_native_64.c | 30 +++++++++++++----
> arch/powerpc/platforms/ps3/htab.c | 2 +-
> arch/powerpc/platforms/pseries/lpar.c | 2 +-
> 4 files changed, 65 insertions(+), 16 deletions(-)
>
> diff --git a/arch/powerpc/include/asm/book3s/64/mmu-hash.h b/arch/powerpc/include/asm/book3s/64/mmu-hash.h
> index e407af2..2e6a823 100644
> --- a/arch/powerpc/include/asm/book3s/64/mmu-hash.h
> +++ b/arch/powerpc/include/asm/book3s/64/mmu-hash.h
> @@ -70,7 +70,9 @@
>
> #define HPTE_V_SSIZE_SHIFT 62
> #define HPTE_V_AVPN_SHIFT 7
> +#define HPTE_V_COMMON_BITS ASM_CONST(0x000fffffffffffff)
> #define HPTE_V_AVPN ASM_CONST(0x3fffffffffffff80)
> +#define HPTE_V_AVPN_3_0 ASM_CONST(0x000fffffffffff80)
> #define HPTE_V_AVPN_VAL(x) (((x) & HPTE_V_AVPN) >> HPTE_V_AVPN_SHIFT)
> #define HPTE_V_COMPARE(x,y) (!(((x) ^ (y)) & 0xffffffffffffff80UL))
> #define HPTE_V_BOLTED ASM_CONST(0x0000000000000010)
> @@ -80,14 +82,16 @@
> #define HPTE_V_VALID ASM_CONST(0x0000000000000001)
>
> /*
> - * ISA 3.0 have a different HPTE format.
> + * ISA 3.0 has a different HPTE format.
> */
> #define HPTE_R_3_0_SSIZE_SHIFT 58
> +#define HPTE_R_3_0_SSIZE_MASK (3ull << HPTE_R_3_0_SSIZE_SHIFT)
> #define HPTE_R_PP0 ASM_CONST(0x8000000000000000)
> #define HPTE_R_TS ASM_CONST(0x4000000000000000)
> #define HPTE_R_KEY_HI ASM_CONST(0x3000000000000000)
> #define HPTE_R_RPN_SHIFT 12
> #define HPTE_R_RPN ASM_CONST(0x0ffffffffffff000)
> +#define HPTE_R_RPN_3_0 ASM_CONST(0x01fffffffffff000)
> #define HPTE_R_PP ASM_CONST(0x0000000000000003)
> #define HPTE_R_PPP ASM_CONST(0x8000000000000003)
> #define HPTE_R_N ASM_CONST(0x0000000000000004)
> @@ -316,12 +320,43 @@ static inline unsigned long hpte_encode_avpn(unsigned long vpn, int psize,
> */
> v = (vpn >> (23 - VPN_SHIFT)) & ~(mmu_psize_defs[psize].avpnm);
> v <<= HPTE_V_AVPN_SHIFT;
> - if (!cpu_has_feature(CPU_FTR_ARCH_300))
> - v |= ((unsigned long) ssize) << HPTE_V_SSIZE_SHIFT;
> + v |= ((unsigned long) ssize) << HPTE_V_SSIZE_SHIFT;
> return v;
> }
>
> /*
> + * ISA v3.0 defines a new HPTE format, which differs from the old
> + * format in having smaller AVPN and ARPN fields, and the B field
> + * in the second dword instead of the first.
> + */
> +static inline unsigned long hpte_old_to_new_v(unsigned long v)
How about
static inline unsigned long hpte_old_to_3_0_v(unsigned long v)
> +{
> + /* trim AVPN, drop B */
> + return v & HPTE_V_COMMON_BITS;
> +}
> +
> +static inline unsigned long hpte_old_to_new_r(unsigned long v, unsigned long r)
> +{
> + /* move B field from 1st to 2nd dword, trim ARPN */
> + return (r & ~HPTE_R_3_0_SSIZE_MASK) |
> + (((v) >> HPTE_V_SSIZE_SHIFT) << HPTE_R_3_0_SSIZE_SHIFT);
> +}
> +
> +static inline unsigned long hpte_new_to_old_v(unsigned long v, unsigned long r)
> +{
> + /* insert B field */
> + return (v & HPTE_V_COMMON_BITS) |
> + ((r & HPTE_R_3_0_SSIZE_MASK) <<
> + (HPTE_V_SSIZE_SHIFT - HPTE_R_3_0_SSIZE_SHIFT));
> +}
> +
> +static inline unsigned long hpte_new_to_old_r(unsigned long r)
> +{
> + /* clear out B field */
> + return r & ~HPTE_R_3_0_SSIZE_MASK;
> +}
> +
> +/*
> * This function sets the AVPN and L fields of the HPTE appropriately
> * using the base page size and actual page size.
> */
> @@ -341,12 +376,8 @@ static inline unsigned long hpte_encode_v(unsigned long vpn, int base_psize,
> * aligned for the requested page size
> */
> static inline unsigned long hpte_encode_r(unsigned long pa, int base_psize,
> - int actual_psize, int ssize)
> + int actual_psize)
> {
> -
> - if (cpu_has_feature(CPU_FTR_ARCH_300))
> - pa |= ((unsigned long) ssize) << HPTE_R_3_0_SSIZE_SHIFT;
> -
> /* A 4K page needs no special encoding */
> if (actual_psize == MMU_PAGE_4K)
> return pa & HPTE_R_RPN;
> diff --git a/arch/powerpc/mm/hash_native_64.c b/arch/powerpc/mm/hash_native_64.c
> index 83ddc0e..ad9fd52 100644
> --- a/arch/powerpc/mm/hash_native_64.c
> +++ b/arch/powerpc/mm/hash_native_64.c
> @@ -221,13 +221,18 @@ static long native_hpte_insert(unsigned long hpte_group, unsigned long vpn,
> return -1;
>
> hpte_v = hpte_encode_v(vpn, psize, apsize, ssize) | vflags | HPTE_V_VALID;
> - hpte_r = hpte_encode_r(pa, psize, apsize, ssize) | rflags;
> + hpte_r = hpte_encode_r(pa, psize, apsize) | rflags;
>
> if (!(vflags & HPTE_V_BOLTED)) {
> DBG_LOW(" i=%x hpte_v=%016lx, hpte_r=%016lx\n",
> i, hpte_v, hpte_r);
> }
>
> + if (cpu_has_feature(CPU_FTR_ARCH_300)) {
> + hpte_r = hpte_old_to_new_r(hpte_v, hpte_r);
> + hpte_v = hpte_old_to_new_v(hpte_v);
> + }
> +
> hptep->r = cpu_to_be64(hpte_r);
> /* Guarantee the second dword is visible before the valid bit */
> eieio();
> @@ -295,6 +300,8 @@ static long native_hpte_updatepp(unsigned long slot, unsigned long newpp,
> vpn, want_v & HPTE_V_AVPN, slot, newpp);
>
> hpte_v = be64_to_cpu(hptep->v);
> + if (cpu_has_feature(CPU_FTR_ARCH_300))
> + hpte_v = hpte_new_to_old_v(hpte_v, be64_to_cpu(hptep->r));
> /*
> * We need to invalidate the TLB always because hpte_remove doesn't do
> * a tlb invalidate. If a hash bucket gets full, we "evict" a more/less
> @@ -309,6 +316,8 @@ static long native_hpte_updatepp(unsigned long slot, unsigned long newpp,
> native_lock_hpte(hptep);
> /* recheck with locks held */
> hpte_v = be64_to_cpu(hptep->v);
> + if (cpu_has_feature(CPU_FTR_ARCH_300))
> + hpte_v = hpte_new_to_old_v(hpte_v, be64_to_cpu(hptep->r));
> if (unlikely(!HPTE_V_COMPARE(hpte_v, want_v) ||
> !(hpte_v & HPTE_V_VALID))) {
> ret = -1;
> @@ -350,6 +359,8 @@ static long native_hpte_find(unsigned long vpn, int psize, int ssize)
> for (i = 0; i < HPTES_PER_GROUP; i++) {
> hptep = htab_address + slot;
> hpte_v = be64_to_cpu(hptep->v);
> + if (cpu_has_feature(CPU_FTR_ARCH_300))
> + hpte_v = hpte_new_to_old_v(hpte_v, be64_to_cpu(hptep->r));
>
> if (HPTE_V_COMPARE(hpte_v, want_v) && (hpte_v & HPTE_V_VALID))
> /* HPTE matches */
> @@ -409,6 +420,8 @@ static void native_hpte_invalidate(unsigned long slot, unsigned long vpn,
> want_v = hpte_encode_avpn(vpn, bpsize, ssize);
> native_lock_hpte(hptep);
> hpte_v = be64_to_cpu(hptep->v);
> + if (cpu_has_feature(CPU_FTR_ARCH_300))
> + hpte_v = hpte_new_to_old_v(hpte_v, be64_to_cpu(hptep->r));
>
> /*
> * We need to invalidate the TLB always because hpte_remove doesn't do
> @@ -467,6 +480,8 @@ static void native_hugepage_invalidate(unsigned long vsid,
> want_v = hpte_encode_avpn(vpn, psize, ssize);
> native_lock_hpte(hptep);
> hpte_v = be64_to_cpu(hptep->v);
> + if (cpu_has_feature(CPU_FTR_ARCH_300))
> + hpte_v = hpte_new_to_old_v(hpte_v, be64_to_cpu(hptep->r));
>
> /* Even if we miss, we need to invalidate the TLB */
> if (!HPTE_V_COMPARE(hpte_v, want_v) || !(hpte_v & HPTE_V_VALID))
> @@ -504,6 +519,10 @@ static void hpte_decode(struct hash_pte *hpte, unsigned long slot,
> /* Look at the 8 bit LP value */
> unsigned int lp = (hpte_r >> LP_SHIFT) & ((1 << LP_BITS) - 1);
>
> + if (cpu_has_feature(CPU_FTR_ARCH_300)) {
> + hpte_v = hpte_new_to_old_v(hpte_v, hpte_r);
> + hpte_r = hpte_new_to_old_r(hpte_r);
> + }
> if (!(hpte_v & HPTE_V_LARGE)) {
> size = MMU_PAGE_4K;
> a_size = MMU_PAGE_4K;
> @@ -512,11 +531,7 @@ static void hpte_decode(struct hash_pte *hpte, unsigned long slot,
> a_size = hpte_page_sizes[lp] >> 4;
> }
> /* This works for all page sizes, and for 256M and 1T segments */
> - if (cpu_has_feature(CPU_FTR_ARCH_300))
> - *ssize = hpte_r >> HPTE_R_3_0_SSIZE_SHIFT;
> - else
> - *ssize = hpte_v >> HPTE_V_SSIZE_SHIFT;
> -
> + *ssize = hpte_v >> HPTE_V_SSIZE_SHIFT;
> shift = mmu_psize_defs[size].shift;
>
> avpn = (HPTE_V_AVPN_VAL(hpte_v) & ~mmu_psize_defs[size].avpnm);
> @@ -639,6 +654,9 @@ static void native_flush_hash_range(unsigned long number, int local)
> want_v = hpte_encode_avpn(vpn, psize, ssize);
> native_lock_hpte(hptep);
> hpte_v = be64_to_cpu(hptep->v);
> + if (cpu_has_feature(CPU_FTR_ARCH_300))
> + hpte_v = hpte_new_to_old_v(hpte_v,
> + be64_to_cpu(hptep->r));
> if (!HPTE_V_COMPARE(hpte_v, want_v) ||
> !(hpte_v & HPTE_V_VALID))
> native_unlock_hpte(hptep);
> diff --git a/arch/powerpc/platforms/ps3/htab.c b/arch/powerpc/platforms/ps3/htab.c
> index cb3c503..cc2b281 100644
> --- a/arch/powerpc/platforms/ps3/htab.c
> +++ b/arch/powerpc/platforms/ps3/htab.c
> @@ -63,7 +63,7 @@ static long ps3_hpte_insert(unsigned long hpte_group, unsigned long vpn,
> vflags &= ~HPTE_V_SECONDARY;
>
> hpte_v = hpte_encode_v(vpn, psize, apsize, ssize) | vflags | HPTE_V_VALID;
> - hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize, ssize) | rflags;
> + hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize) | rflags;
>
> spin_lock_irqsave(&ps3_htab_lock, flags);
>
> diff --git a/arch/powerpc/platforms/pseries/lpar.c b/arch/powerpc/platforms/pseries/lpar.c
> index aa35245..f2c98f6 100644
> --- a/arch/powerpc/platforms/pseries/lpar.c
> +++ b/arch/powerpc/platforms/pseries/lpar.c
> @@ -145,7 +145,7 @@ static long pSeries_lpar_hpte_insert(unsigned long hpte_group,
> hpte_group, vpn, pa, rflags, vflags, psize);
>
> hpte_v = hpte_encode_v(vpn, psize, apsize, ssize) | vflags | HPTE_V_VALID;
> - hpte_r = hpte_encode_r(pa, psize, apsize, ssize) | rflags;
> + hpte_r = hpte_encode_r(pa, psize, apsize) | rflags;
>
> if (!(vflags & HPTE_V_BOLTED))
> pr_devel(" hpte_v=%016lx, hpte_r=%016lx\n", hpte_v, hpte_r);
> --
> 2.10.1
More information about the Linuxppc-dev
mailing list