[PATCH v2] powerpc/64s/radix: do not flush TLB when relaxing access
Aneesh Kumar K.V
aneesh.kumar at linux.vnet.ibm.com
Mon May 21 16:02:02 AEST 2018
Nicholas Piggin <npiggin at gmail.com> writes:
> Radix flushes the TLB when updating ptes to increase permissiveness
> of protection (increase access authority). Book3S does not require
> TLB flushing in this case, and it is not done on hash. This patch
> avoids the flush for radix.
>
> From Power ISA v3.0B, p.1090:
>
> Setting a Reference or Change Bit or Upgrading Access Authority
> (PTE Subject to Atomic Hardware Updates)
>
> If the only change being made to a valid PTE that is subject to
> atomic hardware updates is to set the Reference or Change bit to 1
> or to add access authorities, a simpler sequence suffices because
> the translation hardware will refetch the PTE if an access is
> attempted for which the only problems were reference and/or change
> bits needing to be set or insufficient access authority.
>
> The nest MMU on POWER9 does not re-fetch the PTE after such an access
> attempt before faulting, so address spaces with a coprocessor
> attached will continue to flush in these cases.
>
> This reduces tlbies for a kernel compile workload from 1.28M to 0.95M,
> tlbiels from 20.17M 19.68M.
>
> fork --fork --exec benchmark improved 2.77% (12000->12300).
>
Reviewed-by: Aneesh Kumar K.V <aneesh.kumar at linux.vnet.ibm.com>
> Signed-off-by: Nicholas Piggin <npiggin at gmail.com>
> ---
> Oops I missed this patch, it's supposed to go as the first patch in
> the "Various TLB and PTE improvements" patch.
>
> arch/powerpc/mm/pgtable-book3s64.c | 10 +++++++---
> arch/powerpc/mm/pgtable.c | 29 ++++++++++++++++++++++++++---
> 2 files changed, 33 insertions(+), 6 deletions(-)
>
> diff --git a/arch/powerpc/mm/pgtable-book3s64.c b/arch/powerpc/mm/pgtable-book3s64.c
> index 518518fb7c45..994492453f0e 100644
> --- a/arch/powerpc/mm/pgtable-book3s64.c
> +++ b/arch/powerpc/mm/pgtable-book3s64.c
> @@ -31,16 +31,20 @@ int (*register_process_table)(unsigned long base, unsigned long page_size,
> int pmdp_set_access_flags(struct vm_area_struct *vma, unsigned long address,
> pmd_t *pmdp, pmd_t entry, int dirty)
> {
> + struct mm_struct *mm = vma->vm_mm;
> int changed;
> #ifdef CONFIG_DEBUG_VM
> WARN_ON(!pmd_trans_huge(*pmdp) && !pmd_devmap(*pmdp));
> - assert_spin_locked(&vma->vm_mm->page_table_lock);
> + assert_spin_locked(&mm->page_table_lock);
> #endif
> changed = !pmd_same(*(pmdp), entry);
> if (changed) {
> - __ptep_set_access_flags(vma->vm_mm, pmdp_ptep(pmdp),
> + __ptep_set_access_flags(mm, pmdp_ptep(pmdp),
> pmd_pte(entry), address);
> - flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
> + /* See ptep_set_access_flags comments */
> + if (atomic_read(&mm->context.copros) > 0)
> + flush_pmd_tlb_range(vma, address,
> + address + HPAGE_PMD_SIZE);
> }
> return changed;
> }
> diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
> index 9f361ae571e9..525ec4656a55 100644
> --- a/arch/powerpc/mm/pgtable.c
> +++ b/arch/powerpc/mm/pgtable.c
> @@ -217,14 +217,37 @@ void set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep,
> int ptep_set_access_flags(struct vm_area_struct *vma, unsigned long address,
> pte_t *ptep, pte_t entry, int dirty)
> {
> + struct mm_struct *mm = vma->vm_mm;
> int changed;
> +
> entry = set_access_flags_filter(entry, vma, dirty);
> changed = !pte_same(*(ptep), entry);
> if (changed) {
> if (!is_vm_hugetlb_page(vma))
> - assert_pte_locked(vma->vm_mm, address);
> - __ptep_set_access_flags(vma->vm_mm, ptep, entry, address);
> - flush_tlb_page(vma, address);
> + assert_pte_locked(mm, address);
> + __ptep_set_access_flags(mm, ptep, entry, address);
> + if (IS_ENABLED(CONFIG_PPC_BOOK3S_64)) {
> + /*
> + * Book3S does not require a TLB flush when relaxing
> + * access restrictions because the core MMU will reload
> + * the pte after taking an access fault. However the
> + * NMMU on POWER9 does not re-load the pte, so flush
> + * if we have a coprocessor attached to this address
> + * space.
> + *
> + * This could be further refined and pushed out to
> + * NMMU drivers so TLBIEs are only done for NMMU
> + * faults, but this is a more minimal fix. The NMMU
> + * fault handler does a get_user_pages_remote or
> + * similar to bring the page tables in, and this
> + * flush_tlb_page will do a global TLBIE because the
> + * coprocessor is attached to the address space.
> + */
> + if (atomic_read(&mm->context.copros) > 0)
> + flush_tlb_page(vma, address);
> + } else {
> + flush_tlb_page(vma, address);
> + }
> }
> return changed;
> }
> --
> 2.17.0
More information about the Linuxppc-dev
mailing list