[PATCH v5 1/7] powerpc/mm: update ptep_set_access_flag to not do full mm tlb flush
Aneesh Kumar K.V
aneesh.kumar at linux.vnet.ibm.com
Thu Nov 24 01:36:00 AEDT 2016
Balbir Singh <bsingharora at gmail.com> writes:
> On 23/11/16 22:53, Aneesh Kumar K.V wrote:
>> Balbir Singh <bsingharora at gmail.com> writes:
>>
>>> On 23/11/16 22:09, Aneesh Kumar K.V wrote:
>>>> When we are updating pte, we just need to flush the tlb mapping for
>>>> that pte. Right now we do a full mm flush because we don't track page
>>>> size. Update the interface to track the page size and use that to
>>>> do the right tlb flush.
>>>>
>>>
>>> Could you also clarify the scope -- this seems to be _radix_ only.
>>> The problem statement is not very clear and why doesn't the flush_tlb_page()
>>> following ptep_set_access_flags() work? What else do we need to do?
>>
>> Yes it modifies only radix part. Don't understand the flush_tlb_page()
>> part of the comment above. We are modifying the tlbflush that we need to do in the pte update
>> sequence for DD1. ie, we need to do the flush before we can set the pte
>> with new value.
>>
>> Also in this specific case, we can idealy drop that flush_tlb_page,
>> because relaxing an access really don't need a tlb flush from generic
>> architecture point of view. I left it there to make sure, we measure and
>> get the invalidate path correct before going ahead with that
>> optimization.
>>
>
> OK.. here is my untested solution. I've only compiled it.
> It breaks the 64/hash/radix abstractions, but it makes the
> changes much simpler
>
> Signed-off-by: Balbir Singh <bsingharora at gmail.com>
I find the below one more confusing and complicated, spreading the
details of DD1 around the code. I am not sure what extra i could have
done to simplify the code. We have done the arch pte updates such that
most of the update use the pte_update() interface and the one which relax
the access bits get to ptep_set_access_flag. All pte updated rules are
contained there. What you did below is that you moved the dd1 sequence
out to a place where page size is available. What I did in my patch is to
pass page size around. IMHO it is a matter of style. I also want to pass
page size around so that we keep huge_pte_update, pte_update,
ptep_set_access_flags all similar.
>
> diff --git a/arch/powerpc/include/asm/book3s/64/radix.h b/arch/powerpc/include/asm/book3s/64/radix.h
> index 2a46dea..2454217 100644
> --- a/arch/powerpc/include/asm/book3s/64/radix.h
> +++ b/arch/powerpc/include/asm/book3s/64/radix.h
> @@ -162,6 +162,30 @@ static inline unsigned long radix__pte_update(struct mm_struct *mm,
> return old_pte;
> }
>
> +static inline void radix__ptep_dd1_set_access_flags(struct mm_struct *mm,
> + unsigned long addr,
> + pte_t *ptep, pte_t entry,
> + unsigned long page_size)
> +{
> +
> + unsigned long set = pte_val(entry) & (_PAGE_DIRTY | _PAGE_ACCESSED |
> + _PAGE_RW | _PAGE_EXEC);
> +
> + unsigned long old_pte, new_pte;
> +
> + old_pte = __radix_pte_update(ptep, ~0, 0);
> + asm volatile("ptesync" : : : "memory");
> + /*
> + * new value of pte
> + */
> + new_pte = old_pte | set;
> +
> + radix__flush_tlb_page_psize(mm, addr, page_size);
> + __radix_pte_update(ptep, 0, new_pte);
> +
> + asm volatile("ptesync" : : : "memory");
> +}
> +
> /*
> * Set the dirty and/or accessed bits atomically in a linux PTE, this
> * function doesn't need to invalidate tlb.
> @@ -173,26 +197,7 @@ static inline void radix__ptep_set_access_flags(struct mm_struct *mm,
> unsigned long set = pte_val(entry) & (_PAGE_DIRTY | _PAGE_ACCESSED |
> _PAGE_RW | _PAGE_EXEC);
>
> - if (cpu_has_feature(CPU_FTR_POWER9_DD1)) {
> -
> - unsigned long old_pte, new_pte;
> -
> - old_pte = __radix_pte_update(ptep, ~0, 0);
> - asm volatile("ptesync" : : : "memory");
> - /*
> - * new value of pte
> - */
> - new_pte = old_pte | set;
> -
> - /*
> - * For now let's do heavy pid flush
> - * radix__flush_tlb_page_psize(mm, addr, mmu_virtual_psize);
> - */
> - radix__flush_tlb_mm(mm);
> -
> - __radix_pte_update(ptep, 0, new_pte);
> - } else
> - __radix_pte_update(ptep, 0, set);
> + __radix_pte_update(ptep, 0, set);
> asm volatile("ptesync" : : : "memory");
> }
>
> diff --git a/arch/powerpc/mm/pgtable-book3s64.c b/arch/powerpc/mm/pgtable-book3s64.c
> index f4f437c..0c7ee0e 100644
> --- a/arch/powerpc/mm/pgtable-book3s64.c
> +++ b/arch/powerpc/mm/pgtable-book3s64.c
> @@ -7,12 +7,14 @@
> * 2 of the License, or (at your option) any later version.
> */
>
> +#include <linux/kernel.h>
> #include <linux/sched.h>
> #include <asm/pgalloc.h>
> #include <asm/tlb.h>
>
> #include "mmu_decl.h"
> #include <trace/events/thp.h>
> +#include <linux/hugetlb.h>
>
> int (*register_process_table)(unsigned long base, unsigned long page_size,
> unsigned long tbl_size);
> @@ -35,7 +37,15 @@ int pmdp_set_access_flags(struct vm_area_struct *vma, unsigned long address,
> #endif
> changed = !pmd_same(*(pmdp), entry);
> if (changed) {
> - __ptep_set_access_flags(vma->vm_mm, pmdp_ptep(pmdp), pmd_pte(entry));
> + if (radix_enabled() && cpu_has_feature(CPU_FTR_POWER9_DD1)) {
> + unsigned long page_size;
> + page_size = is_vm_hugetlb_page(vma) ? PAGE_SIZE :
> + huge_page_size(hstate_vma(vma));
> +
> + radix__ptep_dd1_set_access_flags(vma->vm_mm, address,
> + pmdp_ptep(pmdp), pmd_pte(entry), page_size);
> + } else
> + __ptep_set_access_flags(vma->vm_mm, pmdp_ptep(pmdp), pmd_pte(entry));
> flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
> }
> return changed;
> diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
> index 911fdfb..ebb464e 100644
> --- a/arch/powerpc/mm/pgtable.c
> +++ b/arch/powerpc/mm/pgtable.c
> @@ -224,7 +224,16 @@ int ptep_set_access_flags(struct vm_area_struct *vma, unsigned long address,
> if (changed) {
> if (!is_vm_hugetlb_page(vma))
> assert_pte_locked(vma->vm_mm, address);
> - __ptep_set_access_flags(vma->vm_mm, ptep, entry);
> + /* Workaround to call radix update directly for DD1 */
> + if (radix_enabled() && cpu_has_feature(CPU_FTR_POWER9_DD1)) {
> + unsigned long page_size;
> + page_size = is_vm_hugetlb_page(vma) ? PAGE_SIZE :
> + huge_page_size(hstate_vma(vma));
> +
> + radix__ptep_dd1_set_access_flags(vma->vm_mm, address,
> + ptep, entry, page_size);
> + } else
> + __ptep_set_access_flags(vma->vm_mm, ptep, entry);
> flush_tlb_page(vma, address);
> }
> return changed;
More information about the Linuxppc-dev
mailing list