[RFC PATCH] powerpc/mm: Reduce memory usage for mm_context_t for radix

Nicholas Piggin npiggin at gmail.com
Fri Apr 5 03:13:30 AEDT 2019


Christophe Leroy's on April 3, 2019 4:31 am:
> 
> 
> Le 02/04/2019 à 16:34, Aneesh Kumar K.V a écrit :
>> Currently, our mm_context_t on book3s64 include all hash specific
>> context details like slice mask, subpage protection details. We
>> can skip allocating those on radix. This will help us to save
>> 8K per mm_context with radix translation.
>> 
>> With the patch applied we have
>> 
>> sizeof(mm_context_t)  = 136
>> sizeof(struct hash_mm_context)  = 8288
>> 
>> Signed-off-by: Aneesh Kumar K.V <aneesh.kumar at linux.ibm.com>
>> ---
>> NOTE:
>> 
>> If we want to do this, I am still trying to figure out how best we can do this
>> without all the #ifdef and other overhead for 8xx book3e
>> 
>> 
>>   arch/powerpc/include/asm/book3s/64/mmu-hash.h |  2 +-
>>   arch/powerpc/include/asm/book3s/64/mmu.h      | 48 +++++++++++--------
>>   arch/powerpc/include/asm/book3s/64/slice.h    |  6 +--
>>   arch/powerpc/kernel/paca.c                    |  9 ++--
>>   arch/powerpc/kernel/setup-common.c            |  7 ++-
>>   arch/powerpc/mm/hash_utils_64.c               | 10 ++--
>>   arch/powerpc/mm/mmu_context_book3s64.c        | 16 ++++++-
>>   arch/powerpc/mm/slb.c                         |  2 +-
>>   arch/powerpc/mm/slice.c                       | 48 +++++++++----------
>>   arch/powerpc/mm/subpage-prot.c                |  8 ++--
>>   10 files changed, 91 insertions(+), 65 deletions(-)
>> 
>> diff --git a/arch/powerpc/include/asm/book3s/64/mmu-hash.h b/arch/powerpc/include/asm/book3s/64/mmu-hash.h
>> index a28a28079edb..d801be977623 100644
>> --- a/arch/powerpc/include/asm/book3s/64/mmu-hash.h
>> +++ b/arch/powerpc/include/asm/book3s/64/mmu-hash.h
>> @@ -657,7 +657,7 @@ extern void slb_set_size(u16 size);
>>   
>>   /* 4 bits per slice and we have one slice per 1TB */
>>   #define SLICE_ARRAY_SIZE	(H_PGTABLE_RANGE >> 41)
>> -#define TASK_SLICE_ARRAY_SZ(x)	((x)->context.slb_addr_limit >> 41)
>> +#define TASK_SLICE_ARRAY_SZ(x)	((x)->context.hash_context->slb_addr_limit >> 41)
>>   
>>   #ifndef __ASSEMBLY__
>>   
>> diff --git a/arch/powerpc/include/asm/book3s/64/mmu.h b/arch/powerpc/include/asm/book3s/64/mmu.h
>> index a809bdd77322..07e76e304a3b 100644
>> --- a/arch/powerpc/include/asm/book3s/64/mmu.h
>> +++ b/arch/powerpc/include/asm/book3s/64/mmu.h
>> @@ -114,6 +114,33 @@ struct slice_mask {
>>   	DECLARE_BITMAP(high_slices, SLICE_NUM_HIGH);
>>   };
>>   
>> +struct hash_mm_context {
>> +
>> +	u16 user_psize; /* page size index */
> 
> Could we keep that in mm_context_t ?

Why do you want it there?


>> @@ -155,15 +155,15 @@ static struct slice_mask *slice_mask_for_size(struct mm_struct *mm, int psize)
>>   {
>>   #ifdef CONFIG_PPC_64K_PAGES
>>   	if (psize == MMU_PAGE_64K)
>> -		return &mm->context.mask_64k;
>> +		return &mm->context.hash_context->mask_64k;
> 
> You should take the two patches below, that would help:
> https://patchwork.ozlabs.org/patch/1059056/
> https://patchwork.ozlabs.org/patch/1059058/

Above patches seem good I think. What I think Aneesh should have is
is a macro or inline that gives a pointer to the hash_mm_context from
an mm_context pointer.

Architectures which always want it should just put the hash struct
in their mm_context struct and that avoids the pointer overhead
completely.

Thanks,
Nick



More information about the Linuxppc-dev mailing list