[PATCH] powerpc/64s/slice: Use addr limit when computing slice mask

Michael Ellerman mpe at ellerman.id.au
Fri Nov 10 20:44:38 AEDT 2017


"Aneesh Kumar K.V" <aneesh.kumar at linux.vnet.ibm.com> writes:

> While computing slice mask for the free area we need make sure we only search
> in the addr limit applicable for this mmap. We update the slb_addr_limit
> after we request for a mmap above 128TB. But the following mmap request
> with hint addr below 128TB should still limit its search to below 128TB. ie.
> we should not use slb_addr_limit to compute slice mask in this case. Instead,
> we should derive high addr limit based on the mmap hint addr value.
>
> Signed-off-by: Aneesh Kumar K.V <aneesh.kumar at linux.vnet.ibm.com>
> ---
>  arch/powerpc/mm/slice.c | 34 ++++++++++++++++++++++------------
>  1 file changed, 22 insertions(+), 12 deletions(-)

How does this relate to the fixes Nick has sent?

cheers

> diff --git a/arch/powerpc/mm/slice.c b/arch/powerpc/mm/slice.c
> index 564fff06f5c1..23ec2c5e3b78 100644
> --- a/arch/powerpc/mm/slice.c
> +++ b/arch/powerpc/mm/slice.c
> @@ -122,7 +122,8 @@ static int slice_high_has_vma(struct mm_struct *mm, unsigned long slice)
>  	return !slice_area_is_free(mm, start, end - start);
>  }
>  
> -static void slice_mask_for_free(struct mm_struct *mm, struct slice_mask *ret)
> +static void slice_mask_for_free(struct mm_struct *mm, struct slice_mask *ret,
> +				unsigned long high_limit)
>  {
>  	unsigned long i;
>  
> @@ -133,15 +134,16 @@ static void slice_mask_for_free(struct mm_struct *mm, struct slice_mask *ret)
>  		if (!slice_low_has_vma(mm, i))
>  			ret->low_slices |= 1u << i;
>  
> -	if (mm->context.slb_addr_limit <= SLICE_LOW_TOP)
> +	if (high_limit <= SLICE_LOW_TOP)
>  		return;
>  
> -	for (i = 0; i < GET_HIGH_SLICE_INDEX(mm->context.slb_addr_limit); i++)
> +	for (i = 0; i < GET_HIGH_SLICE_INDEX(high_limit); i++)
>  		if (!slice_high_has_vma(mm, i))
>  			__set_bit(i, ret->high_slices);
>  }
>  
> -static void slice_mask_for_size(struct mm_struct *mm, int psize, struct slice_mask *ret)
> +static void slice_mask_for_size(struct mm_struct *mm, int psize, struct slice_mask *ret,
> +				unsigned long high_limit)
>  {
>  	unsigned char *hpsizes;
>  	int index, mask_index;
> @@ -156,8 +158,11 @@ static void slice_mask_for_size(struct mm_struct *mm, int psize, struct slice_ma
>  		if (((lpsizes >> (i * 4)) & 0xf) == psize)
>  			ret->low_slices |= 1u << i;
>  
> +	if (high_limit <= SLICE_LOW_TOP)
> +		return;
> +
>  	hpsizes = mm->context.high_slices_psize;
> -	for (i = 0; i < GET_HIGH_SLICE_INDEX(mm->context.slb_addr_limit); i++) {
> +	for (i = 0; i < GET_HIGH_SLICE_INDEX(high_limit); i++) {
>  		mask_index = i & 0x1;
>  		index = i >> 1;
>  		if (((hpsizes[index] >> (mask_index * 4)) & 0xf) == psize)
> @@ -169,6 +174,10 @@ static int slice_check_fit(struct mm_struct *mm,
>  			   struct slice_mask mask, struct slice_mask available)
>  {
>  	DECLARE_BITMAP(result, SLICE_NUM_HIGH);
> +	/*
> +	 * Make sure we just do bit compare only to the max
> +	 * addr limit and not the full bit map size.
> +	 */
>  	unsigned long slice_count = GET_HIGH_SLICE_INDEX(mm->context.slb_addr_limit);
>  
>  	bitmap_and(result, mask.high_slices,
> @@ -472,7 +481,7 @@ unsigned long slice_get_unmapped_area(unsigned long addr, unsigned long len,
>  	/* First make up a "good" mask of slices that have the right size
>  	 * already
>  	 */
> -	slice_mask_for_size(mm, psize, &good_mask);
> +	slice_mask_for_size(mm, psize, &good_mask, high_limit);
>  	slice_print_mask(" good_mask", good_mask);
>  
>  	/*
> @@ -497,7 +506,7 @@ unsigned long slice_get_unmapped_area(unsigned long addr, unsigned long len,
>  #ifdef CONFIG_PPC_64K_PAGES
>  	/* If we support combo pages, we can allow 64k pages in 4k slices */
>  	if (psize == MMU_PAGE_64K) {
> -		slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask);
> +		slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask, high_limit);
>  		if (fixed)
>  			slice_or_mask(&good_mask, &compat_mask);
>  	}
> @@ -530,11 +539,11 @@ unsigned long slice_get_unmapped_area(unsigned long addr, unsigned long len,
>  			return newaddr;
>  		}
>  	}
> -
> -	/* We don't fit in the good mask, check what other slices are
> +	/*
> +	 * We don't fit in the good mask, check what other slices are
>  	 * empty and thus can be converted
>  	 */
> -	slice_mask_for_free(mm, &potential_mask);
> +	slice_mask_for_free(mm, &potential_mask, high_limit);
>  	slice_or_mask(&potential_mask, &good_mask);
>  	slice_print_mask(" potential", potential_mask);
>  
> @@ -744,17 +753,18 @@ int is_hugepage_only_range(struct mm_struct *mm, unsigned long addr,
>  {
>  	struct slice_mask mask, available;
>  	unsigned int psize = mm->context.user_psize;
> +	unsigned long high_limit = mm->context.slb_addr_limit;
>  
>  	if (radix_enabled())
>  		return 0;
>  
>  	slice_range_to_mask(addr, len, &mask);
> -	slice_mask_for_size(mm, psize, &available);
> +	slice_mask_for_size(mm, psize, &available, high_limit);
>  #ifdef CONFIG_PPC_64K_PAGES
>  	/* We need to account for 4k slices too */
>  	if (psize == MMU_PAGE_64K) {
>  		struct slice_mask compat_mask;
> -		slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask);
> +		slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask, high_limit);
>  		slice_or_mask(&available, &compat_mask);
>  	}
>  #endif
> -- 
> 2.13.6


More information about the Linuxppc-dev mailing list