[PATCH 14/18] KVM: PPC: Book3S HV: MMU notifier callbacks for radix guests

Suraj Jitindar Singh sjitindarsingh at gmail.com
Mon Jan 23 14:21:00 AEDT 2017


On Thu, 2017-01-12 at 20:07 +1100, Paul Mackerras wrote:
> This adapts our implementations of the MMU notifier callbacks
> (unmap_hva, unmap_hva_range, age_hva, test_age_hva, set_spte_hva)
> to call radix functions when the guest is using radix.  These
> implementations are much simpler than for HPT guests because we
> have only one PTE to deal with, so we don't need to traverse
> rmap chains.
> 
> Signed-off-by: Paul Mackerras <paulus at ozlabs.org>
> ---
>  arch/powerpc/include/asm/kvm_book3s.h  |  6 ++++
>  arch/powerpc/kvm/book3s_64_mmu_hv.c    | 64 +++++++++++++++++++++++-
> ----------
>  arch/powerpc/kvm/book3s_64_mmu_radix.c | 54
> ++++++++++++++++++++++++++++
>  3 files changed, 103 insertions(+), 21 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/kvm_book3s.h
> b/arch/powerpc/include/asm/kvm_book3s.h
> index ff5cd5c..952cc4b 100644
> --- a/arch/powerpc/include/asm/kvm_book3s.h
> +++ b/arch/powerpc/include/asm/kvm_book3s.h
> @@ -192,6 +192,12 @@ extern int kvmppc_mmu_radix_xlate(struct
> kvm_vcpu *vcpu, gva_t eaddr,
>  extern void kvmppc_free_radix(struct kvm *kvm);
>  extern int kvmppc_radix_init(void);
>  extern void kvmppc_radix_exit(void);
> +extern int kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +			unsigned long gfn);
> +extern int kvm_age_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +			unsigned long gfn);
> +extern int kvm_test_age_radix(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> +			unsigned long gfn);
>  
>  /* XXX remove this export when load_last_inst() is generic */
>  extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size,
> void *ptr, bool data);
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c
> b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> index 57690c2..fbb3de4 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> @@ -701,12 +701,13 @@ static void kvmppc_rmap_reset(struct kvm *kvm)
>  	srcu_read_unlock(&kvm->srcu, srcu_idx);
>  }
>  
> +typedef int (*hva_handler_fn)(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> +			      unsigned long gfn);
> +
>  static int kvm_handle_hva_range(struct kvm *kvm,
>  				unsigned long start,
>  				unsigned long end,
> -				int (*handler)(struct kvm *kvm,
> -					       unsigned long *rmapp,
> -					       unsigned long gfn))
> +				hva_handler_fn handler)
>  {
>  	int ret;
>  	int retval = 0;
> @@ -731,9 +732,7 @@ static int kvm_handle_hva_range(struct kvm *kvm,
>  		gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE -
> 1, memslot);
>  
>  		for (; gfn < gfn_end; ++gfn) {
> -			gfn_t gfn_offset = gfn - memslot->base_gfn;
> -
> -			ret = handler(kvm, &memslot-
> >arch.rmap[gfn_offset], gfn);
> +			ret = handler(kvm, memslot, gfn);
>  			retval |= ret;
>  		}
>  	}
> @@ -742,20 +741,21 @@ static int kvm_handle_hva_range(struct kvm
> *kvm,
>  }
>  
>  static int kvm_handle_hva(struct kvm *kvm, unsigned long hva,
> -			  int (*handler)(struct kvm *kvm, unsigned
> long *rmapp,
> -					 unsigned long gfn))
> +			  hva_handler_fn handler)
>  {
>  	return kvm_handle_hva_range(kvm, hva, hva + 1, handler);
>  }
>  
> -static int kvm_unmap_rmapp(struct kvm *kvm, unsigned long *rmapp,
> +static int kvm_unmap_rmapp(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
>  			   unsigned long gfn)
>  {
>  	struct revmap_entry *rev = kvm->arch.revmap;
>  	unsigned long h, i, j;
>  	__be64 *hptep;
>  	unsigned long ptel, psize, rcbits;
> +	unsigned long *rmapp;
>  
> +	rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
>  	for (;;) {
>  		lock_rmap(rmapp);
>  		if (!(*rmapp & KVMPPC_RMAP_PRESENT)) {
> @@ -816,26 +816,36 @@ static int kvm_unmap_rmapp(struct kvm *kvm,
> unsigned long *rmapp,
>  
>  int kvm_unmap_hva_hv(struct kvm *kvm, unsigned long hva)
>  {
> -	kvm_handle_hva(kvm, hva, kvm_unmap_rmapp);
> +	hva_handler_fn handler;
> +
> +	handler = kvm->arch.radix ? kvm_unmap_radix : 
kvm_is_radix() for consistency?
> kvm_unmap_rmapp;
> +	kvm_handle_hva(kvm, hva, handler);
>  	return 0;
>  }
>  
>  int kvm_unmap_hva_range_hv(struct kvm *kvm, unsigned long start,
> unsigned long end)
>  {
> -	kvm_handle_hva_range(kvm, start, end, kvm_unmap_rmapp);
> +	hva_handler_fn handler;
> +
> +	handler = kvm->arch.radix ? kvm_unmap_radix : 
ditto
> kvm_unmap_rmapp;
> +	kvm_handle_hva_range(kvm, start, end, handler);
>  	return 0;
>  }
>  
>  void kvmppc_core_flush_memslot_hv(struct kvm *kvm,
>  				  struct kvm_memory_slot *memslot)
>  {
> -	unsigned long *rmapp;
>  	unsigned long gfn;
>  	unsigned long n;
> +	unsigned long *rmapp;
>  
> -	rmapp = memslot->arch.rmap;
>  	gfn = memslot->base_gfn;
> -	for (n = memslot->npages; n; --n) {
> +	rmapp = memslot->arch.rmap;
> +	for (n = memslot->npages; n; --n, ++gfn) {
> +		if (kvm->arch.radix) {
ditto
> +			kvm_unmap_radix(kvm, memslot, gfn);
> +			continue;
> +		}
>  		/*
>  		 * Testing the present bit without locking is OK
> because
>  		 * the memslot has been marked invalid already, and
> hence
> @@ -843,20 +853,21 @@ void kvmppc_core_flush_memslot_hv(struct kvm
> *kvm,
>  		 * thus the present bit can't go from 0 to 1.
>  		 */
>  		if (*rmapp & KVMPPC_RMAP_PRESENT)
> -			kvm_unmap_rmapp(kvm, rmapp, gfn);
> +			kvm_unmap_rmapp(kvm, memslot, gfn);
>  		++rmapp;
> -		++gfn;
>  	}
>  }
>  
> -static int kvm_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
> +static int kvm_age_rmapp(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
>  			 unsigned long gfn)
>  {
>  	struct revmap_entry *rev = kvm->arch.revmap;
>  	unsigned long head, i, j;
>  	__be64 *hptep;
>  	int ret = 0;
> +	unsigned long *rmapp;
>  
> +	rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
>   retry:
>  	lock_rmap(rmapp);
>  	if (*rmapp & KVMPPC_RMAP_REFERENCED) {
> @@ -904,17 +915,22 @@ static int kvm_age_rmapp(struct kvm *kvm,
> unsigned long *rmapp,
>  
>  int kvm_age_hva_hv(struct kvm *kvm, unsigned long start, unsigned
> long end)
>  {
> -	return kvm_handle_hva_range(kvm, start, end, kvm_age_rmapp);
> +	hva_handler_fn handler;
> +
> +	handler = kvm->arch.radix ? kvm_age_radix : kvm_age_rmapp;
ditto
> +	return kvm_handle_hva_range(kvm, start, end, handler);
>  }
>  
> -static int kvm_test_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
> +static int kvm_test_age_rmapp(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
>  			      unsigned long gfn)
>  {
>  	struct revmap_entry *rev = kvm->arch.revmap;
>  	unsigned long head, i, j;
>  	unsigned long *hp;
>  	int ret = 1;
> +	unsigned long *rmapp;
>  
> +	rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
>  	if (*rmapp & KVMPPC_RMAP_REFERENCED)
>  		return 1;
>  
> @@ -940,12 +956,18 @@ static int kvm_test_age_rmapp(struct kvm *kvm,
> unsigned long *rmapp,
>  
>  int kvm_test_age_hva_hv(struct kvm *kvm, unsigned long hva)
>  {
> -	return kvm_handle_hva(kvm, hva, kvm_test_age_rmapp);
> +	hva_handler_fn handler;
> +
> +	handler = kvm->arch.radix ? kvm_test_age_radix : 
ditto
> kvm_test_age_rmapp;
> +	return kvm_handle_hva(kvm, hva, handler);
>  }
>  
>  void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t
> pte)
>  {
> -	kvm_handle_hva(kvm, hva, kvm_unmap_rmapp);
> +	hva_handler_fn handler;
> +
> +	handler = kvm->arch.radix ? kvm_unmap_radix : 
ditto
> kvm_unmap_rmapp;
> +	kvm_handle_hva(kvm, hva, handler);
>  }
>  
>  static int vcpus_running(struct kvm *kvm)
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> index 865ea9b..69cabad 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> @@ -463,6 +463,60 @@ int kvmppc_book3s_radix_page_fault(struct
> kvm_run *run, struct kvm_vcpu *vcpu,
>  	return ret;
>  }
>  
> +/* Called with kvm->lock held */
> +int kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +		    unsigned long gfn)
> +{
> +	pte_t *ptep;
> +	unsigned long gpa = gfn << PAGE_SHIFT;
> +	unsigned int shift;
> +
> +	ptep = __find_linux_pte_or_hugepte(kvm->arch.pgtable, gpa,
> +					   NULL, &shift);
> +	if (ptep && pte_present(*ptep)) {
> +		kvmppc_radix_update_pte(kvm, ptep, _PAGE_PRESENT, 0,
> +					gpa, shift);
> +		kvmppc_radix_tlbie_page(kvm, gpa, shift);
> +	}
> +	return 0;				
> +}
> +
> +/* Called with kvm->lock held */
> +int kvm_age_radix(struct kvm *kvm, struct kvm_memory_slot *memslot,
> +		  unsigned long gfn)
> +{
> +	pte_t *ptep;
> +	unsigned long gpa = gfn << PAGE_SHIFT;
> +	unsigned int shift;
> +	int ref = 0;
> +
> +	ptep = __find_linux_pte_or_hugepte(kvm->arch.pgtable, gpa,
> +					   NULL, &shift);
> +	if (ptep && pte_present(*ptep) && pte_young(*ptep)) {
> +		kvmppc_radix_update_pte(kvm, ptep, _PAGE_ACCESSED,
> 0,
> +					gpa, shift);
> +		/* XXX need to flush tlb here? */
> +		ref = 1;
> +	}
> +	return ref;
> +}
> +
> +/* Called with kvm->lock held */
> +int kvm_test_age_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +		       unsigned long gfn)
> +{
> +	pte_t *ptep;
> +	unsigned long gpa = gfn << PAGE_SHIFT;
> +	unsigned int shift;
> +	int ref = 0;
> +
> +	ptep = __find_linux_pte_or_hugepte(kvm->arch.pgtable, gpa,
> +					   NULL, &shift);
> +	if (ptep && pte_present(*ptep) && pte_young(*ptep))
> +		ref = 1;
> +	return ref;
> +}
> +
>  void kvmppc_free_radix(struct kvm *kvm)
>  {
>  	unsigned long ig, iu, im;


More information about the Linuxppc-dev mailing list