[PATCH v4 09/22] powerpc/kvm/book3s: Add helper to walk partition scoped linux page table.

Aneesh Kumar K.V aneesh.kumar at linux.ibm.com
Thu May 28 16:01:04 AEST 2020


On 5/28/20 7:13 AM, Paul Mackerras wrote:
> On Tue, May 05, 2020 at 12:47:16PM +0530, Aneesh Kumar K.V wrote:
>> The locking rules for walking partition scoped table is different from process
>> scoped table. Hence add a helper for secondary linux page table walk and also
>> add check whether we are holding the right locks.
> 
> This patch is causing new warnings to appear when testing migration,
> like this:
> 
> [  142.090159] ------------[ cut here ]------------
> [  142.090160] find_kvm_secondary_pte called with kvm mmu_lock not held
> [  142.090176] WARNING: CPU: 23 PID: 5341 at arch/powerpc/include/asm/kvm_book3s_64.h:644 kvmppc_hv_get_dirty_log_radix+0x2e4/0x340 [kvm_hv]
> [  142.090177] Modules linked in: xt_conntrack nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 libcrc32c bridge stp llc ebtable_filter ebtables ip6table_filter ip6_tables bpfilter overlay binfmt_misc input_leds raid_class scsi_transport_sas sch_fq_codel sunrpc kvm_hv kvm
> [  142.090188] CPU: 23 PID: 5341 Comm: qemu-system-ppc Tainted: G        W         5.7.0-rc5-kvm-00211-g9ccf10d6d088 #432
> [  142.090189] NIP:  c008000000fe848c LR: c008000000fe8488 CTR: 0000000000000000
> [  142.090190] REGS: c000001e19f077e0 TRAP: 0700   Tainted: G        W          (5.7.0-rc5-kvm-00211-g9ccf10d6d088)
> [  142.090191] MSR:  9000000000029033 <SF,HV,EE,ME,IR,DR,RI,LE>  CR: 42222422  XER: 20040000
> [  142.090196] CFAR: c00000000012f5ac IRQMASK: 0
>                 GPR00: c008000000fe8488 c000001e19f07a70 c008000000ffe200 0000000000000039
>                 GPR04: 0000000000000001 c000001ffc8b4900 0000000000018840 0000000000000007
>                 GPR08: 0000000000000003 0000000000000001 0000000000000007 0000000000000001
>                 GPR12: 0000000000002000 c000001fff6d9400 000000011f884678 00007fff70b70000
>                 GPR16: 00007fff7137cb90 00007fff7dcb4410 0000000000000001 0000000000000000
>                 GPR20: 000000000ffe0000 0000000000000000 0000000000000001 0000000000000000
>                 GPR24: 8000000000000000 0000000000000001 c000001e1f67e600 c000001e1fd82410
>                 GPR28: 0000000000001000 c000001e2e410000 0000000000000fff 0000000000000ffe
> [  142.090217] NIP [c008000000fe848c] kvmppc_hv_get_dirty_log_radix+0x2e4/0x340 [kvm_hv]
> [  142.090223] LR [c008000000fe8488] kvmppc_hv_get_dirty_log_radix+0x2e0/0x340 [kvm_hv]
> [  142.090224] Call Trace:
> [  142.090230] [c000001e19f07a70] [c008000000fe8488] kvmppc_hv_get_dirty_log_radix+0x2e0/0x340 [kvm_hv] (unreliable)
> [  142.090236] [c000001e19f07b50] [c008000000fd42e4] kvm_vm_ioctl_get_dirty_log_hv+0x33c/0x3c0 [kvm_hv]
> [  142.090292] [c000001e19f07be0] [c008000000eea878] kvm_vm_ioctl_get_dirty_log+0x30/0x50 [kvm]
> [  142.090300] [c000001e19f07c00] [c008000000edc818] kvm_vm_ioctl+0x2b0/0xc00 [kvm]
> [  142.090302] [c000001e19f07d50] [c00000000046e148] ksys_ioctl+0xf8/0x150
> [  142.090305] [c000001e19f07da0] [c00000000046e1c8] sys_ioctl+0x28/0x80
> [  142.090307] [c000001e19f07dc0] [c00000000003652c] system_call_exception+0x16c/0x240
> [  142.090309] [c000001e19f07e20] [c00000000000d070] system_call_common+0xf0/0x278
> [  142.090310] Instruction dump:
> [  142.090312] 7d3a512a 4200ffd0 7ffefb78 4bfffdc4 60000000 3c820000 e8848468 3c620000
> [  142.090317] e86384a8 38840010 4800673d e8410018 <0fe00000> 4bfffdd4 60000000 60000000
> [  142.090322] ---[ end trace 619d45057b6919e0 ]---
> 
> Indeed, kvm_radix_test_clear_dirty() tests the PTE dirty bit
> locklessly, and only takes the kvm->mmu_lock once it finds a dirty
> PTE.  I think that is important for performance, since on any given
> scan of the guest real address space we may only find a small
> proportion of the guest pages to be dirty.
> 
> Are you now relying on the kvm->mmu_lock to protect the existence of
> the PTEs, or just their content?
> 

The patch series should not change any rules w.r.t kvm partition scoped 
page table walk. We only added helpers to make it explicit that this is 
different from regular linux page table walk. And kvm->mmu_lock is what 
was used to protect the partition scoped table walk earlier.

In this specific case, what we need probably is an open coded kvm 
partition scoped walk with a comment around explaining why is it ok to 
walk that partition scoped table without taking kvm->mmu_lock.

What happens when a parallel invalidate happens to Qemu address space? 
Since we are not holding kvm->mmu_lock mmu notifier will complete and we 
will go ahead with unmapping partition scoped table.

Do we need a change like below?

@@ -1040,7 +1040,7 @@ static int kvm_radix_test_clear_dirty(struct kvm *kvm,
  {
  	unsigned long gfn = memslot->base_gfn + pagenum;
  	unsigned long gpa = gfn << PAGE_SHIFT;
-	pte_t *ptep;
+	pte_t *ptep, pte;
  	unsigned int shift;
  	int ret = 0;
  	unsigned long old, *rmapp;
@@ -1049,11 +1049,23 @@ static int kvm_radix_test_clear_dirty(struct kvm 
*kvm,
  		return ret;

  	ptep = find_kvm_secondary_pte(kvm, gpa, &shift);
-	if (ptep && pte_present(*ptep) && pte_dirty(*ptep)) {
+	if (!ptep)
+		return 0;
+
+	pte = READ_ONCE(*ptep);
+	if (pte_present(pte) && pte_dirty(pte)) {
  		ret = 1;
  		if (shift)
  			ret = 1 << (shift - PAGE_SHIFT);
  		spin_lock(&kvm->mmu_lock);
+		/*
+		 * Recheck the pte again
+		 */
+		if (pte_val(pte) != pte_val(*ptep)) {
+			spin_unlock(&kvm->mmu_lock);
+			return 0;
+		}
+
  		old = kvmppc_radix_update_pte(kvm, ptep, _PAGE_DIRTY, 0,
  					      gpa, shift);
  		kvmppc_radix_tlbie_page(kvm, gpa, shift, kvm->arch.lpid);



More information about the Linuxppc-dev mailing list