]> git.ipfire.org Git - thirdparty/linux.git/commitdiff
KVM: s390: use __kvm_faultin_pfn()
authorClaudio Imbrenda <imbrenda@linux.ibm.com>
Thu, 23 Jan 2025 14:46:18 +0000 (15:46 +0100)
committerClaudio Imbrenda <imbrenda@linux.ibm.com>
Fri, 31 Jan 2025 11:03:52 +0000 (12:03 +0100)
Refactor the existing page fault handling code to use __kvm_faultin_pfn().

This possible now that memslots are always present.

Acked-by: Janosch Frank <frankja@linux.ibm.com>
Reviewed-by: Christoph Schlameuss <schlameuss@linux.ibm.com>
Link: https://lore.kernel.org/r/20250123144627.312456-7-imbrenda@linux.ibm.com
Signed-off-by: Claudio Imbrenda <imbrenda@linux.ibm.com>
Message-ID: <20250123144627.312456-7-imbrenda@linux.ibm.com>

arch/s390/kvm/kvm-s390.c
arch/s390/kvm/kvm-s390.h
arch/s390/mm/gmap.c

index a25ca440760f1ff27e28180331455cbf473b62a6..70c98bf12754114efa0cb533f0cf53b4fbba68e0 100644 (file)
@@ -4786,11 +4786,104 @@ static void kvm_s390_assert_primary_as(struct kvm_vcpu *vcpu)
                current->thread.gmap_int_code, current->thread.gmap_teid.val);
 }
 
+/*
+ * __kvm_s390_handle_dat_fault() - handle a dat fault for the gmap of a vcpu
+ * @vcpu: the vCPU whose gmap is to be fixed up
+ * @gfn: the guest frame number used for memslots (including fake memslots)
+ * @gaddr: the gmap address, does not have to match @gfn for ucontrol gmaps
+ * @flags: FOLL_* flags
+ *
+ * Return: 0 on success, < 0 in case of error.
+ * Context: The mm lock must not be held before calling. May sleep.
+ */
+int __kvm_s390_handle_dat_fault(struct kvm_vcpu *vcpu, gfn_t gfn, gpa_t gaddr, unsigned int flags)
+{
+       struct kvm_memory_slot *slot;
+       unsigned int fault_flags;
+       bool writable, unlocked;
+       unsigned long vmaddr;
+       struct page *page;
+       kvm_pfn_t pfn;
+       int rc;
+
+       slot = kvm_vcpu_gfn_to_memslot(vcpu, gfn);
+       if (!slot || slot->flags & KVM_MEMSLOT_INVALID)
+               return vcpu_post_run_addressing_exception(vcpu);
+
+       fault_flags = flags & FOLL_WRITE ? FAULT_FLAG_WRITE : 0;
+       if (vcpu->arch.gmap->pfault_enabled)
+               flags |= FOLL_NOWAIT;
+       vmaddr = __gfn_to_hva_memslot(slot, gfn);
+
+try_again:
+       pfn = __kvm_faultin_pfn(slot, gfn, flags, &writable, &page);
+
+       /* Access outside memory, inject addressing exception */
+       if (is_noslot_pfn(pfn))
+               return vcpu_post_run_addressing_exception(vcpu);
+       /* Signal pending: try again */
+       if (pfn == KVM_PFN_ERR_SIGPENDING)
+               return -EAGAIN;
+
+       /* Needs I/O, try to setup async pfault (only possible with FOLL_NOWAIT) */
+       if (pfn == KVM_PFN_ERR_NEEDS_IO) {
+               trace_kvm_s390_major_guest_pfault(vcpu);
+               if (kvm_arch_setup_async_pf(vcpu))
+                       return 0;
+               vcpu->stat.pfault_sync++;
+               /* Could not setup async pfault, try again synchronously */
+               flags &= ~FOLL_NOWAIT;
+               goto try_again;
+       }
+       /* Any other error */
+       if (is_error_pfn(pfn))
+               return -EFAULT;
+
+       /* Success */
+       mmap_read_lock(vcpu->arch.gmap->mm);
+       /* Mark the userspace PTEs as young and/or dirty, to avoid page fault loops */
+       rc = fixup_user_fault(vcpu->arch.gmap->mm, vmaddr, fault_flags, &unlocked);
+       if (!rc)
+               rc = __gmap_link(vcpu->arch.gmap, gaddr, vmaddr);
+       scoped_guard(spinlock, &vcpu->kvm->mmu_lock) {
+               kvm_release_faultin_page(vcpu->kvm, page, false, writable);
+       }
+       mmap_read_unlock(vcpu->arch.gmap->mm);
+       return rc;
+}
+
+static int vcpu_dat_fault_handler(struct kvm_vcpu *vcpu, unsigned long gaddr, unsigned int flags)
+{
+       unsigned long gaddr_tmp;
+       gfn_t gfn;
+
+       gfn = gpa_to_gfn(gaddr);
+       if (kvm_is_ucontrol(vcpu->kvm)) {
+               /*
+                * This translates the per-vCPU guest address into a
+                * fake guest address, which can then be used with the
+                * fake memslots that are identity mapping userspace.
+                * This allows ucontrol VMs to use the normal fault
+                * resolution path, like normal VMs.
+                */
+               mmap_read_lock(vcpu->arch.gmap->mm);
+               gaddr_tmp = __gmap_translate(vcpu->arch.gmap, gaddr);
+               mmap_read_unlock(vcpu->arch.gmap->mm);
+               if (gaddr_tmp == -EFAULT) {
+                       vcpu->run->exit_reason = KVM_EXIT_S390_UCONTROL;
+                       vcpu->run->s390_ucontrol.trans_exc_code = gaddr;
+                       vcpu->run->s390_ucontrol.pgm_code = PGM_SEGMENT_TRANSLATION;
+                       return -EREMOTE;
+               }
+               gfn = gpa_to_gfn(gaddr_tmp);
+       }
+       return __kvm_s390_handle_dat_fault(vcpu, gfn, gaddr, flags);
+}
+
 static int vcpu_post_run_handle_fault(struct kvm_vcpu *vcpu)
 {
        unsigned int flags = 0;
        unsigned long gaddr;
-       int rc = 0;
 
        gaddr = current->thread.gmap_teid.addr * PAGE_SIZE;
        if (kvm_s390_cur_gmap_fault_is_write())
@@ -4842,37 +4935,14 @@ static int vcpu_post_run_handle_fault(struct kvm_vcpu *vcpu)
        case PGM_REGION_SECOND_TRANS:
        case PGM_REGION_THIRD_TRANS:
                kvm_s390_assert_primary_as(vcpu);
-               if (vcpu->arch.gmap->pfault_enabled) {
-                       rc = gmap_fault(vcpu->arch.gmap, gaddr, flags | FAULT_FLAG_RETRY_NOWAIT);
-                       if (rc == -EFAULT)
-                               return vcpu_post_run_addressing_exception(vcpu);
-                       if (rc == -EAGAIN) {
-                               trace_kvm_s390_major_guest_pfault(vcpu);
-                               if (kvm_arch_setup_async_pf(vcpu))
-                                       return 0;
-                               vcpu->stat.pfault_sync++;
-                       } else {
-                               return rc;
-                       }
-               }
-               rc = gmap_fault(vcpu->arch.gmap, gaddr, flags);
-               if (rc == -EFAULT) {
-                       if (kvm_is_ucontrol(vcpu->kvm)) {
-                               vcpu->run->exit_reason = KVM_EXIT_S390_UCONTROL;
-                               vcpu->run->s390_ucontrol.trans_exc_code = gaddr;
-                               vcpu->run->s390_ucontrol.pgm_code = 0x10;
-                               return -EREMOTE;
-                       }
-                       return vcpu_post_run_addressing_exception(vcpu);
-               }
-               break;
+               return vcpu_dat_fault_handler(vcpu, gaddr, flags);
        default:
                KVM_BUG(1, vcpu->kvm, "Unexpected program interrupt 0x%x, TEID 0x%016lx",
                        current->thread.gmap_int_code, current->thread.gmap_teid.val);
                send_sig(SIGSEGV, current, 0);
                break;
        }
-       return rc;
+       return 0;
 }
 
 static int vcpu_post_run(struct kvm_vcpu *vcpu, int exit_reason)
@@ -5751,7 +5821,9 @@ long kvm_arch_vcpu_ioctl(struct file *filp,
        }
 #endif
        case KVM_S390_VCPU_FAULT: {
-               r = gmap_fault(vcpu->arch.gmap, arg, 0);
+               idx = srcu_read_lock(&vcpu->kvm->srcu);
+               r = vcpu_dat_fault_handler(vcpu, arg, 0);
+               srcu_read_unlock(&vcpu->kvm->srcu, idx);
                break;
        }
        case KVM_ENABLE_CAP:
index 30736ac16f848f2b3c2c98b45c97fdab59d3ac7a..3be5291723c8e14a0a73fc0ea1500ae51a7659b0 100644 (file)
@@ -410,6 +410,12 @@ void kvm_s390_vcpu_unsetup_cmma(struct kvm_vcpu *vcpu);
 void kvm_s390_set_cpu_timer(struct kvm_vcpu *vcpu, __u64 cputm);
 __u64 kvm_s390_get_cpu_timer(struct kvm_vcpu *vcpu);
 int kvm_s390_cpus_from_pv(struct kvm *kvm, u16 *rc, u16 *rrc);
+int __kvm_s390_handle_dat_fault(struct kvm_vcpu *vcpu, gfn_t gfn, gpa_t gaddr, unsigned int flags);
+
+static inline int kvm_s390_handle_dat_fault(struct kvm_vcpu *vcpu, gpa_t gaddr, unsigned int flags)
+{
+       return __kvm_s390_handle_dat_fault(vcpu, gpa_to_gfn(gaddr), gaddr, flags);
+}
 
 /* implemented in diag.c */
 int kvm_s390_handle_diag(struct kvm_vcpu *vcpu);
index 3e6e25119a964dcf50f20fbb3076eb157a60f989..bfaba773330675f319b81af2ccdc2ff05eda9d70 100644 (file)
@@ -605,6 +605,7 @@ int __gmap_link(struct gmap *gmap, unsigned long gaddr, unsigned long vmaddr)
        radix_tree_preload_end();
        return rc;
 }
+EXPORT_SYMBOL(__gmap_link);
 
 /**
  * fixup_user_fault_nowait - manually resolve a user page fault without waiting