aboutsummaryrefslogtreecommitdiffstats
path: root/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71-e3000/0082-KVM-SVM-limit-kvm_handle_page_fault-to-PF-handling.patch
diff options
context:
space:
mode:
Diffstat (limited to 'meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71-e3000/0082-KVM-SVM-limit-kvm_handle_page_fault-to-PF-handling.patch')
-rw-r--r--meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71-e3000/0082-KVM-SVM-limit-kvm_handle_page_fault-to-PF-handling.patch111
1 files changed, 111 insertions, 0 deletions
diff --git a/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71-e3000/0082-KVM-SVM-limit-kvm_handle_page_fault-to-PF-handling.patch b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71-e3000/0082-KVM-SVM-limit-kvm_handle_page_fault-to-PF-handling.patch
new file mode 100644
index 00000000..f3775839
--- /dev/null
+++ b/meta-amd-bsp/recipes-kernel/linux/linux-yocto-4.14.71-e3000/0082-KVM-SVM-limit-kvm_handle_page_fault-to-PF-handling.patch
@@ -0,0 +1,111 @@
+From 656ae55f24476e152600b6109dbdb517f883dc77 Mon Sep 17 00:00:00 2001
+From: Sudheesh Mavila <sudheesh.mavila@amd.com>
+Date: Tue, 14 Aug 2018 22:48:00 +0530
+Subject: [PATCH 82/95] KVM: SVM: limit kvm_handle_page_fault to #PF handling
+
+ commit d0006530576f1c7a49b2010eac7afdcb5a3613ae
+
+ It has always annoyed me a bit how SVM_EXIT_NPF is handled by
+ pf_interception. This is also the only reason behind the
+ under-documented need_unprotect argument to kvm_handle_page_fault.
+ Let NPF go straight to kvm_mmu_page_fault, just like VMX
+ does in handle_ept_violation and handle_ept_misconfig.
+
+ Reviewed-by: Brijesh Singh <brijesh.singh@amd.com>
+ Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
+
+Signed-off-by: Sudheesh Mavila <sudheesh.mavila@amd.com>
+---
+ arch/x86/kvm/mmu.c | 5 ++---
+ arch/x86/kvm/mmu.h | 3 +--
+ arch/x86/kvm/svm.c | 15 +++++++++++++--
+ arch/x86/kvm/vmx.c | 3 +--
+ 4 files changed, 17 insertions(+), 9 deletions(-)
+
+diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
+index 1dfb808..1408d79 100644
+--- a/arch/x86/kvm/mmu.c
++++ b/arch/x86/kvm/mmu.c
+@@ -3853,8 +3853,7 @@ static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn,
+ }
+
+ int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code,
+- u64 fault_address, char *insn, int insn_len,
+- bool need_unprotect)
++ u64 fault_address, char *insn, int insn_len)
+ {
+ int r = 1;
+
+@@ -3863,7 +3862,7 @@ int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code,
+ default:
+ trace_kvm_page_fault(fault_address, error_code);
+
+- if (need_unprotect && kvm_event_needs_reinjection(vcpu))
++ if (kvm_event_needs_reinjection(vcpu))
+ kvm_mmu_unprotect_page_virt(vcpu, fault_address);
+ r = kvm_mmu_page_fault(vcpu, fault_address, error_code, insn,
+ insn_len);
+diff --git a/arch/x86/kvm/mmu.h b/arch/x86/kvm/mmu.h
+index efc8576..5b408c0 100644
+--- a/arch/x86/kvm/mmu.h
++++ b/arch/x86/kvm/mmu.h
+@@ -66,8 +66,7 @@ void kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, bool execonly,
+ bool accessed_dirty);
+ bool kvm_can_do_async_pf(struct kvm_vcpu *vcpu);
+ int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code,
+- u64 fault_address, char *insn, int insn_len,
+- bool need_unprotect);
++ u64 fault_address, char *insn, int insn_len);
+
+ static inline unsigned int kvm_mmu_available_pages(struct kvm *kvm)
+ {
+diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
+index 2dcfd9f..364531e 100755
+--- a/arch/x86/kvm/svm.c
++++ b/arch/x86/kvm/svm.c
+@@ -2488,7 +2488,18 @@ static int pf_interception(struct vcpu_svm *svm)
+
+ return kvm_handle_page_fault(&svm->vcpu, error_code, fault_address,
+ svm->vmcb->control.insn_bytes,
+- svm->vmcb->control.insn_len, !npt_enabled);
++ svm->vmcb->control.insn_len);
++}
++
++static int npf_interception(struct vcpu_svm *svm)
++{
++ u64 fault_address = svm->vmcb->control.exit_info_2;
++ u64 error_code = svm->vmcb->control.exit_info_1;
++
++ trace_kvm_page_fault(fault_address, error_code);
++ return kvm_mmu_page_fault(&svm->vcpu, fault_address, error_code,
++ svm->vmcb->control.insn_bytes,
++ svm->vmcb->control.insn_len);
+ }
+
+ static int db_interception(struct vcpu_svm *svm)
+@@ -4586,7 +4597,7 @@ static int (*const svm_exit_handlers[])(struct vcpu_svm *svm) = {
+ [SVM_EXIT_MONITOR] = monitor_interception,
+ [SVM_EXIT_MWAIT] = mwait_interception,
+ [SVM_EXIT_XSETBV] = xsetbv_interception,
+- [SVM_EXIT_NPF] = pf_interception,
++ [SVM_EXIT_NPF] = npf_interception,
+ [SVM_EXIT_RSM] = emulate_on_interception,
+ [SVM_EXIT_AVIC_INCOMPLETE_IPI] = avic_incomplete_ipi_interception,
+ [SVM_EXIT_AVIC_UNACCELERATED_ACCESS] = avic_unaccelerated_access_interception,
+diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
+index fd46d89..815ff09 100644
+--- a/arch/x86/kvm/vmx.c
++++ b/arch/x86/kvm/vmx.c
+@@ -6344,8 +6344,7 @@ static int handle_exception(struct kvm_vcpu *vcpu)
+ cr2 = vmcs_readl(EXIT_QUALIFICATION);
+ /* EPT won't cause page fault directly */
+ WARN_ON_ONCE(!vcpu->arch.apf.host_apf_reason && enable_ept);
+- return kvm_handle_page_fault(vcpu, error_code, cr2, NULL, 0,
+- true);
++ return kvm_handle_page_fault(vcpu, error_code, cr2, NULL, 0);
+ }
+
+ ex_no = intr_info & INTR_INFO_VECTOR_MASK;
+--
+2.7.4
+