x86/KVM/VMX: Add find_msr() helper function
authorKonrad Rzeszutek Wilk <konrad.wilk@oracle.com>
Thu, 21 Jun 2018 00:11:39 +0000 (20:11 -0400)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Wed, 15 Aug 2018 16:12:54 +0000 (18:12 +0200)
commit ca83b4a7f2d068da79a029d323024aa45decb250 upstream

.. to help find the MSR on either the guest or host MSR list.

Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
arch/x86/kvm/vmx.c

index 36779522acd5411bfd66295d888aabe06d974b52..44b0e0026c418088b568ea15e801719aca75e913 100644 (file)
@@ -2022,9 +2022,20 @@ static void clear_atomic_switch_msr_special(struct vcpu_vmx *vmx,
        vm_exit_controls_clearbit(vmx, exit);
 }
 
+static int find_msr(struct vmx_msrs *m, unsigned int msr)
+{
+       unsigned int i;
+
+       for (i = 0; i < m->nr; ++i) {
+               if (m->val[i].index == msr)
+                       return i;
+       }
+       return -ENOENT;
+}
+
 static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr)
 {
-       unsigned i;
+       int i;
        struct msr_autoload *m = &vmx->msr_autoload;
 
        switch (msr) {
@@ -2045,11 +2056,8 @@ static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr)
                }
                break;
        }
-       for (i = 0; i < m->guest.nr; ++i)
-               if (m->guest.val[i].index == msr)
-                       break;
-
-       if (i == m->guest.nr)
+       i = find_msr(&m->guest, msr);
+       if (i < 0)
                return;
        --m->guest.nr;
        --m->host.nr;
@@ -2073,7 +2081,7 @@ static void add_atomic_switch_msr_special(struct vcpu_vmx *vmx,
 static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr,
                                  u64 guest_val, u64 host_val)
 {
-       unsigned i;
+       int i;
        struct msr_autoload *m = &vmx->msr_autoload;
 
        switch (msr) {
@@ -2108,16 +2116,13 @@ static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr,
                wrmsrl(MSR_IA32_PEBS_ENABLE, 0);
        }
 
-       for (i = 0; i < m->guest.nr; ++i)
-               if (m->guest.val[i].index == msr)
-                       break;
-
+       i = find_msr(&m->guest, msr);
        if (i == NR_AUTOLOAD_MSRS) {
                printk_once(KERN_WARNING "Not enough msr switch entries. "
                                "Can't add msr %x\n", msr);
                return;
-       } else if (i == m->guest.nr) {
-               ++m->guest.nr;
+       } else if (i < 0) {
+               i = m->guest.nr++;
                ++m->host.nr;
                vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, m->guest.nr);
                vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, m->host.nr);