KVM: x86: Generate set of VMX feature MSRs using first/last definitions
authorSean Christopherson <seanjc@google.com>
Sat, 11 Mar 2023 00:46:01 +0000 (16:46 -0800)
committerSean Christopherson <seanjc@google.com>
Thu, 6 Apr 2023 21:57:23 +0000 (14:57 -0700)
Add VMX MSRs to the runtime list of feature MSRs by iterating over the
range of emulated MSRs instead of manually defining each MSR in the "all"
list.  Using the range definition reduces the cost of emulating a new VMX
MSR, e.g. prevents forgetting to add an MSR to the list.

Extracting the VMX MSRs from the "all" list, which is a compile-time
constant, also shrinks the list to the point where the compiler can
heavily optimize code that iterates over the list.

No functional change intended.

Reviewed-by: Xiaoyao Li <xiaoyao.li@intel.com>
Link: https://lore.kernel.org/r/20230311004618.920745-5-seanjc@google.com
Signed-off-by: Sean Christopherson <seanjc@google.com>
arch/x86/kvm/x86.c

index 087497aebdedf6e70095e7677419a18b25c0ed76..6b667c651951dc40dc78984ba034af01f644d6ef 100644 (file)
@@ -1543,36 +1543,19 @@ static u32 emulated_msrs[ARRAY_SIZE(emulated_msrs_all)];
 static unsigned num_emulated_msrs;
 
 /*
- * List of msr numbers which are used to expose MSR-based features that
- * can be used by a hypervisor to validate requested CPU features.
+ * List of MSRs that control the existence of MSR-based features, i.e. MSRs
+ * that are effectively CPUID leafs.  VMX MSRs are also included in the set of
+ * feature MSRs, but are handled separately to allow expedited lookups.
  */
-static const u32 msr_based_features_all[] = {
-       MSR_IA32_VMX_BASIC,
-       MSR_IA32_VMX_TRUE_PINBASED_CTLS,
-       MSR_IA32_VMX_PINBASED_CTLS,
-       MSR_IA32_VMX_TRUE_PROCBASED_CTLS,
-       MSR_IA32_VMX_PROCBASED_CTLS,
-       MSR_IA32_VMX_TRUE_EXIT_CTLS,
-       MSR_IA32_VMX_EXIT_CTLS,
-       MSR_IA32_VMX_TRUE_ENTRY_CTLS,
-       MSR_IA32_VMX_ENTRY_CTLS,
-       MSR_IA32_VMX_MISC,
-       MSR_IA32_VMX_CR0_FIXED0,
-       MSR_IA32_VMX_CR0_FIXED1,
-       MSR_IA32_VMX_CR4_FIXED0,
-       MSR_IA32_VMX_CR4_FIXED1,
-       MSR_IA32_VMX_VMCS_ENUM,
-       MSR_IA32_VMX_PROCBASED_CTLS2,
-       MSR_IA32_VMX_EPT_VPID_CAP,
-       MSR_IA32_VMX_VMFUNC,
-
+static const u32 msr_based_features_all_except_vmx[] = {
        MSR_AMD64_DE_CFG,
        MSR_IA32_UCODE_REV,
        MSR_IA32_ARCH_CAPABILITIES,
        MSR_IA32_PERF_CAPABILITIES,
 };
 
-static u32 msr_based_features[ARRAY_SIZE(msr_based_features_all)];
+static u32 msr_based_features[ARRAY_SIZE(msr_based_features_all_except_vmx) +
+                             (KVM_LAST_EMULATED_VMX_MSR - KVM_FIRST_EMULATED_VMX_MSR + 1)];
 static unsigned int num_msr_based_features;
 
 /*
@@ -7009,6 +6992,18 @@ out:
        return r;
 }
 
+static void kvm_probe_feature_msr(u32 msr_index)
+{
+       struct kvm_msr_entry msr = {
+               .index = msr_index,
+       };
+
+       if (kvm_get_msr_feature(&msr))
+               return;
+
+       msr_based_features[num_msr_based_features++] = msr_index;
+}
+
 static void kvm_probe_msr_to_save(u32 msr_index)
 {
        u32 dummy[2];
@@ -7110,15 +7105,11 @@ static void kvm_init_msr_lists(void)
                emulated_msrs[num_emulated_msrs++] = emulated_msrs_all[i];
        }
 
-       for (i = 0; i < ARRAY_SIZE(msr_based_features_all); i++) {
-               struct kvm_msr_entry msr;
+       for (i = KVM_FIRST_EMULATED_VMX_MSR; i <= KVM_LAST_EMULATED_VMX_MSR; i++)
+               kvm_probe_feature_msr(i);
 
-               msr.index = msr_based_features_all[i];
-               if (kvm_get_msr_feature(&msr))
-                       continue;
-
-               msr_based_features[num_msr_based_features++] = msr_based_features_all[i];
-       }
+       for (i = 0; i < ARRAY_SIZE(msr_based_features_all_except_vmx); i++)
+               kvm_probe_feature_msr(msr_based_features_all_except_vmx[i]);
 }
 
 static int vcpu_mmio_write(struct kvm_vcpu *vcpu, gpa_t addr, int len,