KVM: SVM: Optimize nested svm msrpm merging

This patch optimizes the way the msrpm of the host and the
guest are merged. The old code merged the 2 msrpm pages
completly. This code needed to touch 24kb of memory for that
operation. The optimized variant this patch introduces
merges only the parts where the host msrpm may contain zero
bits. This reduces the amount of memory which is touched to
48 bytes.

Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>
Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
This commit is contained in:
Joerg Roedel 2010-03-01 15:34:37 +01:00 committed by Avi Kivity
parent ac72a9b733
commit 323c3d809b

View file

@ -93,6 +93,9 @@ struct nested_state {
}; };
#define MSRPM_OFFSETS 16
static u32 msrpm_offsets[MSRPM_OFFSETS] __read_mostly;
struct vcpu_svm { struct vcpu_svm {
struct kvm_vcpu vcpu; struct kvm_vcpu vcpu;
struct vmcb *vmcb; struct vmcb *vmcb;
@ -510,6 +513,49 @@ static void svm_vcpu_init_msrpm(u32 *msrpm)
} }
} }
static void add_msr_offset(u32 offset)
{
int i;
for (i = 0; i < MSRPM_OFFSETS; ++i) {
/* Offset already in list? */
if (msrpm_offsets[i] == offset)
return;
/* Slot used by another offset? */
if (msrpm_offsets[i] != MSR_INVALID)
continue;
/* Add offset to list */
msrpm_offsets[i] = offset;
return;
}
/*
* If this BUG triggers the msrpm_offsets table has an overflow. Just
* increase MSRPM_OFFSETS in this case.
*/
BUG();
}
static void init_msrpm_offsets(void)
{
int i;
memset(msrpm_offsets, 0xff, sizeof(msrpm_offsets));
for (i = 0; direct_access_msrs[i].index != MSR_INVALID; i++) {
u32 offset;
offset = svm_msrpm_offset(direct_access_msrs[i].index);
BUG_ON(offset == MSR_INVALID);
add_msr_offset(offset);
}
}
static void svm_enable_lbrv(struct vcpu_svm *svm) static void svm_enable_lbrv(struct vcpu_svm *svm)
{ {
u32 *msrpm = svm->msrpm; u32 *msrpm = svm->msrpm;
@ -548,6 +594,8 @@ static __init int svm_hardware_setup(void)
memset(iopm_va, 0xff, PAGE_SIZE * (1 << IOPM_ALLOC_ORDER)); memset(iopm_va, 0xff, PAGE_SIZE * (1 << IOPM_ALLOC_ORDER));
iopm_base = page_to_pfn(iopm_pages) << PAGE_SHIFT; iopm_base = page_to_pfn(iopm_pages) << PAGE_SHIFT;
init_msrpm_offsets();
if (boot_cpu_has(X86_FEATURE_NX)) if (boot_cpu_has(X86_FEATURE_NX))
kvm_enable_efer_bits(EFER_NX); kvm_enable_efer_bits(EFER_NX);
@ -811,6 +859,7 @@ static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id)
svm_vcpu_init_msrpm(svm->msrpm); svm_vcpu_init_msrpm(svm->msrpm);
svm->nested.msrpm = page_address(nested_msrpm_pages); svm->nested.msrpm = page_address(nested_msrpm_pages);
svm_vcpu_init_msrpm(svm->nested.msrpm);
svm->vmcb = page_address(page); svm->vmcb = page_address(page);
clear_page(svm->vmcb); clear_page(svm->vmcb);
@ -1888,21 +1937,34 @@ static int nested_svm_vmexit(struct vcpu_svm *svm)
static bool nested_svm_vmrun_msrpm(struct vcpu_svm *svm) static bool nested_svm_vmrun_msrpm(struct vcpu_svm *svm)
{ {
u32 *nested_msrpm; /*
struct page *page; * This function merges the msr permission bitmaps of kvm and the
* nested vmcb. It is omptimized in that it only merges the parts where
* the kvm msr permission bitmap may contain zero bits
*/
int i; int i;
nested_msrpm = nested_svm_map(svm, svm->nested.vmcb_msrpm, &page); if (!(svm->nested.intercept & (1ULL << INTERCEPT_MSR_PROT)))
if (!nested_msrpm) return true;
return false;
for (i = 0; i < PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER) / 4; i++) for (i = 0; i < MSRPM_OFFSETS; i++) {
svm->nested.msrpm[i] = svm->msrpm[i] | nested_msrpm[i]; u32 value, p;
u64 offset;
if (msrpm_offsets[i] == 0xffffffff)
break;
offset = svm->nested.vmcb_msrpm + msrpm_offsets[i];
p = msrpm_offsets[i] / 4;
if (kvm_read_guest(svm->vcpu.kvm, offset, &value, 4))
return false;
svm->nested.msrpm[p] = svm->msrpm[p] | value;
}
svm->vmcb->control.msrpm_base_pa = __pa(svm->nested.msrpm); svm->vmcb->control.msrpm_base_pa = __pa(svm->nested.msrpm);
nested_svm_unmap(page);
return true; return true;
} }