Changes for KVM for arm/arm64 for 3.18
This includes a bunch of changes: - Support read-only memory slots on arm/arm64 - Various changes to fix Sparse warnings - Correctly detect write vs. read Stage-2 faults - Various VGIC cleanups and fixes - Dynamic VGIC data strcuture sizing - Fix SGI set_clear_pend offset bug - Fix VTTBR_BADDR Mask - Correctly report the FSC on Stage-2 faults -----BEGIN PGP SIGNATURE----- Version: GnuPG v1 iQEcBAABAgAGBQJUJWAdAAoJEEtpOizt6ddy9cMH+gIoUPnRJLe+PPcOOyxOx6pr +CnD/zAd0sLvxZLP/LBOzu99H3YrbO5kwI/172/8G1zUNI2hp6YxEEJaBCTHrz6l RwgLy7a3EMMY51nJo5w2dkFUo8cUX9MsHqMpl2Xb7Dvo2ZHp+nDqRjwRY6yi+t4V dWSJTRG6X+DIWyysij6jBtfKU6MpU+4NW3Zdk1fapf8QDkn+cBtV5X2QcmERCaIe A1j9hiGi43KA3XWeeePU3aVaxC2XUhTayP8VsfVxoNG2manaS6lqjmbif5ghs/0h rw7R3/Aj0MJny2zT016MkvKJKRukuVRD6e1lcYghqnSJhL2FossowZ9fHRADpqU= =QgU8 -----END PGP SIGNATURE----- Merge tag 'kvm-arm-for-3.18' of git://git.kernel.org/pub/scm/linux/kernel/git/kvmarm/kvmarm into kvm-next Changes for KVM for arm/arm64 for 3.18 This includes a bunch of changes: - Support read-only memory slots on arm/arm64 - Various changes to fix Sparse warnings - Correctly detect write vs. read Stage-2 faults - Various VGIC cleanups and fixes - Dynamic VGIC data strcuture sizing - Fix SGI set_clear_pend offset bug - Fix VTTBR_BADDR Mask - Correctly report the FSC on Stage-2 faults Conflicts: virt/kvm/eventfd.c [duplicate, different patch where the kvm-arm version broke x86. The kvm tree instead has the right one]
This commit is contained in:
commit
e77d99d4a4
20 changed files with 677 additions and 214 deletions
|
@ -71,3 +71,13 @@ Groups:
|
||||||
Errors:
|
Errors:
|
||||||
-ENODEV: Getting or setting this register is not yet supported
|
-ENODEV: Getting or setting this register is not yet supported
|
||||||
-EBUSY: One or more VCPUs are running
|
-EBUSY: One or more VCPUs are running
|
||||||
|
|
||||||
|
KVM_DEV_ARM_VGIC_GRP_NR_IRQS
|
||||||
|
Attributes:
|
||||||
|
A value describing the number of interrupts (SGI, PPI and SPI) for
|
||||||
|
this GIC instance, ranging from 64 to 1024, in increments of 32.
|
||||||
|
|
||||||
|
Errors:
|
||||||
|
-EINVAL: Value set is out of the expected range
|
||||||
|
-EBUSY: Value has already be set, or GIC has already been initialized
|
||||||
|
with default values.
|
||||||
|
|
|
@ -148,6 +148,11 @@ static inline bool kvm_vcpu_trap_is_iabt(struct kvm_vcpu *vcpu)
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline u8 kvm_vcpu_trap_get_fault(struct kvm_vcpu *vcpu)
|
static inline u8 kvm_vcpu_trap_get_fault(struct kvm_vcpu *vcpu)
|
||||||
|
{
|
||||||
|
return kvm_vcpu_get_hsr(vcpu) & HSR_FSC;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline u8 kvm_vcpu_trap_get_fault_type(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
return kvm_vcpu_get_hsr(vcpu) & HSR_FSC_TYPE;
|
return kvm_vcpu_get_hsr(vcpu) & HSR_FSC_TYPE;
|
||||||
}
|
}
|
||||||
|
|
|
@ -43,7 +43,7 @@
|
||||||
#include <kvm/arm_vgic.h>
|
#include <kvm/arm_vgic.h>
|
||||||
|
|
||||||
u32 *kvm_vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num, u32 mode);
|
u32 *kvm_vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num, u32 mode);
|
||||||
int kvm_target_cpu(void);
|
int __attribute_const__ kvm_target_cpu(void);
|
||||||
int kvm_reset_vcpu(struct kvm_vcpu *vcpu);
|
int kvm_reset_vcpu(struct kvm_vcpu *vcpu);
|
||||||
void kvm_reset_coprocs(struct kvm_vcpu *vcpu);
|
void kvm_reset_coprocs(struct kvm_vcpu *vcpu);
|
||||||
|
|
||||||
|
|
|
@ -78,17 +78,6 @@ static inline void kvm_set_pte(pte_t *pte, pte_t new_pte)
|
||||||
flush_pmd_entry(pte);
|
flush_pmd_entry(pte);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline bool kvm_is_write_fault(unsigned long hsr)
|
|
||||||
{
|
|
||||||
unsigned long hsr_ec = hsr >> HSR_EC_SHIFT;
|
|
||||||
if (hsr_ec == HSR_EC_IABT)
|
|
||||||
return false;
|
|
||||||
else if ((hsr & HSR_ISV) && !(hsr & HSR_WNR))
|
|
||||||
return false;
|
|
||||||
else
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline void kvm_clean_pgd(pgd_t *pgd)
|
static inline void kvm_clean_pgd(pgd_t *pgd)
|
||||||
{
|
{
|
||||||
clean_dcache_area(pgd, PTRS_PER_S2_PGD * sizeof(pgd_t));
|
clean_dcache_area(pgd, PTRS_PER_S2_PGD * sizeof(pgd_t));
|
||||||
|
|
|
@ -25,6 +25,7 @@
|
||||||
|
|
||||||
#define __KVM_HAVE_GUEST_DEBUG
|
#define __KVM_HAVE_GUEST_DEBUG
|
||||||
#define __KVM_HAVE_IRQ_LINE
|
#define __KVM_HAVE_IRQ_LINE
|
||||||
|
#define __KVM_HAVE_READONLY_MEM
|
||||||
|
|
||||||
#define KVM_REG_SIZE(id) \
|
#define KVM_REG_SIZE(id) \
|
||||||
(1U << (((id) & KVM_REG_SIZE_MASK) >> KVM_REG_SIZE_SHIFT))
|
(1U << (((id) & KVM_REG_SIZE_MASK) >> KVM_REG_SIZE_SHIFT))
|
||||||
|
@ -173,6 +174,7 @@ struct kvm_arch_memory_slot {
|
||||||
#define KVM_DEV_ARM_VGIC_CPUID_MASK (0xffULL << KVM_DEV_ARM_VGIC_CPUID_SHIFT)
|
#define KVM_DEV_ARM_VGIC_CPUID_MASK (0xffULL << KVM_DEV_ARM_VGIC_CPUID_SHIFT)
|
||||||
#define KVM_DEV_ARM_VGIC_OFFSET_SHIFT 0
|
#define KVM_DEV_ARM_VGIC_OFFSET_SHIFT 0
|
||||||
#define KVM_DEV_ARM_VGIC_OFFSET_MASK (0xffffffffULL << KVM_DEV_ARM_VGIC_OFFSET_SHIFT)
|
#define KVM_DEV_ARM_VGIC_OFFSET_MASK (0xffffffffULL << KVM_DEV_ARM_VGIC_OFFSET_SHIFT)
|
||||||
|
#define KVM_DEV_ARM_VGIC_GRP_NR_IRQS 3
|
||||||
|
|
||||||
/* KVM_IRQ_LINE irq field index values */
|
/* KVM_IRQ_LINE irq field index values */
|
||||||
#define KVM_ARM_IRQ_TYPE_SHIFT 24
|
#define KVM_ARM_IRQ_TYPE_SHIFT 24
|
||||||
|
|
|
@ -82,7 +82,7 @@ struct kvm_vcpu *kvm_arm_get_running_vcpu(void)
|
||||||
/**
|
/**
|
||||||
* kvm_arm_get_running_vcpus - get the per-CPU array of currently running vcpus.
|
* kvm_arm_get_running_vcpus - get the per-CPU array of currently running vcpus.
|
||||||
*/
|
*/
|
||||||
struct kvm_vcpu __percpu **kvm_get_running_vcpus(void)
|
struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void)
|
||||||
{
|
{
|
||||||
return &kvm_arm_running_vcpu;
|
return &kvm_arm_running_vcpu;
|
||||||
}
|
}
|
||||||
|
@ -161,6 +161,8 @@ void kvm_arch_destroy_vm(struct kvm *kvm)
|
||||||
kvm->vcpus[i] = NULL;
|
kvm->vcpus[i] = NULL;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
kvm_vgic_destroy(kvm);
|
||||||
}
|
}
|
||||||
|
|
||||||
int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
|
int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
|
||||||
|
@ -177,6 +179,7 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext)
|
||||||
case KVM_CAP_ONE_REG:
|
case KVM_CAP_ONE_REG:
|
||||||
case KVM_CAP_ARM_PSCI:
|
case KVM_CAP_ARM_PSCI:
|
||||||
case KVM_CAP_ARM_PSCI_0_2:
|
case KVM_CAP_ARM_PSCI_0_2:
|
||||||
|
case KVM_CAP_READONLY_MEM:
|
||||||
r = 1;
|
r = 1;
|
||||||
break;
|
break;
|
||||||
case KVM_CAP_COALESCED_MMIO:
|
case KVM_CAP_COALESCED_MMIO:
|
||||||
|
@ -242,6 +245,7 @@ void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
kvm_mmu_free_memory_caches(vcpu);
|
kvm_mmu_free_memory_caches(vcpu);
|
||||||
kvm_timer_vcpu_terminate(vcpu);
|
kvm_timer_vcpu_terminate(vcpu);
|
||||||
|
kvm_vgic_vcpu_destroy(vcpu);
|
||||||
kmem_cache_free(kvm_vcpu_cache, vcpu);
|
kmem_cache_free(kvm_vcpu_cache, vcpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -257,16 +261,9 @@ int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
|
||||||
|
|
||||||
int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu)
|
int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
int ret;
|
|
||||||
|
|
||||||
/* Force users to call KVM_ARM_VCPU_INIT */
|
/* Force users to call KVM_ARM_VCPU_INIT */
|
||||||
vcpu->arch.target = -1;
|
vcpu->arch.target = -1;
|
||||||
|
|
||||||
/* Set up VGIC */
|
|
||||||
ret = kvm_vgic_vcpu_init(vcpu);
|
|
||||||
if (ret)
|
|
||||||
return ret;
|
|
||||||
|
|
||||||
/* Set up the timer */
|
/* Set up the timer */
|
||||||
kvm_timer_vcpu_init(vcpu);
|
kvm_timer_vcpu_init(vcpu);
|
||||||
|
|
||||||
|
@ -413,9 +410,9 @@ static void update_vttbr(struct kvm *kvm)
|
||||||
|
|
||||||
/* update vttbr to be used with the new vmid */
|
/* update vttbr to be used with the new vmid */
|
||||||
pgd_phys = virt_to_phys(kvm->arch.pgd);
|
pgd_phys = virt_to_phys(kvm->arch.pgd);
|
||||||
|
BUG_ON(pgd_phys & ~VTTBR_BADDR_MASK);
|
||||||
vmid = ((u64)(kvm->arch.vmid) << VTTBR_VMID_SHIFT) & VTTBR_VMID_MASK;
|
vmid = ((u64)(kvm->arch.vmid) << VTTBR_VMID_SHIFT) & VTTBR_VMID_MASK;
|
||||||
kvm->arch.vttbr = pgd_phys & VTTBR_BADDR_MASK;
|
kvm->arch.vttbr = pgd_phys | vmid;
|
||||||
kvm->arch.vttbr |= vmid;
|
|
||||||
|
|
||||||
spin_unlock(&kvm_vmid_lock);
|
spin_unlock(&kvm_vmid_lock);
|
||||||
}
|
}
|
||||||
|
|
|
@ -791,7 +791,7 @@ static bool is_valid_cache(u32 val)
|
||||||
u32 level, ctype;
|
u32 level, ctype;
|
||||||
|
|
||||||
if (val >= CSSELR_MAX)
|
if (val >= CSSELR_MAX)
|
||||||
return -ENOENT;
|
return false;
|
||||||
|
|
||||||
/* Bottom bit is Instruction or Data bit. Next 3 bits are level. */
|
/* Bottom bit is Instruction or Data bit. Next 3 bits are level. */
|
||||||
level = (val >> 1);
|
level = (val >> 1);
|
||||||
|
|
|
@ -163,7 +163,7 @@ static int set_timer_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
|
||||||
|
|
||||||
ret = copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id));
|
ret = copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id));
|
||||||
if (ret != 0)
|
if (ret != 0)
|
||||||
return ret;
|
return -EFAULT;
|
||||||
|
|
||||||
return kvm_arm_timer_set_reg(vcpu, reg->id, val);
|
return kvm_arm_timer_set_reg(vcpu, reg->id, val);
|
||||||
}
|
}
|
||||||
|
|
|
@ -746,22 +746,29 @@ static bool transparent_hugepage_adjust(pfn_t *pfnp, phys_addr_t *ipap)
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static bool kvm_is_write_fault(struct kvm_vcpu *vcpu)
|
||||||
|
{
|
||||||
|
if (kvm_vcpu_trap_is_iabt(vcpu))
|
||||||
|
return false;
|
||||||
|
|
||||||
|
return kvm_vcpu_dabt_iswrite(vcpu);
|
||||||
|
}
|
||||||
|
|
||||||
static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
|
static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
|
||||||
struct kvm_memory_slot *memslot,
|
struct kvm_memory_slot *memslot, unsigned long hva,
|
||||||
unsigned long fault_status)
|
unsigned long fault_status)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
bool write_fault, writable, hugetlb = false, force_pte = false;
|
bool write_fault, writable, hugetlb = false, force_pte = false;
|
||||||
unsigned long mmu_seq;
|
unsigned long mmu_seq;
|
||||||
gfn_t gfn = fault_ipa >> PAGE_SHIFT;
|
gfn_t gfn = fault_ipa >> PAGE_SHIFT;
|
||||||
unsigned long hva = gfn_to_hva(vcpu->kvm, gfn);
|
|
||||||
struct kvm *kvm = vcpu->kvm;
|
struct kvm *kvm = vcpu->kvm;
|
||||||
struct kvm_mmu_memory_cache *memcache = &vcpu->arch.mmu_page_cache;
|
struct kvm_mmu_memory_cache *memcache = &vcpu->arch.mmu_page_cache;
|
||||||
struct vm_area_struct *vma;
|
struct vm_area_struct *vma;
|
||||||
pfn_t pfn;
|
pfn_t pfn;
|
||||||
pgprot_t mem_type = PAGE_S2;
|
pgprot_t mem_type = PAGE_S2;
|
||||||
|
|
||||||
write_fault = kvm_is_write_fault(kvm_vcpu_get_hsr(vcpu));
|
write_fault = kvm_is_write_fault(vcpu);
|
||||||
if (fault_status == FSC_PERM && !write_fault) {
|
if (fault_status == FSC_PERM && !write_fault) {
|
||||||
kvm_err("Unexpected L2 read permission error\n");
|
kvm_err("Unexpected L2 read permission error\n");
|
||||||
return -EFAULT;
|
return -EFAULT;
|
||||||
|
@ -863,7 +870,8 @@ int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run)
|
||||||
unsigned long fault_status;
|
unsigned long fault_status;
|
||||||
phys_addr_t fault_ipa;
|
phys_addr_t fault_ipa;
|
||||||
struct kvm_memory_slot *memslot;
|
struct kvm_memory_slot *memslot;
|
||||||
bool is_iabt;
|
unsigned long hva;
|
||||||
|
bool is_iabt, write_fault, writable;
|
||||||
gfn_t gfn;
|
gfn_t gfn;
|
||||||
int ret, idx;
|
int ret, idx;
|
||||||
|
|
||||||
|
@ -874,17 +882,22 @@ int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run)
|
||||||
kvm_vcpu_get_hfar(vcpu), fault_ipa);
|
kvm_vcpu_get_hfar(vcpu), fault_ipa);
|
||||||
|
|
||||||
/* Check the stage-2 fault is trans. fault or write fault */
|
/* Check the stage-2 fault is trans. fault or write fault */
|
||||||
fault_status = kvm_vcpu_trap_get_fault(vcpu);
|
fault_status = kvm_vcpu_trap_get_fault_type(vcpu);
|
||||||
if (fault_status != FSC_FAULT && fault_status != FSC_PERM) {
|
if (fault_status != FSC_FAULT && fault_status != FSC_PERM) {
|
||||||
kvm_err("Unsupported fault status: EC=%#x DFCS=%#lx\n",
|
kvm_err("Unsupported FSC: EC=%#x xFSC=%#lx ESR_EL2=%#lx\n",
|
||||||
kvm_vcpu_trap_get_class(vcpu), fault_status);
|
kvm_vcpu_trap_get_class(vcpu),
|
||||||
|
(unsigned long)kvm_vcpu_trap_get_fault(vcpu),
|
||||||
|
(unsigned long)kvm_vcpu_get_hsr(vcpu));
|
||||||
return -EFAULT;
|
return -EFAULT;
|
||||||
}
|
}
|
||||||
|
|
||||||
idx = srcu_read_lock(&vcpu->kvm->srcu);
|
idx = srcu_read_lock(&vcpu->kvm->srcu);
|
||||||
|
|
||||||
gfn = fault_ipa >> PAGE_SHIFT;
|
gfn = fault_ipa >> PAGE_SHIFT;
|
||||||
if (!kvm_is_visible_gfn(vcpu->kvm, gfn)) {
|
memslot = gfn_to_memslot(vcpu->kvm, gfn);
|
||||||
|
hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable);
|
||||||
|
write_fault = kvm_is_write_fault(vcpu);
|
||||||
|
if (kvm_is_error_hva(hva) || (write_fault && !writable)) {
|
||||||
if (is_iabt) {
|
if (is_iabt) {
|
||||||
/* Prefetch Abort on I/O address */
|
/* Prefetch Abort on I/O address */
|
||||||
kvm_inject_pabt(vcpu, kvm_vcpu_get_hfar(vcpu));
|
kvm_inject_pabt(vcpu, kvm_vcpu_get_hfar(vcpu));
|
||||||
|
@ -892,13 +905,6 @@ int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run)
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (fault_status != FSC_FAULT) {
|
|
||||||
kvm_err("Unsupported fault status on io memory: %#lx\n",
|
|
||||||
fault_status);
|
|
||||||
ret = -EFAULT;
|
|
||||||
goto out_unlock;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The IPA is reported as [MAX:12], so we need to
|
* The IPA is reported as [MAX:12], so we need to
|
||||||
* complement it with the bottom 12 bits from the
|
* complement it with the bottom 12 bits from the
|
||||||
|
@ -910,9 +916,7 @@ int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run)
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
}
|
}
|
||||||
|
|
||||||
memslot = gfn_to_memslot(vcpu->kvm, gfn);
|
ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status);
|
||||||
|
|
||||||
ret = user_mem_abort(vcpu, fault_ipa, memslot, fault_status);
|
|
||||||
if (ret == 0)
|
if (ret == 0)
|
||||||
ret = 1;
|
ret = 1;
|
||||||
out_unlock:
|
out_unlock:
|
||||||
|
|
|
@ -122,6 +122,17 @@
|
||||||
#define VTCR_EL2_T0SZ_MASK 0x3f
|
#define VTCR_EL2_T0SZ_MASK 0x3f
|
||||||
#define VTCR_EL2_T0SZ_40B 24
|
#define VTCR_EL2_T0SZ_40B 24
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We configure the Stage-2 page tables to always restrict the IPA space to be
|
||||||
|
* 40 bits wide (T0SZ = 24). Systems with a PARange smaller than 40 bits are
|
||||||
|
* not known to exist and will break with this configuration.
|
||||||
|
*
|
||||||
|
* Note that when using 4K pages, we concatenate two first level page tables
|
||||||
|
* together.
|
||||||
|
*
|
||||||
|
* The magic numbers used for VTTBR_X in this patch can be found in Tables
|
||||||
|
* D4-23 and D4-25 in ARM DDI 0487A.b.
|
||||||
|
*/
|
||||||
#ifdef CONFIG_ARM64_64K_PAGES
|
#ifdef CONFIG_ARM64_64K_PAGES
|
||||||
/*
|
/*
|
||||||
* Stage2 translation configuration:
|
* Stage2 translation configuration:
|
||||||
|
@ -149,7 +160,7 @@
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#define VTTBR_BADDR_SHIFT (VTTBR_X - 1)
|
#define VTTBR_BADDR_SHIFT (VTTBR_X - 1)
|
||||||
#define VTTBR_BADDR_MASK (((1LLU << (40 - VTTBR_X)) - 1) << VTTBR_BADDR_SHIFT)
|
#define VTTBR_BADDR_MASK (((1LLU << (PHYS_MASK_SHIFT - VTTBR_X)) - 1) << VTTBR_BADDR_SHIFT)
|
||||||
#define VTTBR_VMID_SHIFT (48LLU)
|
#define VTTBR_VMID_SHIFT (48LLU)
|
||||||
#define VTTBR_VMID_MASK (0xffLLU << VTTBR_VMID_SHIFT)
|
#define VTTBR_VMID_MASK (0xffLLU << VTTBR_VMID_SHIFT)
|
||||||
|
|
||||||
|
|
|
@ -173,6 +173,11 @@ static inline bool kvm_vcpu_trap_is_iabt(const struct kvm_vcpu *vcpu)
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline u8 kvm_vcpu_trap_get_fault(const struct kvm_vcpu *vcpu)
|
static inline u8 kvm_vcpu_trap_get_fault(const struct kvm_vcpu *vcpu)
|
||||||
|
{
|
||||||
|
return kvm_vcpu_get_hsr(vcpu) & ESR_EL2_FSC;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline u8 kvm_vcpu_trap_get_fault_type(const struct kvm_vcpu *vcpu)
|
||||||
{
|
{
|
||||||
return kvm_vcpu_get_hsr(vcpu) & ESR_EL2_FSC_TYPE;
|
return kvm_vcpu_get_hsr(vcpu) & ESR_EL2_FSC_TYPE;
|
||||||
}
|
}
|
||||||
|
|
|
@ -43,7 +43,7 @@
|
||||||
|
|
||||||
#define KVM_VCPU_MAX_FEATURES 3
|
#define KVM_VCPU_MAX_FEATURES 3
|
||||||
|
|
||||||
int kvm_target_cpu(void);
|
int __attribute_const__ kvm_target_cpu(void);
|
||||||
int kvm_reset_vcpu(struct kvm_vcpu *vcpu);
|
int kvm_reset_vcpu(struct kvm_vcpu *vcpu);
|
||||||
int kvm_arch_dev_ioctl_check_extension(long ext);
|
int kvm_arch_dev_ioctl_check_extension(long ext);
|
||||||
|
|
||||||
|
@ -197,7 +197,7 @@ static inline void kvm_arch_mmu_notifier_invalidate_page(struct kvm *kvm,
|
||||||
}
|
}
|
||||||
|
|
||||||
struct kvm_vcpu *kvm_arm_get_running_vcpu(void);
|
struct kvm_vcpu *kvm_arm_get_running_vcpu(void);
|
||||||
struct kvm_vcpu __percpu **kvm_get_running_vcpus(void);
|
struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void);
|
||||||
|
|
||||||
u64 kvm_call_hyp(void *hypfn, ...);
|
u64 kvm_call_hyp(void *hypfn, ...);
|
||||||
|
|
||||||
|
|
|
@ -59,10 +59,9 @@
|
||||||
#define KERN_TO_HYP(kva) ((unsigned long)kva - PAGE_OFFSET + HYP_PAGE_OFFSET)
|
#define KERN_TO_HYP(kva) ((unsigned long)kva - PAGE_OFFSET + HYP_PAGE_OFFSET)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Align KVM with the kernel's view of physical memory. Should be
|
* We currently only support a 40bit IPA.
|
||||||
* 40bit IPA, with PGD being 8kB aligned in the 4KB page configuration.
|
|
||||||
*/
|
*/
|
||||||
#define KVM_PHYS_SHIFT PHYS_MASK_SHIFT
|
#define KVM_PHYS_SHIFT (40)
|
||||||
#define KVM_PHYS_SIZE (1UL << KVM_PHYS_SHIFT)
|
#define KVM_PHYS_SIZE (1UL << KVM_PHYS_SHIFT)
|
||||||
#define KVM_PHYS_MASK (KVM_PHYS_SIZE - 1UL)
|
#define KVM_PHYS_MASK (KVM_PHYS_SIZE - 1UL)
|
||||||
|
|
||||||
|
@ -93,19 +92,6 @@ void kvm_clear_hyp_idmap(void);
|
||||||
#define kvm_set_pte(ptep, pte) set_pte(ptep, pte)
|
#define kvm_set_pte(ptep, pte) set_pte(ptep, pte)
|
||||||
#define kvm_set_pmd(pmdp, pmd) set_pmd(pmdp, pmd)
|
#define kvm_set_pmd(pmdp, pmd) set_pmd(pmdp, pmd)
|
||||||
|
|
||||||
static inline bool kvm_is_write_fault(unsigned long esr)
|
|
||||||
{
|
|
||||||
unsigned long esr_ec = esr >> ESR_EL2_EC_SHIFT;
|
|
||||||
|
|
||||||
if (esr_ec == ESR_EL2_EC_IABT)
|
|
||||||
return false;
|
|
||||||
|
|
||||||
if ((esr & ESR_EL2_ISV) && !(esr & ESR_EL2_WNR))
|
|
||||||
return false;
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline void kvm_clean_pgd(pgd_t *pgd) {}
|
static inline void kvm_clean_pgd(pgd_t *pgd) {}
|
||||||
static inline void kvm_clean_pmd_entry(pmd_t *pmd) {}
|
static inline void kvm_clean_pmd_entry(pmd_t *pmd) {}
|
||||||
static inline void kvm_clean_pte(pte_t *pte) {}
|
static inline void kvm_clean_pte(pte_t *pte) {}
|
||||||
|
|
|
@ -37,6 +37,7 @@
|
||||||
|
|
||||||
#define __KVM_HAVE_GUEST_DEBUG
|
#define __KVM_HAVE_GUEST_DEBUG
|
||||||
#define __KVM_HAVE_IRQ_LINE
|
#define __KVM_HAVE_IRQ_LINE
|
||||||
|
#define __KVM_HAVE_READONLY_MEM
|
||||||
|
|
||||||
#define KVM_REG_SIZE(id) \
|
#define KVM_REG_SIZE(id) \
|
||||||
(1U << (((id) & KVM_REG_SIZE_MASK) >> KVM_REG_SIZE_SHIFT))
|
(1U << (((id) & KVM_REG_SIZE_MASK) >> KVM_REG_SIZE_SHIFT))
|
||||||
|
@ -159,6 +160,7 @@ struct kvm_arch_memory_slot {
|
||||||
#define KVM_DEV_ARM_VGIC_CPUID_MASK (0xffULL << KVM_DEV_ARM_VGIC_CPUID_SHIFT)
|
#define KVM_DEV_ARM_VGIC_CPUID_MASK (0xffULL << KVM_DEV_ARM_VGIC_CPUID_SHIFT)
|
||||||
#define KVM_DEV_ARM_VGIC_OFFSET_SHIFT 0
|
#define KVM_DEV_ARM_VGIC_OFFSET_SHIFT 0
|
||||||
#define KVM_DEV_ARM_VGIC_OFFSET_MASK (0xffffffffULL << KVM_DEV_ARM_VGIC_OFFSET_SHIFT)
|
#define KVM_DEV_ARM_VGIC_OFFSET_MASK (0xffffffffULL << KVM_DEV_ARM_VGIC_OFFSET_SHIFT)
|
||||||
|
#define KVM_DEV_ARM_VGIC_GRP_NR_IRQS 3
|
||||||
|
|
||||||
/* KVM_IRQ_LINE irq field index values */
|
/* KVM_IRQ_LINE irq field index values */
|
||||||
#define KVM_ARM_IRQ_TYPE_SHIFT 24
|
#define KVM_ARM_IRQ_TYPE_SHIFT 24
|
||||||
|
|
|
@ -174,7 +174,7 @@ static int set_timer_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
|
||||||
|
|
||||||
ret = copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id));
|
ret = copy_from_user(&val, uaddr, KVM_REG_SIZE(reg->id));
|
||||||
if (ret != 0)
|
if (ret != 0)
|
||||||
return ret;
|
return -EFAULT;
|
||||||
|
|
||||||
return kvm_arm_timer_set_reg(vcpu, reg->id, val);
|
return kvm_arm_timer_set_reg(vcpu, reg->id, val);
|
||||||
}
|
}
|
||||||
|
|
|
@ -1218,7 +1218,7 @@ static bool is_valid_cache(u32 val)
|
||||||
u32 level, ctype;
|
u32 level, ctype;
|
||||||
|
|
||||||
if (val >= CSSELR_MAX)
|
if (val >= CSSELR_MAX)
|
||||||
return -ENOENT;
|
return false;
|
||||||
|
|
||||||
/* Bottom bit is Instruction or Data bit. Next 3 bits are level. */
|
/* Bottom bit is Instruction or Data bit. Next 3 bits are level. */
|
||||||
level = (val >> 1);
|
level = (val >> 1);
|
||||||
|
|
|
@ -25,26 +25,25 @@
|
||||||
#include <linux/spinlock.h>
|
#include <linux/spinlock.h>
|
||||||
#include <linux/types.h>
|
#include <linux/types.h>
|
||||||
|
|
||||||
#define VGIC_NR_IRQS 256
|
#define VGIC_NR_IRQS_LEGACY 256
|
||||||
#define VGIC_NR_SGIS 16
|
#define VGIC_NR_SGIS 16
|
||||||
#define VGIC_NR_PPIS 16
|
#define VGIC_NR_PPIS 16
|
||||||
#define VGIC_NR_PRIVATE_IRQS (VGIC_NR_SGIS + VGIC_NR_PPIS)
|
#define VGIC_NR_PRIVATE_IRQS (VGIC_NR_SGIS + VGIC_NR_PPIS)
|
||||||
#define VGIC_NR_SHARED_IRQS (VGIC_NR_IRQS - VGIC_NR_PRIVATE_IRQS)
|
|
||||||
#define VGIC_MAX_CPUS KVM_MAX_VCPUS
|
|
||||||
|
|
||||||
#define VGIC_V2_MAX_LRS (1 << 6)
|
#define VGIC_V2_MAX_LRS (1 << 6)
|
||||||
#define VGIC_V3_MAX_LRS 16
|
#define VGIC_V3_MAX_LRS 16
|
||||||
|
#define VGIC_MAX_IRQS 1024
|
||||||
|
|
||||||
/* Sanity checks... */
|
/* Sanity checks... */
|
||||||
#if (VGIC_MAX_CPUS > 8)
|
#if (KVM_MAX_VCPUS > 8)
|
||||||
#error Invalid number of CPU interfaces
|
#error Invalid number of CPU interfaces
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (VGIC_NR_IRQS & 31)
|
#if (VGIC_NR_IRQS_LEGACY & 31)
|
||||||
#error "VGIC_NR_IRQS must be a multiple of 32"
|
#error "VGIC_NR_IRQS must be a multiple of 32"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (VGIC_NR_IRQS > 1024)
|
#if (VGIC_NR_IRQS_LEGACY > VGIC_MAX_IRQS)
|
||||||
#error "VGIC_NR_IRQS must be <= 1024"
|
#error "VGIC_NR_IRQS must be <= 1024"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -54,19 +53,33 @@
|
||||||
* - a bunch of shared interrupts (SPI)
|
* - a bunch of shared interrupts (SPI)
|
||||||
*/
|
*/
|
||||||
struct vgic_bitmap {
|
struct vgic_bitmap {
|
||||||
union {
|
/*
|
||||||
u32 reg[VGIC_NR_PRIVATE_IRQS / 32];
|
* - One UL per VCPU for private interrupts (assumes UL is at
|
||||||
DECLARE_BITMAP(reg_ul, VGIC_NR_PRIVATE_IRQS);
|
* least 32 bits)
|
||||||
} percpu[VGIC_MAX_CPUS];
|
* - As many UL as necessary for shared interrupts.
|
||||||
union {
|
*
|
||||||
u32 reg[VGIC_NR_SHARED_IRQS / 32];
|
* The private interrupts are accessed via the "private"
|
||||||
DECLARE_BITMAP(reg_ul, VGIC_NR_SHARED_IRQS);
|
* field, one UL per vcpu (the state for vcpu n is in
|
||||||
} shared;
|
* private[n]). The shared interrupts are accessed via the
|
||||||
|
* "shared" pointer (IRQn state is at bit n-32 in the bitmap).
|
||||||
|
*/
|
||||||
|
unsigned long *private;
|
||||||
|
unsigned long *shared;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct vgic_bytemap {
|
struct vgic_bytemap {
|
||||||
u32 percpu[VGIC_MAX_CPUS][VGIC_NR_PRIVATE_IRQS / 4];
|
/*
|
||||||
u32 shared[VGIC_NR_SHARED_IRQS / 4];
|
* - 8 u32 per VCPU for private interrupts
|
||||||
|
* - As many u32 as necessary for shared interrupts.
|
||||||
|
*
|
||||||
|
* The private interrupts are accessed via the "private"
|
||||||
|
* field, (the state for vcpu n is in private[n*8] to
|
||||||
|
* private[n*8 + 7]). The shared interrupts are accessed via
|
||||||
|
* the "shared" pointer (IRQn state is at byte (n-32)%4 of the
|
||||||
|
* shared[(n-32)/4] word).
|
||||||
|
*/
|
||||||
|
u32 *private;
|
||||||
|
u32 *shared;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct kvm_vcpu;
|
struct kvm_vcpu;
|
||||||
|
@ -127,6 +140,9 @@ struct vgic_dist {
|
||||||
bool in_kernel;
|
bool in_kernel;
|
||||||
bool ready;
|
bool ready;
|
||||||
|
|
||||||
|
int nr_cpus;
|
||||||
|
int nr_irqs;
|
||||||
|
|
||||||
/* Virtual control interface mapping */
|
/* Virtual control interface mapping */
|
||||||
void __iomem *vctrl_base;
|
void __iomem *vctrl_base;
|
||||||
|
|
||||||
|
@ -140,11 +156,25 @@ struct vgic_dist {
|
||||||
/* Interrupt enabled (one bit per IRQ) */
|
/* Interrupt enabled (one bit per IRQ) */
|
||||||
struct vgic_bitmap irq_enabled;
|
struct vgic_bitmap irq_enabled;
|
||||||
|
|
||||||
/* Interrupt 'pin' level */
|
/* Level-triggered interrupt external input is asserted */
|
||||||
struct vgic_bitmap irq_state;
|
struct vgic_bitmap irq_level;
|
||||||
|
|
||||||
/* Level-triggered interrupt in progress */
|
/*
|
||||||
struct vgic_bitmap irq_active;
|
* Interrupt state is pending on the distributor
|
||||||
|
*/
|
||||||
|
struct vgic_bitmap irq_pending;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Tracks writes to GICD_ISPENDRn and GICD_ICPENDRn for level-triggered
|
||||||
|
* interrupts. Essentially holds the state of the flip-flop in
|
||||||
|
* Figure 4-10 on page 4-101 in ARM IHI 0048B.b.
|
||||||
|
* Once set, it is only cleared for level-triggered interrupts on
|
||||||
|
* guest ACKs (when we queue it) or writes to GICD_ICPENDRn.
|
||||||
|
*/
|
||||||
|
struct vgic_bitmap irq_soft_pend;
|
||||||
|
|
||||||
|
/* Level-triggered interrupt queued on VCPU interface */
|
||||||
|
struct vgic_bitmap irq_queued;
|
||||||
|
|
||||||
/* Interrupt priority. Not used yet. */
|
/* Interrupt priority. Not used yet. */
|
||||||
struct vgic_bytemap irq_priority;
|
struct vgic_bytemap irq_priority;
|
||||||
|
@ -152,15 +182,36 @@ struct vgic_dist {
|
||||||
/* Level/edge triggered */
|
/* Level/edge triggered */
|
||||||
struct vgic_bitmap irq_cfg;
|
struct vgic_bitmap irq_cfg;
|
||||||
|
|
||||||
/* Source CPU per SGI and target CPU */
|
/*
|
||||||
u8 irq_sgi_sources[VGIC_MAX_CPUS][VGIC_NR_SGIS];
|
* Source CPU per SGI and target CPU:
|
||||||
|
*
|
||||||
|
* Each byte represent a SGI observable on a VCPU, each bit of
|
||||||
|
* this byte indicating if the corresponding VCPU has
|
||||||
|
* generated this interrupt. This is a GICv2 feature only.
|
||||||
|
*
|
||||||
|
* For VCPUn (n < 8), irq_sgi_sources[n*16] to [n*16 + 15] are
|
||||||
|
* the SGIs observable on VCPUn.
|
||||||
|
*/
|
||||||
|
u8 *irq_sgi_sources;
|
||||||
|
|
||||||
/* Target CPU for each IRQ */
|
/*
|
||||||
u8 irq_spi_cpu[VGIC_NR_SHARED_IRQS];
|
* Target CPU for each SPI:
|
||||||
struct vgic_bitmap irq_spi_target[VGIC_MAX_CPUS];
|
*
|
||||||
|
* Array of available SPI, each byte indicating the target
|
||||||
|
* VCPU for SPI. IRQn (n >=32) is at irq_spi_cpu[n-32].
|
||||||
|
*/
|
||||||
|
u8 *irq_spi_cpu;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Reverse lookup of irq_spi_cpu for faster compute pending:
|
||||||
|
*
|
||||||
|
* Array of bitmaps, one per VCPU, describing if IRQn is
|
||||||
|
* routed to a particular VCPU.
|
||||||
|
*/
|
||||||
|
struct vgic_bitmap *irq_spi_target;
|
||||||
|
|
||||||
/* Bitmap indicating which CPU has something pending */
|
/* Bitmap indicating which CPU has something pending */
|
||||||
unsigned long irq_pending_on_cpu;
|
unsigned long *irq_pending_on_cpu;
|
||||||
#endif
|
#endif
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -190,11 +241,11 @@ struct vgic_v3_cpu_if {
|
||||||
struct vgic_cpu {
|
struct vgic_cpu {
|
||||||
#ifdef CONFIG_KVM_ARM_VGIC
|
#ifdef CONFIG_KVM_ARM_VGIC
|
||||||
/* per IRQ to LR mapping */
|
/* per IRQ to LR mapping */
|
||||||
u8 vgic_irq_lr_map[VGIC_NR_IRQS];
|
u8 *vgic_irq_lr_map;
|
||||||
|
|
||||||
/* Pending interrupts on this VCPU */
|
/* Pending interrupts on this VCPU */
|
||||||
DECLARE_BITMAP( pending_percpu, VGIC_NR_PRIVATE_IRQS);
|
DECLARE_BITMAP( pending_percpu, VGIC_NR_PRIVATE_IRQS);
|
||||||
DECLARE_BITMAP( pending_shared, VGIC_NR_SHARED_IRQS);
|
unsigned long *pending_shared;
|
||||||
|
|
||||||
/* Bitmap of used/free list registers */
|
/* Bitmap of used/free list registers */
|
||||||
DECLARE_BITMAP( lr_used, VGIC_V2_MAX_LRS);
|
DECLARE_BITMAP( lr_used, VGIC_V2_MAX_LRS);
|
||||||
|
@ -225,7 +276,8 @@ int kvm_vgic_addr(struct kvm *kvm, unsigned long type, u64 *addr, bool write);
|
||||||
int kvm_vgic_hyp_init(void);
|
int kvm_vgic_hyp_init(void);
|
||||||
int kvm_vgic_init(struct kvm *kvm);
|
int kvm_vgic_init(struct kvm *kvm);
|
||||||
int kvm_vgic_create(struct kvm *kvm);
|
int kvm_vgic_create(struct kvm *kvm);
|
||||||
int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu);
|
void kvm_vgic_destroy(struct kvm *kvm);
|
||||||
|
void kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu);
|
||||||
void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu);
|
void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu);
|
||||||
void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu);
|
void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu);
|
||||||
int kvm_vgic_inject_irq(struct kvm *kvm, int cpuid, unsigned int irq_num,
|
int kvm_vgic_inject_irq(struct kvm *kvm, int cpuid, unsigned int irq_num,
|
||||||
|
|
|
@ -536,6 +536,8 @@ struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
|
||||||
unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
|
unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
|
||||||
unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
|
unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
|
||||||
unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
|
unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
|
||||||
|
unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
|
||||||
|
bool *writable);
|
||||||
void kvm_release_page_clean(struct page *page);
|
void kvm_release_page_clean(struct page *page);
|
||||||
void kvm_release_page_dirty(struct page *page);
|
void kvm_release_page_dirty(struct page *page);
|
||||||
void kvm_set_page_accessed(struct page *page);
|
void kvm_set_page_accessed(struct page *page);
|
||||||
|
|
File diff suppressed because it is too large
Load diff
|
@ -1095,9 +1095,9 @@ EXPORT_SYMBOL_GPL(gfn_to_hva);
|
||||||
* If writable is set to false, the hva returned by this function is only
|
* If writable is set to false, the hva returned by this function is only
|
||||||
* allowed to be read.
|
* allowed to be read.
|
||||||
*/
|
*/
|
||||||
unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable)
|
unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot,
|
||||||
|
gfn_t gfn, bool *writable)
|
||||||
{
|
{
|
||||||
struct kvm_memory_slot *slot = gfn_to_memslot(kvm, gfn);
|
|
||||||
unsigned long hva = __gfn_to_hva_many(slot, gfn, NULL, false);
|
unsigned long hva = __gfn_to_hva_many(slot, gfn, NULL, false);
|
||||||
|
|
||||||
if (!kvm_is_error_hva(hva) && writable)
|
if (!kvm_is_error_hva(hva) && writable)
|
||||||
|
@ -1106,6 +1106,13 @@ unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable)
|
||||||
return hva;
|
return hva;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable)
|
||||||
|
{
|
||||||
|
struct kvm_memory_slot *slot = gfn_to_memslot(kvm, gfn);
|
||||||
|
|
||||||
|
return gfn_to_hva_memslot_prot(slot, gfn, writable);
|
||||||
|
}
|
||||||
|
|
||||||
static int kvm_read_hva(void *data, void __user *hva, int len)
|
static int kvm_read_hva(void *data, void __user *hva, int len)
|
||||||
{
|
{
|
||||||
return __copy_from_user(data, hva, len);
|
return __copy_from_user(data, hva, len);
|
||||||
|
|
Loading…
Add table
Reference in a new issue