|
|
|
@@ -10,6 +10,7 @@
|
|
|
|
|
#ifndef __MIPS_KVM_HOST_H__
|
|
|
|
|
#define __MIPS_KVM_HOST_H__
|
|
|
|
|
|
|
|
|
|
#include <linux/cpumask.h>
|
|
|
|
|
#include <linux/mutex.h>
|
|
|
|
|
#include <linux/hrtimer.h>
|
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
@@ -33,12 +34,23 @@
|
|
|
|
|
#define KVM_REG_MIPS_CP0_ENTRYLO0 MIPS_CP0_64(2, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_ENTRYLO1 MIPS_CP0_64(3, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_CONTEXT MIPS_CP0_64(4, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_CONTEXTCONFIG MIPS_CP0_32(4, 1)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_USERLOCAL MIPS_CP0_64(4, 2)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_XCONTEXTCONFIG MIPS_CP0_64(4, 3)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_PAGEMASK MIPS_CP0_32(5, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_PAGEGRAIN MIPS_CP0_32(5, 1)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_SEGCTL0 MIPS_CP0_64(5, 2)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_SEGCTL1 MIPS_CP0_64(5, 3)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_SEGCTL2 MIPS_CP0_64(5, 4)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_PWBASE MIPS_CP0_64(5, 5)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_PWFIELD MIPS_CP0_64(5, 6)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_PWSIZE MIPS_CP0_64(5, 7)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_WIRED MIPS_CP0_32(6, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_PWCTL MIPS_CP0_32(6, 6)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_HWRENA MIPS_CP0_32(7, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_BADVADDR MIPS_CP0_64(8, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_BADINSTR MIPS_CP0_32(8, 1)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_BADINSTRP MIPS_CP0_32(8, 2)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_COUNT MIPS_CP0_32(9, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_ENTRYHI MIPS_CP0_64(10, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_COMPARE MIPS_CP0_32(11, 0)
|
|
|
|
@@ -55,6 +67,7 @@
|
|
|
|
|
#define KVM_REG_MIPS_CP0_CONFIG4 MIPS_CP0_32(16, 4)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_CONFIG5 MIPS_CP0_32(16, 5)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_CONFIG7 MIPS_CP0_32(16, 7)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_MAARI MIPS_CP0_64(17, 2)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_XCONTEXT MIPS_CP0_64(20, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_ERROREPC MIPS_CP0_64(30, 0)
|
|
|
|
|
#define KVM_REG_MIPS_CP0_KSCRATCH1 MIPS_CP0_64(31, 2)
|
|
|
|
@@ -70,9 +83,13 @@
|
|
|
|
|
/* memory slots that does not exposed to userspace */
|
|
|
|
|
#define KVM_PRIVATE_MEM_SLOTS 0
|
|
|
|
|
|
|
|
|
|
#define KVM_COALESCED_MMIO_PAGE_OFFSET 1
|
|
|
|
|
#define KVM_HALT_POLL_NS_DEFAULT 500000
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
extern unsigned long GUESTID_MASK;
|
|
|
|
|
extern unsigned long GUESTID_FIRST_VERSION;
|
|
|
|
|
extern unsigned long GUESTID_VERSION_MASK;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
@@ -145,6 +162,16 @@ struct kvm_vcpu_stat {
|
|
|
|
|
u64 fpe_exits;
|
|
|
|
|
u64 msa_disabled_exits;
|
|
|
|
|
u64 flush_dcache_exits;
|
|
|
|
|
#ifdef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
u64 vz_gpsi_exits;
|
|
|
|
|
u64 vz_gsfc_exits;
|
|
|
|
|
u64 vz_hc_exits;
|
|
|
|
|
u64 vz_grr_exits;
|
|
|
|
|
u64 vz_gva_exits;
|
|
|
|
|
u64 vz_ghfc_exits;
|
|
|
|
|
u64 vz_gpa_exits;
|
|
|
|
|
u64 vz_resvd_exits;
|
|
|
|
|
#endif
|
|
|
|
|
u64 halt_successful_poll;
|
|
|
|
|
u64 halt_attempted_poll;
|
|
|
|
|
u64 halt_poll_invalid;
|
|
|
|
@@ -157,6 +184,8 @@ struct kvm_arch_memory_slot {
|
|
|
|
|
struct kvm_arch {
|
|
|
|
|
/* Guest physical mm */
|
|
|
|
|
struct mm_struct gpa_mm;
|
|
|
|
|
/* Mask of CPUs needing GPA ASID flush */
|
|
|
|
|
cpumask_t asid_flush_mask;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
#define N_MIPS_COPROC_REGS 32
|
|
|
|
@@ -214,6 +243,11 @@ struct mips_coproc {
|
|
|
|
|
#define MIPS_CP0_CONFIG4_SEL 4
|
|
|
|
|
#define MIPS_CP0_CONFIG5_SEL 5
|
|
|
|
|
|
|
|
|
|
#define MIPS_CP0_GUESTCTL2 10
|
|
|
|
|
#define MIPS_CP0_GUESTCTL2_SEL 5
|
|
|
|
|
#define MIPS_CP0_GTOFFSET 12
|
|
|
|
|
#define MIPS_CP0_GTOFFSET_SEL 7
|
|
|
|
|
|
|
|
|
|
/* Resume Flags */
|
|
|
|
|
#define RESUME_FLAG_DR (1<<0) /* Reload guest nonvolatile state? */
|
|
|
|
|
#define RESUME_FLAG_HOST (1<<1) /* Resume host? */
|
|
|
|
@@ -229,6 +263,7 @@ enum emulation_result {
|
|
|
|
|
EMULATE_WAIT, /* WAIT instruction */
|
|
|
|
|
EMULATE_PRIV_FAIL,
|
|
|
|
|
EMULATE_EXCEPT, /* A guest exception has been generated */
|
|
|
|
|
EMULATE_HYPERCALL, /* HYPCALL instruction */
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
#define mips3_paddr_to_tlbpfn(x) \
|
|
|
|
@@ -276,13 +311,18 @@ struct kvm_mmu_memory_cache {
|
|
|
|
|
struct kvm_vcpu_arch {
|
|
|
|
|
void *guest_ebase;
|
|
|
|
|
int (*vcpu_run)(struct kvm_run *run, struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
/* Host registers preserved across guest mode execution */
|
|
|
|
|
unsigned long host_stack;
|
|
|
|
|
unsigned long host_gp;
|
|
|
|
|
unsigned long host_pgd;
|
|
|
|
|
unsigned long host_entryhi;
|
|
|
|
|
|
|
|
|
|
/* Host CP0 registers used when handling exits from guest */
|
|
|
|
|
unsigned long host_cp0_badvaddr;
|
|
|
|
|
unsigned long host_cp0_epc;
|
|
|
|
|
u32 host_cp0_cause;
|
|
|
|
|
u32 host_cp0_guestctl0;
|
|
|
|
|
u32 host_cp0_badinstr;
|
|
|
|
|
u32 host_cp0_badinstrp;
|
|
|
|
|
|
|
|
|
@@ -340,7 +380,23 @@ struct kvm_vcpu_arch {
|
|
|
|
|
/* Cache some mmu pages needed inside spinlock regions */
|
|
|
|
|
struct kvm_mmu_memory_cache mmu_page_cache;
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
/* vcpu's vzguestid is different on each host cpu in an smp system */
|
|
|
|
|
u32 vzguestid[NR_CPUS];
|
|
|
|
|
|
|
|
|
|
/* wired guest TLB entries */
|
|
|
|
|
struct kvm_mips_tlb *wired_tlb;
|
|
|
|
|
unsigned int wired_tlb_limit;
|
|
|
|
|
unsigned int wired_tlb_used;
|
|
|
|
|
|
|
|
|
|
/* emulated guest MAAR registers */
|
|
|
|
|
unsigned long maar[6];
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Last CPU the VCPU state was loaded on */
|
|
|
|
|
int last_sched_cpu;
|
|
|
|
|
/* Last CPU the VCPU actually executed guest code on */
|
|
|
|
|
int last_exec_cpu;
|
|
|
|
|
|
|
|
|
|
/* WAIT executed */
|
|
|
|
|
int wait;
|
|
|
|
@@ -349,78 +405,6 @@ struct kvm_vcpu_arch {
|
|
|
|
|
u8 msa_enabled;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#define kvm_read_c0_guest_index(cop0) (cop0->reg[MIPS_CP0_TLB_INDEX][0])
|
|
|
|
|
#define kvm_write_c0_guest_index(cop0, val) (cop0->reg[MIPS_CP0_TLB_INDEX][0] = val)
|
|
|
|
|
#define kvm_read_c0_guest_entrylo0(cop0) (cop0->reg[MIPS_CP0_TLB_LO0][0])
|
|
|
|
|
#define kvm_write_c0_guest_entrylo0(cop0, val) (cop0->reg[MIPS_CP0_TLB_LO0][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_entrylo1(cop0) (cop0->reg[MIPS_CP0_TLB_LO1][0])
|
|
|
|
|
#define kvm_write_c0_guest_entrylo1(cop0, val) (cop0->reg[MIPS_CP0_TLB_LO1][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_context(cop0) (cop0->reg[MIPS_CP0_TLB_CONTEXT][0])
|
|
|
|
|
#define kvm_write_c0_guest_context(cop0, val) (cop0->reg[MIPS_CP0_TLB_CONTEXT][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_userlocal(cop0) (cop0->reg[MIPS_CP0_TLB_CONTEXT][2])
|
|
|
|
|
#define kvm_write_c0_guest_userlocal(cop0, val) (cop0->reg[MIPS_CP0_TLB_CONTEXT][2] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_pagemask(cop0) (cop0->reg[MIPS_CP0_TLB_PG_MASK][0])
|
|
|
|
|
#define kvm_write_c0_guest_pagemask(cop0, val) (cop0->reg[MIPS_CP0_TLB_PG_MASK][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_wired(cop0) (cop0->reg[MIPS_CP0_TLB_WIRED][0])
|
|
|
|
|
#define kvm_write_c0_guest_wired(cop0, val) (cop0->reg[MIPS_CP0_TLB_WIRED][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_hwrena(cop0) (cop0->reg[MIPS_CP0_HWRENA][0])
|
|
|
|
|
#define kvm_write_c0_guest_hwrena(cop0, val) (cop0->reg[MIPS_CP0_HWRENA][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_badvaddr(cop0) (cop0->reg[MIPS_CP0_BAD_VADDR][0])
|
|
|
|
|
#define kvm_write_c0_guest_badvaddr(cop0, val) (cop0->reg[MIPS_CP0_BAD_VADDR][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_count(cop0) (cop0->reg[MIPS_CP0_COUNT][0])
|
|
|
|
|
#define kvm_write_c0_guest_count(cop0, val) (cop0->reg[MIPS_CP0_COUNT][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_entryhi(cop0) (cop0->reg[MIPS_CP0_TLB_HI][0])
|
|
|
|
|
#define kvm_write_c0_guest_entryhi(cop0, val) (cop0->reg[MIPS_CP0_TLB_HI][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_compare(cop0) (cop0->reg[MIPS_CP0_COMPARE][0])
|
|
|
|
|
#define kvm_write_c0_guest_compare(cop0, val) (cop0->reg[MIPS_CP0_COMPARE][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_status(cop0) (cop0->reg[MIPS_CP0_STATUS][0])
|
|
|
|
|
#define kvm_write_c0_guest_status(cop0, val) (cop0->reg[MIPS_CP0_STATUS][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_intctl(cop0) (cop0->reg[MIPS_CP0_STATUS][1])
|
|
|
|
|
#define kvm_write_c0_guest_intctl(cop0, val) (cop0->reg[MIPS_CP0_STATUS][1] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_cause(cop0) (cop0->reg[MIPS_CP0_CAUSE][0])
|
|
|
|
|
#define kvm_write_c0_guest_cause(cop0, val) (cop0->reg[MIPS_CP0_CAUSE][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_epc(cop0) (cop0->reg[MIPS_CP0_EXC_PC][0])
|
|
|
|
|
#define kvm_write_c0_guest_epc(cop0, val) (cop0->reg[MIPS_CP0_EXC_PC][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_prid(cop0) (cop0->reg[MIPS_CP0_PRID][0])
|
|
|
|
|
#define kvm_write_c0_guest_prid(cop0, val) (cop0->reg[MIPS_CP0_PRID][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_ebase(cop0) (cop0->reg[MIPS_CP0_PRID][1])
|
|
|
|
|
#define kvm_write_c0_guest_ebase(cop0, val) (cop0->reg[MIPS_CP0_PRID][1] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_config(cop0) (cop0->reg[MIPS_CP0_CONFIG][0])
|
|
|
|
|
#define kvm_read_c0_guest_config1(cop0) (cop0->reg[MIPS_CP0_CONFIG][1])
|
|
|
|
|
#define kvm_read_c0_guest_config2(cop0) (cop0->reg[MIPS_CP0_CONFIG][2])
|
|
|
|
|
#define kvm_read_c0_guest_config3(cop0) (cop0->reg[MIPS_CP0_CONFIG][3])
|
|
|
|
|
#define kvm_read_c0_guest_config4(cop0) (cop0->reg[MIPS_CP0_CONFIG][4])
|
|
|
|
|
#define kvm_read_c0_guest_config5(cop0) (cop0->reg[MIPS_CP0_CONFIG][5])
|
|
|
|
|
#define kvm_read_c0_guest_config7(cop0) (cop0->reg[MIPS_CP0_CONFIG][7])
|
|
|
|
|
#define kvm_write_c0_guest_config(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][0] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_config1(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][1] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_config2(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][2] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_config3(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][3] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_config4(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][4] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_config5(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][5] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_config7(cop0, val) (cop0->reg[MIPS_CP0_CONFIG][7] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_errorepc(cop0) (cop0->reg[MIPS_CP0_ERROR_PC][0])
|
|
|
|
|
#define kvm_write_c0_guest_errorepc(cop0, val) (cop0->reg[MIPS_CP0_ERROR_PC][0] = (val))
|
|
|
|
|
#define kvm_read_c0_guest_kscratch1(cop0) (cop0->reg[MIPS_CP0_DESAVE][2])
|
|
|
|
|
#define kvm_read_c0_guest_kscratch2(cop0) (cop0->reg[MIPS_CP0_DESAVE][3])
|
|
|
|
|
#define kvm_read_c0_guest_kscratch3(cop0) (cop0->reg[MIPS_CP0_DESAVE][4])
|
|
|
|
|
#define kvm_read_c0_guest_kscratch4(cop0) (cop0->reg[MIPS_CP0_DESAVE][5])
|
|
|
|
|
#define kvm_read_c0_guest_kscratch5(cop0) (cop0->reg[MIPS_CP0_DESAVE][6])
|
|
|
|
|
#define kvm_read_c0_guest_kscratch6(cop0) (cop0->reg[MIPS_CP0_DESAVE][7])
|
|
|
|
|
#define kvm_write_c0_guest_kscratch1(cop0, val) (cop0->reg[MIPS_CP0_DESAVE][2] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_kscratch2(cop0, val) (cop0->reg[MIPS_CP0_DESAVE][3] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_kscratch3(cop0, val) (cop0->reg[MIPS_CP0_DESAVE][4] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_kscratch4(cop0, val) (cop0->reg[MIPS_CP0_DESAVE][5] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_kscratch5(cop0, val) (cop0->reg[MIPS_CP0_DESAVE][6] = (val))
|
|
|
|
|
#define kvm_write_c0_guest_kscratch6(cop0, val) (cop0->reg[MIPS_CP0_DESAVE][7] = (val))
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Some of the guest registers may be modified asynchronously (e.g. from a
|
|
|
|
|
* hrtimer callback in hard irq context) and therefore need stronger atomicity
|
|
|
|
|
* guarantees than other registers.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
static inline void _kvm_atomic_set_c0_guest_reg(unsigned long *reg,
|
|
|
|
|
unsigned long val)
|
|
|
|
|
{
|
|
|
|
@@ -471,26 +455,286 @@ static inline void _kvm_atomic_change_c0_guest_reg(unsigned long *reg,
|
|
|
|
|
} while (unlikely(!temp));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#define kvm_set_c0_guest_status(cop0, val) (cop0->reg[MIPS_CP0_STATUS][0] |= (val))
|
|
|
|
|
#define kvm_clear_c0_guest_status(cop0, val) (cop0->reg[MIPS_CP0_STATUS][0] &= ~(val))
|
|
|
|
|
/* Guest register types, used in accessor build below */
|
|
|
|
|
#define __KVMT32 u32
|
|
|
|
|
#define __KVMTl unsigned long
|
|
|
|
|
|
|
|
|
|
/* Cause can be modified asynchronously from hardirq hrtimer callback */
|
|
|
|
|
#define kvm_set_c0_guest_cause(cop0, val) \
|
|
|
|
|
_kvm_atomic_set_c0_guest_reg(&cop0->reg[MIPS_CP0_CAUSE][0], val)
|
|
|
|
|
#define kvm_clear_c0_guest_cause(cop0, val) \
|
|
|
|
|
_kvm_atomic_clear_c0_guest_reg(&cop0->reg[MIPS_CP0_CAUSE][0], val)
|
|
|
|
|
#define kvm_change_c0_guest_cause(cop0, change, val) \
|
|
|
|
|
_kvm_atomic_change_c0_guest_reg(&cop0->reg[MIPS_CP0_CAUSE][0], \
|
|
|
|
|
change, val)
|
|
|
|
|
/*
|
|
|
|
|
* __BUILD_KVM_$ops_SAVED(): kvm_$op_sw_gc0_$reg()
|
|
|
|
|
* These operate on the saved guest C0 state in RAM.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#define kvm_set_c0_guest_ebase(cop0, val) (cop0->reg[MIPS_CP0_PRID][1] |= (val))
|
|
|
|
|
#define kvm_clear_c0_guest_ebase(cop0, val) (cop0->reg[MIPS_CP0_PRID][1] &= ~(val))
|
|
|
|
|
#define kvm_change_c0_guest_ebase(cop0, change, val) \
|
|
|
|
|
/* Generate saved context simple accessors */
|
|
|
|
|
#define __BUILD_KVM_RW_SAVED(name, type, _reg, sel) \
|
|
|
|
|
static inline __KVMT##type kvm_read_sw_gc0_##name(struct mips_coproc *cop0) \
|
|
|
|
|
{ \
|
|
|
|
|
kvm_clear_c0_guest_ebase(cop0, change); \
|
|
|
|
|
kvm_set_c0_guest_ebase(cop0, ((val) & (change))); \
|
|
|
|
|
return cop0->reg[(_reg)][(sel)]; \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_write_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
cop0->reg[(_reg)][(sel)] = val; \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate saved context bitwise modifiers */
|
|
|
|
|
#define __BUILD_KVM_SET_SAVED(name, type, _reg, sel) \
|
|
|
|
|
static inline void kvm_set_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
cop0->reg[(_reg)][(sel)] |= val; \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_clear_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
cop0->reg[(_reg)][(sel)] &= ~val; \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_change_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type mask, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
unsigned long _mask = mask; \
|
|
|
|
|
cop0->reg[(_reg)][(sel)] &= ~_mask; \
|
|
|
|
|
cop0->reg[(_reg)][(sel)] |= val & _mask; \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate saved context atomic bitwise modifiers */
|
|
|
|
|
#define __BUILD_KVM_ATOMIC_SAVED(name, type, _reg, sel) \
|
|
|
|
|
static inline void kvm_set_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
_kvm_atomic_set_c0_guest_reg(&cop0->reg[(_reg)][(sel)], val); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_clear_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
_kvm_atomic_clear_c0_guest_reg(&cop0->reg[(_reg)][(sel)], val); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_change_sw_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type mask, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
_kvm_atomic_change_c0_guest_reg(&cop0->reg[(_reg)][(sel)], mask, \
|
|
|
|
|
val); \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* __BUILD_KVM_$ops_VZ(): kvm_$op_vz_gc0_$reg()
|
|
|
|
|
* These operate on the VZ guest C0 context in hardware.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
/* Generate VZ guest context simple accessors */
|
|
|
|
|
#define __BUILD_KVM_RW_VZ(name, type, _reg, sel) \
|
|
|
|
|
static inline __KVMT##type kvm_read_vz_gc0_##name(struct mips_coproc *cop0) \
|
|
|
|
|
{ \
|
|
|
|
|
return read_gc0_##name(); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_write_vz_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
write_gc0_##name(val); \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate VZ guest context bitwise modifiers */
|
|
|
|
|
#define __BUILD_KVM_SET_VZ(name, type, _reg, sel) \
|
|
|
|
|
static inline void kvm_set_vz_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
set_gc0_##name(val); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_clear_vz_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
clear_gc0_##name(val); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_change_vz_gc0_##name(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type mask, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
change_gc0_##name(mask, val); \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate VZ guest context save/restore to/from saved context */
|
|
|
|
|
#define __BUILD_KVM_SAVE_VZ(name, _reg, sel) \
|
|
|
|
|
static inline void kvm_restore_gc0_##name(struct mips_coproc *cop0) \
|
|
|
|
|
{ \
|
|
|
|
|
write_gc0_##name(cop0->reg[(_reg)][(sel)]); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_save_gc0_##name(struct mips_coproc *cop0) \
|
|
|
|
|
{ \
|
|
|
|
|
cop0->reg[(_reg)][(sel)] = read_gc0_##name(); \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* __BUILD_KVM_$ops_WRAP(): kvm_$op_$name1() -> kvm_$op_$name2()
|
|
|
|
|
* These wrap a set of operations to provide them with a different name.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
/* Generate simple accessor wrapper */
|
|
|
|
|
#define __BUILD_KVM_RW_WRAP(name1, name2, type) \
|
|
|
|
|
static inline __KVMT##type kvm_read_##name1(struct mips_coproc *cop0) \
|
|
|
|
|
{ \
|
|
|
|
|
return kvm_read_##name2(cop0); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_write_##name1(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
kvm_write_##name2(cop0, val); \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Generate bitwise modifier wrapper */
|
|
|
|
|
#define __BUILD_KVM_SET_WRAP(name1, name2, type) \
|
|
|
|
|
static inline void kvm_set_##name1(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
kvm_set_##name2(cop0, val); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_clear_##name1(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
kvm_clear_##name2(cop0, val); \
|
|
|
|
|
} \
|
|
|
|
|
static inline void kvm_change_##name1(struct mips_coproc *cop0, \
|
|
|
|
|
__KVMT##type mask, \
|
|
|
|
|
__KVMT##type val) \
|
|
|
|
|
{ \
|
|
|
|
|
kvm_change_##name2(cop0, mask, val); \
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* __BUILD_KVM_$ops_SW(): kvm_$op_c0_guest_$reg() -> kvm_$op_sw_gc0_$reg()
|
|
|
|
|
* These generate accessors operating on the saved context in RAM, and wrap them
|
|
|
|
|
* with the common guest C0 accessors (for use by common emulation code).
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#define __BUILD_KVM_RW_SW(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_RW_SAVED(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_RW_WRAP(c0_guest_##name, sw_gc0_##name, type)
|
|
|
|
|
|
|
|
|
|
#define __BUILD_KVM_SET_SW(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_SET_SAVED(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_SET_WRAP(c0_guest_##name, sw_gc0_##name, type)
|
|
|
|
|
|
|
|
|
|
#define __BUILD_KVM_ATOMIC_SW(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_ATOMIC_SAVED(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_SET_WRAP(c0_guest_##name, sw_gc0_##name, type)
|
|
|
|
|
|
|
|
|
|
#ifndef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* T&E (trap & emulate software based virtualisation)
|
|
|
|
|
* We generate the common accessors operating exclusively on the saved context
|
|
|
|
|
* in RAM.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#define __BUILD_KVM_RW_HW __BUILD_KVM_RW_SW
|
|
|
|
|
#define __BUILD_KVM_SET_HW __BUILD_KVM_SET_SW
|
|
|
|
|
#define __BUILD_KVM_ATOMIC_HW __BUILD_KVM_ATOMIC_SW
|
|
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* VZ (hardware assisted virtualisation)
|
|
|
|
|
* These macros use the active guest state in VZ mode (hardware registers),
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* __BUILD_KVM_$ops_HW(): kvm_$op_c0_guest_$reg() -> kvm_$op_vz_gc0_$reg()
|
|
|
|
|
* These generate accessors operating on the VZ guest context in hardware, and
|
|
|
|
|
* wrap them with the common guest C0 accessors (for use by common emulation
|
|
|
|
|
* code).
|
|
|
|
|
*
|
|
|
|
|
* Accessors operating on the saved context in RAM are also generated to allow
|
|
|
|
|
* convenient explicit saving and restoring of the state.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#define __BUILD_KVM_RW_HW(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_RW_SAVED(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_RW_VZ(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_RW_WRAP(c0_guest_##name, vz_gc0_##name, type) \
|
|
|
|
|
__BUILD_KVM_SAVE_VZ(name, _reg, sel)
|
|
|
|
|
|
|
|
|
|
#define __BUILD_KVM_SET_HW(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_SET_SAVED(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_SET_VZ(name, type, _reg, sel) \
|
|
|
|
|
__BUILD_KVM_SET_WRAP(c0_guest_##name, vz_gc0_##name, type)
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* We can't do atomic modifications of COP0 state if hardware can modify it.
|
|
|
|
|
* Races must be handled explicitly.
|
|
|
|
|
*/
|
|
|
|
|
#define __BUILD_KVM_ATOMIC_HW __BUILD_KVM_SET_HW
|
|
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Define accessors for CP0 registers that are accessible to the guest. These
|
|
|
|
|
* are primarily used by common emulation code, which may need to access the
|
|
|
|
|
* registers differently depending on the implementation.
|
|
|
|
|
*
|
|
|
|
|
* fns_hw/sw name type reg num select
|
|
|
|
|
*/
|
|
|
|
|
__BUILD_KVM_RW_HW(index, 32, MIPS_CP0_TLB_INDEX, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(entrylo0, l, MIPS_CP0_TLB_LO0, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(entrylo1, l, MIPS_CP0_TLB_LO1, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(context, l, MIPS_CP0_TLB_CONTEXT, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(contextconfig, 32, MIPS_CP0_TLB_CONTEXT, 1)
|
|
|
|
|
__BUILD_KVM_RW_HW(userlocal, l, MIPS_CP0_TLB_CONTEXT, 2)
|
|
|
|
|
__BUILD_KVM_RW_HW(xcontextconfig, l, MIPS_CP0_TLB_CONTEXT, 3)
|
|
|
|
|
__BUILD_KVM_RW_HW(pagemask, l, MIPS_CP0_TLB_PG_MASK, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(pagegrain, 32, MIPS_CP0_TLB_PG_MASK, 1)
|
|
|
|
|
__BUILD_KVM_RW_HW(segctl0, l, MIPS_CP0_TLB_PG_MASK, 2)
|
|
|
|
|
__BUILD_KVM_RW_HW(segctl1, l, MIPS_CP0_TLB_PG_MASK, 3)
|
|
|
|
|
__BUILD_KVM_RW_HW(segctl2, l, MIPS_CP0_TLB_PG_MASK, 4)
|
|
|
|
|
__BUILD_KVM_RW_HW(pwbase, l, MIPS_CP0_TLB_PG_MASK, 5)
|
|
|
|
|
__BUILD_KVM_RW_HW(pwfield, l, MIPS_CP0_TLB_PG_MASK, 6)
|
|
|
|
|
__BUILD_KVM_RW_HW(pwsize, l, MIPS_CP0_TLB_PG_MASK, 7)
|
|
|
|
|
__BUILD_KVM_RW_HW(wired, 32, MIPS_CP0_TLB_WIRED, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(pwctl, 32, MIPS_CP0_TLB_WIRED, 6)
|
|
|
|
|
__BUILD_KVM_RW_HW(hwrena, 32, MIPS_CP0_HWRENA, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(badvaddr, l, MIPS_CP0_BAD_VADDR, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(badinstr, 32, MIPS_CP0_BAD_VADDR, 1)
|
|
|
|
|
__BUILD_KVM_RW_HW(badinstrp, 32, MIPS_CP0_BAD_VADDR, 2)
|
|
|
|
|
__BUILD_KVM_RW_SW(count, 32, MIPS_CP0_COUNT, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(entryhi, l, MIPS_CP0_TLB_HI, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(compare, 32, MIPS_CP0_COMPARE, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(status, 32, MIPS_CP0_STATUS, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(intctl, 32, MIPS_CP0_STATUS, 1)
|
|
|
|
|
__BUILD_KVM_RW_HW(cause, 32, MIPS_CP0_CAUSE, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(epc, l, MIPS_CP0_EXC_PC, 0)
|
|
|
|
|
__BUILD_KVM_RW_SW(prid, 32, MIPS_CP0_PRID, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(ebase, l, MIPS_CP0_PRID, 1)
|
|
|
|
|
__BUILD_KVM_RW_HW(config, 32, MIPS_CP0_CONFIG, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(config1, 32, MIPS_CP0_CONFIG, 1)
|
|
|
|
|
__BUILD_KVM_RW_HW(config2, 32, MIPS_CP0_CONFIG, 2)
|
|
|
|
|
__BUILD_KVM_RW_HW(config3, 32, MIPS_CP0_CONFIG, 3)
|
|
|
|
|
__BUILD_KVM_RW_HW(config4, 32, MIPS_CP0_CONFIG, 4)
|
|
|
|
|
__BUILD_KVM_RW_HW(config5, 32, MIPS_CP0_CONFIG, 5)
|
|
|
|
|
__BUILD_KVM_RW_HW(config6, 32, MIPS_CP0_CONFIG, 6)
|
|
|
|
|
__BUILD_KVM_RW_HW(config7, 32, MIPS_CP0_CONFIG, 7)
|
|
|
|
|
__BUILD_KVM_RW_SW(maari, l, MIPS_CP0_LLADDR, 2)
|
|
|
|
|
__BUILD_KVM_RW_HW(xcontext, l, MIPS_CP0_TLB_XCONTEXT, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(errorepc, l, MIPS_CP0_ERROR_PC, 0)
|
|
|
|
|
__BUILD_KVM_RW_HW(kscratch1, l, MIPS_CP0_DESAVE, 2)
|
|
|
|
|
__BUILD_KVM_RW_HW(kscratch2, l, MIPS_CP0_DESAVE, 3)
|
|
|
|
|
__BUILD_KVM_RW_HW(kscratch3, l, MIPS_CP0_DESAVE, 4)
|
|
|
|
|
__BUILD_KVM_RW_HW(kscratch4, l, MIPS_CP0_DESAVE, 5)
|
|
|
|
|
__BUILD_KVM_RW_HW(kscratch5, l, MIPS_CP0_DESAVE, 6)
|
|
|
|
|
__BUILD_KVM_RW_HW(kscratch6, l, MIPS_CP0_DESAVE, 7)
|
|
|
|
|
|
|
|
|
|
/* Bitwise operations (on HW state) */
|
|
|
|
|
__BUILD_KVM_SET_HW(status, 32, MIPS_CP0_STATUS, 0)
|
|
|
|
|
/* Cause can be modified asynchronously from hardirq hrtimer callback */
|
|
|
|
|
__BUILD_KVM_ATOMIC_HW(cause, 32, MIPS_CP0_CAUSE, 0)
|
|
|
|
|
__BUILD_KVM_SET_HW(ebase, l, MIPS_CP0_PRID, 1)
|
|
|
|
|
|
|
|
|
|
/* Bitwise operations (on saved state) */
|
|
|
|
|
__BUILD_KVM_SET_SAVED(config, 32, MIPS_CP0_CONFIG, 0)
|
|
|
|
|
__BUILD_KVM_SET_SAVED(config1, 32, MIPS_CP0_CONFIG, 1)
|
|
|
|
|
__BUILD_KVM_SET_SAVED(config2, 32, MIPS_CP0_CONFIG, 2)
|
|
|
|
|
__BUILD_KVM_SET_SAVED(config3, 32, MIPS_CP0_CONFIG, 3)
|
|
|
|
|
__BUILD_KVM_SET_SAVED(config4, 32, MIPS_CP0_CONFIG, 4)
|
|
|
|
|
__BUILD_KVM_SET_SAVED(config5, 32, MIPS_CP0_CONFIG, 5)
|
|
|
|
|
|
|
|
|
|
/* Helpers */
|
|
|
|
|
|
|
|
|
|
static inline bool kvm_mips_guest_can_have_fpu(struct kvm_vcpu_arch *vcpu)
|
|
|
|
@@ -531,6 +775,10 @@ struct kvm_mips_callbacks {
|
|
|
|
|
int (*handle_msa_fpe)(struct kvm_vcpu *vcpu);
|
|
|
|
|
int (*handle_fpe)(struct kvm_vcpu *vcpu);
|
|
|
|
|
int (*handle_msa_disabled)(struct kvm_vcpu *vcpu);
|
|
|
|
|
int (*handle_guest_exit)(struct kvm_vcpu *vcpu);
|
|
|
|
|
int (*hardware_enable)(void);
|
|
|
|
|
void (*hardware_disable)(void);
|
|
|
|
|
int (*check_extension)(struct kvm *kvm, long ext);
|
|
|
|
|
int (*vcpu_init)(struct kvm_vcpu *vcpu);
|
|
|
|
|
void (*vcpu_uninit)(struct kvm_vcpu *vcpu);
|
|
|
|
|
int (*vcpu_setup)(struct kvm_vcpu *vcpu);
|
|
|
|
@@ -599,6 +847,10 @@ u32 kvm_get_user_asid(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
u32 kvm_get_commpage_asid (struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
int kvm_mips_handle_vz_root_tlb_fault(unsigned long badvaddr,
|
|
|
|
|
struct kvm_vcpu *vcpu, bool write_fault);
|
|
|
|
|
#endif
|
|
|
|
|
extern int kvm_mips_handle_kseg0_tlb_fault(unsigned long badbaddr,
|
|
|
|
|
struct kvm_vcpu *vcpu,
|
|
|
|
|
bool write_fault);
|
|
|
|
@@ -625,6 +877,18 @@ extern int kvm_mips_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi,
|
|
|
|
|
extern int kvm_mips_guest_tlb_lookup(struct kvm_vcpu *vcpu,
|
|
|
|
|
unsigned long entryhi);
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi);
|
|
|
|
|
int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva,
|
|
|
|
|
unsigned long *gpa);
|
|
|
|
|
void kvm_vz_local_flush_roottlb_all_guests(void);
|
|
|
|
|
void kvm_vz_local_flush_guesttlb_all(void);
|
|
|
|
|
void kvm_vz_save_guesttlb(struct kvm_mips_tlb *buf, unsigned int index,
|
|
|
|
|
unsigned int count);
|
|
|
|
|
void kvm_vz_load_guesttlb(const struct kvm_mips_tlb *buf, unsigned int index,
|
|
|
|
|
unsigned int count);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
void kvm_mips_suspend_mm(int cpu);
|
|
|
|
|
void kvm_mips_resume_mm(int cpu);
|
|
|
|
|
|
|
|
|
@@ -795,7 +1059,7 @@ extern enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu,
|
|
|
|
|
u32 kvm_mips_read_count(struct kvm_vcpu *vcpu);
|
|
|
|
|
void kvm_mips_write_count(struct kvm_vcpu *vcpu, u32 count);
|
|
|
|
|
void kvm_mips_write_compare(struct kvm_vcpu *vcpu, u32 compare, bool ack);
|
|
|
|
|
void kvm_mips_init_count(struct kvm_vcpu *vcpu);
|
|
|
|
|
void kvm_mips_init_count(struct kvm_vcpu *vcpu, unsigned long count_hz);
|
|
|
|
|
int kvm_mips_set_count_ctl(struct kvm_vcpu *vcpu, s64 count_ctl);
|
|
|
|
|
int kvm_mips_set_count_resume(struct kvm_vcpu *vcpu, s64 count_resume);
|
|
|
|
|
int kvm_mips_set_count_hz(struct kvm_vcpu *vcpu, s64 count_hz);
|
|
|
|
@@ -803,6 +1067,20 @@ void kvm_mips_count_enable_cause(struct kvm_vcpu *vcpu);
|
|
|
|
|
void kvm_mips_count_disable_cause(struct kvm_vcpu *vcpu);
|
|
|
|
|
enum hrtimer_restart kvm_mips_count_timeout(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
/* fairly internal functions requiring some care to use */
|
|
|
|
|
int kvm_mips_count_disabled(struct kvm_vcpu *vcpu);
|
|
|
|
|
ktime_t kvm_mips_freeze_hrtimer(struct kvm_vcpu *vcpu, u32 *count);
|
|
|
|
|
int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before,
|
|
|
|
|
u32 count, int min_drift);
|
|
|
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_MIPS_VZ
|
|
|
|
|
void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu);
|
|
|
|
|
void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu);
|
|
|
|
|
#else
|
|
|
|
|
static inline void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu) {}
|
|
|
|
|
static inline void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu) {}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
enum emulation_result kvm_mips_check_privilege(u32 cause,
|
|
|
|
|
u32 *opc,
|
|
|
|
|
struct kvm_run *run,
|
|
|
|
@@ -827,11 +1105,20 @@ enum emulation_result kvm_mips_emulate_load(union mips_instruction inst,
|
|
|
|
|
struct kvm_run *run,
|
|
|
|
|
struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
/* COP0 */
|
|
|
|
|
enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
unsigned int kvm_mips_config1_wrmask(struct kvm_vcpu *vcpu);
|
|
|
|
|
unsigned int kvm_mips_config3_wrmask(struct kvm_vcpu *vcpu);
|
|
|
|
|
unsigned int kvm_mips_config4_wrmask(struct kvm_vcpu *vcpu);
|
|
|
|
|
unsigned int kvm_mips_config5_wrmask(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
/* Hypercalls (hypcall.c) */
|
|
|
|
|
|
|
|
|
|
enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu,
|
|
|
|
|
union mips_instruction inst);
|
|
|
|
|
int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
|
|
/* Dynamic binary translation */
|
|
|
|
|
extern int kvm_mips_trans_cache_index(union mips_instruction inst,
|
|
|
|
|
u32 *opc, struct kvm_vcpu *vcpu);
|
|
|
|
@@ -846,7 +1133,6 @@ extern int kvm_mips_trans_mtc0(union mips_instruction inst, u32 *opc,
|
|
|
|
|
extern void kvm_mips_dump_stats(struct kvm_vcpu *vcpu);
|
|
|
|
|
extern unsigned long kvm_mips_get_ramsize(struct kvm *kvm);
|
|
|
|
|
|
|
|
|
|
static inline void kvm_arch_hardware_disable(void) {}
|
|
|
|
|
static inline void kvm_arch_hardware_unsetup(void) {}
|
|
|
|
|
static inline void kvm_arch_sync_events(struct kvm *kvm) {}
|
|
|
|
|
static inline void kvm_arch_free_memslot(struct kvm *kvm,
|
|
|
|
|