Merge branch 'kvm-updates/3.4' of git://git.kernel.org/pub/scm/virt/kvm/kvm
Pull kvm updates from Avi Kivity: "Changes include timekeeping improvements, support for assigning host PCI devices that share interrupt lines, s390 user-controlled guests, a large ppc update, and random fixes." This is with the sign-off's fixed, hopefully next merge window we won't have rebased commits. * 'kvm-updates/3.4' of git://git.kernel.org/pub/scm/virt/kvm/kvm: (130 commits) KVM: Convert intx_mask_lock to spin lock KVM: x86: fix kvm_write_tsc() TSC matching thinko x86: kvmclock: abstract save/restore sched_clock_state KVM: nVMX: Fix erroneous exception bitmap check KVM: Ignore the writes to MSR_K7_HWCR(3) KVM: MMU: make use of ->root_level in reset_rsvds_bits_mask KVM: PMU: add proper support for fixed counter 2 KVM: PMU: Fix raw event check KVM: PMU: warn when pin control is set in eventsel msr KVM: VMX: Fix delayed load of shared MSRs KVM: use correct tlbs dirty type in cmpxchg KVM: Allow host IRQ sharing for assigned PCI 2.3 devices KVM: Ensure all vcpus are consistent with in-kernel irqchip settings KVM: x86 emulator: Allow PM/VM86 switch during task switch KVM: SVM: Fix CPL updates KVM: x86 emulator: VM86 segments must have DPL 3 KVM: x86 emulator: Fix task switch privilege checks arch/powerpc/kvm/book3s_hv.c: included linux/sched.h twice KVM: x86 emulator: correctly mask pmc index bits in RDPMC instruction emulation KVM: mmu_notifier: Flush TLBs before releasing mmu_lock ...
This commit is contained in:
@@ -412,16 +412,23 @@ int main(void)
|
||||
DEFINE(VCPU_SPRG2, offsetof(struct kvm_vcpu, arch.shregs.sprg2));
|
||||
DEFINE(VCPU_SPRG3, offsetof(struct kvm_vcpu, arch.shregs.sprg3));
|
||||
#endif
|
||||
DEFINE(VCPU_SPRG4, offsetof(struct kvm_vcpu, arch.sprg4));
|
||||
DEFINE(VCPU_SPRG5, offsetof(struct kvm_vcpu, arch.sprg5));
|
||||
DEFINE(VCPU_SPRG6, offsetof(struct kvm_vcpu, arch.sprg6));
|
||||
DEFINE(VCPU_SPRG7, offsetof(struct kvm_vcpu, arch.sprg7));
|
||||
DEFINE(VCPU_SHARED_SPRG4, offsetof(struct kvm_vcpu_arch_shared, sprg4));
|
||||
DEFINE(VCPU_SHARED_SPRG5, offsetof(struct kvm_vcpu_arch_shared, sprg5));
|
||||
DEFINE(VCPU_SHARED_SPRG6, offsetof(struct kvm_vcpu_arch_shared, sprg6));
|
||||
DEFINE(VCPU_SHARED_SPRG7, offsetof(struct kvm_vcpu_arch_shared, sprg7));
|
||||
DEFINE(VCPU_SHADOW_PID, offsetof(struct kvm_vcpu, arch.shadow_pid));
|
||||
DEFINE(VCPU_SHADOW_PID1, offsetof(struct kvm_vcpu, arch.shadow_pid1));
|
||||
DEFINE(VCPU_SHARED, offsetof(struct kvm_vcpu, arch.shared));
|
||||
DEFINE(VCPU_SHARED_MSR, offsetof(struct kvm_vcpu_arch_shared, msr));
|
||||
DEFINE(VCPU_SHADOW_MSR, offsetof(struct kvm_vcpu, arch.shadow_msr));
|
||||
|
||||
DEFINE(VCPU_SHARED_MAS0, offsetof(struct kvm_vcpu_arch_shared, mas0));
|
||||
DEFINE(VCPU_SHARED_MAS1, offsetof(struct kvm_vcpu_arch_shared, mas1));
|
||||
DEFINE(VCPU_SHARED_MAS2, offsetof(struct kvm_vcpu_arch_shared, mas2));
|
||||
DEFINE(VCPU_SHARED_MAS7_3, offsetof(struct kvm_vcpu_arch_shared, mas7_3));
|
||||
DEFINE(VCPU_SHARED_MAS4, offsetof(struct kvm_vcpu_arch_shared, mas4));
|
||||
DEFINE(VCPU_SHARED_MAS6, offsetof(struct kvm_vcpu_arch_shared, mas6));
|
||||
|
||||
/* book3s */
|
||||
#ifdef CONFIG_KVM_BOOK3S_64_HV
|
||||
DEFINE(KVM_LPID, offsetof(struct kvm, arch.lpid));
|
||||
@@ -434,6 +441,7 @@ int main(void)
|
||||
DEFINE(KVM_LAST_VCPU, offsetof(struct kvm, arch.last_vcpu));
|
||||
DEFINE(KVM_LPCR, offsetof(struct kvm, arch.lpcr));
|
||||
DEFINE(KVM_RMOR, offsetof(struct kvm, arch.rmor));
|
||||
DEFINE(KVM_VRMA_SLB_V, offsetof(struct kvm, arch.vrma_slb_v));
|
||||
DEFINE(VCPU_DSISR, offsetof(struct kvm_vcpu, arch.shregs.dsisr));
|
||||
DEFINE(VCPU_DAR, offsetof(struct kvm_vcpu, arch.shregs.dar));
|
||||
#endif
|
||||
|
@@ -101,14 +101,14 @@ data_access_not_stab:
|
||||
END_MMU_FTR_SECTION_IFCLR(MMU_FTR_SLB)
|
||||
#endif
|
||||
EXCEPTION_PROLOG_PSERIES(PACA_EXGEN, data_access_common, EXC_STD,
|
||||
KVMTEST_PR, 0x300)
|
||||
KVMTEST, 0x300)
|
||||
|
||||
. = 0x380
|
||||
.globl data_access_slb_pSeries
|
||||
data_access_slb_pSeries:
|
||||
HMT_MEDIUM
|
||||
SET_SCRATCH0(r13)
|
||||
EXCEPTION_PROLOG_1(PACA_EXSLB, KVMTEST_PR, 0x380)
|
||||
EXCEPTION_PROLOG_1(PACA_EXSLB, KVMTEST, 0x380)
|
||||
std r3,PACA_EXSLB+EX_R3(r13)
|
||||
mfspr r3,SPRN_DAR
|
||||
#ifdef __DISABLED__
|
||||
@@ -330,8 +330,8 @@ do_stab_bolted_pSeries:
|
||||
EXCEPTION_PROLOG_PSERIES_1(.do_stab_bolted, EXC_STD)
|
||||
#endif /* CONFIG_POWER4_ONLY */
|
||||
|
||||
KVM_HANDLER_PR_SKIP(PACA_EXGEN, EXC_STD, 0x300)
|
||||
KVM_HANDLER_PR_SKIP(PACA_EXSLB, EXC_STD, 0x380)
|
||||
KVM_HANDLER_SKIP(PACA_EXGEN, EXC_STD, 0x300)
|
||||
KVM_HANDLER_SKIP(PACA_EXSLB, EXC_STD, 0x380)
|
||||
KVM_HANDLER_PR(PACA_EXGEN, EXC_STD, 0x400)
|
||||
KVM_HANDLER_PR(PACA_EXSLB, EXC_STD, 0x480)
|
||||
KVM_HANDLER_PR(PACA_EXGEN, EXC_STD, 0x900)
|
||||
|
@@ -1,5 +1,6 @@
|
||||
/*
|
||||
* Copyright (C) 2010 SUSE Linux Products GmbH. All rights reserved.
|
||||
* Copyright 2010-2011 Freescale Semiconductor, Inc.
|
||||
*
|
||||
* Authors:
|
||||
* Alexander Graf <agraf@suse.de>
|
||||
@@ -29,6 +30,7 @@
|
||||
#include <asm/sections.h>
|
||||
#include <asm/cacheflush.h>
|
||||
#include <asm/disassemble.h>
|
||||
#include <asm/ppc-opcode.h>
|
||||
|
||||
#define KVM_MAGIC_PAGE (-4096L)
|
||||
#define magic_var(x) KVM_MAGIC_PAGE + offsetof(struct kvm_vcpu_arch_shared, x)
|
||||
@@ -41,34 +43,30 @@
|
||||
#define KVM_INST_B 0x48000000
|
||||
#define KVM_INST_B_MASK 0x03ffffff
|
||||
#define KVM_INST_B_MAX 0x01ffffff
|
||||
#define KVM_INST_LI 0x38000000
|
||||
|
||||
#define KVM_MASK_RT 0x03e00000
|
||||
#define KVM_RT_30 0x03c00000
|
||||
#define KVM_MASK_RB 0x0000f800
|
||||
#define KVM_INST_MFMSR 0x7c0000a6
|
||||
#define KVM_INST_MFSPR_SPRG0 0x7c1042a6
|
||||
#define KVM_INST_MFSPR_SPRG1 0x7c1142a6
|
||||
#define KVM_INST_MFSPR_SPRG2 0x7c1242a6
|
||||
#define KVM_INST_MFSPR_SPRG3 0x7c1342a6
|
||||
#define KVM_INST_MFSPR_SRR0 0x7c1a02a6
|
||||
#define KVM_INST_MFSPR_SRR1 0x7c1b02a6
|
||||
#define KVM_INST_MFSPR_DAR 0x7c1302a6
|
||||
#define KVM_INST_MFSPR_DSISR 0x7c1202a6
|
||||
|
||||
#define KVM_INST_MTSPR_SPRG0 0x7c1043a6
|
||||
#define KVM_INST_MTSPR_SPRG1 0x7c1143a6
|
||||
#define KVM_INST_MTSPR_SPRG2 0x7c1243a6
|
||||
#define KVM_INST_MTSPR_SPRG3 0x7c1343a6
|
||||
#define KVM_INST_MTSPR_SRR0 0x7c1a03a6
|
||||
#define KVM_INST_MTSPR_SRR1 0x7c1b03a6
|
||||
#define KVM_INST_MTSPR_DAR 0x7c1303a6
|
||||
#define KVM_INST_MTSPR_DSISR 0x7c1203a6
|
||||
#define SPR_FROM 0
|
||||
#define SPR_TO 0x100
|
||||
|
||||
#define KVM_INST_SPR(sprn, moveto) (0x7c0002a6 | \
|
||||
(((sprn) & 0x1f) << 16) | \
|
||||
(((sprn) & 0x3e0) << 6) | \
|
||||
(moveto))
|
||||
|
||||
#define KVM_INST_MFSPR(sprn) KVM_INST_SPR(sprn, SPR_FROM)
|
||||
#define KVM_INST_MTSPR(sprn) KVM_INST_SPR(sprn, SPR_TO)
|
||||
|
||||
#define KVM_INST_TLBSYNC 0x7c00046c
|
||||
#define KVM_INST_MTMSRD_L0 0x7c000164
|
||||
#define KVM_INST_MTMSRD_L1 0x7c010164
|
||||
#define KVM_INST_MTMSR 0x7c000124
|
||||
|
||||
#define KVM_INST_WRTEE 0x7c000106
|
||||
#define KVM_INST_WRTEEI_0 0x7c000146
|
||||
#define KVM_INST_WRTEEI_1 0x7c008146
|
||||
|
||||
@@ -270,26 +268,27 @@ static void kvm_patch_ins_mtmsr(u32 *inst, u32 rt)
|
||||
|
||||
#ifdef CONFIG_BOOKE
|
||||
|
||||
extern u32 kvm_emulate_wrteei_branch_offs;
|
||||
extern u32 kvm_emulate_wrteei_ee_offs;
|
||||
extern u32 kvm_emulate_wrteei_len;
|
||||
extern u32 kvm_emulate_wrteei[];
|
||||
extern u32 kvm_emulate_wrtee_branch_offs;
|
||||
extern u32 kvm_emulate_wrtee_reg_offs;
|
||||
extern u32 kvm_emulate_wrtee_orig_ins_offs;
|
||||
extern u32 kvm_emulate_wrtee_len;
|
||||
extern u32 kvm_emulate_wrtee[];
|
||||
|
||||
static void kvm_patch_ins_wrteei(u32 *inst)
|
||||
static void kvm_patch_ins_wrtee(u32 *inst, u32 rt, int imm_one)
|
||||
{
|
||||
u32 *p;
|
||||
int distance_start;
|
||||
int distance_end;
|
||||
ulong next_inst;
|
||||
|
||||
p = kvm_alloc(kvm_emulate_wrteei_len * 4);
|
||||
p = kvm_alloc(kvm_emulate_wrtee_len * 4);
|
||||
if (!p)
|
||||
return;
|
||||
|
||||
/* Find out where we are and put everything there */
|
||||
distance_start = (ulong)p - (ulong)inst;
|
||||
next_inst = ((ulong)inst + 4);
|
||||
distance_end = next_inst - (ulong)&p[kvm_emulate_wrteei_branch_offs];
|
||||
distance_end = next_inst - (ulong)&p[kvm_emulate_wrtee_branch_offs];
|
||||
|
||||
/* Make sure we only write valid b instructions */
|
||||
if (distance_start > KVM_INST_B_MAX) {
|
||||
@@ -298,10 +297,65 @@ static void kvm_patch_ins_wrteei(u32 *inst)
|
||||
}
|
||||
|
||||
/* Modify the chunk to fit the invocation */
|
||||
memcpy(p, kvm_emulate_wrteei, kvm_emulate_wrteei_len * 4);
|
||||
p[kvm_emulate_wrteei_branch_offs] |= distance_end & KVM_INST_B_MASK;
|
||||
p[kvm_emulate_wrteei_ee_offs] |= (*inst & MSR_EE);
|
||||
flush_icache_range((ulong)p, (ulong)p + kvm_emulate_wrteei_len * 4);
|
||||
memcpy(p, kvm_emulate_wrtee, kvm_emulate_wrtee_len * 4);
|
||||
p[kvm_emulate_wrtee_branch_offs] |= distance_end & KVM_INST_B_MASK;
|
||||
|
||||
if (imm_one) {
|
||||
p[kvm_emulate_wrtee_reg_offs] =
|
||||
KVM_INST_LI | __PPC_RT(30) | MSR_EE;
|
||||
} else {
|
||||
/* Make clobbered registers work too */
|
||||
switch (get_rt(rt)) {
|
||||
case 30:
|
||||
kvm_patch_ins_ll(&p[kvm_emulate_wrtee_reg_offs],
|
||||
magic_var(scratch2), KVM_RT_30);
|
||||
break;
|
||||
case 31:
|
||||
kvm_patch_ins_ll(&p[kvm_emulate_wrtee_reg_offs],
|
||||
magic_var(scratch1), KVM_RT_30);
|
||||
break;
|
||||
default:
|
||||
p[kvm_emulate_wrtee_reg_offs] |= rt;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
p[kvm_emulate_wrtee_orig_ins_offs] = *inst;
|
||||
flush_icache_range((ulong)p, (ulong)p + kvm_emulate_wrtee_len * 4);
|
||||
|
||||
/* Patch the invocation */
|
||||
kvm_patch_ins_b(inst, distance_start);
|
||||
}
|
||||
|
||||
extern u32 kvm_emulate_wrteei_0_branch_offs;
|
||||
extern u32 kvm_emulate_wrteei_0_len;
|
||||
extern u32 kvm_emulate_wrteei_0[];
|
||||
|
||||
static void kvm_patch_ins_wrteei_0(u32 *inst)
|
||||
{
|
||||
u32 *p;
|
||||
int distance_start;
|
||||
int distance_end;
|
||||
ulong next_inst;
|
||||
|
||||
p = kvm_alloc(kvm_emulate_wrteei_0_len * 4);
|
||||
if (!p)
|
||||
return;
|
||||
|
||||
/* Find out where we are and put everything there */
|
||||
distance_start = (ulong)p - (ulong)inst;
|
||||
next_inst = ((ulong)inst + 4);
|
||||
distance_end = next_inst - (ulong)&p[kvm_emulate_wrteei_0_branch_offs];
|
||||
|
||||
/* Make sure we only write valid b instructions */
|
||||
if (distance_start > KVM_INST_B_MAX) {
|
||||
kvm_patching_worked = false;
|
||||
return;
|
||||
}
|
||||
|
||||
memcpy(p, kvm_emulate_wrteei_0, kvm_emulate_wrteei_0_len * 4);
|
||||
p[kvm_emulate_wrteei_0_branch_offs] |= distance_end & KVM_INST_B_MASK;
|
||||
flush_icache_range((ulong)p, (ulong)p + kvm_emulate_wrteei_0_len * 4);
|
||||
|
||||
/* Patch the invocation */
|
||||
kvm_patch_ins_b(inst, distance_start);
|
||||
@@ -380,56 +434,191 @@ static void kvm_check_ins(u32 *inst, u32 features)
|
||||
case KVM_INST_MFMSR:
|
||||
kvm_patch_ins_ld(inst, magic_var(msr), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_SPRG0:
|
||||
case KVM_INST_MFSPR(SPRN_SPRG0):
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg0), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_SPRG1:
|
||||
case KVM_INST_MFSPR(SPRN_SPRG1):
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg1), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_SPRG2:
|
||||
case KVM_INST_MFSPR(SPRN_SPRG2):
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg2), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_SPRG3:
|
||||
case KVM_INST_MFSPR(SPRN_SPRG3):
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg3), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_SRR0:
|
||||
case KVM_INST_MFSPR(SPRN_SRR0):
|
||||
kvm_patch_ins_ld(inst, magic_var(srr0), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_SRR1:
|
||||
case KVM_INST_MFSPR(SPRN_SRR1):
|
||||
kvm_patch_ins_ld(inst, magic_var(srr1), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_DAR:
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MFSPR(SPRN_DEAR):
|
||||
#else
|
||||
case KVM_INST_MFSPR(SPRN_DAR):
|
||||
#endif
|
||||
kvm_patch_ins_ld(inst, magic_var(dar), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR_DSISR:
|
||||
case KVM_INST_MFSPR(SPRN_DSISR):
|
||||
kvm_patch_ins_lwz(inst, magic_var(dsisr), inst_rt);
|
||||
break;
|
||||
|
||||
#ifdef CONFIG_PPC_BOOK3E_MMU
|
||||
case KVM_INST_MFSPR(SPRN_MAS0):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(mas0), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_MAS1):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(mas1), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_MAS2):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_ld(inst, magic_var(mas2), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_MAS3):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(mas7_3) + 4, inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_MAS4):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(mas4), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_MAS6):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(mas6), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_MAS7):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(mas7_3), inst_rt);
|
||||
break;
|
||||
#endif /* CONFIG_PPC_BOOK3E_MMU */
|
||||
|
||||
case KVM_INST_MFSPR(SPRN_SPRG4):
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MFSPR(SPRN_SPRG4R):
|
||||
#endif
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg4), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_SPRG5):
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MFSPR(SPRN_SPRG5R):
|
||||
#endif
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg5), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_SPRG6):
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MFSPR(SPRN_SPRG6R):
|
||||
#endif
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg6), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MFSPR(SPRN_SPRG7):
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MFSPR(SPRN_SPRG7R):
|
||||
#endif
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_ld(inst, magic_var(sprg7), inst_rt);
|
||||
break;
|
||||
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MFSPR(SPRN_ESR):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(esr), inst_rt);
|
||||
break;
|
||||
#endif
|
||||
|
||||
case KVM_INST_MFSPR(SPRN_PIR):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_lwz(inst, magic_var(pir), inst_rt);
|
||||
break;
|
||||
|
||||
|
||||
/* Stores */
|
||||
case KVM_INST_MTSPR_SPRG0:
|
||||
case KVM_INST_MTSPR(SPRN_SPRG0):
|
||||
kvm_patch_ins_std(inst, magic_var(sprg0), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_SPRG1:
|
||||
case KVM_INST_MTSPR(SPRN_SPRG1):
|
||||
kvm_patch_ins_std(inst, magic_var(sprg1), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_SPRG2:
|
||||
case KVM_INST_MTSPR(SPRN_SPRG2):
|
||||
kvm_patch_ins_std(inst, magic_var(sprg2), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_SPRG3:
|
||||
case KVM_INST_MTSPR(SPRN_SPRG3):
|
||||
kvm_patch_ins_std(inst, magic_var(sprg3), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_SRR0:
|
||||
case KVM_INST_MTSPR(SPRN_SRR0):
|
||||
kvm_patch_ins_std(inst, magic_var(srr0), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_SRR1:
|
||||
case KVM_INST_MTSPR(SPRN_SRR1):
|
||||
kvm_patch_ins_std(inst, magic_var(srr1), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_DAR:
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MTSPR(SPRN_DEAR):
|
||||
#else
|
||||
case KVM_INST_MTSPR(SPRN_DAR):
|
||||
#endif
|
||||
kvm_patch_ins_std(inst, magic_var(dar), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR_DSISR:
|
||||
case KVM_INST_MTSPR(SPRN_DSISR):
|
||||
kvm_patch_ins_stw(inst, magic_var(dsisr), inst_rt);
|
||||
break;
|
||||
#ifdef CONFIG_PPC_BOOK3E_MMU
|
||||
case KVM_INST_MTSPR(SPRN_MAS0):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(mas0), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_MAS1):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(mas1), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_MAS2):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_std(inst, magic_var(mas2), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_MAS3):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(mas7_3) + 4, inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_MAS4):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(mas4), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_MAS6):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(mas6), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_MAS7):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(mas7_3), inst_rt);
|
||||
break;
|
||||
#endif /* CONFIG_PPC_BOOK3E_MMU */
|
||||
|
||||
case KVM_INST_MTSPR(SPRN_SPRG4):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_std(inst, magic_var(sprg4), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_SPRG5):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_std(inst, magic_var(sprg5), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_SPRG6):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_std(inst, magic_var(sprg6), inst_rt);
|
||||
break;
|
||||
case KVM_INST_MTSPR(SPRN_SPRG7):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_std(inst, magic_var(sprg7), inst_rt);
|
||||
break;
|
||||
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_MTSPR(SPRN_ESR):
|
||||
if (features & KVM_MAGIC_FEAT_MAS0_TO_SPRG7)
|
||||
kvm_patch_ins_stw(inst, magic_var(esr), inst_rt);
|
||||
break;
|
||||
#endif
|
||||
|
||||
/* Nops */
|
||||
case KVM_INST_TLBSYNC:
|
||||
@@ -444,6 +633,11 @@ static void kvm_check_ins(u32 *inst, u32 features)
|
||||
case KVM_INST_MTMSRD_L0:
|
||||
kvm_patch_ins_mtmsr(inst, inst_rt);
|
||||
break;
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_WRTEE:
|
||||
kvm_patch_ins_wrtee(inst, inst_rt, 0);
|
||||
break;
|
||||
#endif
|
||||
}
|
||||
|
||||
switch (inst_no_rt & ~KVM_MASK_RB) {
|
||||
@@ -461,13 +655,19 @@ static void kvm_check_ins(u32 *inst, u32 features)
|
||||
switch (_inst) {
|
||||
#ifdef CONFIG_BOOKE
|
||||
case KVM_INST_WRTEEI_0:
|
||||
kvm_patch_ins_wrteei_0(inst);
|
||||
break;
|
||||
|
||||
case KVM_INST_WRTEEI_1:
|
||||
kvm_patch_ins_wrteei(inst);
|
||||
kvm_patch_ins_wrtee(inst, 0, 1);
|
||||
break;
|
||||
#endif
|
||||
}
|
||||
}
|
||||
|
||||
extern u32 kvm_template_start[];
|
||||
extern u32 kvm_template_end[];
|
||||
|
||||
static void kvm_use_magic_page(void)
|
||||
{
|
||||
u32 *p;
|
||||
@@ -488,8 +688,23 @@ static void kvm_use_magic_page(void)
|
||||
start = (void*)_stext;
|
||||
end = (void*)_etext;
|
||||
|
||||
for (p = start; p < end; p++)
|
||||
/*
|
||||
* Being interrupted in the middle of patching would
|
||||
* be bad for SPRG4-7, which KVM can't keep in sync
|
||||
* with emulated accesses because reads don't trap.
|
||||
*/
|
||||
local_irq_disable();
|
||||
|
||||
for (p = start; p < end; p++) {
|
||||
/* Avoid patching the template code */
|
||||
if (p >= kvm_template_start && p < kvm_template_end) {
|
||||
p = kvm_template_end - 1;
|
||||
continue;
|
||||
}
|
||||
kvm_check_ins(p, features);
|
||||
}
|
||||
|
||||
local_irq_enable();
|
||||
|
||||
printk(KERN_INFO "KVM: Live patching for a fast VM %s\n",
|
||||
kvm_patching_worked ? "worked" : "failed");
|
||||
|
@@ -13,6 +13,7 @@
|
||||
* Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
||||
*
|
||||
* Copyright SUSE Linux Products GmbH 2010
|
||||
* Copyright 2010-2011 Freescale Semiconductor, Inc.
|
||||
*
|
||||
* Authors: Alexander Graf <agraf@suse.de>
|
||||
*/
|
||||
@@ -65,6 +66,9 @@ kvm_hypercall_start:
|
||||
shared->critical == r1 and r2 is always != r1 */ \
|
||||
STL64(r2, KVM_MAGIC_PAGE + KVM_MAGIC_CRITICAL, 0);
|
||||
|
||||
.global kvm_template_start
|
||||
kvm_template_start:
|
||||
|
||||
.global kvm_emulate_mtmsrd
|
||||
kvm_emulate_mtmsrd:
|
||||
|
||||
@@ -167,6 +171,9 @@ maybe_stay_in_guest:
|
||||
kvm_emulate_mtmsr_reg2:
|
||||
ori r30, r0, 0
|
||||
|
||||
/* Put MSR into magic page because we don't call mtmsr */
|
||||
STL64(r30, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
||||
|
||||
/* Check if we have to fetch an interrupt */
|
||||
lwz r31, (KVM_MAGIC_PAGE + KVM_MAGIC_INT)(0)
|
||||
cmpwi r31, 0
|
||||
@@ -174,15 +181,10 @@ kvm_emulate_mtmsr_reg2:
|
||||
|
||||
/* Check if we may trigger an interrupt */
|
||||
andi. r31, r30, MSR_EE
|
||||
beq no_mtmsr
|
||||
|
||||
b do_mtmsr
|
||||
bne do_mtmsr
|
||||
|
||||
no_mtmsr:
|
||||
|
||||
/* Put MSR into magic page because we don't call mtmsr */
|
||||
STL64(r30, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
||||
|
||||
SCRATCH_RESTORE
|
||||
|
||||
/* Go back to caller */
|
||||
@@ -210,24 +212,80 @@ kvm_emulate_mtmsr_orig_ins_offs:
|
||||
kvm_emulate_mtmsr_len:
|
||||
.long (kvm_emulate_mtmsr_end - kvm_emulate_mtmsr) / 4
|
||||
|
||||
|
||||
|
||||
.global kvm_emulate_wrteei
|
||||
kvm_emulate_wrteei:
|
||||
/* also used for wrteei 1 */
|
||||
.global kvm_emulate_wrtee
|
||||
kvm_emulate_wrtee:
|
||||
|
||||
SCRATCH_SAVE
|
||||
|
||||
/* Fetch old MSR in r31 */
|
||||
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
||||
|
||||
/* Remove MSR_EE from old MSR */
|
||||
li r30, 0
|
||||
ori r30, r30, MSR_EE
|
||||
andc r31, r31, r30
|
||||
/* Insert new MSR[EE] */
|
||||
kvm_emulate_wrtee_reg:
|
||||
ori r30, r0, 0
|
||||
rlwimi r31, r30, 0, MSR_EE
|
||||
|
||||
/* OR new MSR_EE onto the old MSR */
|
||||
kvm_emulate_wrteei_ee:
|
||||
ori r31, r31, 0
|
||||
/*
|
||||
* If MSR[EE] is now set, check for a pending interrupt.
|
||||
* We could skip this if MSR[EE] was already on, but that
|
||||
* should be rare, so don't bother.
|
||||
*/
|
||||
andi. r30, r30, MSR_EE
|
||||
|
||||
/* Put MSR into magic page because we don't call wrtee */
|
||||
STL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
||||
|
||||
beq no_wrtee
|
||||
|
||||
/* Check if we have to fetch an interrupt */
|
||||
lwz r30, (KVM_MAGIC_PAGE + KVM_MAGIC_INT)(0)
|
||||
cmpwi r30, 0
|
||||
bne do_wrtee
|
||||
|
||||
no_wrtee:
|
||||
SCRATCH_RESTORE
|
||||
|
||||
/* Go back to caller */
|
||||
kvm_emulate_wrtee_branch:
|
||||
b .
|
||||
|
||||
do_wrtee:
|
||||
SCRATCH_RESTORE
|
||||
|
||||
/* Just fire off the wrtee if it's critical */
|
||||
kvm_emulate_wrtee_orig_ins:
|
||||
wrtee r0
|
||||
|
||||
b kvm_emulate_wrtee_branch
|
||||
|
||||
kvm_emulate_wrtee_end:
|
||||
|
||||
.global kvm_emulate_wrtee_branch_offs
|
||||
kvm_emulate_wrtee_branch_offs:
|
||||
.long (kvm_emulate_wrtee_branch - kvm_emulate_wrtee) / 4
|
||||
|
||||
.global kvm_emulate_wrtee_reg_offs
|
||||
kvm_emulate_wrtee_reg_offs:
|
||||
.long (kvm_emulate_wrtee_reg - kvm_emulate_wrtee) / 4
|
||||
|
||||
.global kvm_emulate_wrtee_orig_ins_offs
|
||||
kvm_emulate_wrtee_orig_ins_offs:
|
||||
.long (kvm_emulate_wrtee_orig_ins - kvm_emulate_wrtee) / 4
|
||||
|
||||
.global kvm_emulate_wrtee_len
|
||||
kvm_emulate_wrtee_len:
|
||||
.long (kvm_emulate_wrtee_end - kvm_emulate_wrtee) / 4
|
||||
|
||||
.global kvm_emulate_wrteei_0
|
||||
kvm_emulate_wrteei_0:
|
||||
SCRATCH_SAVE
|
||||
|
||||
/* Fetch old MSR in r31 */
|
||||
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
||||
|
||||
/* Remove MSR_EE from old MSR */
|
||||
rlwinm r31, r31, 0, ~MSR_EE
|
||||
|
||||
/* Write new MSR value back */
|
||||
STL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
||||
@@ -235,22 +293,17 @@ kvm_emulate_wrteei_ee:
|
||||
SCRATCH_RESTORE
|
||||
|
||||
/* Go back to caller */
|
||||
kvm_emulate_wrteei_branch:
|
||||
kvm_emulate_wrteei_0_branch:
|
||||
b .
|
||||
kvm_emulate_wrteei_end:
|
||||
kvm_emulate_wrteei_0_end:
|
||||
|
||||
.global kvm_emulate_wrteei_branch_offs
|
||||
kvm_emulate_wrteei_branch_offs:
|
||||
.long (kvm_emulate_wrteei_branch - kvm_emulate_wrteei) / 4
|
||||
|
||||
.global kvm_emulate_wrteei_ee_offs
|
||||
kvm_emulate_wrteei_ee_offs:
|
||||
.long (kvm_emulate_wrteei_ee - kvm_emulate_wrteei) / 4
|
||||
|
||||
.global kvm_emulate_wrteei_len
|
||||
kvm_emulate_wrteei_len:
|
||||
.long (kvm_emulate_wrteei_end - kvm_emulate_wrteei) / 4
|
||||
.global kvm_emulate_wrteei_0_branch_offs
|
||||
kvm_emulate_wrteei_0_branch_offs:
|
||||
.long (kvm_emulate_wrteei_0_branch - kvm_emulate_wrteei_0) / 4
|
||||
|
||||
.global kvm_emulate_wrteei_0_len
|
||||
kvm_emulate_wrteei_0_len:
|
||||
.long (kvm_emulate_wrteei_0_end - kvm_emulate_wrteei_0) / 4
|
||||
|
||||
.global kvm_emulate_mtsrin
|
||||
kvm_emulate_mtsrin:
|
||||
@@ -300,3 +353,6 @@ kvm_emulate_mtsrin_orig_ins_offs:
|
||||
.global kvm_emulate_mtsrin_len
|
||||
kvm_emulate_mtsrin_len:
|
||||
.long (kvm_emulate_mtsrin_end - kvm_emulate_mtsrin) / 4
|
||||
|
||||
.global kvm_template_end
|
||||
kvm_template_end:
|
||||
|
@@ -598,7 +598,7 @@ void __init setup_arch(char **cmdline_p)
|
||||
/* Initialize the MMU context management stuff */
|
||||
mmu_context_init();
|
||||
|
||||
kvm_rma_init();
|
||||
kvm_linear_init();
|
||||
|
||||
ppc64_boot_msg(0x15, "Setup Done");
|
||||
}
|
||||
|
Reference in New Issue
Block a user