forked from Minki/linux
caff52dc0b
Some of the templates used for KVM patching are only used on certain platforms, but currently they are always built-in, fix that. Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/20190911115746.12433-4-mpe@ellerman.id.au
355 lines
8.0 KiB
ArmAsm
355 lines
8.0 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
*
|
|
* Copyright SUSE Linux Products GmbH 2010
|
|
* Copyright 2010-2011 Freescale Semiconductor, Inc.
|
|
*
|
|
* Authors: Alexander Graf <agraf@suse.de>
|
|
*/
|
|
|
|
#include <asm/ppc_asm.h>
|
|
#include <asm/kvm_asm.h>
|
|
#include <asm/reg.h>
|
|
#include <asm/page.h>
|
|
#include <asm/asm-offsets.h>
|
|
#include <asm/asm-compat.h>
|
|
|
|
#define KVM_MAGIC_PAGE (-4096)
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define LL64(reg, offs, reg2) ld reg, (offs)(reg2)
|
|
#define STL64(reg, offs, reg2) std reg, (offs)(reg2)
|
|
#else
|
|
#define LL64(reg, offs, reg2) lwz reg, (offs + 4)(reg2)
|
|
#define STL64(reg, offs, reg2) stw reg, (offs + 4)(reg2)
|
|
#endif
|
|
|
|
#define SCRATCH_SAVE \
|
|
/* Enable critical section. We are critical if \
|
|
shared->critical == r1 */ \
|
|
STL64(r1, KVM_MAGIC_PAGE + KVM_MAGIC_CRITICAL, 0); \
|
|
\
|
|
/* Save state */ \
|
|
PPC_STL r31, (KVM_MAGIC_PAGE + KVM_MAGIC_SCRATCH1)(0); \
|
|
PPC_STL r30, (KVM_MAGIC_PAGE + KVM_MAGIC_SCRATCH2)(0); \
|
|
mfcr r31; \
|
|
stw r31, (KVM_MAGIC_PAGE + KVM_MAGIC_SCRATCH3)(0);
|
|
|
|
#define SCRATCH_RESTORE \
|
|
/* Restore state */ \
|
|
PPC_LL r31, (KVM_MAGIC_PAGE + KVM_MAGIC_SCRATCH1)(0); \
|
|
lwz r30, (KVM_MAGIC_PAGE + KVM_MAGIC_SCRATCH3)(0); \
|
|
mtcr r30; \
|
|
PPC_LL r30, (KVM_MAGIC_PAGE + KVM_MAGIC_SCRATCH2)(0); \
|
|
\
|
|
/* Disable critical section. We are critical if \
|
|
shared->critical == r1 and r2 is always != r1 */ \
|
|
STL64(r2, KVM_MAGIC_PAGE + KVM_MAGIC_CRITICAL, 0);
|
|
|
|
.global kvm_template_start
|
|
kvm_template_start:
|
|
|
|
.global kvm_emulate_mtmsrd
|
|
kvm_emulate_mtmsrd:
|
|
|
|
SCRATCH_SAVE
|
|
|
|
/* Put MSR & ~(MSR_EE|MSR_RI) in r31 */
|
|
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
lis r30, (~(MSR_EE | MSR_RI))@h
|
|
ori r30, r30, (~(MSR_EE | MSR_RI))@l
|
|
and r31, r31, r30
|
|
|
|
/* OR the register's (MSR_EE|MSR_RI) on MSR */
|
|
kvm_emulate_mtmsrd_reg:
|
|
ori r30, r0, 0
|
|
andi. r30, r30, (MSR_EE|MSR_RI)
|
|
or r31, r31, r30
|
|
|
|
/* Put MSR back into magic page */
|
|
STL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
/* Check if we have to fetch an interrupt */
|
|
lwz r31, (KVM_MAGIC_PAGE + KVM_MAGIC_INT)(0)
|
|
cmpwi r31, 0
|
|
beq+ no_check
|
|
|
|
/* Check if we may trigger an interrupt */
|
|
andi. r30, r30, MSR_EE
|
|
beq no_check
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
/* Nag hypervisor */
|
|
kvm_emulate_mtmsrd_orig_ins:
|
|
tlbsync
|
|
|
|
b kvm_emulate_mtmsrd_branch
|
|
|
|
no_check:
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
/* Go back to caller */
|
|
kvm_emulate_mtmsrd_branch:
|
|
b .
|
|
kvm_emulate_mtmsrd_end:
|
|
|
|
.global kvm_emulate_mtmsrd_branch_offs
|
|
kvm_emulate_mtmsrd_branch_offs:
|
|
.long (kvm_emulate_mtmsrd_branch - kvm_emulate_mtmsrd) / 4
|
|
|
|
.global kvm_emulate_mtmsrd_reg_offs
|
|
kvm_emulate_mtmsrd_reg_offs:
|
|
.long (kvm_emulate_mtmsrd_reg - kvm_emulate_mtmsrd) / 4
|
|
|
|
.global kvm_emulate_mtmsrd_orig_ins_offs
|
|
kvm_emulate_mtmsrd_orig_ins_offs:
|
|
.long (kvm_emulate_mtmsrd_orig_ins - kvm_emulate_mtmsrd) / 4
|
|
|
|
.global kvm_emulate_mtmsrd_len
|
|
kvm_emulate_mtmsrd_len:
|
|
.long (kvm_emulate_mtmsrd_end - kvm_emulate_mtmsrd) / 4
|
|
|
|
|
|
#define MSR_SAFE_BITS (MSR_EE | MSR_RI)
|
|
#define MSR_CRITICAL_BITS ~MSR_SAFE_BITS
|
|
|
|
.global kvm_emulate_mtmsr
|
|
kvm_emulate_mtmsr:
|
|
|
|
SCRATCH_SAVE
|
|
|
|
/* Fetch old MSR in r31 */
|
|
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
/* Find the changed bits between old and new MSR */
|
|
kvm_emulate_mtmsr_reg1:
|
|
ori r30, r0, 0
|
|
xor r31, r30, r31
|
|
|
|
/* Check if we need to really do mtmsr */
|
|
LOAD_REG_IMMEDIATE(r30, MSR_CRITICAL_BITS)
|
|
and. r31, r31, r30
|
|
|
|
/* No critical bits changed? Maybe we can stay in the guest. */
|
|
beq maybe_stay_in_guest
|
|
|
|
do_mtmsr:
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
/* Just fire off the mtmsr if it's critical */
|
|
kvm_emulate_mtmsr_orig_ins:
|
|
mtmsr r0
|
|
|
|
b kvm_emulate_mtmsr_branch
|
|
|
|
maybe_stay_in_guest:
|
|
|
|
/* Get the target register in r30 */
|
|
kvm_emulate_mtmsr_reg2:
|
|
ori r30, r0, 0
|
|
|
|
/* Put MSR into magic page because we don't call mtmsr */
|
|
STL64(r30, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
/* Check if we have to fetch an interrupt */
|
|
lwz r31, (KVM_MAGIC_PAGE + KVM_MAGIC_INT)(0)
|
|
cmpwi r31, 0
|
|
beq+ no_mtmsr
|
|
|
|
/* Check if we may trigger an interrupt */
|
|
andi. r31, r30, MSR_EE
|
|
bne do_mtmsr
|
|
|
|
no_mtmsr:
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
/* Go back to caller */
|
|
kvm_emulate_mtmsr_branch:
|
|
b .
|
|
kvm_emulate_mtmsr_end:
|
|
|
|
.global kvm_emulate_mtmsr_branch_offs
|
|
kvm_emulate_mtmsr_branch_offs:
|
|
.long (kvm_emulate_mtmsr_branch - kvm_emulate_mtmsr) / 4
|
|
|
|
.global kvm_emulate_mtmsr_reg1_offs
|
|
kvm_emulate_mtmsr_reg1_offs:
|
|
.long (kvm_emulate_mtmsr_reg1 - kvm_emulate_mtmsr) / 4
|
|
|
|
.global kvm_emulate_mtmsr_reg2_offs
|
|
kvm_emulate_mtmsr_reg2_offs:
|
|
.long (kvm_emulate_mtmsr_reg2 - kvm_emulate_mtmsr) / 4
|
|
|
|
.global kvm_emulate_mtmsr_orig_ins_offs
|
|
kvm_emulate_mtmsr_orig_ins_offs:
|
|
.long (kvm_emulate_mtmsr_orig_ins - kvm_emulate_mtmsr) / 4
|
|
|
|
.global kvm_emulate_mtmsr_len
|
|
kvm_emulate_mtmsr_len:
|
|
.long (kvm_emulate_mtmsr_end - kvm_emulate_mtmsr) / 4
|
|
|
|
#ifdef CONFIG_BOOKE
|
|
|
|
/* also used for wrteei 1 */
|
|
.global kvm_emulate_wrtee
|
|
kvm_emulate_wrtee:
|
|
|
|
SCRATCH_SAVE
|
|
|
|
/* Fetch old MSR in r31 */
|
|
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
/* Insert new MSR[EE] */
|
|
kvm_emulate_wrtee_reg:
|
|
ori r30, r0, 0
|
|
rlwimi r31, r30, 0, MSR_EE
|
|
|
|
/*
|
|
* If MSR[EE] is now set, check for a pending interrupt.
|
|
* We could skip this if MSR[EE] was already on, but that
|
|
* should be rare, so don't bother.
|
|
*/
|
|
andi. r30, r30, MSR_EE
|
|
|
|
/* Put MSR into magic page because we don't call wrtee */
|
|
STL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
beq no_wrtee
|
|
|
|
/* Check if we have to fetch an interrupt */
|
|
lwz r30, (KVM_MAGIC_PAGE + KVM_MAGIC_INT)(0)
|
|
cmpwi r30, 0
|
|
bne do_wrtee
|
|
|
|
no_wrtee:
|
|
SCRATCH_RESTORE
|
|
|
|
/* Go back to caller */
|
|
kvm_emulate_wrtee_branch:
|
|
b .
|
|
|
|
do_wrtee:
|
|
SCRATCH_RESTORE
|
|
|
|
/* Just fire off the wrtee if it's critical */
|
|
kvm_emulate_wrtee_orig_ins:
|
|
wrtee r0
|
|
|
|
b kvm_emulate_wrtee_branch
|
|
|
|
kvm_emulate_wrtee_end:
|
|
|
|
.global kvm_emulate_wrtee_branch_offs
|
|
kvm_emulate_wrtee_branch_offs:
|
|
.long (kvm_emulate_wrtee_branch - kvm_emulate_wrtee) / 4
|
|
|
|
.global kvm_emulate_wrtee_reg_offs
|
|
kvm_emulate_wrtee_reg_offs:
|
|
.long (kvm_emulate_wrtee_reg - kvm_emulate_wrtee) / 4
|
|
|
|
.global kvm_emulate_wrtee_orig_ins_offs
|
|
kvm_emulate_wrtee_orig_ins_offs:
|
|
.long (kvm_emulate_wrtee_orig_ins - kvm_emulate_wrtee) / 4
|
|
|
|
.global kvm_emulate_wrtee_len
|
|
kvm_emulate_wrtee_len:
|
|
.long (kvm_emulate_wrtee_end - kvm_emulate_wrtee) / 4
|
|
|
|
.global kvm_emulate_wrteei_0
|
|
kvm_emulate_wrteei_0:
|
|
SCRATCH_SAVE
|
|
|
|
/* Fetch old MSR in r31 */
|
|
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
/* Remove MSR_EE from old MSR */
|
|
rlwinm r31, r31, 0, ~MSR_EE
|
|
|
|
/* Write new MSR value back */
|
|
STL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
/* Go back to caller */
|
|
kvm_emulate_wrteei_0_branch:
|
|
b .
|
|
kvm_emulate_wrteei_0_end:
|
|
|
|
.global kvm_emulate_wrteei_0_branch_offs
|
|
kvm_emulate_wrteei_0_branch_offs:
|
|
.long (kvm_emulate_wrteei_0_branch - kvm_emulate_wrteei_0) / 4
|
|
|
|
.global kvm_emulate_wrteei_0_len
|
|
kvm_emulate_wrteei_0_len:
|
|
.long (kvm_emulate_wrteei_0_end - kvm_emulate_wrteei_0) / 4
|
|
|
|
#endif /* CONFIG_BOOKE */
|
|
|
|
#ifdef CONFIG_PPC_BOOK3S_32
|
|
|
|
.global kvm_emulate_mtsrin
|
|
kvm_emulate_mtsrin:
|
|
|
|
SCRATCH_SAVE
|
|
|
|
LL64(r31, KVM_MAGIC_PAGE + KVM_MAGIC_MSR, 0)
|
|
andi. r31, r31, MSR_DR | MSR_IR
|
|
beq kvm_emulate_mtsrin_reg1
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
kvm_emulate_mtsrin_orig_ins:
|
|
nop
|
|
b kvm_emulate_mtsrin_branch
|
|
|
|
kvm_emulate_mtsrin_reg1:
|
|
/* rX >> 26 */
|
|
rlwinm r30,r0,6,26,29
|
|
|
|
kvm_emulate_mtsrin_reg2:
|
|
stw r0, (KVM_MAGIC_PAGE + KVM_MAGIC_SR)(r30)
|
|
|
|
SCRATCH_RESTORE
|
|
|
|
/* Go back to caller */
|
|
kvm_emulate_mtsrin_branch:
|
|
b .
|
|
kvm_emulate_mtsrin_end:
|
|
|
|
.global kvm_emulate_mtsrin_branch_offs
|
|
kvm_emulate_mtsrin_branch_offs:
|
|
.long (kvm_emulate_mtsrin_branch - kvm_emulate_mtsrin) / 4
|
|
|
|
.global kvm_emulate_mtsrin_reg1_offs
|
|
kvm_emulate_mtsrin_reg1_offs:
|
|
.long (kvm_emulate_mtsrin_reg1 - kvm_emulate_mtsrin) / 4
|
|
|
|
.global kvm_emulate_mtsrin_reg2_offs
|
|
kvm_emulate_mtsrin_reg2_offs:
|
|
.long (kvm_emulate_mtsrin_reg2 - kvm_emulate_mtsrin) / 4
|
|
|
|
.global kvm_emulate_mtsrin_orig_ins_offs
|
|
kvm_emulate_mtsrin_orig_ins_offs:
|
|
.long (kvm_emulate_mtsrin_orig_ins - kvm_emulate_mtsrin) / 4
|
|
|
|
.global kvm_emulate_mtsrin_len
|
|
kvm_emulate_mtsrin_len:
|
|
.long (kvm_emulate_mtsrin_end - kvm_emulate_mtsrin) / 4
|
|
|
|
#endif /* CONFIG_PPC_BOOK3S_32 */
|
|
|
|
.balign 4
|
|
.global kvm_tmp
|
|
kvm_tmp:
|
|
.space (64 * 1024)
|
|
|
|
.global kvm_tmp_end
|
|
kvm_tmp_end:
|
|
|
|
.global kvm_template_end
|
|
kvm_template_end:
|