mirror of
https://github.com/Motorhead1991/qemu.git
synced 2025-08-15 22:21:57 -06:00

Move the global function name to a hook on TCGCPUOps. Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
247 lines
7.3 KiB
C
247 lines
7.3 KiB
C
/*
|
|
* i386 TCG cpu class initialization
|
|
*
|
|
* Copyright (c) 2003 Fabrice Bellard
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "cpu.h"
|
|
#include "helper-tcg.h"
|
|
#include "qemu/accel.h"
|
|
#include "accel/accel-cpu-target.h"
|
|
#include "exec/translation-block.h"
|
|
#include "exec/target_page.h"
|
|
#include "accel/tcg/cpu-ops.h"
|
|
#include "tcg-cpu.h"
|
|
|
|
/* Frob eflags into and out of the CPU temporary format. */
|
|
|
|
static void x86_cpu_exec_enter(CPUState *cs)
|
|
{
|
|
X86CPU *cpu = X86_CPU(cs);
|
|
CPUX86State *env = &cpu->env;
|
|
|
|
CC_SRC = env->eflags & (CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
|
|
env->df = 1 - (2 * ((env->eflags >> 10) & 1));
|
|
CC_OP = CC_OP_EFLAGS;
|
|
env->eflags &= ~(DF_MASK | CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
|
|
}
|
|
|
|
static void x86_cpu_exec_exit(CPUState *cs)
|
|
{
|
|
X86CPU *cpu = X86_CPU(cs);
|
|
CPUX86State *env = &cpu->env;
|
|
|
|
env->eflags = cpu_compute_eflags(env);
|
|
}
|
|
|
|
static TCGTBCPUState x86_get_tb_cpu_state(CPUState *cs)
|
|
{
|
|
CPUX86State *env = cpu_env(cs);
|
|
uint32_t flags, cs_base;
|
|
vaddr pc;
|
|
|
|
flags = env->hflags |
|
|
(env->eflags & (IOPL_MASK | TF_MASK | RF_MASK | VM_MASK | AC_MASK));
|
|
if (env->hflags & HF_CS64_MASK) {
|
|
cs_base = 0;
|
|
pc = env->eip;
|
|
} else {
|
|
cs_base = env->segs[R_CS].base;
|
|
pc = (uint32_t)(cs_base + env->eip);
|
|
}
|
|
|
|
return (TCGTBCPUState){ .pc = pc, .flags = flags, .cs_base = cs_base };
|
|
}
|
|
|
|
static void x86_cpu_synchronize_from_tb(CPUState *cs,
|
|
const TranslationBlock *tb)
|
|
{
|
|
/* The instruction pointer is always up to date with CF_PCREL. */
|
|
if (!(tb_cflags(tb) & CF_PCREL)) {
|
|
CPUX86State *env = cpu_env(cs);
|
|
|
|
if (tb->flags & HF_CS64_MASK) {
|
|
env->eip = tb->pc;
|
|
} else {
|
|
env->eip = (uint32_t)(tb->pc - tb->cs_base);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void x86_restore_state_to_opc(CPUState *cs,
|
|
const TranslationBlock *tb,
|
|
const uint64_t *data)
|
|
{
|
|
X86CPU *cpu = X86_CPU(cs);
|
|
CPUX86State *env = &cpu->env;
|
|
int cc_op = data[1];
|
|
uint64_t new_pc;
|
|
|
|
if (tb_cflags(tb) & CF_PCREL) {
|
|
/*
|
|
* data[0] in PC-relative TBs is also a linear address, i.e. an address with
|
|
* the CS base added, because it is not guaranteed that EIP bits 12 and higher
|
|
* stay the same across the translation block. Add the CS base back before
|
|
* replacing the low bits, and subtract it below just like for !CF_PCREL.
|
|
*/
|
|
uint64_t pc = env->eip + tb->cs_base;
|
|
new_pc = (pc & TARGET_PAGE_MASK) | data[0];
|
|
} else {
|
|
new_pc = data[0];
|
|
}
|
|
if (tb->flags & HF_CS64_MASK) {
|
|
env->eip = new_pc;
|
|
} else {
|
|
env->eip = (uint32_t)(new_pc - tb->cs_base);
|
|
}
|
|
|
|
if (cc_op != CC_OP_DYNAMIC) {
|
|
env->cc_op = cc_op;
|
|
}
|
|
}
|
|
|
|
int x86_mmu_index_pl(CPUX86State *env, unsigned pl)
|
|
{
|
|
int mmu_index_32 = (env->hflags & HF_CS64_MASK) ? 0 : 1;
|
|
int mmu_index_base =
|
|
pl == 3 ? MMU_USER64_IDX :
|
|
!(env->hflags & HF_SMAP_MASK) ? MMU_KNOSMAP64_IDX :
|
|
(env->eflags & AC_MASK) ? MMU_KNOSMAP64_IDX : MMU_KSMAP64_IDX;
|
|
|
|
return mmu_index_base + mmu_index_32;
|
|
}
|
|
|
|
static int x86_cpu_mmu_index(CPUState *cs, bool ifetch)
|
|
{
|
|
CPUX86State *env = cpu_env(cs);
|
|
return x86_mmu_index_pl(env, env->hflags & HF_CPL_MASK);
|
|
}
|
|
|
|
#ifndef CONFIG_USER_ONLY
|
|
static bool x86_debug_check_breakpoint(CPUState *cs)
|
|
{
|
|
X86CPU *cpu = X86_CPU(cs);
|
|
CPUX86State *env = &cpu->env;
|
|
|
|
/* RF disables all architectural breakpoints. */
|
|
return !(env->eflags & RF_MASK);
|
|
}
|
|
|
|
static void x86_cpu_exec_reset(CPUState *cs)
|
|
{
|
|
CPUArchState *env = cpu_env(cs);
|
|
|
|
cpu_svm_check_intercept_param(env, SVM_EXIT_INIT, 0, 0);
|
|
do_cpu_init(env_archcpu(env));
|
|
cs->exception_index = EXCP_HALTED;
|
|
}
|
|
#endif
|
|
|
|
const TCGCPUOps x86_tcg_ops = {
|
|
.mttcg_supported = true,
|
|
.precise_smc = true,
|
|
/*
|
|
* The x86 has a strong memory model with some store-after-load re-ordering
|
|
*/
|
|
.guest_default_memory_order = TCG_MO_ALL & ~TCG_MO_ST_LD,
|
|
.initialize = tcg_x86_init,
|
|
.translate_code = x86_translate_code,
|
|
.get_tb_cpu_state = x86_get_tb_cpu_state,
|
|
.synchronize_from_tb = x86_cpu_synchronize_from_tb,
|
|
.restore_state_to_opc = x86_restore_state_to_opc,
|
|
.mmu_index = x86_cpu_mmu_index,
|
|
.cpu_exec_enter = x86_cpu_exec_enter,
|
|
.cpu_exec_exit = x86_cpu_exec_exit,
|
|
#ifdef CONFIG_USER_ONLY
|
|
.fake_user_interrupt = x86_cpu_do_interrupt,
|
|
.record_sigsegv = x86_cpu_record_sigsegv,
|
|
.record_sigbus = x86_cpu_record_sigbus,
|
|
#else
|
|
.tlb_fill = x86_cpu_tlb_fill,
|
|
.do_interrupt = x86_cpu_do_interrupt,
|
|
.cpu_exec_halt = x86_cpu_exec_halt,
|
|
.cpu_exec_interrupt = x86_cpu_exec_interrupt,
|
|
.cpu_exec_reset = x86_cpu_exec_reset,
|
|
.do_unaligned_access = x86_cpu_do_unaligned_access,
|
|
.debug_excp_handler = breakpoint_handler,
|
|
.debug_check_breakpoint = x86_debug_check_breakpoint,
|
|
.need_replay_interrupt = x86_need_replay_interrupt,
|
|
#endif /* !CONFIG_USER_ONLY */
|
|
};
|
|
|
|
static void x86_tcg_cpu_xsave_init(void)
|
|
{
|
|
#define XO(bit, field) \
|
|
x86_ext_save_areas[bit].offset = offsetof(X86XSaveArea, field);
|
|
|
|
XO(XSTATE_FP_BIT, legacy);
|
|
XO(XSTATE_SSE_BIT, legacy);
|
|
XO(XSTATE_YMM_BIT, avx_state);
|
|
XO(XSTATE_BNDREGS_BIT, bndreg_state);
|
|
XO(XSTATE_BNDCSR_BIT, bndcsr_state);
|
|
XO(XSTATE_OPMASK_BIT, opmask_state);
|
|
XO(XSTATE_ZMM_Hi256_BIT, zmm_hi256_state);
|
|
XO(XSTATE_Hi16_ZMM_BIT, hi16_zmm_state);
|
|
XO(XSTATE_PKRU_BIT, pkru_state);
|
|
|
|
#undef XO
|
|
}
|
|
|
|
/*
|
|
* TCG-specific defaults that override cpudef models when using TCG.
|
|
* Only for builtin_x86_defs models initialized with x86_register_cpudef_types.
|
|
*/
|
|
static PropValue x86_tcg_default_props[] = {
|
|
{ "vme", "off" },
|
|
{ NULL, NULL },
|
|
};
|
|
|
|
static void x86_tcg_cpu_instance_init(CPUState *cs)
|
|
{
|
|
X86CPU *cpu = X86_CPU(cs);
|
|
X86CPUClass *xcc = X86_CPU_GET_CLASS(cpu);
|
|
|
|
if (xcc->model) {
|
|
/* Special cases not set in the X86CPUDefinition structs: */
|
|
x86_cpu_apply_props(cpu, x86_tcg_default_props);
|
|
}
|
|
|
|
x86_tcg_cpu_xsave_init();
|
|
}
|
|
|
|
static void x86_tcg_cpu_accel_class_init(ObjectClass *oc, const void *data)
|
|
{
|
|
AccelCPUClass *acc = ACCEL_CPU_CLASS(oc);
|
|
|
|
#ifndef CONFIG_USER_ONLY
|
|
acc->cpu_target_realize = tcg_cpu_realizefn;
|
|
#endif /* CONFIG_USER_ONLY */
|
|
|
|
acc->cpu_instance_init = x86_tcg_cpu_instance_init;
|
|
}
|
|
static const TypeInfo x86_tcg_cpu_accel_type_info = {
|
|
.name = ACCEL_CPU_NAME("tcg"),
|
|
|
|
.parent = TYPE_ACCEL_CPU,
|
|
.class_init = x86_tcg_cpu_accel_class_init,
|
|
.abstract = true,
|
|
};
|
|
static void x86_tcg_cpu_accel_register_types(void)
|
|
{
|
|
type_register_static(&x86_tcg_cpu_accel_type_info);
|
|
}
|
|
type_init(x86_tcg_cpu_accel_register_types);
|