1
2
3
4
5
6
7
8
9
10
11
12#ifndef _ASM_X86_KVM_X86_EMULATE_H
13#define _ASM_X86_KVM_X86_EMULATE_H
14
15#include <asm/desc_defs.h>
16#include "fpu.h"
17
18struct x86_emulate_ctxt;
19enum x86_intercept;
20enum x86_intercept_stage;
21
22struct x86_exception {
23 u8 vector;
24 bool error_code_valid;
25 u16 error_code;
26 bool nested_page_fault;
27 u64 address;
28 u8 async_page_fault;
29};
30
31
32
33
34
35
36struct x86_instruction_info {
37 u8 intercept;
38 u8 rep_prefix;
39 u8 modrm_mod;
40 u8 modrm_reg;
41 u8 modrm_rm;
42 u64 src_val;
43 u64 dst_val;
44 u8 src_bytes;
45 u8 dst_bytes;
46 u8 ad_bytes;
47 u64 next_rip;
48};
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81#define X86EMUL_CONTINUE 0
82
83#define X86EMUL_UNHANDLEABLE 1
84
85#define X86EMUL_PROPAGATE_FAULT 2
86#define X86EMUL_RETRY_INSTR 3
87#define X86EMUL_CMPXCHG_FAILED 4
88#define X86EMUL_IO_NEEDED 5
89#define X86EMUL_INTERCEPTED 6
90
91struct x86_emulate_ops {
92
93
94
95
96
97 ulong (*read_gpr)(struct x86_emulate_ctxt *ctxt, unsigned reg);
98
99
100
101
102
103
104 void (*write_gpr)(struct x86_emulate_ctxt *ctxt, unsigned reg, ulong val);
105
106
107
108
109
110
111
112
113 int (*read_std)(struct x86_emulate_ctxt *ctxt,
114 unsigned long addr, void *val,
115 unsigned int bytes,
116 struct x86_exception *fault, bool system);
117
118
119
120
121
122
123
124
125 int (*read_phys)(struct x86_emulate_ctxt *ctxt, unsigned long addr,
126 void *val, unsigned int bytes);
127
128
129
130
131
132
133
134
135
136 int (*write_std)(struct x86_emulate_ctxt *ctxt,
137 unsigned long addr, void *val, unsigned int bytes,
138 struct x86_exception *fault, bool system);
139
140
141
142
143
144
145
146 int (*fetch)(struct x86_emulate_ctxt *ctxt,
147 unsigned long addr, void *val, unsigned int bytes,
148 struct x86_exception *fault);
149
150
151
152
153
154
155
156 int (*read_emulated)(struct x86_emulate_ctxt *ctxt,
157 unsigned long addr, void *val, unsigned int bytes,
158 struct x86_exception *fault);
159
160
161
162
163
164
165
166
167 int (*write_emulated)(struct x86_emulate_ctxt *ctxt,
168 unsigned long addr, const void *val,
169 unsigned int bytes,
170 struct x86_exception *fault);
171
172
173
174
175
176
177
178
179
180 int (*cmpxchg_emulated)(struct x86_emulate_ctxt *ctxt,
181 unsigned long addr,
182 const void *old,
183 const void *new,
184 unsigned int bytes,
185 struct x86_exception *fault);
186 void (*invlpg)(struct x86_emulate_ctxt *ctxt, ulong addr);
187
188 int (*pio_in_emulated)(struct x86_emulate_ctxt *ctxt,
189 int size, unsigned short port, void *val,
190 unsigned int count);
191
192 int (*pio_out_emulated)(struct x86_emulate_ctxt *ctxt,
193 int size, unsigned short port, const void *val,
194 unsigned int count);
195
196 bool (*get_segment)(struct x86_emulate_ctxt *ctxt, u16 *selector,
197 struct desc_struct *desc, u32 *base3, int seg);
198 void (*set_segment)(struct x86_emulate_ctxt *ctxt, u16 selector,
199 struct desc_struct *desc, u32 base3, int seg);
200 unsigned long (*get_cached_segment_base)(struct x86_emulate_ctxt *ctxt,
201 int seg);
202 void (*get_gdt)(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt);
203 void (*get_idt)(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt);
204 void (*set_gdt)(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt);
205 void (*set_idt)(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt);
206 ulong (*get_cr)(struct x86_emulate_ctxt *ctxt, int cr);
207 int (*set_cr)(struct x86_emulate_ctxt *ctxt, int cr, ulong val);
208 int (*cpl)(struct x86_emulate_ctxt *ctxt);
209 void (*get_dr)(struct x86_emulate_ctxt *ctxt, int dr, ulong *dest);
210 int (*set_dr)(struct x86_emulate_ctxt *ctxt, int dr, ulong value);
211 u64 (*get_smbase)(struct x86_emulate_ctxt *ctxt);
212 void (*set_smbase)(struct x86_emulate_ctxt *ctxt, u64 smbase);
213 int (*set_msr_with_filter)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data);
214 int (*get_msr_with_filter)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata);
215 int (*set_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data);
216 int (*get_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata);
217 int (*check_pmc)(struct x86_emulate_ctxt *ctxt, u32 pmc);
218 int (*read_pmc)(struct x86_emulate_ctxt *ctxt, u32 pmc, u64 *pdata);
219 void (*halt)(struct x86_emulate_ctxt *ctxt);
220 void (*wbinvd)(struct x86_emulate_ctxt *ctxt);
221 int (*fix_hypercall)(struct x86_emulate_ctxt *ctxt);
222 int (*intercept)(struct x86_emulate_ctxt *ctxt,
223 struct x86_instruction_info *info,
224 enum x86_intercept_stage stage);
225
226 bool (*get_cpuid)(struct x86_emulate_ctxt *ctxt, u32 *eax, u32 *ebx,
227 u32 *ecx, u32 *edx, bool exact_only);
228 bool (*guest_has_long_mode)(struct x86_emulate_ctxt *ctxt);
229 bool (*guest_has_movbe)(struct x86_emulate_ctxt *ctxt);
230 bool (*guest_has_fxsr)(struct x86_emulate_ctxt *ctxt);
231 bool (*guest_has_rdpid)(struct x86_emulate_ctxt *ctxt);
232
233 void (*set_nmi_mask)(struct x86_emulate_ctxt *ctxt, bool masked);
234
235 unsigned (*get_hflags)(struct x86_emulate_ctxt *ctxt);
236 void (*exiting_smm)(struct x86_emulate_ctxt *ctxt);
237 int (*leave_smm)(struct x86_emulate_ctxt *ctxt, const char *smstate);
238 void (*triple_fault)(struct x86_emulate_ctxt *ctxt);
239 int (*set_xcr)(struct x86_emulate_ctxt *ctxt, u32 index, u64 xcr);
240};
241
242
243struct operand {
244 enum { OP_REG, OP_MEM, OP_MEM_STR, OP_IMM, OP_XMM, OP_MM, OP_NONE } type;
245 unsigned int bytes;
246 unsigned int count;
247 union {
248 unsigned long orig_val;
249 u64 orig_val64;
250 };
251 union {
252 unsigned long *reg;
253 struct segmented_address {
254 ulong ea;
255 unsigned seg;
256 } mem;
257 unsigned xmm;
258 unsigned mm;
259 } addr;
260 union {
261 unsigned long val;
262 u64 val64;
263 char valptr[sizeof(sse128_t)];
264 sse128_t vec_val;
265 u64 mm_val;
266 void *data;
267 };
268};
269
270struct fetch_cache {
271 u8 data[15];
272 u8 *ptr;
273 u8 *end;
274};
275
276struct read_cache {
277 u8 data[1024];
278 unsigned long pos;
279 unsigned long end;
280};
281
282
283enum x86emul_mode {
284 X86EMUL_MODE_REAL,
285 X86EMUL_MODE_VM86,
286 X86EMUL_MODE_PROT16,
287 X86EMUL_MODE_PROT32,
288 X86EMUL_MODE_PROT64,
289};
290
291
292#define X86EMUL_GUEST_MASK (1 << 5)
293#define X86EMUL_SMM_MASK (1 << 6)
294#define X86EMUL_SMM_INSIDE_NMI_MASK (1 << 7)
295
296
297
298
299
300struct fastop;
301
302typedef void (*fastop_t)(struct fastop *);
303
304struct x86_emulate_ctxt {
305 void *vcpu;
306 const struct x86_emulate_ops *ops;
307
308
309 unsigned long eflags;
310 unsigned long eip;
311
312 enum x86emul_mode mode;
313
314
315 int interruptibility;
316
317 bool perm_ok;
318 bool tf;
319
320 bool have_exception;
321 struct x86_exception exception;
322
323
324 bool gpa_available;
325 gpa_t gpa_val;
326
327
328
329
330
331
332 u8 opcode_len;
333 u8 b;
334 u8 intercept;
335 u8 op_bytes;
336 u8 ad_bytes;
337 union {
338 int (*execute)(struct x86_emulate_ctxt *ctxt);
339 fastop_t fop;
340 };
341 int (*check_perm)(struct x86_emulate_ctxt *ctxt);
342
343 bool rip_relative;
344 u8 rex_prefix;
345 u8 lock_prefix;
346 u8 rep_prefix;
347
348 u32 regs_valid;
349
350 u32 regs_dirty;
351
352 u8 modrm;
353 u8 modrm_mod;
354 u8 modrm_reg;
355 u8 modrm_rm;
356 u8 modrm_seg;
357 u8 seg_override;
358 u64 d;
359 unsigned long _eip;
360
361
362 struct operand src;
363 struct operand src2;
364 struct operand dst;
365 struct operand memop;
366 unsigned long _regs[NR_VCPU_REGS];
367 struct operand *memopp;
368 struct fetch_cache fetch;
369 struct read_cache io_read;
370 struct read_cache mem_read;
371 bool is_branch;
372};
373
374
375#define REPE_PREFIX 0xf3
376#define REPNE_PREFIX 0xf2
377
378
379#define X86EMUL_CPUID_VENDOR_AuthenticAMD_ebx 0x68747541
380#define X86EMUL_CPUID_VENDOR_AuthenticAMD_ecx 0x444d4163
381#define X86EMUL_CPUID_VENDOR_AuthenticAMD_edx 0x69746e65
382
383#define X86EMUL_CPUID_VENDOR_AMDisbetterI_ebx 0x69444d41
384#define X86EMUL_CPUID_VENDOR_AMDisbetterI_ecx 0x21726574
385#define X86EMUL_CPUID_VENDOR_AMDisbetterI_edx 0x74656273
386
387#define X86EMUL_CPUID_VENDOR_HygonGenuine_ebx 0x6f677948
388#define X86EMUL_CPUID_VENDOR_HygonGenuine_ecx 0x656e6975
389#define X86EMUL_CPUID_VENDOR_HygonGenuine_edx 0x6e65476e
390
391#define X86EMUL_CPUID_VENDOR_GenuineIntel_ebx 0x756e6547
392#define X86EMUL_CPUID_VENDOR_GenuineIntel_ecx 0x6c65746e
393#define X86EMUL_CPUID_VENDOR_GenuineIntel_edx 0x49656e69
394
395#define X86EMUL_CPUID_VENDOR_CentaurHauls_ebx 0x746e6543
396#define X86EMUL_CPUID_VENDOR_CentaurHauls_ecx 0x736c7561
397#define X86EMUL_CPUID_VENDOR_CentaurHauls_edx 0x48727561
398
399static inline bool is_guest_vendor_intel(u32 ebx, u32 ecx, u32 edx)
400{
401 return ebx == X86EMUL_CPUID_VENDOR_GenuineIntel_ebx &&
402 ecx == X86EMUL_CPUID_VENDOR_GenuineIntel_ecx &&
403 edx == X86EMUL_CPUID_VENDOR_GenuineIntel_edx;
404}
405
406static inline bool is_guest_vendor_amd(u32 ebx, u32 ecx, u32 edx)
407{
408 return (ebx == X86EMUL_CPUID_VENDOR_AuthenticAMD_ebx &&
409 ecx == X86EMUL_CPUID_VENDOR_AuthenticAMD_ecx &&
410 edx == X86EMUL_CPUID_VENDOR_AuthenticAMD_edx) ||
411 (ebx == X86EMUL_CPUID_VENDOR_AMDisbetterI_ebx &&
412 ecx == X86EMUL_CPUID_VENDOR_AMDisbetterI_ecx &&
413 edx == X86EMUL_CPUID_VENDOR_AMDisbetterI_edx);
414}
415
416static inline bool is_guest_vendor_hygon(u32 ebx, u32 ecx, u32 edx)
417{
418 return ebx == X86EMUL_CPUID_VENDOR_HygonGenuine_ebx &&
419 ecx == X86EMUL_CPUID_VENDOR_HygonGenuine_ecx &&
420 edx == X86EMUL_CPUID_VENDOR_HygonGenuine_edx;
421}
422
423enum x86_intercept_stage {
424 X86_ICTP_NONE = 0,
425 X86_ICPT_PRE_EXCEPT,
426 X86_ICPT_POST_EXCEPT,
427 X86_ICPT_POST_MEMACCESS,
428};
429
430enum x86_intercept {
431 x86_intercept_none,
432 x86_intercept_cr_read,
433 x86_intercept_cr_write,
434 x86_intercept_clts,
435 x86_intercept_lmsw,
436 x86_intercept_smsw,
437 x86_intercept_dr_read,
438 x86_intercept_dr_write,
439 x86_intercept_lidt,
440 x86_intercept_sidt,
441 x86_intercept_lgdt,
442 x86_intercept_sgdt,
443 x86_intercept_lldt,
444 x86_intercept_sldt,
445 x86_intercept_ltr,
446 x86_intercept_str,
447 x86_intercept_rdtsc,
448 x86_intercept_rdpmc,
449 x86_intercept_pushf,
450 x86_intercept_popf,
451 x86_intercept_cpuid,
452 x86_intercept_rsm,
453 x86_intercept_iret,
454 x86_intercept_intn,
455 x86_intercept_invd,
456 x86_intercept_pause,
457 x86_intercept_hlt,
458 x86_intercept_invlpg,
459 x86_intercept_invlpga,
460 x86_intercept_vmrun,
461 x86_intercept_vmload,
462 x86_intercept_vmsave,
463 x86_intercept_vmmcall,
464 x86_intercept_stgi,
465 x86_intercept_clgi,
466 x86_intercept_skinit,
467 x86_intercept_rdtscp,
468 x86_intercept_rdpid,
469 x86_intercept_icebp,
470 x86_intercept_wbinvd,
471 x86_intercept_monitor,
472 x86_intercept_mwait,
473 x86_intercept_rdmsr,
474 x86_intercept_wrmsr,
475 x86_intercept_in,
476 x86_intercept_ins,
477 x86_intercept_out,
478 x86_intercept_outs,
479 x86_intercept_xsetbv,
480
481 nr_x86_intercepts
482};
483
484
485#if defined(CONFIG_X86_32)
486#define X86EMUL_MODE_HOST X86EMUL_MODE_PROT32
487#elif defined(CONFIG_X86_64)
488#define X86EMUL_MODE_HOST X86EMUL_MODE_PROT64
489#endif
490
491int x86_decode_insn(struct x86_emulate_ctxt *ctxt, void *insn, int insn_len, int emulation_type);
492bool x86_page_table_writing_insn(struct x86_emulate_ctxt *ctxt);
493#define EMULATION_FAILED -1
494#define EMULATION_OK 0
495#define EMULATION_RESTART 1
496#define EMULATION_INTERCEPTED 2
497void init_decode_cache(struct x86_emulate_ctxt *ctxt);
498int x86_emulate_insn(struct x86_emulate_ctxt *ctxt);
499int emulator_task_switch(struct x86_emulate_ctxt *ctxt,
500 u16 tss_selector, int idt_index, int reason,
501 bool has_error_code, u32 error_code);
502int emulate_int_real(struct x86_emulate_ctxt *ctxt, int irq);
503void emulator_invalidate_register_cache(struct x86_emulate_ctxt *ctxt);
504void emulator_writeback_register_cache(struct x86_emulate_ctxt *ctxt);
505bool emulator_can_use_gpa(struct x86_emulate_ctxt *ctxt);
506
507#endif
508