1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20#include "qemu/osdep.h"
21#include "qemu.h"
22#include "qemu/timer.h"
23#include "user-internals.h"
24#include "cpu_loop-common.h"
25#include "signal-common.h"
26#include "user-mmap.h"
27
28
29
30
31uint64_t cpu_get_tsc(CPUX86State *env)
32{
33 return cpu_get_host_ticks();
34}
35
36static void write_dt(void *ptr, unsigned long addr, unsigned long limit,
37 int flags)
38{
39 unsigned int e1, e2;
40 uint32_t *p;
41 e1 = (addr << 16) | (limit & 0xffff);
42 e2 = ((addr >> 16) & 0xff) | (addr & 0xff000000) | (limit & 0x000f0000);
43 e2 |= flags;
44 p = ptr;
45 p[0] = tswap32(e1);
46 p[1] = tswap32(e2);
47}
48
49static uint64_t *idt_table;
50#ifdef TARGET_X86_64
51static void set_gate64(void *ptr, unsigned int type, unsigned int dpl,
52 uint64_t addr, unsigned int sel)
53{
54 uint32_t *p, e1, e2;
55 e1 = (addr & 0xffff) | (sel << 16);
56 e2 = (addr & 0xffff0000) | 0x8000 | (dpl << 13) | (type << 8);
57 p = ptr;
58 p[0] = tswap32(e1);
59 p[1] = tswap32(e2);
60 p[2] = tswap32(addr >> 32);
61 p[3] = 0;
62}
63
64static void set_idt(int n, unsigned int dpl)
65{
66 set_gate64(idt_table + n * 2, 0, dpl, 0, 0);
67}
68#else
69static void set_gate(void *ptr, unsigned int type, unsigned int dpl,
70 uint32_t addr, unsigned int sel)
71{
72 uint32_t *p, e1, e2;
73 e1 = (addr & 0xffff) | (sel << 16);
74 e2 = (addr & 0xffff0000) | 0x8000 | (dpl << 13) | (type << 8);
75 p = ptr;
76 p[0] = tswap32(e1);
77 p[1] = tswap32(e2);
78}
79
80
81static void set_idt(int n, unsigned int dpl)
82{
83 set_gate(idt_table + n, 0, dpl, 0, 0);
84}
85#endif
86
87#ifdef TARGET_X86_64
88static bool write_ok_or_segv(CPUX86State *env, abi_ptr addr, size_t len)
89{
90
91
92
93
94 if (addr == 0 || access_ok(env_cpu(env), VERIFY_WRITE, addr, len)) {
95 return true;
96 }
97
98 env->error_code = PG_ERROR_W_MASK | PG_ERROR_U_MASK;
99 force_sig_fault(TARGET_SIGSEGV, TARGET_SEGV_MAPERR, addr);
100 return false;
101}
102
103
104
105
106
107static void emulate_vsyscall(CPUX86State *env)
108{
109 int syscall;
110 abi_ulong ret;
111 uint64_t caller;
112
113
114
115
116
117 switch (env->eip & ~TARGET_PAGE_MASK) {
118 case 0x000:
119 syscall = TARGET_NR_gettimeofday;
120 break;
121 case 0x400:
122 syscall = TARGET_NR_time;
123 break;
124 case 0x800:
125 syscall = TARGET_NR_getcpu;
126 break;
127 default:
128 goto sigsegv;
129 }
130
131
132
133
134
135 if (get_user_u64(caller, env->regs[R_ESP])) {
136 goto sigsegv;
137 }
138
139
140
141
142 switch (syscall) {
143 case TARGET_NR_gettimeofday:
144 if (!write_ok_or_segv(env, env->regs[R_EDI],
145 sizeof(struct target_timeval)) ||
146 !write_ok_or_segv(env, env->regs[R_ESI],
147 sizeof(struct target_timezone))) {
148 return;
149 }
150 break;
151 case TARGET_NR_time:
152 if (!write_ok_or_segv(env, env->regs[R_EDI], sizeof(abi_long))) {
153 return;
154 }
155 break;
156 case TARGET_NR_getcpu:
157 if (!write_ok_or_segv(env, env->regs[R_EDI], sizeof(uint32_t)) ||
158 !write_ok_or_segv(env, env->regs[R_ESI], sizeof(uint32_t))) {
159 return;
160 }
161 break;
162 default:
163 g_assert_not_reached();
164 }
165
166
167
168
169 ret = do_syscall(env, syscall, env->regs[R_EDI], env->regs[R_ESI],
170 env->regs[R_EDX], env->regs[10], env->regs[8],
171 env->regs[9], 0, 0);
172 g_assert(ret != -QEMU_ERESTARTSYS);
173 g_assert(ret != -QEMU_ESIGRETURN);
174 if (ret == -TARGET_EFAULT) {
175 goto sigsegv;
176 }
177 env->regs[R_EAX] = ret;
178
179
180 env->eip = caller;
181 env->regs[R_ESP] += 8;
182 return;
183
184 sigsegv:
185 force_sig(TARGET_SIGSEGV);
186}
187#endif
188
189static bool maybe_handle_vm86_trap(CPUX86State *env, int trapnr)
190{
191#ifndef TARGET_X86_64
192 if (env->eflags & VM_MASK) {
193 handle_vm86_trap(env, trapnr);
194 return true;
195 }
196#endif
197 return false;
198}
199
200void cpu_loop(CPUX86State *env)
201{
202 CPUState *cs = env_cpu(env);
203 int trapnr;
204 abi_ulong ret;
205
206 for(;;) {
207 cpu_exec_start(cs);
208 trapnr = cpu_exec(cs);
209 cpu_exec_end(cs);
210 process_queued_cpu_work(cs);
211
212 switch(trapnr) {
213 case 0x80:
214
215 ret = do_syscall(env,
216 env->regs[R_EAX],
217 env->regs[R_EBX],
218 env->regs[R_ECX],
219 env->regs[R_EDX],
220 env->regs[R_ESI],
221 env->regs[R_EDI],
222 env->regs[R_EBP],
223 0, 0);
224 if (ret == -QEMU_ERESTARTSYS) {
225 env->eip -= 2;
226 } else if (ret != -QEMU_ESIGRETURN) {
227 env->regs[R_EAX] = ret;
228 }
229 break;
230#ifndef TARGET_ABI32
231 case EXCP_SYSCALL:
232
233 ret = do_syscall(env,
234 env->regs[R_EAX],
235 env->regs[R_EDI],
236 env->regs[R_ESI],
237 env->regs[R_EDX],
238 env->regs[10],
239 env->regs[8],
240 env->regs[9],
241 0, 0);
242 if (ret == -QEMU_ERESTARTSYS) {
243 env->eip -= 2;
244 } else if (ret != -QEMU_ESIGRETURN) {
245 env->regs[R_EAX] = ret;
246 }
247 break;
248#endif
249#ifdef TARGET_X86_64
250 case EXCP_VSYSCALL:
251 emulate_vsyscall(env);
252 break;
253#endif
254 case EXCP0B_NOSEG:
255 case EXCP0C_STACK:
256 force_sig(TARGET_SIGBUS);
257 break;
258 case EXCP0D_GPF:
259
260 if (maybe_handle_vm86_trap(env, trapnr)) {
261 break;
262 }
263 force_sig(TARGET_SIGSEGV);
264 break;
265 case EXCP0E_PAGE:
266 force_sig_fault(TARGET_SIGSEGV,
267 (env->error_code & PG_ERROR_P_MASK ?
268 TARGET_SEGV_ACCERR : TARGET_SEGV_MAPERR),
269 env->cr[2]);
270 break;
271 case EXCP00_DIVZ:
272 if (maybe_handle_vm86_trap(env, trapnr)) {
273 break;
274 }
275 force_sig_fault(TARGET_SIGFPE, TARGET_FPE_INTDIV, env->eip);
276 break;
277 case EXCP01_DB:
278 if (maybe_handle_vm86_trap(env, trapnr)) {
279 break;
280 }
281 force_sig_fault(TARGET_SIGTRAP, TARGET_TRAP_BRKPT, env->eip);
282 break;
283 case EXCP03_INT3:
284 if (maybe_handle_vm86_trap(env, trapnr)) {
285 break;
286 }
287 force_sig(TARGET_SIGTRAP);
288 break;
289 case EXCP04_INTO:
290 case EXCP05_BOUND:
291 if (maybe_handle_vm86_trap(env, trapnr)) {
292 break;
293 }
294 force_sig(TARGET_SIGSEGV);
295 break;
296 case EXCP06_ILLOP:
297 force_sig_fault(TARGET_SIGILL, TARGET_ILL_ILLOPN, env->eip);
298 break;
299 case EXCP_INTERRUPT:
300
301 break;
302 case EXCP_DEBUG:
303 force_sig_fault(TARGET_SIGTRAP, TARGET_TRAP_BRKPT, env->eip);
304 break;
305 case EXCP_ATOMIC:
306 cpu_exec_step_atomic(cs);
307 break;
308 default:
309 EXCP_DUMP(env, "qemu: unhandled CPU exception 0x%x - aborting\n",
310 trapnr);
311 abort();
312 }
313 process_pending_signals(env);
314 }
315}
316
317void target_cpu_copy_regs(CPUArchState *env, struct target_pt_regs *regs)
318{
319 env->cr[0] = CR0_PG_MASK | CR0_WP_MASK | CR0_PE_MASK;
320 env->hflags |= HF_PE_MASK | HF_CPL_MASK;
321 if (env->features[FEAT_1_EDX] & CPUID_SSE) {
322 env->cr[4] |= CR4_OSFXSR_MASK;
323 env->hflags |= HF_OSFXSR_MASK;
324 }
325#ifndef TARGET_ABI32
326
327 if (!(env->features[FEAT_8000_0001_EDX] & CPUID_EXT2_LM)) {
328 fprintf(stderr, "The selected x86 CPU does not support 64 bit mode\n");
329 exit(EXIT_FAILURE);
330 }
331 env->cr[4] |= CR4_PAE_MASK;
332 env->efer |= MSR_EFER_LMA | MSR_EFER_LME;
333 env->hflags |= HF_LMA_MASK;
334#endif
335
336
337 env->eflags |= IF_MASK;
338
339
340#ifndef TARGET_ABI32
341 env->regs[R_EAX] = regs->rax;
342 env->regs[R_EBX] = regs->rbx;
343 env->regs[R_ECX] = regs->rcx;
344 env->regs[R_EDX] = regs->rdx;
345 env->regs[R_ESI] = regs->rsi;
346 env->regs[R_EDI] = regs->rdi;
347 env->regs[R_EBP] = regs->rbp;
348 env->regs[R_ESP] = regs->rsp;
349 env->eip = regs->rip;
350#else
351 env->regs[R_EAX] = regs->eax;
352 env->regs[R_EBX] = regs->ebx;
353 env->regs[R_ECX] = regs->ecx;
354 env->regs[R_EDX] = regs->edx;
355 env->regs[R_ESI] = regs->esi;
356 env->regs[R_EDI] = regs->edi;
357 env->regs[R_EBP] = regs->ebp;
358 env->regs[R_ESP] = regs->esp;
359 env->eip = regs->eip;
360#endif
361
362
363#ifndef TARGET_ABI32
364 env->idt.limit = 511;
365#else
366 env->idt.limit = 255;
367#endif
368 env->idt.base = target_mmap(0, sizeof(uint64_t) * (env->idt.limit + 1),
369 PROT_READ|PROT_WRITE,
370 MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
371 idt_table = g2h_untagged(env->idt.base);
372 set_idt(0, 0);
373 set_idt(1, 0);
374 set_idt(2, 0);
375 set_idt(3, 3);
376 set_idt(4, 3);
377 set_idt(5, 0);
378 set_idt(6, 0);
379 set_idt(7, 0);
380 set_idt(8, 0);
381 set_idt(9, 0);
382 set_idt(10, 0);
383 set_idt(11, 0);
384 set_idt(12, 0);
385 set_idt(13, 0);
386 set_idt(14, 0);
387 set_idt(15, 0);
388 set_idt(16, 0);
389 set_idt(17, 0);
390 set_idt(18, 0);
391 set_idt(19, 0);
392 set_idt(0x80, 3);
393
394
395 {
396 uint64_t *gdt_table;
397 env->gdt.base = target_mmap(0, sizeof(uint64_t) * TARGET_GDT_ENTRIES,
398 PROT_READ|PROT_WRITE,
399 MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
400 env->gdt.limit = sizeof(uint64_t) * TARGET_GDT_ENTRIES - 1;
401 gdt_table = g2h_untagged(env->gdt.base);
402#ifdef TARGET_ABI32
403 write_dt(&gdt_table[__USER_CS >> 3], 0, 0xfffff,
404 DESC_G_MASK | DESC_B_MASK | DESC_P_MASK | DESC_S_MASK |
405 (3 << DESC_DPL_SHIFT) | (0xa << DESC_TYPE_SHIFT));
406#else
407
408 write_dt(&gdt_table[__USER_CS >> 3], 0, 0xfffff,
409 DESC_G_MASK | DESC_B_MASK | DESC_P_MASK | DESC_S_MASK |
410 DESC_L_MASK |
411 (3 << DESC_DPL_SHIFT) | (0xa << DESC_TYPE_SHIFT));
412#endif
413 write_dt(&gdt_table[__USER_DS >> 3], 0, 0xfffff,
414 DESC_G_MASK | DESC_B_MASK | DESC_P_MASK | DESC_S_MASK |
415 (3 << DESC_DPL_SHIFT) | (0x2 << DESC_TYPE_SHIFT));
416 }
417 cpu_x86_load_seg(env, R_CS, __USER_CS);
418 cpu_x86_load_seg(env, R_SS, __USER_DS);
419#ifdef TARGET_ABI32
420 cpu_x86_load_seg(env, R_DS, __USER_DS);
421 cpu_x86_load_seg(env, R_ES, __USER_DS);
422 cpu_x86_load_seg(env, R_FS, __USER_DS);
423 cpu_x86_load_seg(env, R_GS, __USER_DS);
424
425 env->segs[R_FS].selector = 0;
426#else
427 cpu_x86_load_seg(env, R_DS, 0);
428 cpu_x86_load_seg(env, R_ES, 0);
429 cpu_x86_load_seg(env, R_FS, 0);
430 cpu_x86_load_seg(env, R_GS, 0);
431#endif
432}
433