1
2
3
4
5
6
7
8
9
10
11
12
13#include "../tcg-pool.inc.c"
14#include "qemu/bitops.h"
15
16
17
18
19QEMU_BUILD_BUG_ON(TCG_TYPE_I32 != 0 || TCG_TYPE_I64 != 1);
20
21#ifdef CONFIG_DEBUG_TCG
22static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
23 "x0", "x1", "x2", "x3", "x4", "x5", "x6", "x7",
24 "x8", "x9", "x10", "x11", "x12", "x13", "x14", "x15",
25 "x16", "x17", "x18", "x19", "x20", "x21", "x22", "x23",
26 "x24", "x25", "x26", "x27", "x28", "fp", "x30", "sp",
27
28 "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7",
29 "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15",
30 "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23",
31 "v24", "v25", "v26", "v27", "v28", "fp", "v30", "v31",
32};
33#endif
34
35static const int tcg_target_reg_alloc_order[] = {
36 TCG_REG_X20, TCG_REG_X21, TCG_REG_X22, TCG_REG_X23,
37 TCG_REG_X24, TCG_REG_X25, TCG_REG_X26, TCG_REG_X27,
38 TCG_REG_X28,
39
40 TCG_REG_X8, TCG_REG_X9, TCG_REG_X10, TCG_REG_X11,
41 TCG_REG_X12, TCG_REG_X13, TCG_REG_X14, TCG_REG_X15,
42 TCG_REG_X16, TCG_REG_X17,
43
44 TCG_REG_X0, TCG_REG_X1, TCG_REG_X2, TCG_REG_X3,
45 TCG_REG_X4, TCG_REG_X5, TCG_REG_X6, TCG_REG_X7,
46
47
48
49
50
51
52 TCG_REG_V0, TCG_REG_V1, TCG_REG_V2, TCG_REG_V3,
53 TCG_REG_V4, TCG_REG_V5, TCG_REG_V6, TCG_REG_V7,
54
55 TCG_REG_V16, TCG_REG_V17, TCG_REG_V18, TCG_REG_V19,
56 TCG_REG_V20, TCG_REG_V21, TCG_REG_V22, TCG_REG_V23,
57 TCG_REG_V24, TCG_REG_V25, TCG_REG_V26, TCG_REG_V27,
58 TCG_REG_V28, TCG_REG_V29, TCG_REG_V30, TCG_REG_V31,
59};
60
61static const int tcg_target_call_iarg_regs[8] = {
62 TCG_REG_X0, TCG_REG_X1, TCG_REG_X2, TCG_REG_X3,
63 TCG_REG_X4, TCG_REG_X5, TCG_REG_X6, TCG_REG_X7
64};
65static const int tcg_target_call_oarg_regs[1] = {
66 TCG_REG_X0
67};
68
69#define TCG_REG_TMP TCG_REG_X30
70#define TCG_VEC_TMP TCG_REG_V31
71
72#ifndef CONFIG_SOFTMMU
73
74
75
76
77#define USE_GUEST_BASE (guest_base != 0 || TARGET_LONG_BITS == 32)
78#define TCG_REG_GUEST_BASE TCG_REG_X28
79#endif
80
81static inline bool reloc_pc26(tcg_insn_unit *code_ptr, tcg_insn_unit *target)
82{
83 ptrdiff_t offset = target - code_ptr;
84 if (offset == sextract64(offset, 0, 26)) {
85
86
87 *code_ptr = deposit32(*code_ptr, 0, 26, offset);
88 return true;
89 }
90 return false;
91}
92
93static inline bool reloc_pc19(tcg_insn_unit *code_ptr, tcg_insn_unit *target)
94{
95 ptrdiff_t offset = target - code_ptr;
96 if (offset == sextract64(offset, 0, 19)) {
97 *code_ptr = deposit32(*code_ptr, 5, 19, offset);
98 return true;
99 }
100 return false;
101}
102
103static inline bool patch_reloc(tcg_insn_unit *code_ptr, int type,
104 intptr_t value, intptr_t addend)
105{
106 tcg_debug_assert(addend == 0);
107 switch (type) {
108 case R_AARCH64_JUMP26:
109 case R_AARCH64_CALL26:
110 return reloc_pc26(code_ptr, (tcg_insn_unit *)value);
111 case R_AARCH64_CONDBR19:
112 return reloc_pc19(code_ptr, (tcg_insn_unit *)value);
113 default:
114 g_assert_not_reached();
115 }
116}
117
118#define TCG_CT_CONST_AIMM 0x100
119#define TCG_CT_CONST_LIMM 0x200
120#define TCG_CT_CONST_ZERO 0x400
121#define TCG_CT_CONST_MONE 0x800
122#define TCG_CT_CONST_ORRI 0x1000
123#define TCG_CT_CONST_ANDI 0x2000
124
125
126static const char *target_parse_constraint(TCGArgConstraint *ct,
127 const char *ct_str, TCGType type)
128{
129 switch (*ct_str++) {
130 case 'r':
131 ct->ct |= TCG_CT_REG;
132 ct->u.regs |= 0xffffffffu;
133 break;
134 case 'w':
135 ct->ct |= TCG_CT_REG;
136 ct->u.regs |= 0xffffffff00000000ull;
137 break;
138 case 'l':
139 ct->ct |= TCG_CT_REG;
140 ct->u.regs = 0xffffffffu;
141#ifdef CONFIG_SOFTMMU
142
143
144 tcg_regset_reset_reg(ct->u.regs, TCG_REG_X0);
145 tcg_regset_reset_reg(ct->u.regs, TCG_REG_X1);
146 tcg_regset_reset_reg(ct->u.regs, TCG_REG_X2);
147 tcg_regset_reset_reg(ct->u.regs, TCG_REG_X3);
148#endif
149 break;
150 case 'A':
151 ct->ct |= TCG_CT_CONST_AIMM;
152 break;
153 case 'L':
154 ct->ct |= TCG_CT_CONST_LIMM;
155 break;
156 case 'M':
157 ct->ct |= TCG_CT_CONST_MONE;
158 break;
159 case 'O':
160 ct->ct |= TCG_CT_CONST_ORRI;
161 break;
162 case 'N':
163 ct->ct |= TCG_CT_CONST_ANDI;
164 break;
165 case 'Z':
166 ct->ct |= TCG_CT_CONST_ZERO;
167 break;
168 default:
169 return NULL;
170 }
171 return ct_str;
172}
173
174
175static inline bool is_aimm(uint64_t val)
176{
177 return (val & ~0xfff) == 0 || (val & ~0xfff000) == 0;
178}
179
180
181static inline bool is_limm(uint64_t val)
182{
183
184
185
186
187
188
189
190
191 if ((int64_t)val < 0) {
192 val = ~val;
193 }
194 if (val == 0) {
195 return false;
196 }
197 val += val & -val;
198 return (val & (val - 1)) == 0;
199}
200
201
202static bool is_shimm16(uint16_t v16, int *cmode, int *imm8)
203{
204 if (v16 == (v16 & 0xff)) {
205 *cmode = 0x8;
206 *imm8 = v16 & 0xff;
207 return true;
208 } else if (v16 == (v16 & 0xff00)) {
209 *cmode = 0xa;
210 *imm8 = v16 >> 8;
211 return true;
212 }
213 return false;
214}
215
216
217static bool is_shimm32(uint32_t v32, int *cmode, int *imm8)
218{
219 if (v32 == (v32 & 0xff)) {
220 *cmode = 0x0;
221 *imm8 = v32 & 0xff;
222 return true;
223 } else if (v32 == (v32 & 0xff00)) {
224 *cmode = 0x2;
225 *imm8 = (v32 >> 8) & 0xff;
226 return true;
227 } else if (v32 == (v32 & 0xff0000)) {
228 *cmode = 0x4;
229 *imm8 = (v32 >> 16) & 0xff;
230 return true;
231 } else if (v32 == (v32 & 0xff000000)) {
232 *cmode = 0x6;
233 *imm8 = v32 >> 24;
234 return true;
235 }
236 return false;
237}
238
239
240static bool is_soimm32(uint32_t v32, int *cmode, int *imm8)
241{
242 if ((v32 & 0xffff00ff) == 0xff) {
243 *cmode = 0xc;
244 *imm8 = (v32 >> 8) & 0xff;
245 return true;
246 } else if ((v32 & 0xff00ffff) == 0xffff) {
247 *cmode = 0xd;
248 *imm8 = (v32 >> 16) & 0xff;
249 return true;
250 }
251 return false;
252}
253
254
255static bool is_fimm32(uint32_t v32, int *cmode, int *imm8)
256{
257 if (extract32(v32, 0, 19) == 0
258 && (extract32(v32, 25, 6) == 0x20
259 || extract32(v32, 25, 6) == 0x1f)) {
260 *cmode = 0xf;
261 *imm8 = (extract32(v32, 31, 1) << 7)
262 | (extract32(v32, 25, 1) << 6)
263 | extract32(v32, 19, 6);
264 return true;
265 }
266 return false;
267}
268
269
270static bool is_fimm64(uint64_t v64, int *cmode, int *imm8)
271{
272 if (extract64(v64, 0, 48) == 0
273 && (extract64(v64, 54, 9) == 0x100
274 || extract64(v64, 54, 9) == 0x0ff)) {
275 *cmode = 0xf;
276 *imm8 = (extract64(v64, 63, 1) << 7)
277 | (extract64(v64, 54, 1) << 6)
278 | extract64(v64, 48, 6);
279 return true;
280 }
281 return false;
282}
283
284
285
286
287
288
289static int is_shimm32_pair(uint32_t v32, int *cmode, int *imm8)
290{
291 int i;
292
293 for (i = 6; i > 0; i -= 2) {
294
295 uint32_t tmp = v32 & ~(0xffu << (i * 4));
296 if (is_shimm32(tmp, cmode, imm8) ||
297 is_soimm32(tmp, cmode, imm8)) {
298 break;
299 }
300 }
301 return i;
302}
303
304
305static bool is_shimm1632(uint32_t v32, int *cmode, int *imm8)
306{
307 if (v32 == deposit32(v32, 16, 16, v32)) {
308 return is_shimm16(v32, cmode, imm8);
309 } else {
310 return is_shimm32(v32, cmode, imm8);
311 }
312}
313
314static int tcg_target_const_match(tcg_target_long val, TCGType type,
315 const TCGArgConstraint *arg_ct)
316{
317 int ct = arg_ct->ct;
318
319 if (ct & TCG_CT_CONST) {
320 return 1;
321 }
322 if (type == TCG_TYPE_I32) {
323 val = (int32_t)val;
324 }
325 if ((ct & TCG_CT_CONST_AIMM) && (is_aimm(val) || is_aimm(-val))) {
326 return 1;
327 }
328 if ((ct & TCG_CT_CONST_LIMM) && is_limm(val)) {
329 return 1;
330 }
331 if ((ct & TCG_CT_CONST_ZERO) && val == 0) {
332 return 1;
333 }
334 if ((ct & TCG_CT_CONST_MONE) && val == -1) {
335 return 1;
336 }
337
338 switch (ct & (TCG_CT_CONST_ORRI | TCG_CT_CONST_ANDI)) {
339 case 0:
340 break;
341 case TCG_CT_CONST_ANDI:
342 val = ~val;
343
344 case TCG_CT_CONST_ORRI:
345 if (val == deposit64(val, 32, 32, val)) {
346 int cmode, imm8;
347 return is_shimm1632(val, &cmode, &imm8);
348 }
349 break;
350 default:
351
352 g_assert_not_reached();
353 }
354
355 return 0;
356}
357
358enum aarch64_cond_code {
359 COND_EQ = 0x0,
360 COND_NE = 0x1,
361 COND_CS = 0x2,
362 COND_HS = COND_CS,
363 COND_CC = 0x3,
364 COND_LO = COND_CC,
365 COND_MI = 0x4,
366 COND_PL = 0x5,
367 COND_VS = 0x6,
368 COND_VC = 0x7,
369 COND_HI = 0x8,
370 COND_LS = 0x9,
371 COND_GE = 0xa,
372 COND_LT = 0xb,
373 COND_GT = 0xc,
374 COND_LE = 0xd,
375 COND_AL = 0xe,
376 COND_NV = 0xf,
377};
378
379static const enum aarch64_cond_code tcg_cond_to_aarch64[] = {
380 [TCG_COND_EQ] = COND_EQ,
381 [TCG_COND_NE] = COND_NE,
382 [TCG_COND_LT] = COND_LT,
383 [TCG_COND_GE] = COND_GE,
384 [TCG_COND_LE] = COND_LE,
385 [TCG_COND_GT] = COND_GT,
386
387 [TCG_COND_LTU] = COND_LO,
388 [TCG_COND_GTU] = COND_HI,
389 [TCG_COND_GEU] = COND_HS,
390 [TCG_COND_LEU] = COND_LS,
391};
392
393typedef enum {
394 LDST_ST = 0,
395 LDST_LD = 1,
396 LDST_LD_S_X = 2,
397 LDST_LD_S_W = 3,
398} AArch64LdstType;
399
400
401
402
403
404
405typedef enum {
406
407 I3201_CBZ = 0x34000000,
408 I3201_CBNZ = 0x35000000,
409
410
411 I3202_B_C = 0x54000000,
412
413
414 I3206_B = 0x14000000,
415 I3206_BL = 0x94000000,
416
417
418 I3207_BR = 0xd61f0000,
419 I3207_BLR = 0xd63f0000,
420 I3207_RET = 0xd65f0000,
421
422
423 I3303_LD1R = 0x0d40c000,
424
425
426 I3305_LDR = 0x58000000,
427 I3305_LDR_v64 = 0x5c000000,
428 I3305_LDR_v128 = 0x9c000000,
429
430
431
432 I3312_STRB = 0x38000000 | LDST_ST << 22 | MO_8 << 30,
433 I3312_STRH = 0x38000000 | LDST_ST << 22 | MO_16 << 30,
434 I3312_STRW = 0x38000000 | LDST_ST << 22 | MO_32 << 30,
435 I3312_STRX = 0x38000000 | LDST_ST << 22 | MO_64 << 30,
436
437 I3312_LDRB = 0x38000000 | LDST_LD << 22 | MO_8 << 30,
438 I3312_LDRH = 0x38000000 | LDST_LD << 22 | MO_16 << 30,
439 I3312_LDRW = 0x38000000 | LDST_LD << 22 | MO_32 << 30,
440 I3312_LDRX = 0x38000000 | LDST_LD << 22 | MO_64 << 30,
441
442 I3312_LDRSBW = 0x38000000 | LDST_LD_S_W << 22 | MO_8 << 30,
443 I3312_LDRSHW = 0x38000000 | LDST_LD_S_W << 22 | MO_16 << 30,
444
445 I3312_LDRSBX = 0x38000000 | LDST_LD_S_X << 22 | MO_8 << 30,
446 I3312_LDRSHX = 0x38000000 | LDST_LD_S_X << 22 | MO_16 << 30,
447 I3312_LDRSWX = 0x38000000 | LDST_LD_S_X << 22 | MO_32 << 30,
448
449 I3312_LDRVS = 0x3c000000 | LDST_LD << 22 | MO_32 << 30,
450 I3312_STRVS = 0x3c000000 | LDST_ST << 22 | MO_32 << 30,
451
452 I3312_LDRVD = 0x3c000000 | LDST_LD << 22 | MO_64 << 30,
453 I3312_STRVD = 0x3c000000 | LDST_ST << 22 | MO_64 << 30,
454
455 I3312_LDRVQ = 0x3c000000 | 3 << 22 | 0 << 30,
456 I3312_STRVQ = 0x3c000000 | 2 << 22 | 0 << 30,
457
458 I3312_TO_I3310 = 0x00200800,
459 I3312_TO_I3313 = 0x01000000,
460
461
462 I3314_LDP = 0x28400000,
463 I3314_STP = 0x28000000,
464
465
466 I3401_ADDI = 0x11000000,
467 I3401_ADDSI = 0x31000000,
468 I3401_SUBI = 0x51000000,
469 I3401_SUBSI = 0x71000000,
470
471
472 I3402_BFM = 0x33000000,
473 I3402_SBFM = 0x13000000,
474 I3402_UBFM = 0x53000000,
475
476
477 I3403_EXTR = 0x13800000,
478
479
480 I3404_ANDI = 0x12000000,
481 I3404_ORRI = 0x32000000,
482 I3404_EORI = 0x52000000,
483
484
485 I3405_MOVN = 0x12800000,
486 I3405_MOVZ = 0x52800000,
487 I3405_MOVK = 0x72800000,
488
489
490 I3406_ADR = 0x10000000,
491 I3406_ADRP = 0x90000000,
492
493
494 I3502_ADD = 0x0b000000,
495 I3502_ADDS = 0x2b000000,
496 I3502_SUB = 0x4b000000,
497 I3502_SUBS = 0x6b000000,
498
499
500 I3502S_ADD_LSL = I3502_ADD,
501
502
503 I3503_ADC = 0x1a000000,
504 I3503_SBC = 0x5a000000,
505
506
507 I3506_CSEL = 0x1a800000,
508 I3506_CSINC = 0x1a800400,
509 I3506_CSINV = 0x5a800000,
510 I3506_CSNEG = 0x5a800400,
511
512
513 I3507_CLZ = 0x5ac01000,
514 I3507_RBIT = 0x5ac00000,
515 I3507_REV16 = 0x5ac00400,
516 I3507_REV32 = 0x5ac00800,
517 I3507_REV64 = 0x5ac00c00,
518
519
520 I3508_LSLV = 0x1ac02000,
521 I3508_LSRV = 0x1ac02400,
522 I3508_ASRV = 0x1ac02800,
523 I3508_RORV = 0x1ac02c00,
524 I3508_SMULH = 0x9b407c00,
525 I3508_UMULH = 0x9bc07c00,
526 I3508_UDIV = 0x1ac00800,
527 I3508_SDIV = 0x1ac00c00,
528
529
530 I3509_MADD = 0x1b000000,
531 I3509_MSUB = 0x1b008000,
532
533
534 I3510_AND = 0x0a000000,
535 I3510_BIC = 0x0a200000,
536 I3510_ORR = 0x2a000000,
537 I3510_ORN = 0x2a200000,
538 I3510_EOR = 0x4a000000,
539 I3510_EON = 0x4a200000,
540 I3510_ANDS = 0x6a000000,
541
542
543 I3502S_AND_LSR = I3510_AND | (1 << 22),
544
545
546 I3605_DUP = 0x0e000400,
547 I3605_INS = 0x4e001c00,
548 I3605_UMOV = 0x0e003c00,
549
550
551 I3606_MOVI = 0x0f000400,
552 I3606_MVNI = 0x2f000400,
553 I3606_BIC = 0x2f001400,
554 I3606_ORR = 0x0f001400,
555
556
557 I3614_SSHR = 0x0f000400,
558 I3614_SSRA = 0x0f001400,
559 I3614_SHL = 0x0f005400,
560 I3614_SLI = 0x2f005400,
561 I3614_USHR = 0x2f000400,
562 I3614_USRA = 0x2f001400,
563
564
565 I3616_ADD = 0x0e208400,
566 I3616_AND = 0x0e201c00,
567 I3616_BIC = 0x0e601c00,
568 I3616_BIF = 0x2ee01c00,
569 I3616_BIT = 0x2ea01c00,
570 I3616_BSL = 0x2e601c00,
571 I3616_EOR = 0x2e201c00,
572 I3616_MUL = 0x0e209c00,
573 I3616_ORR = 0x0ea01c00,
574 I3616_ORN = 0x0ee01c00,
575 I3616_SUB = 0x2e208400,
576 I3616_CMGT = 0x0e203400,
577 I3616_CMGE = 0x0e203c00,
578 I3616_CMTST = 0x0e208c00,
579 I3616_CMHI = 0x2e203400,
580 I3616_CMHS = 0x2e203c00,
581 I3616_CMEQ = 0x2e208c00,
582 I3616_SMAX = 0x0e206400,
583 I3616_SMIN = 0x0e206c00,
584 I3616_SSHL = 0x0e204400,
585 I3616_SQADD = 0x0e200c00,
586 I3616_SQSUB = 0x0e202c00,
587 I3616_UMAX = 0x2e206400,
588 I3616_UMIN = 0x2e206c00,
589 I3616_UQADD = 0x2e200c00,
590 I3616_UQSUB = 0x2e202c00,
591 I3616_USHL = 0x2e204400,
592
593
594 I3617_CMGT0 = 0x0e208800,
595 I3617_CMEQ0 = 0x0e209800,
596 I3617_CMLT0 = 0x0e20a800,
597 I3617_CMGE0 = 0x2e208800,
598 I3617_CMLE0 = 0x2e20a800,
599 I3617_NOT = 0x2e205800,
600 I3617_ABS = 0x0e20b800,
601 I3617_NEG = 0x2e20b800,
602
603
604 NOP = 0xd503201f,
605 DMB_ISH = 0xd50338bf,
606 DMB_LD = 0x00000100,
607 DMB_ST = 0x00000200,
608} AArch64Insn;
609
610static inline uint32_t tcg_in32(TCGContext *s)
611{
612 uint32_t v = *(uint32_t *)s->code_ptr;
613 return v;
614}
615
616
617#define tcg_out_insn(S, FMT, OP, ...) \
618 glue(tcg_out_insn_,FMT)(S, glue(glue(glue(I,FMT),_),OP), ## __VA_ARGS__)
619
620static void tcg_out_insn_3303(TCGContext *s, AArch64Insn insn, bool q,
621 TCGReg rt, TCGReg rn, unsigned size)
622{
623 tcg_out32(s, insn | (rt & 0x1f) | (rn << 5) | (size << 10) | (q << 30));
624}
625
626static void tcg_out_insn_3305(TCGContext *s, AArch64Insn insn,
627 int imm19, TCGReg rt)
628{
629 tcg_out32(s, insn | (imm19 & 0x7ffff) << 5 | rt);
630}
631
632static void tcg_out_insn_3201(TCGContext *s, AArch64Insn insn, TCGType ext,
633 TCGReg rt, int imm19)
634{
635 tcg_out32(s, insn | ext << 31 | (imm19 & 0x7ffff) << 5 | rt);
636}
637
638static void tcg_out_insn_3202(TCGContext *s, AArch64Insn insn,
639 TCGCond c, int imm19)
640{
641 tcg_out32(s, insn | tcg_cond_to_aarch64[c] | (imm19 & 0x7ffff) << 5);
642}
643
644static void tcg_out_insn_3206(TCGContext *s, AArch64Insn insn, int imm26)
645{
646 tcg_out32(s, insn | (imm26 & 0x03ffffff));
647}
648
649static void tcg_out_insn_3207(TCGContext *s, AArch64Insn insn, TCGReg rn)
650{
651 tcg_out32(s, insn | rn << 5);
652}
653
654static void tcg_out_insn_3314(TCGContext *s, AArch64Insn insn,
655 TCGReg r1, TCGReg r2, TCGReg rn,
656 tcg_target_long ofs, bool pre, bool w)
657{
658 insn |= 1u << 31;
659 insn |= pre << 24;
660 insn |= w << 23;
661
662 tcg_debug_assert(ofs >= -0x200 && ofs < 0x200 && (ofs & 7) == 0);
663 insn |= (ofs & (0x7f << 3)) << (15 - 3);
664
665 tcg_out32(s, insn | r2 << 10 | rn << 5 | r1);
666}
667
668static void tcg_out_insn_3401(TCGContext *s, AArch64Insn insn, TCGType ext,
669 TCGReg rd, TCGReg rn, uint64_t aimm)
670{
671 if (aimm > 0xfff) {
672 tcg_debug_assert((aimm & 0xfff) == 0);
673 aimm >>= 12;
674 tcg_debug_assert(aimm <= 0xfff);
675 aimm |= 1 << 12;
676 }
677 tcg_out32(s, insn | ext << 31 | aimm << 10 | rn << 5 | rd);
678}
679
680
681
682
683static void tcg_out_insn_3402(TCGContext *s, AArch64Insn insn, TCGType ext,
684 TCGReg rd, TCGReg rn, int n, int immr, int imms)
685{
686 tcg_out32(s, insn | ext << 31 | n << 22 | immr << 16 | imms << 10
687 | rn << 5 | rd);
688}
689
690#define tcg_out_insn_3404 tcg_out_insn_3402
691
692static void tcg_out_insn_3403(TCGContext *s, AArch64Insn insn, TCGType ext,
693 TCGReg rd, TCGReg rn, TCGReg rm, int imms)
694{
695 tcg_out32(s, insn | ext << 31 | ext << 22 | rm << 16 | imms << 10
696 | rn << 5 | rd);
697}
698
699
700
701static void tcg_out_insn_3405(TCGContext *s, AArch64Insn insn, TCGType ext,
702 TCGReg rd, uint16_t half, unsigned shift)
703{
704 tcg_debug_assert((shift & ~0x30) == 0);
705 tcg_out32(s, insn | ext << 31 | shift << (21 - 4) | half << 5 | rd);
706}
707
708static void tcg_out_insn_3406(TCGContext *s, AArch64Insn insn,
709 TCGReg rd, int64_t disp)
710{
711 tcg_out32(s, insn | (disp & 3) << 29 | (disp & 0x1ffffc) << (5 - 2) | rd);
712}
713
714
715
716static inline void tcg_out_insn_3502S(TCGContext *s, AArch64Insn insn,
717 TCGType ext, TCGReg rd, TCGReg rn,
718 TCGReg rm, int imm6)
719{
720 tcg_out32(s, insn | ext << 31 | rm << 16 | imm6 << 10 | rn << 5 | rd);
721}
722
723
724
725
726
727static void tcg_out_insn_3502(TCGContext *s, AArch64Insn insn, TCGType ext,
728 TCGReg rd, TCGReg rn, TCGReg rm)
729{
730 tcg_out32(s, insn | ext << 31 | rm << 16 | rn << 5 | rd);
731}
732
733#define tcg_out_insn_3503 tcg_out_insn_3502
734#define tcg_out_insn_3508 tcg_out_insn_3502
735#define tcg_out_insn_3510 tcg_out_insn_3502
736
737static void tcg_out_insn_3506(TCGContext *s, AArch64Insn insn, TCGType ext,
738 TCGReg rd, TCGReg rn, TCGReg rm, TCGCond c)
739{
740 tcg_out32(s, insn | ext << 31 | rm << 16 | rn << 5 | rd
741 | tcg_cond_to_aarch64[c] << 12);
742}
743
744static void tcg_out_insn_3507(TCGContext *s, AArch64Insn insn, TCGType ext,
745 TCGReg rd, TCGReg rn)
746{
747 tcg_out32(s, insn | ext << 31 | rn << 5 | rd);
748}
749
750static void tcg_out_insn_3509(TCGContext *s, AArch64Insn insn, TCGType ext,
751 TCGReg rd, TCGReg rn, TCGReg rm, TCGReg ra)
752{
753 tcg_out32(s, insn | ext << 31 | rm << 16 | ra << 10 | rn << 5 | rd);
754}
755
756static void tcg_out_insn_3605(TCGContext *s, AArch64Insn insn, bool q,
757 TCGReg rd, TCGReg rn, int dst_idx, int src_idx)
758{
759
760
761 tcg_out32(s, insn | q << 30 | (dst_idx << 16) | (src_idx << 11)
762 | (rd & 0x1f) | (~rn & 0x20) << 6 | (rn & 0x1f) << 5);
763}
764
765static void tcg_out_insn_3606(TCGContext *s, AArch64Insn insn, bool q,
766 TCGReg rd, bool op, int cmode, uint8_t imm8)
767{
768 tcg_out32(s, insn | q << 30 | op << 29 | cmode << 12 | (rd & 0x1f)
769 | (imm8 & 0xe0) << (16 - 5) | (imm8 & 0x1f) << 5);
770}
771
772static void tcg_out_insn_3614(TCGContext *s, AArch64Insn insn, bool q,
773 TCGReg rd, TCGReg rn, unsigned immhb)
774{
775 tcg_out32(s, insn | q << 30 | immhb << 16
776 | (rn & 0x1f) << 5 | (rd & 0x1f));
777}
778
779static void tcg_out_insn_3616(TCGContext *s, AArch64Insn insn, bool q,
780 unsigned size, TCGReg rd, TCGReg rn, TCGReg rm)
781{
782 tcg_out32(s, insn | q << 30 | (size << 22) | (rm & 0x1f) << 16
783 | (rn & 0x1f) << 5 | (rd & 0x1f));
784}
785
786static void tcg_out_insn_3617(TCGContext *s, AArch64Insn insn, bool q,
787 unsigned size, TCGReg rd, TCGReg rn)
788{
789 tcg_out32(s, insn | q << 30 | (size << 22)
790 | (rn & 0x1f) << 5 | (rd & 0x1f));
791}
792
793static void tcg_out_insn_3310(TCGContext *s, AArch64Insn insn,
794 TCGReg rd, TCGReg base, TCGType ext,
795 TCGReg regoff)
796{
797
798 tcg_out32(s, insn | I3312_TO_I3310 | regoff << 16 |
799 0x4000 | ext << 13 | base << 5 | (rd & 0x1f));
800}
801
802static void tcg_out_insn_3312(TCGContext *s, AArch64Insn insn,
803 TCGReg rd, TCGReg rn, intptr_t offset)
804{
805 tcg_out32(s, insn | (offset & 0x1ff) << 12 | rn << 5 | (rd & 0x1f));
806}
807
808static void tcg_out_insn_3313(TCGContext *s, AArch64Insn insn,
809 TCGReg rd, TCGReg rn, uintptr_t scaled_uimm)
810{
811
812 tcg_out32(s, insn | I3312_TO_I3313 | scaled_uimm << 10
813 | rn << 5 | (rd & 0x1f));
814}
815
816
817static void tcg_out_movr(TCGContext *s, TCGType ext, TCGReg rd, TCGReg rm)
818{
819 tcg_out_insn(s, 3510, ORR, ext, rd, TCG_REG_XZR, rm);
820}
821
822
823static void tcg_out_movr_sp(TCGContext *s, TCGType ext, TCGReg rd, TCGReg rn)
824{
825 tcg_out_insn(s, 3401, ADDI, ext, rd, rn, 0);
826}
827
828
829
830
831static void tcg_out_logicali(TCGContext *s, AArch64Insn insn, TCGType ext,
832 TCGReg rd, TCGReg rn, uint64_t limm)
833{
834 unsigned h, l, r, c;
835
836 tcg_debug_assert(is_limm(limm));
837
838 h = clz64(limm);
839 l = ctz64(limm);
840 if (l == 0) {
841 r = 0;
842 c = ctz64(~limm) - 1;
843 if (h == 0) {
844 r = clz64(~limm);
845 c += r;
846 }
847 } else {
848 r = 64 - l;
849 c = r - h - 1;
850 }
851 if (ext == TCG_TYPE_I32) {
852 r &= 31;
853 c &= 31;
854 }
855
856 tcg_out_insn_3404(s, insn, ext, rd, rn, ext, r, c);
857}
858
859static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
860 TCGReg rd, tcg_target_long v64)
861{
862 bool q = type == TCG_TYPE_V128;
863 int cmode, imm8, i;
864
865
866 if (v64 == dup_const(MO_8, v64)) {
867 imm8 = (uint8_t)v64;
868 tcg_out_insn(s, 3606, MOVI, q, rd, 0, 0xe, imm8);
869 return;
870 }
871
872
873
874
875
876 for (i = imm8 = 0; i < 8; i++) {
877 uint8_t byte = v64 >> (i * 8);
878 if (byte == 0xff) {
879 imm8 |= 1 << i;
880 } else if (byte != 0) {
881 goto fail_bytes;
882 }
883 }
884 tcg_out_insn(s, 3606, MOVI, q, rd, 1, 0xe, imm8);
885 return;
886 fail_bytes:
887
888
889
890
891
892
893 if (v64 == dup_const(MO_16, v64)) {
894 uint16_t v16 = v64;
895
896 if (is_shimm16(v16, &cmode, &imm8)) {
897 tcg_out_insn(s, 3606, MOVI, q, rd, 0, cmode, imm8);
898 return;
899 }
900 if (is_shimm16(~v16, &cmode, &imm8)) {
901 tcg_out_insn(s, 3606, MVNI, q, rd, 0, cmode, imm8);
902 return;
903 }
904
905
906
907
908
909 tcg_out_insn(s, 3606, MOVI, q, rd, 0, 0x8, v16 & 0xff);
910 tcg_out_insn(s, 3606, ORR, q, rd, 0, 0xa, v16 >> 8);
911 return;
912 } else if (v64 == dup_const(MO_32, v64)) {
913 uint32_t v32 = v64;
914 uint32_t n32 = ~v32;
915
916 if (is_shimm32(v32, &cmode, &imm8) ||
917 is_soimm32(v32, &cmode, &imm8) ||
918 is_fimm32(v32, &cmode, &imm8)) {
919 tcg_out_insn(s, 3606, MOVI, q, rd, 0, cmode, imm8);
920 return;
921 }
922 if (is_shimm32(n32, &cmode, &imm8) ||
923 is_soimm32(n32, &cmode, &imm8)) {
924 tcg_out_insn(s, 3606, MVNI, q, rd, 0, cmode, imm8);
925 return;
926 }
927
928
929
930
931
932 i = is_shimm32_pair(v32, &cmode, &imm8);
933 if (i) {
934 tcg_out_insn(s, 3606, MOVI, q, rd, 0, cmode, imm8);
935 tcg_out_insn(s, 3606, ORR, q, rd, 0, i, extract32(v32, i * 4, 8));
936 return;
937 }
938 i = is_shimm32_pair(n32, &cmode, &imm8);
939 if (i) {
940 tcg_out_insn(s, 3606, MVNI, q, rd, 0, cmode, imm8);
941 tcg_out_insn(s, 3606, BIC, q, rd, 0, i, extract32(n32, i * 4, 8));
942 return;
943 }
944 } else if (is_fimm64(v64, &cmode, &imm8)) {
945 tcg_out_insn(s, 3606, MOVI, q, rd, 1, cmode, imm8);
946 return;
947 }
948
949
950
951
952
953 if (type == TCG_TYPE_V128) {
954 new_pool_l2(s, R_AARCH64_CONDBR19, s->code_ptr, 0, v64, v64);
955 tcg_out_insn(s, 3305, LDR_v128, 0, rd);
956 } else {
957 new_pool_label(s, v64, R_AARCH64_CONDBR19, s->code_ptr, 0);
958 tcg_out_insn(s, 3305, LDR_v64, 0, rd);
959 }
960}
961
962static bool tcg_out_dup_vec(TCGContext *s, TCGType type, unsigned vece,
963 TCGReg rd, TCGReg rs)
964{
965 int is_q = type - TCG_TYPE_V64;
966 tcg_out_insn(s, 3605, DUP, is_q, rd, rs, 1 << vece, 0);
967 return true;
968}
969
970static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
971 TCGReg r, TCGReg base, intptr_t offset)
972{
973 TCGReg temp = TCG_REG_TMP;
974
975 if (offset < -0xffffff || offset > 0xffffff) {
976 tcg_out_movi(s, TCG_TYPE_PTR, temp, offset);
977 tcg_out_insn(s, 3502, ADD, 1, temp, temp, base);
978 base = temp;
979 } else {
980 AArch64Insn add_insn = I3401_ADDI;
981
982 if (offset < 0) {
983 add_insn = I3401_SUBI;
984 offset = -offset;
985 }
986 if (offset & 0xfff000) {
987 tcg_out_insn_3401(s, add_insn, 1, temp, base, offset & 0xfff000);
988 base = temp;
989 }
990 if (offset & 0xfff) {
991 tcg_out_insn_3401(s, add_insn, 1, temp, base, offset & 0xfff);
992 base = temp;
993 }
994 }
995 tcg_out_insn(s, 3303, LD1R, type == TCG_TYPE_V128, r, base, vece);
996 return true;
997}
998
999static void tcg_out_movi(TCGContext *s, TCGType type, TCGReg rd,
1000 tcg_target_long value)
1001{
1002 tcg_target_long svalue = value;
1003 tcg_target_long ivalue = ~value;
1004 tcg_target_long t0, t1, t2;
1005 int s0, s1;
1006 AArch64Insn opc;
1007
1008 switch (type) {
1009 case TCG_TYPE_I32:
1010 case TCG_TYPE_I64:
1011 tcg_debug_assert(rd < 32);
1012 break;
1013
1014 case TCG_TYPE_V64:
1015 case TCG_TYPE_V128:
1016 tcg_debug_assert(rd >= 32);
1017 tcg_out_dupi_vec(s, type, rd, value);
1018 return;
1019
1020 default:
1021 g_assert_not_reached();
1022 }
1023
1024
1025
1026
1027
1028 if (type == TCG_TYPE_I32 || (value & ~0xffffffffull) == 0) {
1029 svalue = (int32_t)value;
1030 value = (uint32_t)value;
1031 ivalue = (uint32_t)ivalue;
1032 type = TCG_TYPE_I32;
1033 }
1034
1035
1036
1037 if ((value & ~0xffffull) == 0) {
1038 tcg_out_insn(s, 3405, MOVZ, type, rd, value, 0);
1039 return;
1040 } else if ((ivalue & ~0xffffull) == 0) {
1041 tcg_out_insn(s, 3405, MOVN, type, rd, ivalue, 0);
1042 return;
1043 }
1044
1045
1046
1047
1048 if (is_limm(svalue)) {
1049 tcg_out_logicali(s, I3404_ORRI, type, rd, TCG_REG_XZR, svalue);
1050 return;
1051 }
1052
1053
1054
1055 if (type == TCG_TYPE_I64) {
1056 tcg_target_long disp = value - (intptr_t)s->code_ptr;
1057 if (disp == sextract64(disp, 0, 21)) {
1058 tcg_out_insn(s, 3406, ADR, rd, disp);
1059 return;
1060 }
1061 disp = (value >> 12) - ((intptr_t)s->code_ptr >> 12);
1062 if (disp == sextract64(disp, 0, 21)) {
1063 tcg_out_insn(s, 3406, ADRP, rd, disp);
1064 if (value & 0xfff) {
1065 tcg_out_insn(s, 3401, ADDI, type, rd, rd, value & 0xfff);
1066 }
1067 return;
1068 }
1069 }
1070
1071
1072 if (ctpop64(value) >= 32) {
1073 t0 = ivalue;
1074 opc = I3405_MOVN;
1075 } else {
1076 t0 = value;
1077 opc = I3405_MOVZ;
1078 }
1079 s0 = ctz64(t0) & (63 & -16);
1080 t1 = t0 & ~(0xffffUL << s0);
1081 s1 = ctz64(t1) & (63 & -16);
1082 t2 = t1 & ~(0xffffUL << s1);
1083 if (t2 == 0) {
1084 tcg_out_insn_3405(s, opc, type, rd, t0 >> s0, s0);
1085 if (t1 != 0) {
1086 tcg_out_insn(s, 3405, MOVK, type, rd, value >> s1, s1);
1087 }
1088 return;
1089 }
1090
1091
1092 new_pool_label(s, value, R_AARCH64_CONDBR19, s->code_ptr, 0);
1093 tcg_out_insn(s, 3305, LDR, 0, rd);
1094}
1095
1096
1097#define tcg_out_ldst_r tcg_out_insn_3310
1098
1099static void tcg_out_ldst(TCGContext *s, AArch64Insn insn, TCGReg rd,
1100 TCGReg rn, intptr_t offset, int lgsize)
1101{
1102
1103
1104 if (offset >= 0 && !(offset & ((1 << lgsize) - 1))) {
1105 uintptr_t scaled_uimm = offset >> lgsize;
1106 if (scaled_uimm <= 0xfff) {
1107 tcg_out_insn_3313(s, insn, rd, rn, scaled_uimm);
1108 return;
1109 }
1110 }
1111
1112
1113 if (offset >= -256 && offset < 256) {
1114 tcg_out_insn_3312(s, insn, rd, rn, offset);
1115 return;
1116 }
1117
1118
1119 tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_TMP, offset);
1120 tcg_out_ldst_r(s, insn, rd, rn, TCG_TYPE_I64, TCG_REG_TMP);
1121}
1122
1123static bool tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg)
1124{
1125 if (ret == arg) {
1126 return true;
1127 }
1128 switch (type) {
1129 case TCG_TYPE_I32:
1130 case TCG_TYPE_I64:
1131 if (ret < 32 && arg < 32) {
1132 tcg_out_movr(s, type, ret, arg);
1133 break;
1134 } else if (ret < 32) {
1135 tcg_out_insn(s, 3605, UMOV, type, ret, arg, 0, 0);
1136 break;
1137 } else if (arg < 32) {
1138 tcg_out_insn(s, 3605, INS, 0, ret, arg, 4 << type, 0);
1139 break;
1140 }
1141
1142
1143 case TCG_TYPE_V64:
1144 tcg_debug_assert(ret >= 32 && arg >= 32);
1145 tcg_out_insn(s, 3616, ORR, 0, 0, ret, arg, arg);
1146 break;
1147 case TCG_TYPE_V128:
1148 tcg_debug_assert(ret >= 32 && arg >= 32);
1149 tcg_out_insn(s, 3616, ORR, 1, 0, ret, arg, arg);
1150 break;
1151
1152 default:
1153 g_assert_not_reached();
1154 }
1155 return true;
1156}
1157
1158static void tcg_out_ld(TCGContext *s, TCGType type, TCGReg ret,
1159 TCGReg base, intptr_t ofs)
1160{
1161 AArch64Insn insn;
1162 int lgsz;
1163
1164 switch (type) {
1165 case TCG_TYPE_I32:
1166 insn = (ret < 32 ? I3312_LDRW : I3312_LDRVS);
1167 lgsz = 2;
1168 break;
1169 case TCG_TYPE_I64:
1170 insn = (ret < 32 ? I3312_LDRX : I3312_LDRVD);
1171 lgsz = 3;
1172 break;
1173 case TCG_TYPE_V64:
1174 insn = I3312_LDRVD;
1175 lgsz = 3;
1176 break;
1177 case TCG_TYPE_V128:
1178 insn = I3312_LDRVQ;
1179 lgsz = 4;
1180 break;
1181 default:
1182 g_assert_not_reached();
1183 }
1184 tcg_out_ldst(s, insn, ret, base, ofs, lgsz);
1185}
1186
1187static void tcg_out_st(TCGContext *s, TCGType type, TCGReg src,
1188 TCGReg base, intptr_t ofs)
1189{
1190 AArch64Insn insn;
1191 int lgsz;
1192
1193 switch (type) {
1194 case TCG_TYPE_I32:
1195 insn = (src < 32 ? I3312_STRW : I3312_STRVS);
1196 lgsz = 2;
1197 break;
1198 case TCG_TYPE_I64:
1199 insn = (src < 32 ? I3312_STRX : I3312_STRVD);
1200 lgsz = 3;
1201 break;
1202 case TCG_TYPE_V64:
1203 insn = I3312_STRVD;
1204 lgsz = 3;
1205 break;
1206 case TCG_TYPE_V128:
1207 insn = I3312_STRVQ;
1208 lgsz = 4;
1209 break;
1210 default:
1211 g_assert_not_reached();
1212 }
1213 tcg_out_ldst(s, insn, src, base, ofs, lgsz);
1214}
1215
1216static inline bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val,
1217 TCGReg base, intptr_t ofs)
1218{
1219 if (type <= TCG_TYPE_I64 && val == 0) {
1220 tcg_out_st(s, type, TCG_REG_XZR, base, ofs);
1221 return true;
1222 }
1223 return false;
1224}
1225
1226static inline void tcg_out_bfm(TCGContext *s, TCGType ext, TCGReg rd,
1227 TCGReg rn, unsigned int a, unsigned int b)
1228{
1229 tcg_out_insn(s, 3402, BFM, ext, rd, rn, ext, a, b);
1230}
1231
1232static inline void tcg_out_ubfm(TCGContext *s, TCGType ext, TCGReg rd,
1233 TCGReg rn, unsigned int a, unsigned int b)
1234{
1235 tcg_out_insn(s, 3402, UBFM, ext, rd, rn, ext, a, b);
1236}
1237
1238static inline void tcg_out_sbfm(TCGContext *s, TCGType ext, TCGReg rd,
1239 TCGReg rn, unsigned int a, unsigned int b)
1240{
1241 tcg_out_insn(s, 3402, SBFM, ext, rd, rn, ext, a, b);
1242}
1243
1244static inline void tcg_out_extr(TCGContext *s, TCGType ext, TCGReg rd,
1245 TCGReg rn, TCGReg rm, unsigned int a)
1246{
1247 tcg_out_insn(s, 3403, EXTR, ext, rd, rn, rm, a);
1248}
1249
1250static inline void tcg_out_shl(TCGContext *s, TCGType ext,
1251 TCGReg rd, TCGReg rn, unsigned int m)
1252{
1253 int bits = ext ? 64 : 32;
1254 int max = bits - 1;
1255 tcg_out_ubfm(s, ext, rd, rn, bits - (m & max), max - (m & max));
1256}
1257
1258static inline void tcg_out_shr(TCGContext *s, TCGType ext,
1259 TCGReg rd, TCGReg rn, unsigned int m)
1260{
1261 int max = ext ? 63 : 31;
1262 tcg_out_ubfm(s, ext, rd, rn, m & max, max);
1263}
1264
1265static inline void tcg_out_sar(TCGContext *s, TCGType ext,
1266 TCGReg rd, TCGReg rn, unsigned int m)
1267{
1268 int max = ext ? 63 : 31;
1269 tcg_out_sbfm(s, ext, rd, rn, m & max, max);
1270}
1271
1272static inline void tcg_out_rotr(TCGContext *s, TCGType ext,
1273 TCGReg rd, TCGReg rn, unsigned int m)
1274{
1275 int max = ext ? 63 : 31;
1276 tcg_out_extr(s, ext, rd, rn, rn, m & max);
1277}
1278
1279static inline void tcg_out_rotl(TCGContext *s, TCGType ext,
1280 TCGReg rd, TCGReg rn, unsigned int m)
1281{
1282 int bits = ext ? 64 : 32;
1283 int max = bits - 1;
1284 tcg_out_extr(s, ext, rd, rn, rn, bits - (m & max));
1285}
1286
1287static inline void tcg_out_dep(TCGContext *s, TCGType ext, TCGReg rd,
1288 TCGReg rn, unsigned lsb, unsigned width)
1289{
1290 unsigned size = ext ? 64 : 32;
1291 unsigned a = (size - lsb) & (size - 1);
1292 unsigned b = width - 1;
1293 tcg_out_bfm(s, ext, rd, rn, a, b);
1294}
1295
1296static void tcg_out_cmp(TCGContext *s, TCGType ext, TCGReg a,
1297 tcg_target_long b, bool const_b)
1298{
1299 if (const_b) {
1300
1301 if (b >= 0) {
1302 tcg_out_insn(s, 3401, SUBSI, ext, TCG_REG_XZR, a, b);
1303 } else {
1304 tcg_out_insn(s, 3401, ADDSI, ext, TCG_REG_XZR, a, -b);
1305 }
1306 } else {
1307
1308 tcg_out_insn(s, 3502, SUBS, ext, TCG_REG_XZR, a, b);
1309 }
1310}
1311
1312static inline void tcg_out_goto(TCGContext *s, tcg_insn_unit *target)
1313{
1314 ptrdiff_t offset = target - s->code_ptr;
1315 tcg_debug_assert(offset == sextract64(offset, 0, 26));
1316 tcg_out_insn(s, 3206, B, offset);
1317}
1318
1319static inline void tcg_out_goto_long(TCGContext *s, tcg_insn_unit *target)
1320{
1321 ptrdiff_t offset = target - s->code_ptr;
1322 if (offset == sextract64(offset, 0, 26)) {
1323 tcg_out_insn(s, 3206, BL, offset);
1324 } else {
1325 tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_TMP, (intptr_t)target);
1326 tcg_out_insn(s, 3207, BR, TCG_REG_TMP);
1327 }
1328}
1329
1330static inline void tcg_out_callr(TCGContext *s, TCGReg reg)
1331{
1332 tcg_out_insn(s, 3207, BLR, reg);
1333}
1334
1335static inline void tcg_out_call(TCGContext *s, tcg_insn_unit *target)
1336{
1337 ptrdiff_t offset = target - s->code_ptr;
1338 if (offset == sextract64(offset, 0, 26)) {
1339 tcg_out_insn(s, 3206, BL, offset);
1340 } else {
1341 tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_TMP, (intptr_t)target);
1342 tcg_out_callr(s, TCG_REG_TMP);
1343 }
1344}
1345
1346void tb_target_set_jmp_target(uintptr_t tc_ptr, uintptr_t jmp_addr,
1347 uintptr_t addr)
1348{
1349 tcg_insn_unit i1, i2;
1350 TCGType rt = TCG_TYPE_I64;
1351 TCGReg rd = TCG_REG_TMP;
1352 uint64_t pair;
1353
1354 ptrdiff_t offset = addr - jmp_addr;
1355
1356 if (offset == sextract64(offset, 0, 26)) {
1357 i1 = I3206_B | ((offset >> 2) & 0x3ffffff);
1358 i2 = NOP;
1359 } else {
1360 offset = (addr >> 12) - (jmp_addr >> 12);
1361
1362
1363 i1 = I3406_ADRP | (offset & 3) << 29 | (offset & 0x1ffffc) << (5 - 2) | rd;
1364
1365 i2 = I3401_ADDI | rt << 31 | (addr & 0xfff) << 10 | rd << 5 | rd;
1366 }
1367 pair = (uint64_t)i2 << 32 | i1;
1368 atomic_set((uint64_t *)jmp_addr, pair);
1369 flush_icache_range(jmp_addr, jmp_addr + 8);
1370}
1371
1372static inline void tcg_out_goto_label(TCGContext *s, TCGLabel *l)
1373{
1374 if (!l->has_value) {
1375 tcg_out_reloc(s, s->code_ptr, R_AARCH64_JUMP26, l, 0);
1376 tcg_out_insn(s, 3206, B, 0);
1377 } else {
1378 tcg_out_goto(s, l->u.value_ptr);
1379 }
1380}
1381
1382static void tcg_out_brcond(TCGContext *s, TCGType ext, TCGCond c, TCGArg a,
1383 TCGArg b, bool b_const, TCGLabel *l)
1384{
1385 intptr_t offset;
1386 bool need_cmp;
1387
1388 if (b_const && b == 0 && (c == TCG_COND_EQ || c == TCG_COND_NE)) {
1389 need_cmp = false;
1390 } else {
1391 need_cmp = true;
1392 tcg_out_cmp(s, ext, a, b, b_const);
1393 }
1394
1395 if (!l->has_value) {
1396 tcg_out_reloc(s, s->code_ptr, R_AARCH64_CONDBR19, l, 0);
1397 offset = tcg_in32(s) >> 5;
1398 } else {
1399 offset = l->u.value_ptr - s->code_ptr;
1400 tcg_debug_assert(offset == sextract64(offset, 0, 19));
1401 }
1402
1403 if (need_cmp) {
1404 tcg_out_insn(s, 3202, B_C, c, offset);
1405 } else if (c == TCG_COND_EQ) {
1406 tcg_out_insn(s, 3201, CBZ, ext, a, offset);
1407 } else {
1408 tcg_out_insn(s, 3201, CBNZ, ext, a, offset);
1409 }
1410}
1411
1412static inline void tcg_out_rev64(TCGContext *s, TCGReg rd, TCGReg rn)
1413{
1414 tcg_out_insn(s, 3507, REV64, TCG_TYPE_I64, rd, rn);
1415}
1416
1417static inline void tcg_out_rev32(TCGContext *s, TCGReg rd, TCGReg rn)
1418{
1419 tcg_out_insn(s, 3507, REV32, TCG_TYPE_I32, rd, rn);
1420}
1421
1422static inline void tcg_out_rev16(TCGContext *s, TCGReg rd, TCGReg rn)
1423{
1424 tcg_out_insn(s, 3507, REV16, TCG_TYPE_I32, rd, rn);
1425}
1426
1427static inline void tcg_out_sxt(TCGContext *s, TCGType ext, MemOp s_bits,
1428 TCGReg rd, TCGReg rn)
1429{
1430
1431 int bits = (8 << s_bits) - 1;
1432 tcg_out_sbfm(s, ext, rd, rn, 0, bits);
1433}
1434
1435static inline void tcg_out_uxt(TCGContext *s, MemOp s_bits,
1436 TCGReg rd, TCGReg rn)
1437{
1438
1439 int bits = (8 << s_bits) - 1;
1440 tcg_out_ubfm(s, 0, rd, rn, 0, bits);
1441}
1442
1443static void tcg_out_addsubi(TCGContext *s, int ext, TCGReg rd,
1444 TCGReg rn, int64_t aimm)
1445{
1446 if (aimm >= 0) {
1447 tcg_out_insn(s, 3401, ADDI, ext, rd, rn, aimm);
1448 } else {
1449 tcg_out_insn(s, 3401, SUBI, ext, rd, rn, -aimm);
1450 }
1451}
1452
1453static inline void tcg_out_addsub2(TCGContext *s, TCGType ext, TCGReg rl,
1454 TCGReg rh, TCGReg al, TCGReg ah,
1455 tcg_target_long bl, tcg_target_long bh,
1456 bool const_bl, bool const_bh, bool sub)
1457{
1458 TCGReg orig_rl = rl;
1459 AArch64Insn insn;
1460
1461 if (rl == ah || (!const_bh && rl == bh)) {
1462 rl = TCG_REG_TMP;
1463 }
1464
1465 if (const_bl) {
1466 insn = I3401_ADDSI;
1467 if ((bl < 0) ^ sub) {
1468 insn = I3401_SUBSI;
1469 bl = -bl;
1470 }
1471 if (unlikely(al == TCG_REG_XZR)) {
1472
1473
1474
1475
1476
1477 al = TCG_REG_TMP;
1478 tcg_out_movi(s, ext, al, 0);
1479 }
1480 tcg_out_insn_3401(s, insn, ext, rl, al, bl);
1481 } else {
1482 tcg_out_insn_3502(s, sub ? I3502_SUBS : I3502_ADDS, ext, rl, al, bl);
1483 }
1484
1485 insn = I3503_ADC;
1486 if (const_bh) {
1487
1488
1489 if ((bh != 0) ^ sub) {
1490 insn = I3503_SBC;
1491 }
1492 bh = TCG_REG_XZR;
1493 } else if (sub) {
1494 insn = I3503_SBC;
1495 }
1496 tcg_out_insn_3503(s, insn, ext, rh, ah, bh);
1497
1498 tcg_out_mov(s, ext, orig_rl, rl);
1499}
1500
1501static inline void tcg_out_mb(TCGContext *s, TCGArg a0)
1502{
1503 static const uint32_t sync[] = {
1504 [0 ... TCG_MO_ALL] = DMB_ISH | DMB_LD | DMB_ST,
1505 [TCG_MO_ST_ST] = DMB_ISH | DMB_ST,
1506 [TCG_MO_LD_LD] = DMB_ISH | DMB_LD,
1507 [TCG_MO_LD_ST] = DMB_ISH | DMB_LD,
1508 [TCG_MO_LD_ST | TCG_MO_LD_LD] = DMB_ISH | DMB_LD,
1509 };
1510 tcg_out32(s, sync[a0 & TCG_MO_ALL]);
1511}
1512
1513static void tcg_out_cltz(TCGContext *s, TCGType ext, TCGReg d,
1514 TCGReg a0, TCGArg b, bool const_b, bool is_ctz)
1515{
1516 TCGReg a1 = a0;
1517 if (is_ctz) {
1518 a1 = TCG_REG_TMP;
1519 tcg_out_insn(s, 3507, RBIT, ext, a1, a0);
1520 }
1521 if (const_b && b == (ext ? 64 : 32)) {
1522 tcg_out_insn(s, 3507, CLZ, ext, d, a1);
1523 } else {
1524 AArch64Insn sel = I3506_CSEL;
1525
1526 tcg_out_cmp(s, ext, a0, 0, 1);
1527 tcg_out_insn(s, 3507, CLZ, ext, TCG_REG_TMP, a1);
1528
1529 if (const_b) {
1530 if (b == -1) {
1531 b = TCG_REG_XZR;
1532 sel = I3506_CSINV;
1533 } else if (b == 0) {
1534 b = TCG_REG_XZR;
1535 } else {
1536 tcg_out_movi(s, ext, d, b);
1537 b = d;
1538 }
1539 }
1540 tcg_out_insn_3506(s, sel, ext, d, TCG_REG_TMP, b, TCG_COND_NE);
1541 }
1542}
1543
1544#ifdef CONFIG_SOFTMMU
1545#include "../tcg-ldst.inc.c"
1546
1547
1548
1549
1550static void * const qemu_ld_helpers[16] = {
1551 [MO_UB] = helper_ret_ldub_mmu,
1552 [MO_LEUW] = helper_le_lduw_mmu,
1553 [MO_LEUL] = helper_le_ldul_mmu,
1554 [MO_LEQ] = helper_le_ldq_mmu,
1555 [MO_BEUW] = helper_be_lduw_mmu,
1556 [MO_BEUL] = helper_be_ldul_mmu,
1557 [MO_BEQ] = helper_be_ldq_mmu,
1558};
1559
1560
1561
1562
1563
1564static void * const qemu_st_helpers[16] = {
1565 [MO_UB] = helper_ret_stb_mmu,
1566 [MO_LEUW] = helper_le_stw_mmu,
1567 [MO_LEUL] = helper_le_stl_mmu,
1568 [MO_LEQ] = helper_le_stq_mmu,
1569 [MO_BEUW] = helper_be_stw_mmu,
1570 [MO_BEUL] = helper_be_stl_mmu,
1571 [MO_BEQ] = helper_be_stq_mmu,
1572};
1573
1574static inline void tcg_out_adr(TCGContext *s, TCGReg rd, void *target)
1575{
1576 ptrdiff_t offset = tcg_pcrel_diff(s, target);
1577 tcg_debug_assert(offset == sextract64(offset, 0, 21));
1578 tcg_out_insn(s, 3406, ADR, rd, offset);
1579}
1580
1581static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
1582{
1583 TCGMemOpIdx oi = lb->oi;
1584 MemOp opc = get_memop(oi);
1585 MemOp size = opc & MO_SIZE;
1586
1587 if (!reloc_pc19(lb->label_ptr[0], s->code_ptr)) {
1588 return false;
1589 }
1590
1591 tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_X0, TCG_AREG0);
1592 tcg_out_mov(s, TARGET_LONG_BITS == 64, TCG_REG_X1, lb->addrlo_reg);
1593 tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_X2, oi);
1594 tcg_out_adr(s, TCG_REG_X3, lb->raddr);
1595 tcg_out_call(s, qemu_ld_helpers[opc & (MO_BSWAP | MO_SIZE)]);
1596 if (opc & MO_SIGN) {
1597 tcg_out_sxt(s, lb->type, size, lb->datalo_reg, TCG_REG_X0);
1598 } else {
1599 tcg_out_mov(s, size == MO_64, lb->datalo_reg, TCG_REG_X0);
1600 }
1601
1602 tcg_out_goto(s, lb->raddr);
1603 return true;
1604}
1605
1606static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
1607{
1608 TCGMemOpIdx oi = lb->oi;
1609 MemOp opc = get_memop(oi);
1610 MemOp size = opc & MO_SIZE;
1611
1612 if (!reloc_pc19(lb->label_ptr[0], s->code_ptr)) {
1613 return false;
1614 }
1615
1616 tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_X0, TCG_AREG0);
1617 tcg_out_mov(s, TARGET_LONG_BITS == 64, TCG_REG_X1, lb->addrlo_reg);
1618 tcg_out_mov(s, size == MO_64, TCG_REG_X2, lb->datalo_reg);
1619 tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_X3, oi);
1620 tcg_out_adr(s, TCG_REG_X4, lb->raddr);
1621 tcg_out_call(s, qemu_st_helpers[opc & (MO_BSWAP | MO_SIZE)]);
1622 tcg_out_goto(s, lb->raddr);
1623 return true;
1624}
1625
1626static void add_qemu_ldst_label(TCGContext *s, bool is_ld, TCGMemOpIdx oi,
1627 TCGType ext, TCGReg data_reg, TCGReg addr_reg,
1628 tcg_insn_unit *raddr, tcg_insn_unit *label_ptr)
1629{
1630 TCGLabelQemuLdst *label = new_ldst_label(s);
1631
1632 label->is_ld = is_ld;
1633 label->oi = oi;
1634 label->type = ext;
1635 label->datalo_reg = data_reg;
1636 label->addrlo_reg = addr_reg;
1637 label->raddr = raddr;
1638 label->label_ptr[0] = label_ptr;
1639}
1640
1641
1642QEMU_BUILD_BUG_ON(TLB_MASK_TABLE_OFS(0) > 0);
1643QEMU_BUILD_BUG_ON(TLB_MASK_TABLE_OFS(0) < -512);
1644
1645
1646QEMU_BUILD_BUG_ON(offsetof(CPUTLBDescFast, mask) != 0);
1647QEMU_BUILD_BUG_ON(offsetof(CPUTLBDescFast, table) != 8);
1648
1649
1650
1651
1652
1653static void tcg_out_tlb_read(TCGContext *s, TCGReg addr_reg, MemOp opc,
1654 tcg_insn_unit **label_ptr, int mem_index,
1655 bool is_read)
1656{
1657 unsigned a_bits = get_alignment_bits(opc);
1658 unsigned s_bits = opc & MO_SIZE;
1659 unsigned a_mask = (1u << a_bits) - 1;
1660 unsigned s_mask = (1u << s_bits) - 1;
1661 TCGReg x3;
1662 TCGType mask_type;
1663 uint64_t compare_mask;
1664
1665 mask_type = (TARGET_PAGE_BITS + CPU_TLB_DYN_MAX_BITS > 32
1666 ? TCG_TYPE_I64 : TCG_TYPE_I32);
1667
1668
1669 tcg_out_insn(s, 3314, LDP, TCG_REG_X0, TCG_REG_X1, TCG_AREG0,
1670 TLB_MASK_TABLE_OFS(mem_index), 1, 0);
1671
1672
1673 tcg_out_insn(s, 3502S, AND_LSR, mask_type == TCG_TYPE_I64,
1674 TCG_REG_X0, TCG_REG_X0, addr_reg,
1675 TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS);
1676
1677
1678 tcg_out_insn(s, 3502, ADD, 1, TCG_REG_X1, TCG_REG_X1, TCG_REG_X0);
1679
1680
1681 tcg_out_ld(s, TCG_TYPE_TL, TCG_REG_X0, TCG_REG_X1, is_read
1682 ? offsetof(CPUTLBEntry, addr_read)
1683 : offsetof(CPUTLBEntry, addr_write));
1684 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_X1, TCG_REG_X1,
1685 offsetof(CPUTLBEntry, addend));
1686
1687
1688
1689
1690 if (a_bits >= s_bits) {
1691 x3 = addr_reg;
1692 } else {
1693 tcg_out_insn(s, 3401, ADDI, TARGET_LONG_BITS == 64,
1694 TCG_REG_X3, addr_reg, s_mask - a_mask);
1695 x3 = TCG_REG_X3;
1696 }
1697 compare_mask = (uint64_t)TARGET_PAGE_MASK | a_mask;
1698
1699
1700 tcg_out_logicali(s, I3404_ANDI, TARGET_LONG_BITS == 64,
1701 TCG_REG_X3, x3, compare_mask);
1702
1703
1704 tcg_out_cmp(s, TARGET_LONG_BITS == 64, TCG_REG_X0, TCG_REG_X3, 0);
1705
1706
1707 *label_ptr = s->code_ptr;
1708 tcg_out_insn(s, 3202, B_C, TCG_COND_NE, 0);
1709}
1710
1711#endif
1712
1713static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp memop, TCGType ext,
1714 TCGReg data_r, TCGReg addr_r,
1715 TCGType otype, TCGReg off_r)
1716{
1717 const MemOp bswap = memop & MO_BSWAP;
1718
1719 switch (memop & MO_SSIZE) {
1720 case MO_UB:
1721 tcg_out_ldst_r(s, I3312_LDRB, data_r, addr_r, otype, off_r);
1722 break;
1723 case MO_SB:
1724 tcg_out_ldst_r(s, ext ? I3312_LDRSBX : I3312_LDRSBW,
1725 data_r, addr_r, otype, off_r);
1726 break;
1727 case MO_UW:
1728 tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
1729 if (bswap) {
1730 tcg_out_rev16(s, data_r, data_r);
1731 }
1732 break;
1733 case MO_SW:
1734 if (bswap) {
1735 tcg_out_ldst_r(s, I3312_LDRH, data_r, addr_r, otype, off_r);
1736 tcg_out_rev16(s, data_r, data_r);
1737 tcg_out_sxt(s, ext, MO_16, data_r, data_r);
1738 } else {
1739 tcg_out_ldst_r(s, (ext ? I3312_LDRSHX : I3312_LDRSHW),
1740 data_r, addr_r, otype, off_r);
1741 }
1742 break;
1743 case MO_UL:
1744 tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
1745 if (bswap) {
1746 tcg_out_rev32(s, data_r, data_r);
1747 }
1748 break;
1749 case MO_SL:
1750 if (bswap) {
1751 tcg_out_ldst_r(s, I3312_LDRW, data_r, addr_r, otype, off_r);
1752 tcg_out_rev32(s, data_r, data_r);
1753 tcg_out_sxt(s, TCG_TYPE_I64, MO_32, data_r, data_r);
1754 } else {
1755 tcg_out_ldst_r(s, I3312_LDRSWX, data_r, addr_r, otype, off_r);
1756 }
1757 break;
1758 case MO_Q:
1759 tcg_out_ldst_r(s, I3312_LDRX, data_r, addr_r, otype, off_r);
1760 if (bswap) {
1761 tcg_out_rev64(s, data_r, data_r);
1762 }
1763 break;
1764 default:
1765 tcg_abort();
1766 }
1767}
1768
1769static void tcg_out_qemu_st_direct(TCGContext *s, MemOp memop,
1770 TCGReg data_r, TCGReg addr_r,
1771 TCGType otype, TCGReg off_r)
1772{
1773 const MemOp bswap = memop & MO_BSWAP;
1774
1775 switch (memop & MO_SIZE) {
1776 case MO_8:
1777 tcg_out_ldst_r(s, I3312_STRB, data_r, addr_r, otype, off_r);
1778 break;
1779 case MO_16:
1780 if (bswap && data_r != TCG_REG_XZR) {
1781 tcg_out_rev16(s, TCG_REG_TMP, data_r);
1782 data_r = TCG_REG_TMP;
1783 }
1784 tcg_out_ldst_r(s, I3312_STRH, data_r, addr_r, otype, off_r);
1785 break;
1786 case MO_32:
1787 if (bswap && data_r != TCG_REG_XZR) {
1788 tcg_out_rev32(s, TCG_REG_TMP, data_r);
1789 data_r = TCG_REG_TMP;
1790 }
1791 tcg_out_ldst_r(s, I3312_STRW, data_r, addr_r, otype, off_r);
1792 break;
1793 case MO_64:
1794 if (bswap && data_r != TCG_REG_XZR) {
1795 tcg_out_rev64(s, TCG_REG_TMP, data_r);
1796 data_r = TCG_REG_TMP;
1797 }
1798 tcg_out_ldst_r(s, I3312_STRX, data_r, addr_r, otype, off_r);
1799 break;
1800 default:
1801 tcg_abort();
1802 }
1803}
1804
1805static void tcg_out_qemu_ld(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
1806 TCGMemOpIdx oi, TCGType ext)
1807{
1808 MemOp memop = get_memop(oi);
1809 const TCGType otype = TARGET_LONG_BITS == 64 ? TCG_TYPE_I64 : TCG_TYPE_I32;
1810#ifdef CONFIG_SOFTMMU
1811 unsigned mem_index = get_mmuidx(oi);
1812 tcg_insn_unit *label_ptr;
1813
1814 tcg_out_tlb_read(s, addr_reg, memop, &label_ptr, mem_index, 1);
1815 tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
1816 TCG_REG_X1, otype, addr_reg);
1817 add_qemu_ldst_label(s, true, oi, ext, data_reg, addr_reg,
1818 s->code_ptr, label_ptr);
1819#else
1820 if (USE_GUEST_BASE) {
1821 tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
1822 TCG_REG_GUEST_BASE, otype, addr_reg);
1823 } else {
1824 tcg_out_qemu_ld_direct(s, memop, ext, data_reg,
1825 addr_reg, TCG_TYPE_I64, TCG_REG_XZR);
1826 }
1827#endif
1828}
1829
1830static void tcg_out_qemu_st(TCGContext *s, TCGReg data_reg, TCGReg addr_reg,
1831 TCGMemOpIdx oi)
1832{
1833 MemOp memop = get_memop(oi);
1834 const TCGType otype = TARGET_LONG_BITS == 64 ? TCG_TYPE_I64 : TCG_TYPE_I32;
1835#ifdef CONFIG_SOFTMMU
1836 unsigned mem_index = get_mmuidx(oi);
1837 tcg_insn_unit *label_ptr;
1838
1839 tcg_out_tlb_read(s, addr_reg, memop, &label_ptr, mem_index, 0);
1840 tcg_out_qemu_st_direct(s, memop, data_reg,
1841 TCG_REG_X1, otype, addr_reg);
1842 add_qemu_ldst_label(s, false, oi, (memop & MO_SIZE)== MO_64,
1843 data_reg, addr_reg, s->code_ptr, label_ptr);
1844#else
1845 if (USE_GUEST_BASE) {
1846 tcg_out_qemu_st_direct(s, memop, data_reg,
1847 TCG_REG_GUEST_BASE, otype, addr_reg);
1848 } else {
1849 tcg_out_qemu_st_direct(s, memop, data_reg,
1850 addr_reg, TCG_TYPE_I64, TCG_REG_XZR);
1851 }
1852#endif
1853}
1854
1855static tcg_insn_unit *tb_ret_addr;
1856
1857static void tcg_out_op(TCGContext *s, TCGOpcode opc,
1858 const TCGArg args[TCG_MAX_OP_ARGS],
1859 const int const_args[TCG_MAX_OP_ARGS])
1860{
1861
1862
1863 TCGType ext = (tcg_op_defs[opc].flags & TCG_OPF_64BIT) != 0;
1864
1865
1866 TCGArg a0 = args[0];
1867 TCGArg a1 = args[1];
1868 TCGArg a2 = args[2];
1869 int c2 = const_args[2];
1870
1871
1872
1873#define REG0(I) (const_args[I] ? TCG_REG_XZR : (TCGReg)args[I])
1874
1875 switch (opc) {
1876 case INDEX_op_exit_tb:
1877
1878 if (a0 == 0) {
1879 tcg_out_goto_long(s, s->code_gen_epilogue);
1880 } else {
1881 tcg_out_movi(s, TCG_TYPE_I64, TCG_REG_X0, a0);
1882 tcg_out_goto_long(s, tb_ret_addr);
1883 }
1884 break;
1885
1886 case INDEX_op_goto_tb:
1887 if (s->tb_jmp_insn_offset != NULL) {
1888
1889
1890
1891 if ((uintptr_t)s->code_ptr & 7) {
1892 tcg_out32(s, NOP);
1893 }
1894 s->tb_jmp_insn_offset[a0] = tcg_current_code_size(s);
1895
1896
1897 tcg_out_insn(s, 3406, ADRP, TCG_REG_TMP, 0);
1898 tcg_out_insn(s, 3401, ADDI, TCG_TYPE_I64, TCG_REG_TMP, TCG_REG_TMP, 0);
1899 } else {
1900
1901 tcg_debug_assert(s->tb_jmp_target_addr != NULL);
1902 intptr_t offset = tcg_pcrel_diff(s, (s->tb_jmp_target_addr + a0)) >> 2;
1903 tcg_out_insn(s, 3305, LDR, offset, TCG_REG_TMP);
1904 }
1905 tcg_out_insn(s, 3207, BR, TCG_REG_TMP);
1906 set_jmp_reset_offset(s, a0);
1907 break;
1908
1909 case INDEX_op_goto_ptr:
1910 tcg_out_insn(s, 3207, BR, a0);
1911 break;
1912
1913 case INDEX_op_br:
1914 tcg_out_goto_label(s, arg_label(a0));
1915 break;
1916
1917 case INDEX_op_ld8u_i32:
1918 case INDEX_op_ld8u_i64:
1919 tcg_out_ldst(s, I3312_LDRB, a0, a1, a2, 0);
1920 break;
1921 case INDEX_op_ld8s_i32:
1922 tcg_out_ldst(s, I3312_LDRSBW, a0, a1, a2, 0);
1923 break;
1924 case INDEX_op_ld8s_i64:
1925 tcg_out_ldst(s, I3312_LDRSBX, a0, a1, a2, 0);
1926 break;
1927 case INDEX_op_ld16u_i32:
1928 case INDEX_op_ld16u_i64:
1929 tcg_out_ldst(s, I3312_LDRH, a0, a1, a2, 1);
1930 break;
1931 case INDEX_op_ld16s_i32:
1932 tcg_out_ldst(s, I3312_LDRSHW, a0, a1, a2, 1);
1933 break;
1934 case INDEX_op_ld16s_i64:
1935 tcg_out_ldst(s, I3312_LDRSHX, a0, a1, a2, 1);
1936 break;
1937 case INDEX_op_ld_i32:
1938 case INDEX_op_ld32u_i64:
1939 tcg_out_ldst(s, I3312_LDRW, a0, a1, a2, 2);
1940 break;
1941 case INDEX_op_ld32s_i64:
1942 tcg_out_ldst(s, I3312_LDRSWX, a0, a1, a2, 2);
1943 break;
1944 case INDEX_op_ld_i64:
1945 tcg_out_ldst(s, I3312_LDRX, a0, a1, a2, 3);
1946 break;
1947
1948 case INDEX_op_st8_i32:
1949 case INDEX_op_st8_i64:
1950 tcg_out_ldst(s, I3312_STRB, REG0(0), a1, a2, 0);
1951 break;
1952 case INDEX_op_st16_i32:
1953 case INDEX_op_st16_i64:
1954 tcg_out_ldst(s, I3312_STRH, REG0(0), a1, a2, 1);
1955 break;
1956 case INDEX_op_st_i32:
1957 case INDEX_op_st32_i64:
1958 tcg_out_ldst(s, I3312_STRW, REG0(0), a1, a2, 2);
1959 break;
1960 case INDEX_op_st_i64:
1961 tcg_out_ldst(s, I3312_STRX, REG0(0), a1, a2, 3);
1962 break;
1963
1964 case INDEX_op_add_i32:
1965 a2 = (int32_t)a2;
1966
1967 case INDEX_op_add_i64:
1968 if (c2) {
1969 tcg_out_addsubi(s, ext, a0, a1, a2);
1970 } else {
1971 tcg_out_insn(s, 3502, ADD, ext, a0, a1, a2);
1972 }
1973 break;
1974
1975 case INDEX_op_sub_i32:
1976 a2 = (int32_t)a2;
1977
1978 case INDEX_op_sub_i64:
1979 if (c2) {
1980 tcg_out_addsubi(s, ext, a0, a1, -a2);
1981 } else {
1982 tcg_out_insn(s, 3502, SUB, ext, a0, a1, a2);
1983 }
1984 break;
1985
1986 case INDEX_op_neg_i64:
1987 case INDEX_op_neg_i32:
1988 tcg_out_insn(s, 3502, SUB, ext, a0, TCG_REG_XZR, a1);
1989 break;
1990
1991 case INDEX_op_and_i32:
1992 a2 = (int32_t)a2;
1993
1994 case INDEX_op_and_i64:
1995 if (c2) {
1996 tcg_out_logicali(s, I3404_ANDI, ext, a0, a1, a2);
1997 } else {
1998 tcg_out_insn(s, 3510, AND, ext, a0, a1, a2);
1999 }
2000 break;
2001
2002 case INDEX_op_andc_i32:
2003 a2 = (int32_t)a2;
2004
2005 case INDEX_op_andc_i64:
2006 if (c2) {
2007 tcg_out_logicali(s, I3404_ANDI, ext, a0, a1, ~a2);
2008 } else {
2009 tcg_out_insn(s, 3510, BIC, ext, a0, a1, a2);
2010 }
2011 break;
2012
2013 case INDEX_op_or_i32:
2014 a2 = (int32_t)a2;
2015
2016 case INDEX_op_or_i64:
2017 if (c2) {
2018 tcg_out_logicali(s, I3404_ORRI, ext, a0, a1, a2);
2019 } else {
2020 tcg_out_insn(s, 3510, ORR, ext, a0, a1, a2);
2021 }
2022 break;
2023
2024 case INDEX_op_orc_i32:
2025 a2 = (int32_t)a2;
2026
2027 case INDEX_op_orc_i64:
2028 if (c2) {
2029 tcg_out_logicali(s, I3404_ORRI, ext, a0, a1, ~a2);
2030 } else {
2031 tcg_out_insn(s, 3510, ORN, ext, a0, a1, a2);
2032 }
2033 break;
2034
2035 case INDEX_op_xor_i32:
2036 a2 = (int32_t)a2;
2037
2038 case INDEX_op_xor_i64:
2039 if (c2) {
2040 tcg_out_logicali(s, I3404_EORI, ext, a0, a1, a2);
2041 } else {
2042 tcg_out_insn(s, 3510, EOR, ext, a0, a1, a2);
2043 }
2044 break;
2045
2046 case INDEX_op_eqv_i32:
2047 a2 = (int32_t)a2;
2048
2049 case INDEX_op_eqv_i64:
2050 if (c2) {
2051 tcg_out_logicali(s, I3404_EORI, ext, a0, a1, ~a2);
2052 } else {
2053 tcg_out_insn(s, 3510, EON, ext, a0, a1, a2);
2054 }
2055 break;
2056
2057 case INDEX_op_not_i64:
2058 case INDEX_op_not_i32:
2059 tcg_out_insn(s, 3510, ORN, ext, a0, TCG_REG_XZR, a1);
2060 break;
2061
2062 case INDEX_op_mul_i64:
2063 case INDEX_op_mul_i32:
2064 tcg_out_insn(s, 3509, MADD, ext, a0, a1, a2, TCG_REG_XZR);
2065 break;
2066
2067 case INDEX_op_div_i64:
2068 case INDEX_op_div_i32:
2069 tcg_out_insn(s, 3508, SDIV, ext, a0, a1, a2);
2070 break;
2071 case INDEX_op_divu_i64:
2072 case INDEX_op_divu_i32:
2073 tcg_out_insn(s, 3508, UDIV, ext, a0, a1, a2);
2074 break;
2075
2076 case INDEX_op_rem_i64:
2077 case INDEX_op_rem_i32:
2078 tcg_out_insn(s, 3508, SDIV, ext, TCG_REG_TMP, a1, a2);
2079 tcg_out_insn(s, 3509, MSUB, ext, a0, TCG_REG_TMP, a2, a1);
2080 break;
2081 case INDEX_op_remu_i64:
2082 case INDEX_op_remu_i32:
2083 tcg_out_insn(s, 3508, UDIV, ext, TCG_REG_TMP, a1, a2);
2084 tcg_out_insn(s, 3509, MSUB, ext, a0, TCG_REG_TMP, a2, a1);
2085 break;
2086
2087 case INDEX_op_shl_i64:
2088 case INDEX_op_shl_i32:
2089 if (c2) {
2090 tcg_out_shl(s, ext, a0, a1, a2);
2091 } else {
2092 tcg_out_insn(s, 3508, LSLV, ext, a0, a1, a2);
2093 }
2094 break;
2095
2096 case INDEX_op_shr_i64:
2097 case INDEX_op_shr_i32:
2098 if (c2) {
2099 tcg_out_shr(s, ext, a0, a1, a2);
2100 } else {
2101 tcg_out_insn(s, 3508, LSRV, ext, a0, a1, a2);
2102 }
2103 break;
2104
2105 case INDEX_op_sar_i64:
2106 case INDEX_op_sar_i32:
2107 if (c2) {
2108 tcg_out_sar(s, ext, a0, a1, a2);
2109 } else {
2110 tcg_out_insn(s, 3508, ASRV, ext, a0, a1, a2);
2111 }
2112 break;
2113
2114 case INDEX_op_rotr_i64:
2115 case INDEX_op_rotr_i32:
2116 if (c2) {
2117 tcg_out_rotr(s, ext, a0, a1, a2);
2118 } else {
2119 tcg_out_insn(s, 3508, RORV, ext, a0, a1, a2);
2120 }
2121 break;
2122
2123 case INDEX_op_rotl_i64:
2124 case INDEX_op_rotl_i32:
2125 if (c2) {
2126 tcg_out_rotl(s, ext, a0, a1, a2);
2127 } else {
2128 tcg_out_insn(s, 3502, SUB, 0, TCG_REG_TMP, TCG_REG_XZR, a2);
2129 tcg_out_insn(s, 3508, RORV, ext, a0, a1, TCG_REG_TMP);
2130 }
2131 break;
2132
2133 case INDEX_op_clz_i64:
2134 case INDEX_op_clz_i32:
2135 tcg_out_cltz(s, ext, a0, a1, a2, c2, false);
2136 break;
2137 case INDEX_op_ctz_i64:
2138 case INDEX_op_ctz_i32:
2139 tcg_out_cltz(s, ext, a0, a1, a2, c2, true);
2140 break;
2141
2142 case INDEX_op_brcond_i32:
2143 a1 = (int32_t)a1;
2144
2145 case INDEX_op_brcond_i64:
2146 tcg_out_brcond(s, ext, a2, a0, a1, const_args[1], arg_label(args[3]));
2147 break;
2148
2149 case INDEX_op_setcond_i32:
2150 a2 = (int32_t)a2;
2151
2152 case INDEX_op_setcond_i64:
2153 tcg_out_cmp(s, ext, a1, a2, c2);
2154
2155 tcg_out_insn(s, 3506, CSINC, TCG_TYPE_I32, a0, TCG_REG_XZR,
2156 TCG_REG_XZR, tcg_invert_cond(args[3]));
2157 break;
2158
2159 case INDEX_op_movcond_i32:
2160 a2 = (int32_t)a2;
2161
2162 case INDEX_op_movcond_i64:
2163 tcg_out_cmp(s, ext, a1, a2, c2);
2164 tcg_out_insn(s, 3506, CSEL, ext, a0, REG0(3), REG0(4), args[5]);
2165 break;
2166
2167 case INDEX_op_qemu_ld_i32:
2168 case INDEX_op_qemu_ld_i64:
2169 tcg_out_qemu_ld(s, a0, a1, a2, ext);
2170 break;
2171 case INDEX_op_qemu_st_i32:
2172 case INDEX_op_qemu_st_i64:
2173 tcg_out_qemu_st(s, REG0(0), a1, a2);
2174 break;
2175
2176 case INDEX_op_bswap64_i64:
2177 tcg_out_rev64(s, a0, a1);
2178 break;
2179 case INDEX_op_bswap32_i64:
2180 case INDEX_op_bswap32_i32:
2181 tcg_out_rev32(s, a0, a1);
2182 break;
2183 case INDEX_op_bswap16_i64:
2184 case INDEX_op_bswap16_i32:
2185 tcg_out_rev16(s, a0, a1);
2186 break;
2187
2188 case INDEX_op_ext8s_i64:
2189 case INDEX_op_ext8s_i32:
2190 tcg_out_sxt(s, ext, MO_8, a0, a1);
2191 break;
2192 case INDEX_op_ext16s_i64:
2193 case INDEX_op_ext16s_i32:
2194 tcg_out_sxt(s, ext, MO_16, a0, a1);
2195 break;
2196 case INDEX_op_ext_i32_i64:
2197 case INDEX_op_ext32s_i64:
2198 tcg_out_sxt(s, TCG_TYPE_I64, MO_32, a0, a1);
2199 break;
2200 case INDEX_op_ext8u_i64:
2201 case INDEX_op_ext8u_i32:
2202 tcg_out_uxt(s, MO_8, a0, a1);
2203 break;
2204 case INDEX_op_ext16u_i64:
2205 case INDEX_op_ext16u_i32:
2206 tcg_out_uxt(s, MO_16, a0, a1);
2207 break;
2208 case INDEX_op_extu_i32_i64:
2209 case INDEX_op_ext32u_i64:
2210 tcg_out_movr(s, TCG_TYPE_I32, a0, a1);
2211 break;
2212
2213 case INDEX_op_deposit_i64:
2214 case INDEX_op_deposit_i32:
2215 tcg_out_dep(s, ext, a0, REG0(2), args[3], args[4]);
2216 break;
2217
2218 case INDEX_op_extract_i64:
2219 case INDEX_op_extract_i32:
2220 tcg_out_ubfm(s, ext, a0, a1, a2, a2 + args[3] - 1);
2221 break;
2222
2223 case INDEX_op_sextract_i64:
2224 case INDEX_op_sextract_i32:
2225 tcg_out_sbfm(s, ext, a0, a1, a2, a2 + args[3] - 1);
2226 break;
2227
2228 case INDEX_op_extract2_i64:
2229 case INDEX_op_extract2_i32:
2230 tcg_out_extr(s, ext, a0, REG0(2), REG0(1), args[3]);
2231 break;
2232
2233 case INDEX_op_add2_i32:
2234 tcg_out_addsub2(s, TCG_TYPE_I32, a0, a1, REG0(2), REG0(3),
2235 (int32_t)args[4], args[5], const_args[4],
2236 const_args[5], false);
2237 break;
2238 case INDEX_op_add2_i64:
2239 tcg_out_addsub2(s, TCG_TYPE_I64, a0, a1, REG0(2), REG0(3), args[4],
2240 args[5], const_args[4], const_args[5], false);
2241 break;
2242 case INDEX_op_sub2_i32:
2243 tcg_out_addsub2(s, TCG_TYPE_I32, a0, a1, REG0(2), REG0(3),
2244 (int32_t)args[4], args[5], const_args[4],
2245 const_args[5], true);
2246 break;
2247 case INDEX_op_sub2_i64:
2248 tcg_out_addsub2(s, TCG_TYPE_I64, a0, a1, REG0(2), REG0(3), args[4],
2249 args[5], const_args[4], const_args[5], true);
2250 break;
2251
2252 case INDEX_op_muluh_i64:
2253 tcg_out_insn(s, 3508, UMULH, TCG_TYPE_I64, a0, a1, a2);
2254 break;
2255 case INDEX_op_mulsh_i64:
2256 tcg_out_insn(s, 3508, SMULH, TCG_TYPE_I64, a0, a1, a2);
2257 break;
2258
2259 case INDEX_op_mb:
2260 tcg_out_mb(s, a0);
2261 break;
2262
2263 case INDEX_op_mov_i32:
2264 case INDEX_op_mov_i64:
2265 case INDEX_op_movi_i32:
2266 case INDEX_op_movi_i64:
2267 case INDEX_op_call:
2268 default:
2269 g_assert_not_reached();
2270 }
2271
2272#undef REG0
2273}
2274
2275static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
2276 unsigned vecl, unsigned vece,
2277 const TCGArg *args, const int *const_args)
2278{
2279 static const AArch64Insn cmp_insn[16] = {
2280 [TCG_COND_EQ] = I3616_CMEQ,
2281 [TCG_COND_GT] = I3616_CMGT,
2282 [TCG_COND_GE] = I3616_CMGE,
2283 [TCG_COND_GTU] = I3616_CMHI,
2284 [TCG_COND_GEU] = I3616_CMHS,
2285 };
2286 static const AArch64Insn cmp0_insn[16] = {
2287 [TCG_COND_EQ] = I3617_CMEQ0,
2288 [TCG_COND_GT] = I3617_CMGT0,
2289 [TCG_COND_GE] = I3617_CMGE0,
2290 [TCG_COND_LT] = I3617_CMLT0,
2291 [TCG_COND_LE] = I3617_CMLE0,
2292 };
2293
2294 TCGType type = vecl + TCG_TYPE_V64;
2295 unsigned is_q = vecl;
2296 TCGArg a0, a1, a2, a3;
2297 int cmode, imm8;
2298
2299 a0 = args[0];
2300 a1 = args[1];
2301 a2 = args[2];
2302
2303 switch (opc) {
2304 case INDEX_op_ld_vec:
2305 tcg_out_ld(s, type, a0, a1, a2);
2306 break;
2307 case INDEX_op_st_vec:
2308 tcg_out_st(s, type, a0, a1, a2);
2309 break;
2310 case INDEX_op_dupm_vec:
2311 tcg_out_dupm_vec(s, type, vece, a0, a1, a2);
2312 break;
2313 case INDEX_op_add_vec:
2314 tcg_out_insn(s, 3616, ADD, is_q, vece, a0, a1, a2);
2315 break;
2316 case INDEX_op_sub_vec:
2317 tcg_out_insn(s, 3616, SUB, is_q, vece, a0, a1, a2);
2318 break;
2319 case INDEX_op_mul_vec:
2320 tcg_out_insn(s, 3616, MUL, is_q, vece, a0, a1, a2);
2321 break;
2322 case INDEX_op_neg_vec:
2323 tcg_out_insn(s, 3617, NEG, is_q, vece, a0, a1);
2324 break;
2325 case INDEX_op_abs_vec:
2326 tcg_out_insn(s, 3617, ABS, is_q, vece, a0, a1);
2327 break;
2328 case INDEX_op_and_vec:
2329 if (const_args[2]) {
2330 is_shimm1632(~a2, &cmode, &imm8);
2331 if (a0 == a1) {
2332 tcg_out_insn(s, 3606, BIC, is_q, a0, 0, cmode, imm8);
2333 return;
2334 }
2335 tcg_out_insn(s, 3606, MVNI, is_q, a0, 0, cmode, imm8);
2336 a2 = a0;
2337 }
2338 tcg_out_insn(s, 3616, AND, is_q, 0, a0, a1, a2);
2339 break;
2340 case INDEX_op_or_vec:
2341 if (const_args[2]) {
2342 is_shimm1632(a2, &cmode, &imm8);
2343 if (a0 == a1) {
2344 tcg_out_insn(s, 3606, ORR, is_q, a0, 0, cmode, imm8);
2345 return;
2346 }
2347 tcg_out_insn(s, 3606, MOVI, is_q, a0, 0, cmode, imm8);
2348 a2 = a0;
2349 }
2350 tcg_out_insn(s, 3616, ORR, is_q, 0, a0, a1, a2);
2351 break;
2352 case INDEX_op_andc_vec:
2353 if (const_args[2]) {
2354 is_shimm1632(a2, &cmode, &imm8);
2355 if (a0 == a1) {
2356 tcg_out_insn(s, 3606, BIC, is_q, a0, 0, cmode, imm8);
2357 return;
2358 }
2359 tcg_out_insn(s, 3606, MOVI, is_q, a0, 0, cmode, imm8);
2360 a2 = a0;
2361 }
2362 tcg_out_insn(s, 3616, BIC, is_q, 0, a0, a1, a2);
2363 break;
2364 case INDEX_op_orc_vec:
2365 if (const_args[2]) {
2366 is_shimm1632(~a2, &cmode, &imm8);
2367 if (a0 == a1) {
2368 tcg_out_insn(s, 3606, ORR, is_q, a0, 0, cmode, imm8);
2369 return;
2370 }
2371 tcg_out_insn(s, 3606, MVNI, is_q, a0, 0, cmode, imm8);
2372 a2 = a0;
2373 }
2374 tcg_out_insn(s, 3616, ORN, is_q, 0, a0, a1, a2);
2375 break;
2376 case INDEX_op_xor_vec:
2377 tcg_out_insn(s, 3616, EOR, is_q, 0, a0, a1, a2);
2378 break;
2379 case INDEX_op_ssadd_vec:
2380 tcg_out_insn(s, 3616, SQADD, is_q, vece, a0, a1, a2);
2381 break;
2382 case INDEX_op_sssub_vec:
2383 tcg_out_insn(s, 3616, SQSUB, is_q, vece, a0, a1, a2);
2384 break;
2385 case INDEX_op_usadd_vec:
2386 tcg_out_insn(s, 3616, UQADD, is_q, vece, a0, a1, a2);
2387 break;
2388 case INDEX_op_ussub_vec:
2389 tcg_out_insn(s, 3616, UQSUB, is_q, vece, a0, a1, a2);
2390 break;
2391 case INDEX_op_smax_vec:
2392 tcg_out_insn(s, 3616, SMAX, is_q, vece, a0, a1, a2);
2393 break;
2394 case INDEX_op_smin_vec:
2395 tcg_out_insn(s, 3616, SMIN, is_q, vece, a0, a1, a2);
2396 break;
2397 case INDEX_op_umax_vec:
2398 tcg_out_insn(s, 3616, UMAX, is_q, vece, a0, a1, a2);
2399 break;
2400 case INDEX_op_umin_vec:
2401 tcg_out_insn(s, 3616, UMIN, is_q, vece, a0, a1, a2);
2402 break;
2403 case INDEX_op_not_vec:
2404 tcg_out_insn(s, 3617, NOT, is_q, 0, a0, a1);
2405 break;
2406 case INDEX_op_shli_vec:
2407 tcg_out_insn(s, 3614, SHL, is_q, a0, a1, a2 + (8 << vece));
2408 break;
2409 case INDEX_op_shri_vec:
2410 tcg_out_insn(s, 3614, USHR, is_q, a0, a1, (16 << vece) - a2);
2411 break;
2412 case INDEX_op_sari_vec:
2413 tcg_out_insn(s, 3614, SSHR, is_q, a0, a1, (16 << vece) - a2);
2414 break;
2415 case INDEX_op_aa64_sli_vec:
2416 tcg_out_insn(s, 3614, SLI, is_q, a0, a2, args[3] + (8 << vece));
2417 break;
2418 case INDEX_op_shlv_vec:
2419 tcg_out_insn(s, 3616, USHL, is_q, vece, a0, a1, a2);
2420 break;
2421 case INDEX_op_aa64_sshl_vec:
2422 tcg_out_insn(s, 3616, SSHL, is_q, vece, a0, a1, a2);
2423 break;
2424 case INDEX_op_cmp_vec:
2425 {
2426 TCGCond cond = args[3];
2427 AArch64Insn insn;
2428
2429 if (cond == TCG_COND_NE) {
2430 if (const_args[2]) {
2431 tcg_out_insn(s, 3616, CMTST, is_q, vece, a0, a1, a1);
2432 } else {
2433 tcg_out_insn(s, 3616, CMEQ, is_q, vece, a0, a1, a2);
2434 tcg_out_insn(s, 3617, NOT, is_q, 0, a0, a0);
2435 }
2436 } else {
2437 if (const_args[2]) {
2438 insn = cmp0_insn[cond];
2439 if (insn) {
2440 tcg_out_insn_3617(s, insn, is_q, vece, a0, a1);
2441 break;
2442 }
2443 tcg_out_dupi_vec(s, type, TCG_VEC_TMP, 0);
2444 a2 = TCG_VEC_TMP;
2445 }
2446 insn = cmp_insn[cond];
2447 if (insn == 0) {
2448 TCGArg t;
2449 t = a1, a1 = a2, a2 = t;
2450 cond = tcg_swap_cond(cond);
2451 insn = cmp_insn[cond];
2452 tcg_debug_assert(insn != 0);
2453 }
2454 tcg_out_insn_3616(s, insn, is_q, vece, a0, a1, a2);
2455 }
2456 }
2457 break;
2458
2459 case INDEX_op_bitsel_vec:
2460 a3 = args[3];
2461 if (a0 == a3) {
2462 tcg_out_insn(s, 3616, BIT, is_q, 0, a0, a2, a1);
2463 } else if (a0 == a2) {
2464 tcg_out_insn(s, 3616, BIF, is_q, 0, a0, a3, a1);
2465 } else {
2466 if (a0 != a1) {
2467 tcg_out_mov(s, type, a0, a1);
2468 }
2469 tcg_out_insn(s, 3616, BSL, is_q, 0, a0, a2, a3);
2470 }
2471 break;
2472
2473 case INDEX_op_mov_vec:
2474 case INDEX_op_dupi_vec:
2475 case INDEX_op_dup_vec:
2476 default:
2477 g_assert_not_reached();
2478 }
2479}
2480
2481int tcg_can_emit_vec_op(TCGOpcode opc, TCGType type, unsigned vece)
2482{
2483 switch (opc) {
2484 case INDEX_op_add_vec:
2485 case INDEX_op_sub_vec:
2486 case INDEX_op_and_vec:
2487 case INDEX_op_or_vec:
2488 case INDEX_op_xor_vec:
2489 case INDEX_op_andc_vec:
2490 case INDEX_op_orc_vec:
2491 case INDEX_op_neg_vec:
2492 case INDEX_op_abs_vec:
2493 case INDEX_op_not_vec:
2494 case INDEX_op_cmp_vec:
2495 case INDEX_op_shli_vec:
2496 case INDEX_op_shri_vec:
2497 case INDEX_op_sari_vec:
2498 case INDEX_op_ssadd_vec:
2499 case INDEX_op_sssub_vec:
2500 case INDEX_op_usadd_vec:
2501 case INDEX_op_ussub_vec:
2502 case INDEX_op_shlv_vec:
2503 case INDEX_op_bitsel_vec:
2504 return 1;
2505 case INDEX_op_rotli_vec:
2506 case INDEX_op_shrv_vec:
2507 case INDEX_op_sarv_vec:
2508 case INDEX_op_rotlv_vec:
2509 case INDEX_op_rotrv_vec:
2510 return -1;
2511 case INDEX_op_mul_vec:
2512 case INDEX_op_smax_vec:
2513 case INDEX_op_smin_vec:
2514 case INDEX_op_umax_vec:
2515 case INDEX_op_umin_vec:
2516 return vece < MO_64;
2517
2518 default:
2519 return 0;
2520 }
2521}
2522
2523void tcg_expand_vec_op(TCGOpcode opc, TCGType type, unsigned vece,
2524 TCGArg a0, ...)
2525{
2526 va_list va;
2527 TCGv_vec v0, v1, v2, t1, t2;
2528 TCGArg a2;
2529
2530 va_start(va, a0);
2531 v0 = temp_tcgv_vec(arg_temp(a0));
2532 v1 = temp_tcgv_vec(arg_temp(va_arg(va, TCGArg)));
2533 a2 = va_arg(va, TCGArg);
2534 v2 = temp_tcgv_vec(arg_temp(a2));
2535
2536 switch (opc) {
2537 case INDEX_op_rotli_vec:
2538 t1 = tcg_temp_new_vec(type);
2539 tcg_gen_shri_vec(vece, t1, v1, -a2 & ((8 << vece) - 1));
2540 vec_gen_4(INDEX_op_aa64_sli_vec, type, vece,
2541 tcgv_vec_arg(v0), tcgv_vec_arg(t1), tcgv_vec_arg(v1), a2);
2542 tcg_temp_free_vec(t1);
2543 break;
2544
2545 case INDEX_op_shrv_vec:
2546 case INDEX_op_sarv_vec:
2547
2548 t1 = tcg_temp_new_vec(type);
2549 tcg_gen_neg_vec(vece, t1, v2);
2550 opc = (opc == INDEX_op_shrv_vec
2551 ? INDEX_op_shlv_vec : INDEX_op_aa64_sshl_vec);
2552 vec_gen_3(opc, type, vece, tcgv_vec_arg(v0),
2553 tcgv_vec_arg(v1), tcgv_vec_arg(t1));
2554 tcg_temp_free_vec(t1);
2555 break;
2556
2557 case INDEX_op_rotlv_vec:
2558 t1 = tcg_temp_new_vec(type);
2559 tcg_gen_dupi_vec(vece, t1, 8 << vece);
2560 tcg_gen_sub_vec(vece, t1, v2, t1);
2561
2562 vec_gen_3(INDEX_op_shlv_vec, type, vece, tcgv_vec_arg(t1),
2563 tcgv_vec_arg(v1), tcgv_vec_arg(t1));
2564 vec_gen_3(INDEX_op_shlv_vec, type, vece, tcgv_vec_arg(v0),
2565 tcgv_vec_arg(v1), tcgv_vec_arg(v2));
2566 tcg_gen_or_vec(vece, v0, v0, t1);
2567 tcg_temp_free_vec(t1);
2568 break;
2569
2570 case INDEX_op_rotrv_vec:
2571 t1 = tcg_temp_new_vec(type);
2572 t2 = tcg_temp_new_vec(type);
2573 tcg_gen_neg_vec(vece, t1, v2);
2574 tcg_gen_dupi_vec(vece, t2, 8 << vece);
2575 tcg_gen_add_vec(vece, t2, t1, t2);
2576
2577 vec_gen_3(INDEX_op_shlv_vec, type, vece, tcgv_vec_arg(t1),
2578 tcgv_vec_arg(v1), tcgv_vec_arg(t1));
2579 vec_gen_3(INDEX_op_shlv_vec, type, vece, tcgv_vec_arg(t2),
2580 tcgv_vec_arg(v1), tcgv_vec_arg(t2));
2581 tcg_gen_or_vec(vece, v0, t1, t2);
2582 tcg_temp_free_vec(t1);
2583 tcg_temp_free_vec(t2);
2584 break;
2585
2586 default:
2587 g_assert_not_reached();
2588 }
2589
2590 va_end(va);
2591}
2592
2593static const TCGTargetOpDef *tcg_target_op_def(TCGOpcode op)
2594{
2595 static const TCGTargetOpDef r = { .args_ct_str = { "r" } };
2596 static const TCGTargetOpDef r_r = { .args_ct_str = { "r", "r" } };
2597 static const TCGTargetOpDef w_w = { .args_ct_str = { "w", "w" } };
2598 static const TCGTargetOpDef w_r = { .args_ct_str = { "w", "r" } };
2599 static const TCGTargetOpDef w_wr = { .args_ct_str = { "w", "wr" } };
2600 static const TCGTargetOpDef r_l = { .args_ct_str = { "r", "l" } };
2601 static const TCGTargetOpDef r_rA = { .args_ct_str = { "r", "rA" } };
2602 static const TCGTargetOpDef rZ_r = { .args_ct_str = { "rZ", "r" } };
2603 static const TCGTargetOpDef lZ_l = { .args_ct_str = { "lZ", "l" } };
2604 static const TCGTargetOpDef r_r_r = { .args_ct_str = { "r", "r", "r" } };
2605 static const TCGTargetOpDef w_w_w = { .args_ct_str = { "w", "w", "w" } };
2606 static const TCGTargetOpDef w_0_w = { .args_ct_str = { "w", "0", "w" } };
2607 static const TCGTargetOpDef w_w_wO = { .args_ct_str = { "w", "w", "wO" } };
2608 static const TCGTargetOpDef w_w_wN = { .args_ct_str = { "w", "w", "wN" } };
2609 static const TCGTargetOpDef w_w_wZ = { .args_ct_str = { "w", "w", "wZ" } };
2610 static const TCGTargetOpDef r_r_ri = { .args_ct_str = { "r", "r", "ri" } };
2611 static const TCGTargetOpDef r_r_rA = { .args_ct_str = { "r", "r", "rA" } };
2612 static const TCGTargetOpDef r_r_rL = { .args_ct_str = { "r", "r", "rL" } };
2613 static const TCGTargetOpDef r_r_rAL
2614 = { .args_ct_str = { "r", "r", "rAL" } };
2615 static const TCGTargetOpDef dep
2616 = { .args_ct_str = { "r", "0", "rZ" } };
2617 static const TCGTargetOpDef ext2
2618 = { .args_ct_str = { "r", "rZ", "rZ" } };
2619 static const TCGTargetOpDef movc
2620 = { .args_ct_str = { "r", "r", "rA", "rZ", "rZ" } };
2621 static const TCGTargetOpDef add2
2622 = { .args_ct_str = { "r", "r", "rZ", "rZ", "rA", "rMZ" } };
2623 static const TCGTargetOpDef w_w_w_w
2624 = { .args_ct_str = { "w", "w", "w", "w" } };
2625
2626 switch (op) {
2627 case INDEX_op_goto_ptr:
2628 return &r;
2629
2630 case INDEX_op_ld8u_i32:
2631 case INDEX_op_ld8s_i32:
2632 case INDEX_op_ld16u_i32:
2633 case INDEX_op_ld16s_i32:
2634 case INDEX_op_ld_i32:
2635 case INDEX_op_ld8u_i64:
2636 case INDEX_op_ld8s_i64:
2637 case INDEX_op_ld16u_i64:
2638 case INDEX_op_ld16s_i64:
2639 case INDEX_op_ld32u_i64:
2640 case INDEX_op_ld32s_i64:
2641 case INDEX_op_ld_i64:
2642 case INDEX_op_neg_i32:
2643 case INDEX_op_neg_i64:
2644 case INDEX_op_not_i32:
2645 case INDEX_op_not_i64:
2646 case INDEX_op_bswap16_i32:
2647 case INDEX_op_bswap32_i32:
2648 case INDEX_op_bswap16_i64:
2649 case INDEX_op_bswap32_i64:
2650 case INDEX_op_bswap64_i64:
2651 case INDEX_op_ext8s_i32:
2652 case INDEX_op_ext16s_i32:
2653 case INDEX_op_ext8u_i32:
2654 case INDEX_op_ext16u_i32:
2655 case INDEX_op_ext8s_i64:
2656 case INDEX_op_ext16s_i64:
2657 case INDEX_op_ext32s_i64:
2658 case INDEX_op_ext8u_i64:
2659 case INDEX_op_ext16u_i64:
2660 case INDEX_op_ext32u_i64:
2661 case INDEX_op_ext_i32_i64:
2662 case INDEX_op_extu_i32_i64:
2663 case INDEX_op_extract_i32:
2664 case INDEX_op_extract_i64:
2665 case INDEX_op_sextract_i32:
2666 case INDEX_op_sextract_i64:
2667 return &r_r;
2668
2669 case INDEX_op_st8_i32:
2670 case INDEX_op_st16_i32:
2671 case INDEX_op_st_i32:
2672 case INDEX_op_st8_i64:
2673 case INDEX_op_st16_i64:
2674 case INDEX_op_st32_i64:
2675 case INDEX_op_st_i64:
2676 return &rZ_r;
2677
2678 case INDEX_op_add_i32:
2679 case INDEX_op_add_i64:
2680 case INDEX_op_sub_i32:
2681 case INDEX_op_sub_i64:
2682 case INDEX_op_setcond_i32:
2683 case INDEX_op_setcond_i64:
2684 return &r_r_rA;
2685
2686 case INDEX_op_mul_i32:
2687 case INDEX_op_mul_i64:
2688 case INDEX_op_div_i32:
2689 case INDEX_op_div_i64:
2690 case INDEX_op_divu_i32:
2691 case INDEX_op_divu_i64:
2692 case INDEX_op_rem_i32:
2693 case INDEX_op_rem_i64:
2694 case INDEX_op_remu_i32:
2695 case INDEX_op_remu_i64:
2696 case INDEX_op_muluh_i64:
2697 case INDEX_op_mulsh_i64:
2698 return &r_r_r;
2699
2700 case INDEX_op_and_i32:
2701 case INDEX_op_and_i64:
2702 case INDEX_op_or_i32:
2703 case INDEX_op_or_i64:
2704 case INDEX_op_xor_i32:
2705 case INDEX_op_xor_i64:
2706 case INDEX_op_andc_i32:
2707 case INDEX_op_andc_i64:
2708 case INDEX_op_orc_i32:
2709 case INDEX_op_orc_i64:
2710 case INDEX_op_eqv_i32:
2711 case INDEX_op_eqv_i64:
2712 return &r_r_rL;
2713
2714 case INDEX_op_shl_i32:
2715 case INDEX_op_shr_i32:
2716 case INDEX_op_sar_i32:
2717 case INDEX_op_rotl_i32:
2718 case INDEX_op_rotr_i32:
2719 case INDEX_op_shl_i64:
2720 case INDEX_op_shr_i64:
2721 case INDEX_op_sar_i64:
2722 case INDEX_op_rotl_i64:
2723 case INDEX_op_rotr_i64:
2724 return &r_r_ri;
2725
2726 case INDEX_op_clz_i32:
2727 case INDEX_op_ctz_i32:
2728 case INDEX_op_clz_i64:
2729 case INDEX_op_ctz_i64:
2730 return &r_r_rAL;
2731
2732 case INDEX_op_brcond_i32:
2733 case INDEX_op_brcond_i64:
2734 return &r_rA;
2735
2736 case INDEX_op_movcond_i32:
2737 case INDEX_op_movcond_i64:
2738 return &movc;
2739
2740 case INDEX_op_qemu_ld_i32:
2741 case INDEX_op_qemu_ld_i64:
2742 return &r_l;
2743 case INDEX_op_qemu_st_i32:
2744 case INDEX_op_qemu_st_i64:
2745 return &lZ_l;
2746
2747 case INDEX_op_deposit_i32:
2748 case INDEX_op_deposit_i64:
2749 return &dep;
2750
2751 case INDEX_op_extract2_i32:
2752 case INDEX_op_extract2_i64:
2753 return &ext2;
2754
2755 case INDEX_op_add2_i32:
2756 case INDEX_op_add2_i64:
2757 case INDEX_op_sub2_i32:
2758 case INDEX_op_sub2_i64:
2759 return &add2;
2760
2761 case INDEX_op_add_vec:
2762 case INDEX_op_sub_vec:
2763 case INDEX_op_mul_vec:
2764 case INDEX_op_xor_vec:
2765 case INDEX_op_ssadd_vec:
2766 case INDEX_op_sssub_vec:
2767 case INDEX_op_usadd_vec:
2768 case INDEX_op_ussub_vec:
2769 case INDEX_op_smax_vec:
2770 case INDEX_op_smin_vec:
2771 case INDEX_op_umax_vec:
2772 case INDEX_op_umin_vec:
2773 case INDEX_op_shlv_vec:
2774 case INDEX_op_shrv_vec:
2775 case INDEX_op_sarv_vec:
2776 case INDEX_op_aa64_sshl_vec:
2777 return &w_w_w;
2778 case INDEX_op_not_vec:
2779 case INDEX_op_neg_vec:
2780 case INDEX_op_abs_vec:
2781 case INDEX_op_shli_vec:
2782 case INDEX_op_shri_vec:
2783 case INDEX_op_sari_vec:
2784 return &w_w;
2785 case INDEX_op_ld_vec:
2786 case INDEX_op_st_vec:
2787 case INDEX_op_dupm_vec:
2788 return &w_r;
2789 case INDEX_op_dup_vec:
2790 return &w_wr;
2791 case INDEX_op_or_vec:
2792 case INDEX_op_andc_vec:
2793 return &w_w_wO;
2794 case INDEX_op_and_vec:
2795 case INDEX_op_orc_vec:
2796 return &w_w_wN;
2797 case INDEX_op_cmp_vec:
2798 return &w_w_wZ;
2799 case INDEX_op_bitsel_vec:
2800 return &w_w_w_w;
2801 case INDEX_op_aa64_sli_vec:
2802 return &w_0_w;
2803
2804 default:
2805 return NULL;
2806 }
2807}
2808
2809static void tcg_target_init(TCGContext *s)
2810{
2811 tcg_target_available_regs[TCG_TYPE_I32] = 0xffffffffu;
2812 tcg_target_available_regs[TCG_TYPE_I64] = 0xffffffffu;
2813 tcg_target_available_regs[TCG_TYPE_V64] = 0xffffffff00000000ull;
2814 tcg_target_available_regs[TCG_TYPE_V128] = 0xffffffff00000000ull;
2815
2816 tcg_target_call_clobber_regs = -1ull;
2817 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X19);
2818 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X20);
2819 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X21);
2820 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X22);
2821 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X23);
2822 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X24);
2823 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X25);
2824 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X26);
2825 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X27);
2826 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X28);
2827 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_X29);
2828 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V8);
2829 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V9);
2830 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V10);
2831 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V11);
2832 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V12);
2833 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V13);
2834 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V14);
2835 tcg_regset_reset_reg(tcg_target_call_clobber_regs, TCG_REG_V15);
2836
2837 s->reserved_regs = 0;
2838 tcg_regset_set_reg(s->reserved_regs, TCG_REG_SP);
2839 tcg_regset_set_reg(s->reserved_regs, TCG_REG_FP);
2840 tcg_regset_set_reg(s->reserved_regs, TCG_REG_TMP);
2841 tcg_regset_set_reg(s->reserved_regs, TCG_REG_X18);
2842 tcg_regset_set_reg(s->reserved_regs, TCG_VEC_TMP);
2843}
2844
2845
2846#define PUSH_SIZE ((30 - 19 + 1) * 8)
2847
2848#define FRAME_SIZE \
2849 ((PUSH_SIZE \
2850 + TCG_STATIC_CALL_ARGS_SIZE \
2851 + CPU_TEMP_BUF_NLONGS * sizeof(long) \
2852 + TCG_TARGET_STACK_ALIGN - 1) \
2853 & ~(TCG_TARGET_STACK_ALIGN - 1))
2854
2855
2856QEMU_BUILD_BUG_ON(FRAME_SIZE >= (1 << 14));
2857
2858
2859QEMU_BUILD_BUG_ON(FRAME_SIZE - PUSH_SIZE > 0xfff);
2860
2861static void tcg_target_qemu_prologue(TCGContext *s)
2862{
2863 TCGReg r;
2864
2865
2866 tcg_out_insn(s, 3314, STP, TCG_REG_FP, TCG_REG_LR,
2867 TCG_REG_SP, -PUSH_SIZE, 1, 1);
2868
2869
2870 tcg_out_movr_sp(s, TCG_TYPE_I64, TCG_REG_FP, TCG_REG_SP);
2871
2872
2873 for (r = TCG_REG_X19; r <= TCG_REG_X27; r += 2) {
2874 int ofs = (r - TCG_REG_X19 + 2) * 8;
2875 tcg_out_insn(s, 3314, STP, r, r + 1, TCG_REG_SP, ofs, 1, 0);
2876 }
2877
2878
2879 tcg_out_insn(s, 3401, SUBI, TCG_TYPE_I64, TCG_REG_SP, TCG_REG_SP,
2880 FRAME_SIZE - PUSH_SIZE);
2881
2882
2883 tcg_set_frame(s, TCG_REG_SP, TCG_STATIC_CALL_ARGS_SIZE,
2884 CPU_TEMP_BUF_NLONGS * sizeof(long));
2885
2886#if !defined(CONFIG_SOFTMMU)
2887 if (USE_GUEST_BASE) {
2888 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_GUEST_BASE, guest_base);
2889 tcg_regset_set_reg(s->reserved_regs, TCG_REG_GUEST_BASE);
2890 }
2891#endif
2892
2893 tcg_out_mov(s, TCG_TYPE_PTR, TCG_AREG0, tcg_target_call_iarg_regs[0]);
2894 tcg_out_insn(s, 3207, BR, tcg_target_call_iarg_regs[1]);
2895
2896
2897
2898
2899
2900 s->code_gen_epilogue = s->code_ptr;
2901 tcg_out_movi(s, TCG_TYPE_REG, TCG_REG_X0, 0);
2902
2903
2904 tb_ret_addr = s->code_ptr;
2905
2906
2907 tcg_out_insn(s, 3401, ADDI, TCG_TYPE_I64, TCG_REG_SP, TCG_REG_SP,
2908 FRAME_SIZE - PUSH_SIZE);
2909
2910
2911 for (r = TCG_REG_X19; r <= TCG_REG_X27; r += 2) {
2912 int ofs = (r - TCG_REG_X19 + 2) * 8;
2913 tcg_out_insn(s, 3314, LDP, r, r + 1, TCG_REG_SP, ofs, 1, 0);
2914 }
2915
2916
2917 tcg_out_insn(s, 3314, LDP, TCG_REG_FP, TCG_REG_LR,
2918 TCG_REG_SP, PUSH_SIZE, 0, 1);
2919 tcg_out_insn(s, 3207, RET, TCG_REG_LR);
2920}
2921
2922static void tcg_out_nop_fill(tcg_insn_unit *p, int count)
2923{
2924 int i;
2925 for (i = 0; i < count; ++i) {
2926 p[i] = NOP;
2927 }
2928}
2929
2930typedef struct {
2931 DebugFrameHeader h;
2932 uint8_t fde_def_cfa[4];
2933 uint8_t fde_reg_ofs[24];
2934} DebugFrame;
2935
2936#define ELF_HOST_MACHINE EM_AARCH64
2937
2938static const DebugFrame debug_frame = {
2939 .h.cie.len = sizeof(DebugFrameCIE)-4,
2940 .h.cie.id = -1,
2941 .h.cie.version = 1,
2942 .h.cie.code_align = 1,
2943 .h.cie.data_align = 0x78,
2944 .h.cie.return_column = TCG_REG_LR,
2945
2946
2947 .h.fde.len = sizeof(DebugFrame) - offsetof(DebugFrame, h.fde.cie_offset),
2948
2949 .fde_def_cfa = {
2950 12, TCG_REG_SP,
2951 (FRAME_SIZE & 0x7f) | 0x80,
2952 (FRAME_SIZE >> 7)
2953 },
2954 .fde_reg_ofs = {
2955 0x80 + 28, 1,
2956 0x80 + 27, 2,
2957 0x80 + 26, 3,
2958 0x80 + 25, 4,
2959 0x80 + 24, 5,
2960 0x80 + 23, 6,
2961 0x80 + 22, 7,
2962 0x80 + 21, 8,
2963 0x80 + 20, 9,
2964 0x80 + 19, 10,
2965 0x80 + 30, 11,
2966 0x80 + 29, 12,
2967 }
2968};
2969
2970void tcg_register_jit(void *buf, size_t buf_size)
2971{
2972 tcg_register_jit_int(buf, buf_size, &debug_frame, sizeof(debug_frame));
2973}
2974