1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26#include "qemu/osdep.h"
27#include "qemu/host-utils.h"
28
29#ifndef CONFIG_INT128
30
31static inline void mul64(uint64_t *plow, uint64_t *phigh,
32 uint64_t a, uint64_t b)
33{
34 typedef union {
35 uint64_t ll;
36 struct {
37#if HOST_BIG_ENDIAN
38 uint32_t high, low;
39#else
40 uint32_t low, high;
41#endif
42 } l;
43 } LL;
44 LL rl, rm, rn, rh, a0, b0;
45 uint64_t c;
46
47 a0.ll = a;
48 b0.ll = b;
49
50 rl.ll = (uint64_t)a0.l.low * b0.l.low;
51 rm.ll = (uint64_t)a0.l.low * b0.l.high;
52 rn.ll = (uint64_t)a0.l.high * b0.l.low;
53 rh.ll = (uint64_t)a0.l.high * b0.l.high;
54
55 c = (uint64_t)rl.l.high + rm.l.low + rn.l.low;
56 rl.l.high = c;
57 c >>= 32;
58 c = c + rm.l.high + rn.l.high + rh.l.low;
59 rh.l.low = c;
60 rh.l.high += (uint32_t)(c >> 32);
61
62 *plow = rl.ll;
63 *phigh = rh.ll;
64}
65
66
67void mulu64 (uint64_t *plow, uint64_t *phigh, uint64_t a, uint64_t b)
68{
69 mul64(plow, phigh, a, b);
70}
71
72
73void muls64 (uint64_t *plow, uint64_t *phigh, int64_t a, int64_t b)
74{
75 uint64_t rh;
76
77 mul64(plow, &rh, a, b);
78
79
80 if (b < 0) {
81 rh -= a;
82 }
83 if (a < 0) {
84 rh -= b;
85 }
86 *phigh = rh;
87}
88
89
90
91
92
93
94
95uint64_t divu128(uint64_t *plow, uint64_t *phigh, uint64_t divisor)
96{
97 uint64_t dhi = *phigh;
98 uint64_t dlo = *plow;
99 uint64_t rem, dhighest;
100 int sh;
101
102 if (divisor == 0 || dhi == 0) {
103 *plow = dlo / divisor;
104 *phigh = 0;
105 return dlo % divisor;
106 } else {
107 sh = clz64(divisor);
108
109 if (dhi < divisor) {
110 if (sh != 0) {
111
112 divisor <<= sh;
113 dhi = (dhi << sh) | (dlo >> (64 - sh));
114 dlo <<= sh;
115 }
116
117 *phigh = 0;
118 *plow = udiv_qrnnd(&rem, dhi, dlo, divisor);
119 } else {
120 if (sh != 0) {
121
122 divisor <<= sh;
123 dhighest = dhi >> (64 - sh);
124 dhi = (dhi << sh) | (dlo >> (64 - sh));
125 dlo <<= sh;
126
127 *phigh = udiv_qrnnd(&dhi, dhighest, dhi, divisor);
128 } else {
129
130
131
132
133
134
135
136
137
138 dhi -= divisor;
139 *phigh = 1;
140 }
141
142 *plow = udiv_qrnnd(&rem, dhi, dlo, divisor);
143 }
144
145
146
147
148
149 return rem >> sh;
150 }
151}
152
153
154
155
156
157
158int64_t divs128(uint64_t *plow, int64_t *phigh, int64_t divisor)
159{
160 bool neg_quotient = false, neg_remainder = false;
161 uint64_t unsig_hi = *phigh, unsig_lo = *plow;
162 uint64_t rem;
163
164 if (*phigh < 0) {
165 neg_quotient = !neg_quotient;
166 neg_remainder = !neg_remainder;
167
168 if (unsig_lo == 0) {
169 unsig_hi = -unsig_hi;
170 } else {
171 unsig_hi = ~unsig_hi;
172 unsig_lo = -unsig_lo;
173 }
174 }
175
176 if (divisor < 0) {
177 neg_quotient = !neg_quotient;
178
179 divisor = -divisor;
180 }
181
182 rem = divu128(&unsig_lo, &unsig_hi, (uint64_t)divisor);
183
184 if (neg_quotient) {
185 if (unsig_lo == 0) {
186 *phigh = -unsig_hi;
187 *plow = 0;
188 } else {
189 *phigh = ~unsig_hi;
190 *plow = -unsig_lo;
191 }
192 } else {
193 *phigh = unsig_hi;
194 *plow = unsig_lo;
195 }
196
197 if (neg_remainder) {
198 return -rem;
199 } else {
200 return rem;
201 }
202}
203#endif
204
205
206
207
208
209
210
211
212
213
214
215
216void urshift(uint64_t *plow, uint64_t *phigh, int32_t shift)
217{
218 shift &= 127;
219 if (shift == 0) {
220 return;
221 }
222
223 uint64_t h = *phigh >> (shift & 63);
224 if (shift >= 64) {
225 *plow = h;
226 *phigh = 0;
227 } else {
228 *plow = (*plow >> (shift & 63)) | (*phigh << (64 - (shift & 63)));
229 *phigh = h;
230 }
231}
232
233
234
235
236
237
238
239
240
241
242
243
244
245void ulshift(uint64_t *plow, uint64_t *phigh, int32_t shift, bool *overflow)
246{
247 uint64_t low = *plow;
248 uint64_t high = *phigh;
249
250 shift &= 127;
251 if (shift == 0) {
252 return;
253 }
254
255
256 urshift(&low, &high, 128 - shift);
257 if (low | high) {
258 *overflow = true;
259 }
260
261 if (shift >= 64) {
262 *phigh = *plow << (shift & 63);
263 *plow = 0;
264 } else {
265 *phigh = (*plow >> (64 - (shift & 63))) | (*phigh << (shift & 63));
266 *plow = *plow << shift;
267 }
268}
269
270
271
272
273
274
275
276
277
278
279
280
281
282static Int128 udiv256_qrnnd(Int128 *r, Int128 n1, Int128 n0, Int128 d)
283{
284 Int128 d0, d1, q0, q1, r1, r0, m;
285 uint64_t mp0, mp1;
286
287 d0 = int128_make64(int128_getlo(d));
288 d1 = int128_make64(int128_gethi(d));
289
290 r1 = int128_remu(n1, d1);
291 q1 = int128_divu(n1, d1);
292 mp0 = int128_getlo(q1);
293 mp1 = int128_gethi(q1);
294 mulu128(&mp0, &mp1, int128_getlo(d0));
295 m = int128_make128(mp0, mp1);
296 r1 = int128_make128(int128_gethi(n0), int128_getlo(r1));
297 if (int128_ult(r1, m)) {
298 q1 = int128_sub(q1, int128_one());
299 r1 = int128_add(r1, d);
300 if (int128_uge(r1, d)) {
301 if (int128_ult(r1, m)) {
302 q1 = int128_sub(q1, int128_one());
303 r1 = int128_add(r1, d);
304 }
305 }
306 }
307 r1 = int128_sub(r1, m);
308
309 r0 = int128_remu(r1, d1);
310 q0 = int128_divu(r1, d1);
311 mp0 = int128_getlo(q0);
312 mp1 = int128_gethi(q0);
313 mulu128(&mp0, &mp1, int128_getlo(d0));
314 m = int128_make128(mp0, mp1);
315 r0 = int128_make128(int128_getlo(n0), int128_getlo(r0));
316 if (int128_ult(r0, m)) {
317 q0 = int128_sub(q0, int128_one());
318 r0 = int128_add(r0, d);
319 if (int128_uge(r0, d)) {
320 if (int128_ult(r0, m)) {
321 q0 = int128_sub(q0, int128_one());
322 r0 = int128_add(r0, d);
323 }
324 }
325 }
326 r0 = int128_sub(r0, m);
327
328 *r = r0;
329 return int128_or(int128_lshift(q1, 64), q0);
330}
331
332
333
334
335
336
337
338Int128 divu256(Int128 *plow, Int128 *phigh, Int128 divisor)
339{
340 Int128 dhi = *phigh;
341 Int128 dlo = *plow;
342 Int128 rem, dhighest;
343 int sh;
344
345 if (!int128_nz(divisor) || !int128_nz(dhi)) {
346 *plow = int128_divu(dlo, divisor);
347 *phigh = int128_zero();
348 return int128_remu(dlo, divisor);
349 } else {
350 sh = clz128(divisor);
351
352 if (int128_ult(dhi, divisor)) {
353 if (sh != 0) {
354
355 divisor = int128_lshift(divisor, sh);
356 dhi = int128_or(int128_lshift(dhi, sh),
357 int128_urshift(dlo, (128 - sh)));
358 dlo = int128_lshift(dlo, sh);
359 }
360
361 *phigh = int128_zero();
362 *plow = udiv256_qrnnd(&rem, dhi, dlo, divisor);
363 } else {
364 if (sh != 0) {
365
366 divisor = int128_lshift(divisor, sh);
367 dhighest = int128_rshift(dhi, (128 - sh));
368 dhi = int128_or(int128_lshift(dhi, sh),
369 int128_urshift(dlo, (128 - sh)));
370 dlo = int128_lshift(dlo, sh);
371
372 *phigh = udiv256_qrnnd(&dhi, dhighest, dhi, divisor);
373 } else {
374
375
376
377
378
379
380
381
382
383 dhi = int128_sub(dhi, divisor);
384 *phigh = int128_one();
385 }
386
387 *plow = udiv256_qrnnd(&rem, dhi, dlo, divisor);
388 }
389
390
391
392
393
394 rem = int128_urshift(rem, sh);
395 return rem;
396 }
397}
398
399
400
401
402
403
404Int128 divs256(Int128 *plow, Int128 *phigh, Int128 divisor)
405{
406 bool neg_quotient = false, neg_remainder = false;
407 Int128 unsig_hi = *phigh, unsig_lo = *plow;
408 Int128 rem;
409
410 if (!int128_nonneg(*phigh)) {
411 neg_quotient = !neg_quotient;
412 neg_remainder = !neg_remainder;
413
414 if (!int128_nz(unsig_lo)) {
415 unsig_hi = int128_neg(unsig_hi);
416 } else {
417 unsig_hi = int128_not(unsig_hi);
418 unsig_lo = int128_neg(unsig_lo);
419 }
420 }
421
422 if (!int128_nonneg(divisor)) {
423 neg_quotient = !neg_quotient;
424
425 divisor = int128_neg(divisor);
426 }
427
428 rem = divu256(&unsig_lo, &unsig_hi, divisor);
429
430 if (neg_quotient) {
431 if (!int128_nz(unsig_lo)) {
432 *phigh = int128_neg(unsig_hi);
433 *plow = int128_zero();
434 } else {
435 *phigh = int128_not(unsig_hi);
436 *plow = int128_neg(unsig_lo);
437 }
438 } else {
439 *phigh = unsig_hi;
440 *plow = unsig_lo;
441 }
442
443 if (neg_remainder) {
444 return int128_neg(rem);
445 } else {
446 return rem;
447 }
448}
449