1
2
3
4
5
6
7
8
9
10
11#include "qemu/osdep.h"
12#include "qemu/log.h"
13#include "qemu/error-report.h"
14#include "hw/misc/aspeed_hace.h"
15#include "qapi/error.h"
16#include "migration/vmstate.h"
17#include "crypto/hash.h"
18#include "hw/qdev-properties.h"
19#include "hw/irq.h"
20
21#define R_CRYPT_CMD (0x10 / 4)
22
23#define R_STATUS (0x1c / 4)
24#define HASH_IRQ BIT(9)
25#define CRYPT_IRQ BIT(12)
26#define TAG_IRQ BIT(15)
27
28#define R_HASH_SRC (0x20 / 4)
29#define R_HASH_DEST (0x24 / 4)
30#define R_HASH_KEY_BUFF (0x28 / 4)
31#define R_HASH_SRC_LEN (0x2c / 4)
32
33#define R_HASH_CMD (0x30 / 4)
34
35#define HASH_ALGO_MASK (BIT(4) | BIT(5) | BIT(6))
36#define HASH_ALGO_MD5 0
37#define HASH_ALGO_SHA1 BIT(5)
38#define HASH_ALGO_SHA224 BIT(6)
39#define HASH_ALGO_SHA256 (BIT(4) | BIT(6))
40#define HASH_ALGO_SHA512_SERIES (BIT(5) | BIT(6))
41
42#define SHA512_HASH_ALGO_MASK (BIT(10) | BIT(11) | BIT(12))
43#define HASH_ALGO_SHA512_SHA512 0
44#define HASH_ALGO_SHA512_SHA384 BIT(10)
45#define HASH_ALGO_SHA512_SHA256 BIT(11)
46#define HASH_ALGO_SHA512_SHA224 (BIT(10) | BIT(11))
47
48#define HASH_HMAC_MASK (BIT(7) | BIT(8))
49#define HASH_DIGEST 0
50#define HASH_DIGEST_HMAC BIT(7)
51#define HASH_DIGEST_ACCUM BIT(8)
52#define HASH_HMAC_KEY (BIT(7) | BIT(8))
53
54#define HASH_ONLY 0
55#define HASH_ONLY2 BIT(0)
56#define HASH_CRYPT_THEN_HASH BIT(1)
57#define HASH_HASH_THEN_CRYPT (BIT(0) | BIT(1))
58
59#define HASH_IRQ_EN BIT(9)
60#define HASH_SG_EN BIT(18)
61
62#define SG_LIST_LEN_SIZE 4
63#define SG_LIST_LEN_MASK 0x0FFFFFFF
64#define SG_LIST_LEN_LAST BIT(31)
65#define SG_LIST_ADDR_SIZE 4
66#define SG_LIST_ADDR_MASK 0x7FFFFFFF
67#define SG_LIST_ENTRY_SIZE (SG_LIST_LEN_SIZE + SG_LIST_ADDR_SIZE)
68
69static const struct {
70 uint32_t mask;
71 QCryptoHashAlgorithm algo;
72} hash_algo_map[] = {
73 { HASH_ALGO_MD5, QCRYPTO_HASH_ALG_MD5 },
74 { HASH_ALGO_SHA1, QCRYPTO_HASH_ALG_SHA1 },
75 { HASH_ALGO_SHA224, QCRYPTO_HASH_ALG_SHA224 },
76 { HASH_ALGO_SHA256, QCRYPTO_HASH_ALG_SHA256 },
77 { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA512, QCRYPTO_HASH_ALG_SHA512 },
78 { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA384, QCRYPTO_HASH_ALG_SHA384 },
79 { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA256, QCRYPTO_HASH_ALG_SHA256 },
80};
81
82static int hash_algo_lookup(uint32_t reg)
83{
84 int i;
85
86 reg &= HASH_ALGO_MASK | SHA512_HASH_ALGO_MASK;
87
88 for (i = 0; i < ARRAY_SIZE(hash_algo_map); i++) {
89 if (reg == hash_algo_map[i].mask) {
90 return hash_algo_map[i].algo;
91 }
92 }
93
94 return -1;
95}
96
97
98
99
100
101
102
103
104
105
106static bool has_padding(AspeedHACEState *s, struct iovec *iov,
107 hwaddr req_len, uint32_t *total_msg_len,
108 uint32_t *pad_offset)
109{
110 *total_msg_len = (uint32_t)(ldq_be_p(iov->iov_base + req_len - 8) / 8);
111
112
113
114
115
116
117
118
119
120
121
122
123 if (*total_msg_len <= s->total_req_len) {
124 uint32_t padding_size = s->total_req_len - *total_msg_len;
125 uint8_t *padding = iov->iov_base;
126 *pad_offset = req_len - padding_size;
127 if (padding[*pad_offset] == 0x80) {
128 return true;
129 }
130 }
131
132 return false;
133}
134
135static int reconstruct_iov(AspeedHACEState *s, struct iovec *iov, int id,
136 uint32_t *pad_offset)
137{
138 int i, iov_count;
139 if (*pad_offset != 0) {
140 s->iov_cache[s->iov_count].iov_base = iov[id].iov_base;
141 s->iov_cache[s->iov_count].iov_len = *pad_offset;
142 ++s->iov_count;
143 }
144 for (i = 0; i < s->iov_count; i++) {
145 iov[i].iov_base = s->iov_cache[i].iov_base;
146 iov[i].iov_len = s->iov_cache[i].iov_len;
147 }
148 iov_count = s->iov_count;
149 s->iov_count = 0;
150 s->total_req_len = 0;
151 return iov_count;
152}
153
154
155
156
157
158
159
160
161
162
163
164static int gen_acc_mode_iov(AspeedHACEState *s, struct iovec *iov, int id,
165 hwaddr *req_len)
166{
167 uint32_t pad_offset;
168 uint32_t total_msg_len;
169 s->total_req_len += *req_len;
170
171 if (has_padding(s, &iov[id], *req_len, &total_msg_len, &pad_offset)) {
172 if (s->iov_count) {
173 return reconstruct_iov(s, iov, id, &pad_offset);
174 }
175
176 *req_len -= s->total_req_len - total_msg_len;
177 s->total_req_len = 0;
178 iov[id].iov_len = *req_len;
179 } else {
180 s->iov_cache[s->iov_count].iov_base = iov->iov_base;
181 s->iov_cache[s->iov_count].iov_len = *req_len;
182 ++s->iov_count;
183 }
184
185 return id + 1;
186}
187
188static void do_hash_operation(AspeedHACEState *s, int algo, bool sg_mode,
189 bool acc_mode)
190{
191 struct iovec iov[ASPEED_HACE_MAX_SG];
192 g_autofree uint8_t *digest_buf;
193 size_t digest_len = 0;
194 int niov = 0;
195 int i;
196
197 if (sg_mode) {
198 uint32_t len = 0;
199
200 for (i = 0; !(len & SG_LIST_LEN_LAST); i++) {
201 uint32_t addr, src;
202 hwaddr plen;
203
204 if (i == ASPEED_HACE_MAX_SG) {
205 qemu_log_mask(LOG_GUEST_ERROR,
206 "aspeed_hace: guest failed to set end of sg list marker\n");
207 break;
208 }
209
210 src = s->regs[R_HASH_SRC] + (i * SG_LIST_ENTRY_SIZE);
211
212 len = address_space_ldl_le(&s->dram_as, src,
213 MEMTXATTRS_UNSPECIFIED, NULL);
214
215 addr = address_space_ldl_le(&s->dram_as, src + SG_LIST_LEN_SIZE,
216 MEMTXATTRS_UNSPECIFIED, NULL);
217 addr &= SG_LIST_ADDR_MASK;
218
219 plen = len & SG_LIST_LEN_MASK;
220 iov[i].iov_base = address_space_map(&s->dram_as, addr, &plen, false,
221 MEMTXATTRS_UNSPECIFIED);
222
223 if (acc_mode) {
224 niov = gen_acc_mode_iov(s, iov, i, &plen);
225
226 } else {
227 iov[i].iov_len = plen;
228 }
229 }
230 } else {
231 hwaddr len = s->regs[R_HASH_SRC_LEN];
232
233 iov[0].iov_len = len;
234 iov[0].iov_base = address_space_map(&s->dram_as, s->regs[R_HASH_SRC],
235 &len, false,
236 MEMTXATTRS_UNSPECIFIED);
237 i = 1;
238
239 if (s->iov_count) {
240
241
242
243
244
245
246 uint32_t total_msg_len;
247 uint32_t pad_offset;
248 s->total_req_len += len;
249 if (has_padding(s, iov, len, &total_msg_len, &pad_offset)) {
250 niov = reconstruct_iov(s, iov, 0, &pad_offset);
251 }
252 }
253 }
254
255 if (niov) {
256 i = niov;
257 }
258
259 if (qcrypto_hash_bytesv(algo, iov, i, &digest_buf, &digest_len, NULL) < 0) {
260 qemu_log_mask(LOG_GUEST_ERROR, "%s: qcrypto failed\n", __func__);
261 return;
262 }
263
264 if (address_space_write(&s->dram_as, s->regs[R_HASH_DEST],
265 MEMTXATTRS_UNSPECIFIED,
266 digest_buf, digest_len)) {
267 qemu_log_mask(LOG_GUEST_ERROR,
268 "aspeed_hace: address space write failed\n");
269 }
270
271 for (; i > 0; i--) {
272 address_space_unmap(&s->dram_as, iov[i - 1].iov_base,
273 iov[i - 1].iov_len, false,
274 iov[i - 1].iov_len);
275 }
276
277
278
279
280
281 s->regs[R_STATUS] |= HASH_IRQ;
282}
283
284static uint64_t aspeed_hace_read(void *opaque, hwaddr addr, unsigned int size)
285{
286 AspeedHACEState *s = ASPEED_HACE(opaque);
287
288 addr >>= 2;
289
290 if (addr >= ASPEED_HACE_NR_REGS) {
291 qemu_log_mask(LOG_GUEST_ERROR,
292 "%s: Out-of-bounds read at offset 0x%" HWADDR_PRIx "\n",
293 __func__, addr << 2);
294 return 0;
295 }
296
297 return s->regs[addr];
298}
299
300static void aspeed_hace_write(void *opaque, hwaddr addr, uint64_t data,
301 unsigned int size)
302{
303 AspeedHACEState *s = ASPEED_HACE(opaque);
304 AspeedHACEClass *ahc = ASPEED_HACE_GET_CLASS(s);
305
306 addr >>= 2;
307
308 if (addr >= ASPEED_HACE_NR_REGS) {
309 qemu_log_mask(LOG_GUEST_ERROR,
310 "%s: Out-of-bounds write at offset 0x%" HWADDR_PRIx "\n",
311 __func__, addr << 2);
312 return;
313 }
314
315 switch (addr) {
316 case R_STATUS:
317 if (data & HASH_IRQ) {
318 data &= ~HASH_IRQ;
319
320 if (s->regs[addr] & HASH_IRQ) {
321 qemu_irq_lower(s->irq);
322 }
323 }
324 break;
325 case R_HASH_SRC:
326 data &= ahc->src_mask;
327 break;
328 case R_HASH_DEST:
329 data &= ahc->dest_mask;
330 break;
331 case R_HASH_KEY_BUFF:
332 data &= ahc->key_mask;
333 break;
334 case R_HASH_SRC_LEN:
335 data &= 0x0FFFFFFF;
336 break;
337 case R_HASH_CMD: {
338 int algo;
339 data &= ahc->hash_mask;
340
341 if ((data & HASH_DIGEST_HMAC)) {
342 qemu_log_mask(LOG_UNIMP,
343 "%s: HMAC mode not implemented\n",
344 __func__);
345 }
346 if (data & BIT(1)) {
347 qemu_log_mask(LOG_UNIMP,
348 "%s: Cascaded mode not implemented\n",
349 __func__);
350 }
351 algo = hash_algo_lookup(data);
352 if (algo < 0) {
353 qemu_log_mask(LOG_GUEST_ERROR,
354 "%s: Invalid hash algorithm selection 0x%"PRIx64"\n",
355 __func__, data & ahc->hash_mask);
356 break;
357 }
358 do_hash_operation(s, algo, data & HASH_SG_EN,
359 ((data & HASH_HMAC_MASK) == HASH_DIGEST_ACCUM));
360
361 if (data & HASH_IRQ_EN) {
362 qemu_irq_raise(s->irq);
363 }
364 break;
365 }
366 case R_CRYPT_CMD:
367 qemu_log_mask(LOG_UNIMP, "%s: Crypt commands not implemented\n",
368 __func__);
369 break;
370 default:
371 break;
372 }
373
374 s->regs[addr] = data;
375}
376
377static const MemoryRegionOps aspeed_hace_ops = {
378 .read = aspeed_hace_read,
379 .write = aspeed_hace_write,
380 .endianness = DEVICE_LITTLE_ENDIAN,
381 .valid = {
382 .min_access_size = 1,
383 .max_access_size = 4,
384 },
385};
386
387static void aspeed_hace_reset(DeviceState *dev)
388{
389 struct AspeedHACEState *s = ASPEED_HACE(dev);
390
391 memset(s->regs, 0, sizeof(s->regs));
392 s->iov_count = 0;
393 s->total_req_len = 0;
394}
395
396static void aspeed_hace_realize(DeviceState *dev, Error **errp)
397{
398 AspeedHACEState *s = ASPEED_HACE(dev);
399 SysBusDevice *sbd = SYS_BUS_DEVICE(dev);
400
401 sysbus_init_irq(sbd, &s->irq);
402
403 memory_region_init_io(&s->iomem, OBJECT(s), &aspeed_hace_ops, s,
404 TYPE_ASPEED_HACE, 0x1000);
405
406 if (!s->dram_mr) {
407 error_setg(errp, TYPE_ASPEED_HACE ": 'dram' link not set");
408 return;
409 }
410
411 address_space_init(&s->dram_as, s->dram_mr, "dram");
412
413 sysbus_init_mmio(sbd, &s->iomem);
414}
415
416static Property aspeed_hace_properties[] = {
417 DEFINE_PROP_LINK("dram", AspeedHACEState, dram_mr,
418 TYPE_MEMORY_REGION, MemoryRegion *),
419 DEFINE_PROP_END_OF_LIST(),
420};
421
422
423static const VMStateDescription vmstate_aspeed_hace = {
424 .name = TYPE_ASPEED_HACE,
425 .version_id = 1,
426 .minimum_version_id = 1,
427 .fields = (VMStateField[]) {
428 VMSTATE_UINT32_ARRAY(regs, AspeedHACEState, ASPEED_HACE_NR_REGS),
429 VMSTATE_UINT32(total_req_len, AspeedHACEState),
430 VMSTATE_UINT32(iov_count, AspeedHACEState),
431 VMSTATE_END_OF_LIST(),
432 }
433};
434
435static void aspeed_hace_class_init(ObjectClass *klass, void *data)
436{
437 DeviceClass *dc = DEVICE_CLASS(klass);
438
439 dc->realize = aspeed_hace_realize;
440 dc->reset = aspeed_hace_reset;
441 device_class_set_props(dc, aspeed_hace_properties);
442 dc->vmsd = &vmstate_aspeed_hace;
443}
444
445static const TypeInfo aspeed_hace_info = {
446 .name = TYPE_ASPEED_HACE,
447 .parent = TYPE_SYS_BUS_DEVICE,
448 .instance_size = sizeof(AspeedHACEState),
449 .class_init = aspeed_hace_class_init,
450 .class_size = sizeof(AspeedHACEClass)
451};
452
453static void aspeed_ast2400_hace_class_init(ObjectClass *klass, void *data)
454{
455 DeviceClass *dc = DEVICE_CLASS(klass);
456 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
457
458 dc->desc = "AST2400 Hash and Crypto Engine";
459
460 ahc->src_mask = 0x0FFFFFFF;
461 ahc->dest_mask = 0x0FFFFFF8;
462 ahc->key_mask = 0x0FFFFFC0;
463 ahc->hash_mask = 0x000003ff;
464}
465
466static const TypeInfo aspeed_ast2400_hace_info = {
467 .name = TYPE_ASPEED_AST2400_HACE,
468 .parent = TYPE_ASPEED_HACE,
469 .class_init = aspeed_ast2400_hace_class_init,
470};
471
472static void aspeed_ast2500_hace_class_init(ObjectClass *klass, void *data)
473{
474 DeviceClass *dc = DEVICE_CLASS(klass);
475 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
476
477 dc->desc = "AST2500 Hash and Crypto Engine";
478
479 ahc->src_mask = 0x3fffffff;
480 ahc->dest_mask = 0x3ffffff8;
481 ahc->key_mask = 0x3FFFFFC0;
482 ahc->hash_mask = 0x000003ff;
483}
484
485static const TypeInfo aspeed_ast2500_hace_info = {
486 .name = TYPE_ASPEED_AST2500_HACE,
487 .parent = TYPE_ASPEED_HACE,
488 .class_init = aspeed_ast2500_hace_class_init,
489};
490
491static void aspeed_ast2600_hace_class_init(ObjectClass *klass, void *data)
492{
493 DeviceClass *dc = DEVICE_CLASS(klass);
494 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
495
496 dc->desc = "AST2600 Hash and Crypto Engine";
497
498 ahc->src_mask = 0x7FFFFFFF;
499 ahc->dest_mask = 0x7FFFFFF8;
500 ahc->key_mask = 0x7FFFFFF8;
501 ahc->hash_mask = 0x00147FFF;
502}
503
504static const TypeInfo aspeed_ast2600_hace_info = {
505 .name = TYPE_ASPEED_AST2600_HACE,
506 .parent = TYPE_ASPEED_HACE,
507 .class_init = aspeed_ast2600_hace_class_init,
508};
509
510static void aspeed_ast1030_hace_class_init(ObjectClass *klass, void *data)
511{
512 DeviceClass *dc = DEVICE_CLASS(klass);
513 AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
514
515 dc->desc = "AST1030 Hash and Crypto Engine";
516
517 ahc->src_mask = 0x7FFFFFFF;
518 ahc->dest_mask = 0x7FFFFFF8;
519 ahc->key_mask = 0x7FFFFFF8;
520 ahc->hash_mask = 0x00147FFF;
521}
522
523static const TypeInfo aspeed_ast1030_hace_info = {
524 .name = TYPE_ASPEED_AST1030_HACE,
525 .parent = TYPE_ASPEED_HACE,
526 .class_init = aspeed_ast1030_hace_class_init,
527};
528
529static void aspeed_hace_register_types(void)
530{
531 type_register_static(&aspeed_ast2400_hace_info);
532 type_register_static(&aspeed_ast2500_hace_info);
533 type_register_static(&aspeed_ast2600_hace_info);
534 type_register_static(&aspeed_ast1030_hace_info);
535 type_register_static(&aspeed_hace_info);
536}
537
538type_init(aspeed_hace_register_types);
539