1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24#include "msgqueue.h"
25#include <engine/falcon.h>
26
27#include <subdev/secboot.h>
28
29
30#define HDR_SIZE sizeof(struct nvkm_msgqueue_hdr)
31#define QUEUE_ALIGNMENT 4
32
33#define MSG_BUF_SIZE 128
34
35static int
36msg_queue_open(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue)
37{
38 struct nvkm_falcon *falcon = priv->falcon;
39
40 mutex_lock(&queue->mutex);
41
42 queue->position = nvkm_falcon_rd32(falcon, queue->tail_reg);
43
44 return 0;
45}
46
47static void
48msg_queue_close(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
49 bool commit)
50{
51 struct nvkm_falcon *falcon = priv->falcon;
52
53 if (commit)
54 nvkm_falcon_wr32(falcon, queue->tail_reg, queue->position);
55
56 mutex_unlock(&queue->mutex);
57}
58
59static bool
60msg_queue_empty(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue)
61{
62 struct nvkm_falcon *falcon = priv->falcon;
63 u32 head, tail;
64
65 head = nvkm_falcon_rd32(falcon, queue->head_reg);
66 tail = nvkm_falcon_rd32(falcon, queue->tail_reg);
67
68 return head == tail;
69}
70
71static int
72msg_queue_pop(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
73 void *data, u32 size)
74{
75 struct nvkm_falcon *falcon = priv->falcon;
76 const struct nvkm_subdev *subdev = priv->falcon->owner;
77 u32 head, tail, available;
78
79 head = nvkm_falcon_rd32(falcon, queue->head_reg);
80
81 if (head < queue->position)
82 queue->position = queue->offset;
83
84 tail = queue->position;
85
86 available = head - tail;
87
88 if (available == 0) {
89 nvkm_warn(subdev, "no message data available\n");
90 return 0;
91 }
92
93 if (size > available) {
94 nvkm_warn(subdev, "message data smaller than read request\n");
95 size = available;
96 }
97
98 nvkm_falcon_read_dmem(priv->falcon, tail, size, 0, data);
99 queue->position += ALIGN(size, QUEUE_ALIGNMENT);
100
101 return size;
102}
103
104static int
105msg_queue_read(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
106 struct nvkm_msgqueue_hdr *hdr)
107{
108 const struct nvkm_subdev *subdev = priv->falcon->owner;
109 int err;
110
111 err = msg_queue_open(priv, queue);
112 if (err) {
113 nvkm_error(subdev, "fail to open queue %d\n", queue->index);
114 return err;
115 }
116
117 if (msg_queue_empty(priv, queue)) {
118 err = 0;
119 goto close;
120 }
121
122 err = msg_queue_pop(priv, queue, hdr, HDR_SIZE);
123 if (err >= 0 && err != HDR_SIZE)
124 err = -EINVAL;
125 if (err < 0) {
126 nvkm_error(subdev, "failed to read message header: %d\n", err);
127 goto close;
128 }
129
130 if (hdr->size > MSG_BUF_SIZE) {
131 nvkm_error(subdev, "message too big (%d bytes)\n", hdr->size);
132 err = -ENOSPC;
133 goto close;
134 }
135
136 if (hdr->size > HDR_SIZE) {
137 u32 read_size = hdr->size - HDR_SIZE;
138
139 err = msg_queue_pop(priv, queue, (hdr + 1), read_size);
140 if (err >= 0 && err != read_size)
141 err = -EINVAL;
142 if (err < 0) {
143 nvkm_error(subdev, "failed to read message: %d\n", err);
144 goto close;
145 }
146 }
147
148close:
149 msg_queue_close(priv, queue, (err >= 0));
150
151 return err;
152}
153
154static bool
155cmd_queue_has_room(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
156 u32 size, bool *rewind)
157{
158 struct nvkm_falcon *falcon = priv->falcon;
159 u32 head, tail, free;
160
161 size = ALIGN(size, QUEUE_ALIGNMENT);
162
163 head = nvkm_falcon_rd32(falcon, queue->head_reg);
164 tail = nvkm_falcon_rd32(falcon, queue->tail_reg);
165
166 if (head >= tail) {
167 free = queue->offset + queue->size - head;
168 free -= HDR_SIZE;
169
170 if (size > free) {
171 *rewind = true;
172 head = queue->offset;
173 }
174 }
175
176 if (head < tail)
177 free = tail - head - 1;
178
179 return size <= free;
180}
181
182static int
183cmd_queue_push(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
184 void *data, u32 size)
185{
186 nvkm_falcon_load_dmem(priv->falcon, data, queue->position, size, 0);
187 queue->position += ALIGN(size, QUEUE_ALIGNMENT);
188
189 return 0;
190}
191
192
193#define MSGQUEUE_UNIT_REWIND 0x00
194
195static void
196cmd_queue_rewind(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue)
197{
198 const struct nvkm_subdev *subdev = priv->falcon->owner;
199 struct nvkm_msgqueue_hdr cmd;
200 int err;
201
202 cmd.unit_id = MSGQUEUE_UNIT_REWIND;
203 cmd.size = sizeof(cmd);
204 err = cmd_queue_push(priv, queue, &cmd, cmd.size);
205 if (err)
206 nvkm_error(subdev, "queue %d rewind failed\n", queue->index);
207 else
208 nvkm_error(subdev, "queue %d rewinded\n", queue->index);
209
210 queue->position = queue->offset;
211}
212
213static int
214cmd_queue_open(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
215 u32 size)
216{
217 struct nvkm_falcon *falcon = priv->falcon;
218 const struct nvkm_subdev *subdev = priv->falcon->owner;
219 bool rewind = false;
220
221 mutex_lock(&queue->mutex);
222
223 if (!cmd_queue_has_room(priv, queue, size, &rewind)) {
224 nvkm_error(subdev, "queue full\n");
225 mutex_unlock(&queue->mutex);
226 return -EAGAIN;
227 }
228
229 queue->position = nvkm_falcon_rd32(falcon, queue->head_reg);
230
231 if (rewind)
232 cmd_queue_rewind(priv, queue);
233
234 return 0;
235}
236
237static void
238cmd_queue_close(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_queue *queue,
239 bool commit)
240{
241 struct nvkm_falcon *falcon = priv->falcon;
242
243 if (commit)
244 nvkm_falcon_wr32(falcon, queue->head_reg, queue->position);
245
246 mutex_unlock(&queue->mutex);
247}
248
249static int
250cmd_write(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_hdr *cmd,
251 struct nvkm_msgqueue_queue *queue)
252{
253 const struct nvkm_subdev *subdev = priv->falcon->owner;
254 static unsigned timeout = 2000;
255 unsigned long end_jiffies = jiffies + msecs_to_jiffies(timeout);
256 int ret = -EAGAIN;
257 bool commit = true;
258
259 while (ret == -EAGAIN && time_before(jiffies, end_jiffies))
260 ret = cmd_queue_open(priv, queue, cmd->size);
261 if (ret) {
262 nvkm_error(subdev, "pmu_queue_open_write failed\n");
263 return ret;
264 }
265
266 ret = cmd_queue_push(priv, queue, cmd, cmd->size);
267 if (ret) {
268 nvkm_error(subdev, "pmu_queue_push failed\n");
269 commit = false;
270 }
271
272 cmd_queue_close(priv, queue, commit);
273
274 return ret;
275}
276
277static struct nvkm_msgqueue_seq *
278msgqueue_seq_acquire(struct nvkm_msgqueue *priv)
279{
280 const struct nvkm_subdev *subdev = priv->falcon->owner;
281 struct nvkm_msgqueue_seq *seq;
282 u32 index;
283
284 mutex_lock(&priv->seq_lock);
285
286 index = find_first_zero_bit(priv->seq_tbl, NVKM_MSGQUEUE_NUM_SEQUENCES);
287
288 if (index >= NVKM_MSGQUEUE_NUM_SEQUENCES) {
289 nvkm_error(subdev, "no free sequence available\n");
290 mutex_unlock(&priv->seq_lock);
291 return ERR_PTR(-EAGAIN);
292 }
293
294 set_bit(index, priv->seq_tbl);
295
296 mutex_unlock(&priv->seq_lock);
297
298 seq = &priv->seq[index];
299 seq->state = SEQ_STATE_PENDING;
300
301 return seq;
302}
303
304static void
305msgqueue_seq_release(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_seq *seq)
306{
307
308 seq->state = SEQ_STATE_FREE;
309 seq->callback = NULL;
310 seq->completion = NULL;
311 clear_bit(seq->id, priv->seq_tbl);
312}
313
314
315#define CMD_FLAGS_STATUS BIT(0)
316
317#define CMD_FLAGS_INTR BIT(1)
318
319int
320nvkm_msgqueue_post(struct nvkm_msgqueue *priv, enum msgqueue_msg_priority prio,
321 struct nvkm_msgqueue_hdr *cmd, nvkm_msgqueue_callback cb,
322 struct completion *completion, bool wait_init)
323{
324 struct nvkm_msgqueue_seq *seq;
325 struct nvkm_msgqueue_queue *queue;
326 int ret;
327
328 if (wait_init && !wait_for_completion_timeout(&priv->init_done,
329 msecs_to_jiffies(1000)))
330 return -ETIMEDOUT;
331
332 queue = priv->func->cmd_queue(priv, prio);
333 if (IS_ERR(queue))
334 return PTR_ERR(queue);
335
336 seq = msgqueue_seq_acquire(priv);
337 if (IS_ERR(seq))
338 return PTR_ERR(seq);
339
340 cmd->seq_id = seq->id;
341 cmd->ctrl_flags = CMD_FLAGS_STATUS | CMD_FLAGS_INTR;
342
343 seq->callback = cb;
344 seq->state = SEQ_STATE_USED;
345 seq->completion = completion;
346
347 ret = cmd_write(priv, cmd, queue);
348 if (ret) {
349 seq->state = SEQ_STATE_PENDING;
350 msgqueue_seq_release(priv, seq);
351 }
352
353 return ret;
354}
355
356static int
357msgqueue_msg_handle(struct nvkm_msgqueue *priv, struct nvkm_msgqueue_hdr *hdr)
358{
359 const struct nvkm_subdev *subdev = priv->falcon->owner;
360 struct nvkm_msgqueue_seq *seq;
361
362 seq = &priv->seq[hdr->seq_id];
363 if (seq->state != SEQ_STATE_USED && seq->state != SEQ_STATE_CANCELLED) {
364 nvkm_error(subdev, "msg for unknown sequence %d", seq->id);
365 return -EINVAL;
366 }
367
368 if (seq->state == SEQ_STATE_USED) {
369 if (seq->callback)
370 seq->callback(priv, hdr);
371 }
372
373 if (seq->completion)
374 complete(seq->completion);
375
376 msgqueue_seq_release(priv, seq);
377
378 return 0;
379}
380
381static int
382msgqueue_handle_init_msg(struct nvkm_msgqueue *priv,
383 struct nvkm_msgqueue_hdr *hdr)
384{
385 struct nvkm_falcon *falcon = priv->falcon;
386 const struct nvkm_subdev *subdev = falcon->owner;
387 u32 tail;
388 u32 tail_reg;
389 int ret;
390
391
392
393
394
395 switch (falcon->owner->index) {
396 case NVKM_SUBDEV_PMU:
397 tail_reg = 0x4cc;
398 break;
399 case NVKM_ENGINE_SEC2:
400 tail_reg = 0xa34;
401 break;
402 default:
403 nvkm_error(subdev, "falcon %s unsupported for msgqueue!\n",
404 nvkm_subdev_name[falcon->owner->index]);
405 return -EINVAL;
406 }
407
408
409
410
411
412 tail = nvkm_falcon_rd32(falcon, tail_reg);
413 nvkm_falcon_read_dmem(falcon, tail, HDR_SIZE, 0, hdr);
414
415 if (hdr->size > MSG_BUF_SIZE) {
416 nvkm_error(subdev, "message too big (%d bytes)\n", hdr->size);
417 return -ENOSPC;
418 }
419
420 nvkm_falcon_read_dmem(falcon, tail + HDR_SIZE, hdr->size - HDR_SIZE, 0,
421 (hdr + 1));
422
423 tail += ALIGN(hdr->size, QUEUE_ALIGNMENT);
424 nvkm_falcon_wr32(falcon, tail_reg, tail);
425
426 ret = priv->func->init_func->init_callback(priv, hdr);
427 if (ret)
428 return ret;
429
430 return 0;
431}
432
433void
434nvkm_msgqueue_process_msgs(struct nvkm_msgqueue *priv,
435 struct nvkm_msgqueue_queue *queue)
436{
437
438
439
440
441 u8 msg_buffer[MSG_BUF_SIZE];
442 struct nvkm_msgqueue_hdr *hdr = (void *)msg_buffer;
443 int ret;
444
445
446 if ((!priv->init_msg_received)) {
447 ret = msgqueue_handle_init_msg(priv, hdr);
448 if (!ret)
449 priv->init_msg_received = true;
450 } else {
451 while (msg_queue_read(priv, queue, hdr) > 0)
452 msgqueue_msg_handle(priv, hdr);
453 }
454}
455
456void
457nvkm_msgqueue_write_cmdline(struct nvkm_msgqueue *queue, void *buf)
458{
459 if (!queue || !queue->func || !queue->func->init_func)
460 return;
461
462 queue->func->init_func->gen_cmdline(queue, buf);
463}
464
465int
466nvkm_msgqueue_acr_boot_falcons(struct nvkm_msgqueue *queue,
467 unsigned long falcon_mask)
468{
469 unsigned long falcon;
470
471 if (!queue || !queue->func->acr_func)
472 return -ENODEV;
473
474
475 if (queue->func->acr_func->boot_multiple_falcons)
476 return queue->func->acr_func->boot_multiple_falcons(queue,
477 falcon_mask);
478
479
480 if (!queue->func->acr_func->boot_falcon)
481 return -ENODEV;
482
483 for_each_set_bit(falcon, &falcon_mask, NVKM_SECBOOT_FALCON_END) {
484 int ret = queue->func->acr_func->boot_falcon(queue, falcon);
485
486 if (ret)
487 return ret;
488 }
489
490 return 0;
491}
492
493int
494nvkm_msgqueue_new(u32 version, struct nvkm_falcon *falcon,
495 const struct nvkm_secboot *sb, struct nvkm_msgqueue **queue)
496{
497 const struct nvkm_subdev *subdev = falcon->owner;
498 int ret = -EINVAL;
499
500 switch (version) {
501 case 0x0137c63d:
502 ret = msgqueue_0137c63d_new(falcon, sb, queue);
503 break;
504 case 0x0137bca5:
505 ret = msgqueue_0137bca5_new(falcon, sb, queue);
506 break;
507 case 0x0148cdec:
508 case 0x015ccf3e:
509 ret = msgqueue_0148cdec_new(falcon, sb, queue);
510 break;
511 default:
512 nvkm_error(subdev, "unhandled firmware version 0x%08x\n",
513 version);
514 break;
515 }
516
517 if (ret == 0) {
518 nvkm_debug(subdev, "firmware version: 0x%08x\n", version);
519 (*queue)->fw_version = version;
520 }
521
522 return ret;
523}
524
525void
526nvkm_msgqueue_del(struct nvkm_msgqueue **queue)
527{
528 if (*queue) {
529 (*queue)->func->dtor(*queue);
530 *queue = NULL;
531 }
532}
533
534void
535nvkm_msgqueue_recv(struct nvkm_msgqueue *queue)
536{
537 if (!queue->func || !queue->func->recv) {
538 const struct nvkm_subdev *subdev = queue->falcon->owner;
539
540 nvkm_warn(subdev, "missing msgqueue recv function\n");
541 return;
542 }
543
544 queue->func->recv(queue);
545}
546
547int
548nvkm_msgqueue_reinit(struct nvkm_msgqueue *queue)
549{
550
551 if (!queue)
552 return 0;
553
554 queue->init_msg_received = false;
555 reinit_completion(&queue->init_done);
556
557 return 0;
558}
559
560void
561nvkm_msgqueue_ctor(const struct nvkm_msgqueue_func *func,
562 struct nvkm_falcon *falcon,
563 struct nvkm_msgqueue *queue)
564{
565 int i;
566
567 queue->func = func;
568 queue->falcon = falcon;
569 mutex_init(&queue->seq_lock);
570 for (i = 0; i < NVKM_MSGQUEUE_NUM_SEQUENCES; i++)
571 queue->seq[i].id = i;
572
573 init_completion(&queue->init_done);
574
575
576}
577