1
2
3
4
5
6
7
8
9
10
11#include <linux/bitops.h>
12#include <linux/clk.h>
13#include <linux/dmaengine.h>
14#include <linux/dma-mapping.h>
15#include <linux/err.h>
16#include <linux/iopoll.h>
17#include <linux/interrupt.h>
18#include <linux/list.h>
19#include <linux/module.h>
20#include <linux/of.h>
21#include <linux/of_device.h>
22#include <linux/of_dma.h>
23#include <linux/platform_device.h>
24#include <linux/pm_runtime.h>
25#include <linux/refcount.h>
26#include <linux/slab.h>
27
28#include "../virt-dma.h"
29
30#define MTK_CQDMA_USEC_POLL 10
31#define MTK_CQDMA_TIMEOUT_POLL 1000
32#define MTK_CQDMA_DMA_BUSWIDTHS BIT(DMA_SLAVE_BUSWIDTH_4_BYTES)
33#define MTK_CQDMA_ALIGN_SIZE 1
34
35
36#define MTK_CQDMA_NR_VCHANS 32
37
38
39#define MTK_CQDMA_NR_PCHANS 3
40
41
42#define MTK_CQDMA_INT_FLAG 0x0
43#define MTK_CQDMA_INT_EN 0x4
44#define MTK_CQDMA_EN 0x8
45#define MTK_CQDMA_RESET 0xc
46#define MTK_CQDMA_FLUSH 0x14
47#define MTK_CQDMA_SRC 0x1c
48#define MTK_CQDMA_DST 0x20
49#define MTK_CQDMA_LEN1 0x24
50#define MTK_CQDMA_LEN2 0x28
51#define MTK_CQDMA_SRC2 0x60
52#define MTK_CQDMA_DST2 0x64
53
54
55#define MTK_CQDMA_EN_BIT BIT(0)
56#define MTK_CQDMA_INT_FLAG_BIT BIT(0)
57#define MTK_CQDMA_INT_EN_BIT BIT(0)
58#define MTK_CQDMA_FLUSH_BIT BIT(0)
59
60#define MTK_CQDMA_WARM_RST_BIT BIT(0)
61#define MTK_CQDMA_HARD_RST_BIT BIT(1)
62
63#define MTK_CQDMA_MAX_LEN GENMASK(27, 0)
64#define MTK_CQDMA_ADDR_LIMIT GENMASK(31, 0)
65#define MTK_CQDMA_ADDR2_SHFIT (32)
66
67
68
69
70
71
72
73
74
75
76
77
78
79struct mtk_cqdma_vdesc {
80 struct virt_dma_desc vd;
81 size_t len;
82 size_t residue;
83 dma_addr_t dest;
84 dma_addr_t src;
85 struct dma_chan *ch;
86
87 struct list_head node;
88 struct mtk_cqdma_vdesc *parent;
89};
90
91
92
93
94
95
96
97
98
99
100
101struct mtk_cqdma_pchan {
102 struct list_head queue;
103 void __iomem *base;
104 u32 irq;
105
106 refcount_t refcnt;
107
108 struct tasklet_struct tasklet;
109
110
111 spinlock_t lock;
112};
113
114
115
116
117
118
119
120
121
122struct mtk_cqdma_vchan {
123 struct virt_dma_chan vc;
124 struct mtk_cqdma_pchan *pc;
125 struct completion issue_completion;
126 bool issue_synchronize;
127};
128
129
130
131
132
133
134
135
136
137
138
139struct mtk_cqdma_device {
140 struct dma_device ddev;
141 struct clk *clk;
142
143 u32 dma_requests;
144 u32 dma_channels;
145 struct mtk_cqdma_vchan *vc;
146 struct mtk_cqdma_pchan **pc;
147};
148
149static struct mtk_cqdma_device *to_cqdma_dev(struct dma_chan *chan)
150{
151 return container_of(chan->device, struct mtk_cqdma_device, ddev);
152}
153
154static struct mtk_cqdma_vchan *to_cqdma_vchan(struct dma_chan *chan)
155{
156 return container_of(chan, struct mtk_cqdma_vchan, vc.chan);
157}
158
159static struct mtk_cqdma_vdesc *to_cqdma_vdesc(struct virt_dma_desc *vd)
160{
161 return container_of(vd, struct mtk_cqdma_vdesc, vd);
162}
163
164static struct device *cqdma2dev(struct mtk_cqdma_device *cqdma)
165{
166 return cqdma->ddev.dev;
167}
168
169static u32 mtk_dma_read(struct mtk_cqdma_pchan *pc, u32 reg)
170{
171 return readl(pc->base + reg);
172}
173
174static void mtk_dma_write(struct mtk_cqdma_pchan *pc, u32 reg, u32 val)
175{
176 writel_relaxed(val, pc->base + reg);
177}
178
179static void mtk_dma_rmw(struct mtk_cqdma_pchan *pc, u32 reg,
180 u32 mask, u32 set)
181{
182 u32 val;
183
184 val = mtk_dma_read(pc, reg);
185 val &= ~mask;
186 val |= set;
187 mtk_dma_write(pc, reg, val);
188}
189
190static void mtk_dma_set(struct mtk_cqdma_pchan *pc, u32 reg, u32 val)
191{
192 mtk_dma_rmw(pc, reg, 0, val);
193}
194
195static void mtk_dma_clr(struct mtk_cqdma_pchan *pc, u32 reg, u32 val)
196{
197 mtk_dma_rmw(pc, reg, val, 0);
198}
199
200static void mtk_cqdma_vdesc_free(struct virt_dma_desc *vd)
201{
202 kfree(to_cqdma_vdesc(vd));
203}
204
205static int mtk_cqdma_poll_engine_done(struct mtk_cqdma_pchan *pc, bool atomic)
206{
207 u32 status = 0;
208
209 if (!atomic)
210 return readl_poll_timeout(pc->base + MTK_CQDMA_EN,
211 status,
212 !(status & MTK_CQDMA_EN_BIT),
213 MTK_CQDMA_USEC_POLL,
214 MTK_CQDMA_TIMEOUT_POLL);
215
216 return readl_poll_timeout_atomic(pc->base + MTK_CQDMA_EN,
217 status,
218 !(status & MTK_CQDMA_EN_BIT),
219 MTK_CQDMA_USEC_POLL,
220 MTK_CQDMA_TIMEOUT_POLL);
221}
222
223static int mtk_cqdma_hard_reset(struct mtk_cqdma_pchan *pc)
224{
225 mtk_dma_set(pc, MTK_CQDMA_RESET, MTK_CQDMA_HARD_RST_BIT);
226 mtk_dma_clr(pc, MTK_CQDMA_RESET, MTK_CQDMA_HARD_RST_BIT);
227
228 return mtk_cqdma_poll_engine_done(pc, false);
229}
230
231static void mtk_cqdma_start(struct mtk_cqdma_pchan *pc,
232 struct mtk_cqdma_vdesc *cvd)
233{
234
235 if (mtk_cqdma_poll_engine_done(pc, true) < 0)
236 dev_err(cqdma2dev(to_cqdma_dev(cvd->ch)), "cqdma wait transaction timeout\n");
237
238
239 mtk_dma_set(pc, MTK_CQDMA_RESET, MTK_CQDMA_WARM_RST_BIT);
240 if (mtk_cqdma_poll_engine_done(pc, true) < 0)
241 dev_err(cqdma2dev(to_cqdma_dev(cvd->ch)), "cqdma warm reset timeout\n");
242
243
244 mtk_dma_set(pc, MTK_CQDMA_SRC, cvd->src & MTK_CQDMA_ADDR_LIMIT);
245#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
246 mtk_dma_set(pc, MTK_CQDMA_SRC2, cvd->src >> MTK_CQDMA_ADDR2_SHFIT);
247#else
248 mtk_dma_set(pc, MTK_CQDMA_SRC2, 0);
249#endif
250
251
252 mtk_dma_set(pc, MTK_CQDMA_DST, cvd->dest & MTK_CQDMA_ADDR_LIMIT);
253#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
254 mtk_dma_set(pc, MTK_CQDMA_DST2, cvd->dest >> MTK_CQDMA_ADDR2_SHFIT);
255#else
256 mtk_dma_set(pc, MTK_CQDMA_SRC2, 0);
257#endif
258
259
260 mtk_dma_set(pc, MTK_CQDMA_LEN1, cvd->len);
261
262
263 mtk_dma_set(pc, MTK_CQDMA_EN, MTK_CQDMA_EN_BIT);
264}
265
266static void mtk_cqdma_issue_vchan_pending(struct mtk_cqdma_vchan *cvc)
267{
268 struct virt_dma_desc *vd, *vd2;
269 struct mtk_cqdma_pchan *pc = cvc->pc;
270 struct mtk_cqdma_vdesc *cvd;
271 bool trigger_engine = false;
272
273 lockdep_assert_held(&cvc->vc.lock);
274 lockdep_assert_held(&pc->lock);
275
276 list_for_each_entry_safe(vd, vd2, &cvc->vc.desc_issued, node) {
277
278 if (list_empty(&pc->queue))
279 trigger_engine = true;
280
281 cvd = to_cqdma_vdesc(vd);
282
283
284 list_add_tail(&cvd->node, &pc->queue);
285
286
287 if (trigger_engine)
288 mtk_cqdma_start(pc, cvd);
289
290
291 list_del(&vd->node);
292 }
293}
294
295
296
297
298
299static bool mtk_cqdma_is_vchan_active(struct mtk_cqdma_vchan *cvc)
300{
301 struct mtk_cqdma_vdesc *cvd;
302
303 list_for_each_entry(cvd, &cvc->pc->queue, node)
304 if (cvc == to_cqdma_vchan(cvd->ch))
305 return true;
306
307 return false;
308}
309
310
311
312
313static struct mtk_cqdma_vdesc
314*mtk_cqdma_consume_work_queue(struct mtk_cqdma_pchan *pc)
315{
316 struct mtk_cqdma_vchan *cvc;
317 struct mtk_cqdma_vdesc *cvd, *ret = NULL;
318
319
320 cvd = list_first_entry_or_null(&pc->queue,
321 struct mtk_cqdma_vdesc, node);
322 if (unlikely(!cvd || !cvd->parent))
323 return NULL;
324
325 cvc = to_cqdma_vchan(cvd->ch);
326 ret = cvd;
327
328
329 cvd->parent->residue -= cvd->len;
330
331
332 list_del(&cvd->node);
333
334 spin_lock(&cvc->vc.lock);
335
336
337 if (!cvd->parent->residue) {
338
339 vchan_cookie_complete(&cvd->parent->vd);
340
341
342 if (cvc->issue_synchronize && !mtk_cqdma_is_vchan_active(cvc)) {
343 complete(&cvc->issue_completion);
344 cvc->issue_synchronize = false;
345 }
346 }
347
348 spin_unlock(&cvc->vc.lock);
349
350
351 cvd = list_first_entry_or_null(&pc->queue,
352 struct mtk_cqdma_vdesc, node);
353 if (cvd)
354 mtk_cqdma_start(pc, cvd);
355
356 return ret;
357}
358
359static void mtk_cqdma_tasklet_cb(unsigned long data)
360{
361 struct mtk_cqdma_pchan *pc = (struct mtk_cqdma_pchan *)data;
362 struct mtk_cqdma_vdesc *cvd = NULL;
363 unsigned long flags;
364
365 spin_lock_irqsave(&pc->lock, flags);
366
367 cvd = mtk_cqdma_consume_work_queue(pc);
368 spin_unlock_irqrestore(&pc->lock, flags);
369
370
371 if (cvd) {
372 dma_run_dependencies(&cvd->vd.tx);
373
374
375
376
377
378 if (cvd->parent != cvd)
379 kfree(cvd);
380 }
381
382
383 enable_irq(pc->irq);
384}
385
386static irqreturn_t mtk_cqdma_irq(int irq, void *devid)
387{
388 struct mtk_cqdma_device *cqdma = devid;
389 irqreturn_t ret = IRQ_NONE;
390 bool schedule_tasklet = false;
391 u32 i;
392
393
394 for (i = 0; i < cqdma->dma_channels; ++i, schedule_tasklet = false) {
395 spin_lock(&cqdma->pc[i]->lock);
396 if (mtk_dma_read(cqdma->pc[i],
397 MTK_CQDMA_INT_FLAG) & MTK_CQDMA_INT_FLAG_BIT) {
398
399 mtk_dma_clr(cqdma->pc[i], MTK_CQDMA_INT_FLAG,
400 MTK_CQDMA_INT_FLAG_BIT);
401
402 schedule_tasklet = true;
403 ret = IRQ_HANDLED;
404 }
405 spin_unlock(&cqdma->pc[i]->lock);
406
407 if (schedule_tasklet) {
408
409 disable_irq_nosync(cqdma->pc[i]->irq);
410
411
412 tasklet_schedule(&cqdma->pc[i]->tasklet);
413 }
414 }
415
416 return ret;
417}
418
419static struct virt_dma_desc *mtk_cqdma_find_active_desc(struct dma_chan *c,
420 dma_cookie_t cookie)
421{
422 struct mtk_cqdma_vchan *cvc = to_cqdma_vchan(c);
423 struct virt_dma_desc *vd;
424 unsigned long flags;
425
426 spin_lock_irqsave(&cvc->pc->lock, flags);
427 list_for_each_entry(vd, &cvc->pc->queue, node)
428 if (vd->tx.cookie == cookie) {
429 spin_unlock_irqrestore(&cvc->pc->lock, flags);
430 return vd;
431 }
432 spin_unlock_irqrestore(&cvc->pc->lock, flags);
433
434 list_for_each_entry(vd, &cvc->vc.desc_issued, node)
435 if (vd->tx.cookie == cookie)
436 return vd;
437
438 return NULL;
439}
440
441static enum dma_status mtk_cqdma_tx_status(struct dma_chan *c,
442 dma_cookie_t cookie,
443 struct dma_tx_state *txstate)
444{
445 struct mtk_cqdma_vchan *cvc = to_cqdma_vchan(c);
446 struct mtk_cqdma_vdesc *cvd;
447 struct virt_dma_desc *vd;
448 enum dma_status ret;
449 unsigned long flags;
450 size_t bytes = 0;
451
452 ret = dma_cookie_status(c, cookie, txstate);
453 if (ret == DMA_COMPLETE || !txstate)
454 return ret;
455
456 spin_lock_irqsave(&cvc->vc.lock, flags);
457 vd = mtk_cqdma_find_active_desc(c, cookie);
458 spin_unlock_irqrestore(&cvc->vc.lock, flags);
459
460 if (vd) {
461 cvd = to_cqdma_vdesc(vd);
462 bytes = cvd->residue;
463 }
464
465 dma_set_residue(txstate, bytes);
466
467 return ret;
468}
469
470static void mtk_cqdma_issue_pending(struct dma_chan *c)
471{
472 struct mtk_cqdma_vchan *cvc = to_cqdma_vchan(c);
473 unsigned long pc_flags;
474 unsigned long vc_flags;
475
476
477 spin_lock_irqsave(&cvc->pc->lock, pc_flags);
478 spin_lock_irqsave(&cvc->vc.lock, vc_flags);
479
480 if (vchan_issue_pending(&cvc->vc))
481 mtk_cqdma_issue_vchan_pending(cvc);
482
483 spin_unlock_irqrestore(&cvc->vc.lock, vc_flags);
484 spin_unlock_irqrestore(&cvc->pc->lock, pc_flags);
485}
486
487static struct dma_async_tx_descriptor *
488mtk_cqdma_prep_dma_memcpy(struct dma_chan *c, dma_addr_t dest,
489 dma_addr_t src, size_t len, unsigned long flags)
490{
491 struct mtk_cqdma_vdesc **cvd;
492 struct dma_async_tx_descriptor *tx = NULL, *prev_tx = NULL;
493 size_t i, tlen, nr_vd;
494
495
496
497
498
499
500
501
502
503
504
505
506 nr_vd = DIV_ROUND_UP(len, MTK_CQDMA_MAX_LEN);
507 cvd = kcalloc(nr_vd, sizeof(*cvd), GFP_NOWAIT);
508 if (!cvd)
509 return NULL;
510
511 for (i = 0; i < nr_vd; ++i) {
512 cvd[i] = kzalloc(sizeof(*cvd[i]), GFP_NOWAIT);
513 if (!cvd[i]) {
514 for (; i > 0; --i)
515 kfree(cvd[i - 1]);
516 return NULL;
517 }
518
519
520 cvd[i]->ch = c;
521
522
523 tlen = (len > MTK_CQDMA_MAX_LEN) ? MTK_CQDMA_MAX_LEN : len;
524 cvd[i]->len = tlen;
525 cvd[i]->src = src;
526 cvd[i]->dest = dest;
527
528
529 tx = vchan_tx_prep(to_virt_chan(c), &cvd[i]->vd, flags);
530 tx->next = NULL;
531
532 if (!i) {
533 cvd[0]->residue = len;
534 } else {
535 prev_tx->next = tx;
536 cvd[i]->residue = tlen;
537 }
538
539 cvd[i]->parent = cvd[0];
540
541
542 src += tlen;
543 dest += tlen;
544 len -= tlen;
545 prev_tx = tx;
546 }
547
548 return &cvd[0]->vd.tx;
549}
550
551static void mtk_cqdma_free_inactive_desc(struct dma_chan *c)
552{
553 struct virt_dma_chan *vc = to_virt_chan(c);
554 unsigned long flags;
555 LIST_HEAD(head);
556
557
558
559
560
561 spin_lock_irqsave(&vc->lock, flags);
562 list_splice_tail_init(&vc->desc_allocated, &head);
563 list_splice_tail_init(&vc->desc_submitted, &head);
564 list_splice_tail_init(&vc->desc_issued, &head);
565 spin_unlock_irqrestore(&vc->lock, flags);
566
567
568 vchan_dma_desc_free_list(vc, &head);
569}
570
571static void mtk_cqdma_free_active_desc(struct dma_chan *c)
572{
573 struct mtk_cqdma_vchan *cvc = to_cqdma_vchan(c);
574 bool sync_needed = false;
575 unsigned long pc_flags;
576 unsigned long vc_flags;
577
578
579 spin_lock_irqsave(&cvc->pc->lock, pc_flags);
580 spin_lock_irqsave(&cvc->vc.lock, vc_flags);
581
582
583 if (mtk_cqdma_is_vchan_active(cvc)) {
584 cvc->issue_synchronize = true;
585 sync_needed = true;
586 }
587
588 spin_unlock_irqrestore(&cvc->vc.lock, vc_flags);
589 spin_unlock_irqrestore(&cvc->pc->lock, pc_flags);
590
591
592 if (sync_needed)
593 wait_for_completion(&cvc->issue_completion);
594
595
596 vchan_synchronize(&cvc->vc);
597
598 WARN_ONCE(!list_empty(&cvc->vc.desc_completed),
599 "Desc pending still in list desc_completed\n");
600}
601
602static int mtk_cqdma_terminate_all(struct dma_chan *c)
603{
604
605 mtk_cqdma_free_inactive_desc(c);
606
607
608 mtk_cqdma_free_active_desc(c);
609
610 return 0;
611}
612
613static int mtk_cqdma_alloc_chan_resources(struct dma_chan *c)
614{
615 struct mtk_cqdma_device *cqdma = to_cqdma_dev(c);
616 struct mtk_cqdma_vchan *vc = to_cqdma_vchan(c);
617 struct mtk_cqdma_pchan *pc = NULL;
618 u32 i, min_refcnt = U32_MAX, refcnt;
619 unsigned long flags;
620
621
622 for (i = 0; i < cqdma->dma_channels; ++i) {
623 refcnt = refcount_read(&cqdma->pc[i]->refcnt);
624 if (refcnt < min_refcnt) {
625 pc = cqdma->pc[i];
626 min_refcnt = refcnt;
627 }
628 }
629
630 if (!pc)
631 return -ENOSPC;
632
633 spin_lock_irqsave(&pc->lock, flags);
634
635 if (!refcount_read(&pc->refcnt)) {
636
637 mtk_cqdma_hard_reset(pc);
638
639
640 mtk_dma_set(pc, MTK_CQDMA_INT_EN, MTK_CQDMA_INT_EN_BIT);
641
642
643
644
645
646 refcount_set(&pc->refcnt, 1);
647 } else {
648 refcount_inc(&pc->refcnt);
649 }
650
651 spin_unlock_irqrestore(&pc->lock, flags);
652
653 vc->pc = pc;
654
655 return 0;
656}
657
658static void mtk_cqdma_free_chan_resources(struct dma_chan *c)
659{
660 struct mtk_cqdma_vchan *cvc = to_cqdma_vchan(c);
661 unsigned long flags;
662
663
664 mtk_cqdma_terminate_all(c);
665
666 spin_lock_irqsave(&cvc->pc->lock, flags);
667
668
669 if (refcount_dec_and_test(&cvc->pc->refcnt)) {
670
671 mtk_dma_set(cvc->pc, MTK_CQDMA_FLUSH, MTK_CQDMA_FLUSH_BIT);
672
673
674 if (mtk_cqdma_poll_engine_done(cvc->pc, false) < 0)
675 dev_err(cqdma2dev(to_cqdma_dev(c)), "cqdma flush timeout\n");
676
677
678 mtk_dma_clr(cvc->pc, MTK_CQDMA_FLUSH, MTK_CQDMA_FLUSH_BIT);
679 mtk_dma_clr(cvc->pc, MTK_CQDMA_INT_FLAG,
680 MTK_CQDMA_INT_FLAG_BIT);
681
682
683 mtk_dma_clr(cvc->pc, MTK_CQDMA_INT_EN, MTK_CQDMA_INT_EN_BIT);
684 }
685
686 spin_unlock_irqrestore(&cvc->pc->lock, flags);
687}
688
689static int mtk_cqdma_hw_init(struct mtk_cqdma_device *cqdma)
690{
691 unsigned long flags;
692 int err;
693 u32 i;
694
695 pm_runtime_enable(cqdma2dev(cqdma));
696 pm_runtime_get_sync(cqdma2dev(cqdma));
697
698 err = clk_prepare_enable(cqdma->clk);
699
700 if (err) {
701 pm_runtime_put_sync(cqdma2dev(cqdma));
702 pm_runtime_disable(cqdma2dev(cqdma));
703 return err;
704 }
705
706
707 for (i = 0; i < cqdma->dma_channels; ++i) {
708 spin_lock_irqsave(&cqdma->pc[i]->lock, flags);
709 if (mtk_cqdma_hard_reset(cqdma->pc[i]) < 0) {
710 dev_err(cqdma2dev(cqdma), "cqdma hard reset timeout\n");
711 spin_unlock_irqrestore(&cqdma->pc[i]->lock, flags);
712
713 clk_disable_unprepare(cqdma->clk);
714 pm_runtime_put_sync(cqdma2dev(cqdma));
715 pm_runtime_disable(cqdma2dev(cqdma));
716 return -EINVAL;
717 }
718 spin_unlock_irqrestore(&cqdma->pc[i]->lock, flags);
719 }
720
721 return 0;
722}
723
724static void mtk_cqdma_hw_deinit(struct mtk_cqdma_device *cqdma)
725{
726 unsigned long flags;
727 u32 i;
728
729
730 for (i = 0; i < cqdma->dma_channels; ++i) {
731 spin_lock_irqsave(&cqdma->pc[i]->lock, flags);
732 if (mtk_cqdma_hard_reset(cqdma->pc[i]) < 0)
733 dev_err(cqdma2dev(cqdma), "cqdma hard reset timeout\n");
734 spin_unlock_irqrestore(&cqdma->pc[i]->lock, flags);
735 }
736
737 clk_disable_unprepare(cqdma->clk);
738
739 pm_runtime_put_sync(cqdma2dev(cqdma));
740 pm_runtime_disable(cqdma2dev(cqdma));
741}
742
743static const struct of_device_id mtk_cqdma_match[] = {
744 { .compatible = "mediatek,mt6765-cqdma" },
745 { }
746};
747MODULE_DEVICE_TABLE(of, mtk_cqdma_match);
748
749static int mtk_cqdma_probe(struct platform_device *pdev)
750{
751 struct mtk_cqdma_device *cqdma;
752 struct mtk_cqdma_vchan *vc;
753 struct dma_device *dd;
754 struct resource *res;
755 int err;
756 u32 i;
757
758 cqdma = devm_kzalloc(&pdev->dev, sizeof(*cqdma), GFP_KERNEL);
759 if (!cqdma)
760 return -ENOMEM;
761
762 dd = &cqdma->ddev;
763
764 cqdma->clk = devm_clk_get(&pdev->dev, "cqdma");
765 if (IS_ERR(cqdma->clk)) {
766 dev_err(&pdev->dev, "No clock for %s\n",
767 dev_name(&pdev->dev));
768 return PTR_ERR(cqdma->clk);
769 }
770
771 dma_cap_set(DMA_MEMCPY, dd->cap_mask);
772
773 dd->copy_align = MTK_CQDMA_ALIGN_SIZE;
774 dd->device_alloc_chan_resources = mtk_cqdma_alloc_chan_resources;
775 dd->device_free_chan_resources = mtk_cqdma_free_chan_resources;
776 dd->device_tx_status = mtk_cqdma_tx_status;
777 dd->device_issue_pending = mtk_cqdma_issue_pending;
778 dd->device_prep_dma_memcpy = mtk_cqdma_prep_dma_memcpy;
779 dd->device_terminate_all = mtk_cqdma_terminate_all;
780 dd->src_addr_widths = MTK_CQDMA_DMA_BUSWIDTHS;
781 dd->dst_addr_widths = MTK_CQDMA_DMA_BUSWIDTHS;
782 dd->directions = BIT(DMA_MEM_TO_MEM);
783 dd->residue_granularity = DMA_RESIDUE_GRANULARITY_SEGMENT;
784 dd->dev = &pdev->dev;
785 INIT_LIST_HEAD(&dd->channels);
786
787 if (pdev->dev.of_node && of_property_read_u32(pdev->dev.of_node,
788 "dma-requests",
789 &cqdma->dma_requests)) {
790 dev_info(&pdev->dev,
791 "Using %u as missing dma-requests property\n",
792 MTK_CQDMA_NR_VCHANS);
793
794 cqdma->dma_requests = MTK_CQDMA_NR_VCHANS;
795 }
796
797 if (pdev->dev.of_node && of_property_read_u32(pdev->dev.of_node,
798 "dma-channels",
799 &cqdma->dma_channels)) {
800 dev_info(&pdev->dev,
801 "Using %u as missing dma-channels property\n",
802 MTK_CQDMA_NR_PCHANS);
803
804 cqdma->dma_channels = MTK_CQDMA_NR_PCHANS;
805 }
806
807 cqdma->pc = devm_kcalloc(&pdev->dev, cqdma->dma_channels,
808 sizeof(*cqdma->pc), GFP_KERNEL);
809 if (!cqdma->pc)
810 return -ENOMEM;
811
812
813 for (i = 0; i < cqdma->dma_channels; ++i) {
814 cqdma->pc[i] = devm_kcalloc(&pdev->dev, 1,
815 sizeof(**cqdma->pc), GFP_KERNEL);
816 if (!cqdma->pc[i])
817 return -ENOMEM;
818
819 INIT_LIST_HEAD(&cqdma->pc[i]->queue);
820 spin_lock_init(&cqdma->pc[i]->lock);
821 refcount_set(&cqdma->pc[i]->refcnt, 0);
822
823 res = platform_get_resource(pdev, IORESOURCE_MEM, i);
824 if (!res) {
825 dev_err(&pdev->dev, "No mem resource for %s\n",
826 dev_name(&pdev->dev));
827 return -EINVAL;
828 }
829
830 cqdma->pc[i]->base = devm_ioremap_resource(&pdev->dev, res);
831 if (IS_ERR(cqdma->pc[i]->base))
832 return PTR_ERR(cqdma->pc[i]->base);
833
834
835 res = platform_get_resource(pdev, IORESOURCE_IRQ, i);
836 if (!res) {
837 dev_err(&pdev->dev, "No irq resource for %s\n",
838 dev_name(&pdev->dev));
839 return -EINVAL;
840 }
841 cqdma->pc[i]->irq = res->start;
842
843 err = devm_request_irq(&pdev->dev, cqdma->pc[i]->irq,
844 mtk_cqdma_irq, 0, dev_name(&pdev->dev),
845 cqdma);
846 if (err) {
847 dev_err(&pdev->dev,
848 "request_irq failed with err %d\n", err);
849 return -EINVAL;
850 }
851 }
852
853
854 cqdma->vc = devm_kcalloc(&pdev->dev, cqdma->dma_requests,
855 sizeof(*cqdma->vc), GFP_KERNEL);
856 if (!cqdma->vc)
857 return -ENOMEM;
858
859 for (i = 0; i < cqdma->dma_requests; i++) {
860 vc = &cqdma->vc[i];
861 vc->vc.desc_free = mtk_cqdma_vdesc_free;
862 vchan_init(&vc->vc, dd);
863 init_completion(&vc->issue_completion);
864 }
865
866 err = dma_async_device_register(dd);
867 if (err)
868 return err;
869
870 err = of_dma_controller_register(pdev->dev.of_node,
871 of_dma_xlate_by_chan_id, cqdma);
872 if (err) {
873 dev_err(&pdev->dev,
874 "MediaTek CQDMA OF registration failed %d\n", err);
875 goto err_unregister;
876 }
877
878 err = mtk_cqdma_hw_init(cqdma);
879 if (err) {
880 dev_err(&pdev->dev,
881 "MediaTek CQDMA HW initialization failed %d\n", err);
882 goto err_unregister;
883 }
884
885 platform_set_drvdata(pdev, cqdma);
886
887
888 for (i = 0; i < cqdma->dma_channels; ++i)
889 tasklet_init(&cqdma->pc[i]->tasklet, mtk_cqdma_tasklet_cb,
890 (unsigned long)cqdma->pc[i]);
891
892 dev_info(&pdev->dev, "MediaTek CQDMA driver registered\n");
893
894 return 0;
895
896err_unregister:
897 dma_async_device_unregister(dd);
898
899 return err;
900}
901
902static int mtk_cqdma_remove(struct platform_device *pdev)
903{
904 struct mtk_cqdma_device *cqdma = platform_get_drvdata(pdev);
905 struct mtk_cqdma_vchan *vc;
906 unsigned long flags;
907 int i;
908
909
910 for (i = 0; i < cqdma->dma_requests; i++) {
911 vc = &cqdma->vc[i];
912
913 list_del(&vc->vc.chan.device_node);
914 tasklet_kill(&vc->vc.task);
915 }
916
917
918 for (i = 0; i < cqdma->dma_channels; i++) {
919 spin_lock_irqsave(&cqdma->pc[i]->lock, flags);
920 mtk_dma_clr(cqdma->pc[i], MTK_CQDMA_INT_EN,
921 MTK_CQDMA_INT_EN_BIT);
922 spin_unlock_irqrestore(&cqdma->pc[i]->lock, flags);
923
924
925 synchronize_irq(cqdma->pc[i]->irq);
926
927 tasklet_kill(&cqdma->pc[i]->tasklet);
928 }
929
930
931 mtk_cqdma_hw_deinit(cqdma);
932
933 dma_async_device_unregister(&cqdma->ddev);
934 of_dma_controller_free(pdev->dev.of_node);
935
936 return 0;
937}
938
939static struct platform_driver mtk_cqdma_driver = {
940 .probe = mtk_cqdma_probe,
941 .remove = mtk_cqdma_remove,
942 .driver = {
943 .name = KBUILD_MODNAME,
944 .of_match_table = mtk_cqdma_match,
945 },
946};
947module_platform_driver(mtk_cqdma_driver);
948
949MODULE_DESCRIPTION("MediaTek CQDMA Controller Driver");
950MODULE_AUTHOR("Shun-Chih Yu <shun-chih.yu@mediatek.com>");
951MODULE_LICENSE("GPL v2");
952