1
2
3
4
5
6
7#ifndef __SG_SW_QM_H
8#define __SG_SW_QM_H
9
10#include <soc/fsl/qman.h>
11#include "regs.h"
12
13static inline void __dma_to_qm_sg(struct qm_sg_entry *qm_sg_ptr, dma_addr_t dma,
14 u16 offset)
15{
16 qm_sg_entry_set64(qm_sg_ptr, dma);
17 qm_sg_ptr->__reserved2 = 0;
18 qm_sg_ptr->bpid = 0;
19 qm_sg_ptr->offset = cpu_to_be16(offset & QM_SG_OFF_MASK);
20}
21
22static inline void dma_to_qm_sg_one(struct qm_sg_entry *qm_sg_ptr,
23 dma_addr_t dma, u32 len, u16 offset)
24{
25 __dma_to_qm_sg(qm_sg_ptr, dma, offset);
26 qm_sg_entry_set_len(qm_sg_ptr, len);
27}
28
29static inline void dma_to_qm_sg_one_last(struct qm_sg_entry *qm_sg_ptr,
30 dma_addr_t dma, u32 len, u16 offset)
31{
32 __dma_to_qm_sg(qm_sg_ptr, dma, offset);
33 qm_sg_entry_set_f(qm_sg_ptr, len);
34}
35
36static inline void dma_to_qm_sg_one_ext(struct qm_sg_entry *qm_sg_ptr,
37 dma_addr_t dma, u32 len, u16 offset)
38{
39 __dma_to_qm_sg(qm_sg_ptr, dma, offset);
40 qm_sg_ptr->cfg = cpu_to_be32(QM_SG_EXT | (len & QM_SG_LEN_MASK));
41}
42
43static inline void dma_to_qm_sg_one_last_ext(struct qm_sg_entry *qm_sg_ptr,
44 dma_addr_t dma, u32 len,
45 u16 offset)
46{
47 __dma_to_qm_sg(qm_sg_ptr, dma, offset);
48 qm_sg_ptr->cfg = cpu_to_be32(QM_SG_EXT | QM_SG_FIN |
49 (len & QM_SG_LEN_MASK));
50}
51
52
53
54
55
56static inline struct qm_sg_entry *
57sg_to_qm_sg(struct scatterlist *sg, int sg_count,
58 struct qm_sg_entry *qm_sg_ptr, u16 offset)
59{
60 while (sg_count && sg) {
61 dma_to_qm_sg_one(qm_sg_ptr, sg_dma_address(sg),
62 sg_dma_len(sg), offset);
63 qm_sg_ptr++;
64 sg = sg_next(sg);
65 sg_count--;
66 }
67 return qm_sg_ptr - 1;
68}
69
70
71
72
73
74static inline void sg_to_qm_sg_last(struct scatterlist *sg, int sg_count,
75 struct qm_sg_entry *qm_sg_ptr, u16 offset)
76{
77 qm_sg_ptr = sg_to_qm_sg(sg, sg_count, qm_sg_ptr, offset);
78 qm_sg_entry_set_f(qm_sg_ptr, qm_sg_entry_get_len(qm_sg_ptr));
79}
80
81#endif
82