1
2
3
4
5
6
7
8
9#include <linux/kernel.h>
10#include <linux/module.h>
11#include <linux/pci.h>
12#include <linux/device.h>
13#include <linux/dma/edma.h>
14#include <linux/pci-epf.h>
15#include <linux/msi.h>
16#include <linux/bitfield.h>
17
18#include "dw-edma-core.h"
19
20#define DW_PCIE_VSEC_DMA_ID 0x6
21#define DW_PCIE_VSEC_DMA_BAR GENMASK(10, 8)
22#define DW_PCIE_VSEC_DMA_MAP GENMASK(2, 0)
23#define DW_PCIE_VSEC_DMA_WR_CH GENMASK(9, 0)
24#define DW_PCIE_VSEC_DMA_RD_CH GENMASK(25, 16)
25
26#define DW_BLOCK(a, b, c) \
27 { \
28 .bar = a, \
29 .off = b, \
30 .sz = c, \
31 },
32
33struct dw_edma_block {
34 enum pci_barno bar;
35 off_t off;
36 size_t sz;
37};
38
39struct dw_edma_pcie_data {
40
41 struct dw_edma_block rg;
42
43 struct dw_edma_block ll_wr[EDMA_MAX_WR_CH];
44 struct dw_edma_block ll_rd[EDMA_MAX_RD_CH];
45
46 struct dw_edma_block dt_wr[EDMA_MAX_WR_CH];
47 struct dw_edma_block dt_rd[EDMA_MAX_RD_CH];
48
49 enum dw_edma_map_format mf;
50 u8 irqs;
51 u16 wr_ch_cnt;
52 u16 rd_ch_cnt;
53};
54
55static const struct dw_edma_pcie_data snps_edda_data = {
56
57 .rg.bar = BAR_0,
58 .rg.off = 0x00001000,
59 .rg.sz = 0x00002000,
60
61 .ll_wr = {
62
63 DW_BLOCK(BAR_2, 0x00000000, 0x00000800)
64
65 DW_BLOCK(BAR_2, 0x00200000, 0x00000800)
66 },
67 .ll_rd = {
68
69 DW_BLOCK(BAR_2, 0x00400000, 0x00000800)
70
71 DW_BLOCK(BAR_2, 0x00600000, 0x00000800)
72 },
73
74 .dt_wr = {
75
76 DW_BLOCK(BAR_2, 0x00800000, 0x00000800)
77
78 DW_BLOCK(BAR_2, 0x00900000, 0x00000800)
79 },
80 .dt_rd = {
81
82 DW_BLOCK(BAR_2, 0x00a00000, 0x00000800)
83
84 DW_BLOCK(BAR_2, 0x00b00000, 0x00000800)
85 },
86
87 .mf = EDMA_MF_EDMA_UNROLL,
88 .irqs = 1,
89 .wr_ch_cnt = 2,
90 .rd_ch_cnt = 2,
91};
92
93static int dw_edma_pcie_irq_vector(struct device *dev, unsigned int nr)
94{
95 return pci_irq_vector(to_pci_dev(dev), nr);
96}
97
98static const struct dw_edma_core_ops dw_edma_pcie_core_ops = {
99 .irq_vector = dw_edma_pcie_irq_vector,
100};
101
102static void dw_edma_pcie_get_vsec_dma_data(struct pci_dev *pdev,
103 struct dw_edma_pcie_data *pdata)
104{
105 u32 val, map;
106 u16 vsec;
107 u64 off;
108
109 vsec = pci_find_vsec_capability(pdev, PCI_VENDOR_ID_SYNOPSYS,
110 DW_PCIE_VSEC_DMA_ID);
111 if (!vsec)
112 return;
113
114 pci_read_config_dword(pdev, vsec + PCI_VNDR_HEADER, &val);
115 if (PCI_VNDR_HEADER_REV(val) != 0x00 ||
116 PCI_VNDR_HEADER_LEN(val) != 0x18)
117 return;
118
119 pci_dbg(pdev, "Detected PCIe Vendor-Specific Extended Capability DMA\n");
120 pci_read_config_dword(pdev, vsec + 0x8, &val);
121 map = FIELD_GET(DW_PCIE_VSEC_DMA_MAP, val);
122 if (map != EDMA_MF_EDMA_LEGACY &&
123 map != EDMA_MF_EDMA_UNROLL &&
124 map != EDMA_MF_HDMA_COMPAT)
125 return;
126
127 pdata->mf = map;
128 pdata->rg.bar = FIELD_GET(DW_PCIE_VSEC_DMA_BAR, val);
129
130 pci_read_config_dword(pdev, vsec + 0xc, &val);
131 pdata->wr_ch_cnt = min_t(u16, pdata->wr_ch_cnt,
132 FIELD_GET(DW_PCIE_VSEC_DMA_WR_CH, val));
133 pdata->rd_ch_cnt = min_t(u16, pdata->rd_ch_cnt,
134 FIELD_GET(DW_PCIE_VSEC_DMA_RD_CH, val));
135
136 pci_read_config_dword(pdev, vsec + 0x14, &val);
137 off = val;
138 pci_read_config_dword(pdev, vsec + 0x10, &val);
139 off <<= 32;
140 off |= val;
141 pdata->rg.off = off;
142}
143
144static int dw_edma_pcie_probe(struct pci_dev *pdev,
145 const struct pci_device_id *pid)
146{
147 struct dw_edma_pcie_data *pdata = (void *)pid->driver_data;
148 struct dw_edma_pcie_data vsec_data;
149 struct device *dev = &pdev->dev;
150 struct dw_edma_chip *chip;
151 struct dw_edma *dw;
152 int err, nr_irqs;
153 int i, mask;
154
155
156 err = pcim_enable_device(pdev);
157 if (err) {
158 pci_err(pdev, "enabling device failed\n");
159 return err;
160 }
161
162 memcpy(&vsec_data, pdata, sizeof(struct dw_edma_pcie_data));
163
164
165
166
167
168 dw_edma_pcie_get_vsec_dma_data(pdev, &vsec_data);
169
170
171 mask = BIT(vsec_data.rg.bar);
172 for (i = 0; i < vsec_data.wr_ch_cnt; i++) {
173 mask |= BIT(vsec_data.ll_wr[i].bar);
174 mask |= BIT(vsec_data.dt_wr[i].bar);
175 }
176 for (i = 0; i < vsec_data.rd_ch_cnt; i++) {
177 mask |= BIT(vsec_data.ll_rd[i].bar);
178 mask |= BIT(vsec_data.dt_rd[i].bar);
179 }
180 err = pcim_iomap_regions(pdev, mask, pci_name(pdev));
181 if (err) {
182 pci_err(pdev, "eDMA BAR I/O remapping failed\n");
183 return err;
184 }
185
186 pci_set_master(pdev);
187
188
189 err = pci_set_dma_mask(pdev, DMA_BIT_MASK(64));
190 if (!err) {
191 err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64));
192 if (err) {
193 pci_err(pdev, "consistent DMA mask 64 set failed\n");
194 return err;
195 }
196 } else {
197 pci_err(pdev, "DMA mask 64 set failed\n");
198
199 err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
200 if (err) {
201 pci_err(pdev, "DMA mask 32 set failed\n");
202 return err;
203 }
204
205 err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
206 if (err) {
207 pci_err(pdev, "consistent DMA mask 32 set failed\n");
208 return err;
209 }
210 }
211
212
213 chip = devm_kzalloc(dev, sizeof(*chip), GFP_KERNEL);
214 if (!chip)
215 return -ENOMEM;
216
217 dw = devm_kzalloc(dev, sizeof(*dw), GFP_KERNEL);
218 if (!dw)
219 return -ENOMEM;
220
221
222 nr_irqs = pci_alloc_irq_vectors(pdev, 1, vsec_data.irqs,
223 PCI_IRQ_MSI | PCI_IRQ_MSIX);
224 if (nr_irqs < 1) {
225 pci_err(pdev, "fail to alloc IRQ vector (number of IRQs=%u)\n",
226 nr_irqs);
227 return -EPERM;
228 }
229
230
231 chip->dw = dw;
232 chip->dev = dev;
233 chip->id = pdev->devfn;
234 chip->irq = pdev->irq;
235
236 dw->mf = vsec_data.mf;
237 dw->nr_irqs = nr_irqs;
238 dw->ops = &dw_edma_pcie_core_ops;
239 dw->wr_ch_cnt = vsec_data.wr_ch_cnt;
240 dw->rd_ch_cnt = vsec_data.rd_ch_cnt;
241
242 dw->rg_region.vaddr = pcim_iomap_table(pdev)[vsec_data.rg.bar];
243 if (!dw->rg_region.vaddr)
244 return -ENOMEM;
245
246 dw->rg_region.vaddr += vsec_data.rg.off;
247 dw->rg_region.paddr = pdev->resource[vsec_data.rg.bar].start;
248 dw->rg_region.paddr += vsec_data.rg.off;
249 dw->rg_region.sz = vsec_data.rg.sz;
250
251 for (i = 0; i < dw->wr_ch_cnt; i++) {
252 struct dw_edma_region *ll_region = &dw->ll_region_wr[i];
253 struct dw_edma_region *dt_region = &dw->dt_region_wr[i];
254 struct dw_edma_block *ll_block = &vsec_data.ll_wr[i];
255 struct dw_edma_block *dt_block = &vsec_data.dt_wr[i];
256
257 ll_region->vaddr = pcim_iomap_table(pdev)[ll_block->bar];
258 if (!ll_region->vaddr)
259 return -ENOMEM;
260
261 ll_region->vaddr += ll_block->off;
262 ll_region->paddr = pdev->resource[ll_block->bar].start;
263 ll_region->paddr += ll_block->off;
264 ll_region->sz = ll_block->sz;
265
266 dt_region->vaddr = pcim_iomap_table(pdev)[dt_block->bar];
267 if (!dt_region->vaddr)
268 return -ENOMEM;
269
270 dt_region->vaddr += dt_block->off;
271 dt_region->paddr = pdev->resource[dt_block->bar].start;
272 dt_region->paddr += dt_block->off;
273 dt_region->sz = dt_block->sz;
274 }
275
276 for (i = 0; i < dw->rd_ch_cnt; i++) {
277 struct dw_edma_region *ll_region = &dw->ll_region_rd[i];
278 struct dw_edma_region *dt_region = &dw->dt_region_rd[i];
279 struct dw_edma_block *ll_block = &vsec_data.ll_rd[i];
280 struct dw_edma_block *dt_block = &vsec_data.dt_rd[i];
281
282 ll_region->vaddr = pcim_iomap_table(pdev)[ll_block->bar];
283 if (!ll_region->vaddr)
284 return -ENOMEM;
285
286 ll_region->vaddr += ll_block->off;
287 ll_region->paddr = pdev->resource[ll_block->bar].start;
288 ll_region->paddr += ll_block->off;
289 ll_region->sz = ll_block->sz;
290
291 dt_region->vaddr = pcim_iomap_table(pdev)[dt_block->bar];
292 if (!dt_region->vaddr)
293 return -ENOMEM;
294
295 dt_region->vaddr += dt_block->off;
296 dt_region->paddr = pdev->resource[dt_block->bar].start;
297 dt_region->paddr += dt_block->off;
298 dt_region->sz = dt_block->sz;
299 }
300
301
302 if (dw->mf == EDMA_MF_EDMA_LEGACY)
303 pci_dbg(pdev, "Version:\teDMA Port Logic (0x%x)\n", dw->mf);
304 else if (dw->mf == EDMA_MF_EDMA_UNROLL)
305 pci_dbg(pdev, "Version:\teDMA Unroll (0x%x)\n", dw->mf);
306 else if (dw->mf == EDMA_MF_HDMA_COMPAT)
307 pci_dbg(pdev, "Version:\tHDMA Compatible (0x%x)\n", dw->mf);
308 else
309 pci_dbg(pdev, "Version:\tUnknown (0x%x)\n", dw->mf);
310
311 pci_dbg(pdev, "Registers:\tBAR=%u, off=0x%.8lx, sz=0x%zx bytes, addr(v=%p, p=%pa)\n",
312 vsec_data.rg.bar, vsec_data.rg.off, vsec_data.rg.sz,
313 dw->rg_region.vaddr, &dw->rg_region.paddr);
314
315
316 for (i = 0; i < dw->wr_ch_cnt; i++) {
317 pci_dbg(pdev, "L. List:\tWRITE CH%.2u, BAR=%u, off=0x%.8lx, sz=0x%zx bytes, addr(v=%p, p=%pa)\n",
318 i, vsec_data.ll_wr[i].bar,
319 vsec_data.ll_wr[i].off, dw->ll_region_wr[i].sz,
320 dw->ll_region_wr[i].vaddr, &dw->ll_region_wr[i].paddr);
321
322 pci_dbg(pdev, "Data:\tWRITE CH%.2u, BAR=%u, off=0x%.8lx, sz=0x%zx bytes, addr(v=%p, p=%pa)\n",
323 i, vsec_data.dt_wr[i].bar,
324 vsec_data.dt_wr[i].off, dw->dt_region_wr[i].sz,
325 dw->dt_region_wr[i].vaddr, &dw->dt_region_wr[i].paddr);
326 }
327
328 for (i = 0; i < dw->rd_ch_cnt; i++) {
329 pci_dbg(pdev, "L. List:\tREAD CH%.2u, BAR=%u, off=0x%.8lx, sz=0x%zx bytes, addr(v=%p, p=%pa)\n",
330 i, vsec_data.ll_rd[i].bar,
331 vsec_data.ll_rd[i].off, dw->ll_region_rd[i].sz,
332 dw->ll_region_rd[i].vaddr, &dw->ll_region_rd[i].paddr);
333
334 pci_dbg(pdev, "Data:\tREAD CH%.2u, BAR=%u, off=0x%.8lx, sz=0x%zx bytes, addr(v=%p, p=%pa)\n",
335 i, vsec_data.dt_rd[i].bar,
336 vsec_data.dt_rd[i].off, dw->dt_region_rd[i].sz,
337 dw->dt_region_rd[i].vaddr, &dw->dt_region_rd[i].paddr);
338 }
339
340 pci_dbg(pdev, "Nr. IRQs:\t%u\n", dw->nr_irqs);
341
342
343 if (!pci_dev_msi_enabled(pdev)) {
344 pci_err(pdev, "enable interrupt failed\n");
345 return -EPERM;
346 }
347
348 dw->irq = devm_kcalloc(dev, nr_irqs, sizeof(*dw->irq), GFP_KERNEL);
349 if (!dw->irq)
350 return -ENOMEM;
351
352
353 err = dw_edma_probe(chip);
354 if (err) {
355 pci_err(pdev, "eDMA probe failed\n");
356 return err;
357 }
358
359
360 pci_set_drvdata(pdev, chip);
361
362 return 0;
363}
364
365static void dw_edma_pcie_remove(struct pci_dev *pdev)
366{
367 struct dw_edma_chip *chip = pci_get_drvdata(pdev);
368 int err;
369
370
371 err = dw_edma_remove(chip);
372 if (err)
373 pci_warn(pdev, "can't remove device properly: %d\n", err);
374
375
376 pci_free_irq_vectors(pdev);
377}
378
379static const struct pci_device_id dw_edma_pcie_id_table[] = {
380 { PCI_DEVICE_DATA(SYNOPSYS, EDDA, &snps_edda_data) },
381 { }
382};
383MODULE_DEVICE_TABLE(pci, dw_edma_pcie_id_table);
384
385static struct pci_driver dw_edma_pcie_driver = {
386 .name = "dw-edma-pcie",
387 .id_table = dw_edma_pcie_id_table,
388 .probe = dw_edma_pcie_probe,
389 .remove = dw_edma_pcie_remove,
390};
391
392module_pci_driver(dw_edma_pcie_driver);
393
394MODULE_LICENSE("GPL v2");
395MODULE_DESCRIPTION("Synopsys DesignWare eDMA PCIe driver");
396MODULE_AUTHOR("Gustavo Pimentel <gustavo.pimentel@synopsys.com>");
397