1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21#include <linux/kernel.h>
22#include <linux/module.h>
23#include <linux/slab.h>
24
25#include "solo6x10.h"
26
27static int multi_p2m;
28module_param(multi_p2m, uint, 0644);
29MODULE_PARM_DESC(multi_p2m,
30 "Use multiple P2M DMA channels (default: no, 6010-only)");
31
32static int desc_mode;
33module_param(desc_mode, uint, 0644);
34MODULE_PARM_DESC(desc_mode,
35 "Allow use of descriptor mode DMA (default: no, 6010-only)");
36
37int solo_p2m_dma(struct solo_dev *solo_dev, int wr,
38 void *sys_addr, u32 ext_addr, u32 size,
39 int repeat, u32 ext_size)
40{
41 dma_addr_t dma_addr;
42 int ret;
43
44 if (WARN_ON_ONCE((unsigned long)sys_addr & 0x03))
45 return -EINVAL;
46 if (WARN_ON_ONCE(!size))
47 return -EINVAL;
48
49 dma_addr = pci_map_single(solo_dev->pdev, sys_addr, size,
50 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE);
51 if (pci_dma_mapping_error(solo_dev->pdev, dma_addr))
52 return -ENOMEM;
53
54 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size,
55 repeat, ext_size);
56
57 pci_unmap_single(solo_dev->pdev, dma_addr, size,
58 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE);
59
60 return ret;
61}
62
63
64int solo_p2m_dma_desc(struct solo_dev *solo_dev,
65 struct solo_p2m_desc *desc, dma_addr_t desc_dma,
66 int desc_cnt)
67{
68 struct solo_p2m_dev *p2m_dev;
69 unsigned int timeout;
70 unsigned int config = 0;
71 int ret = 0;
72 unsigned int p2m_id = 0;
73
74
75 if (solo_dev->type != SOLO_DEV_6110 && multi_p2m)
76 p2m_id = atomic_inc_return(&solo_dev->p2m_count) % SOLO_NR_P2M;
77
78 p2m_dev = &solo_dev->p2m_dev[p2m_id];
79
80 if (mutex_lock_interruptible(&p2m_dev->mutex))
81 return -EINTR;
82
83 reinit_completion(&p2m_dev->completion);
84 p2m_dev->error = 0;
85
86 if (desc_cnt > 1 && solo_dev->type != SOLO_DEV_6110 && desc_mode) {
87
88 p2m_dev->desc_count = p2m_dev->desc_idx = 0;
89 config = solo_reg_read(solo_dev, SOLO_P2M_CONFIG(p2m_id));
90
91 solo_reg_write(solo_dev, SOLO_P2M_DES_ADR(p2m_id), desc_dma);
92 solo_reg_write(solo_dev, SOLO_P2M_DESC_ID(p2m_id), desc_cnt);
93 solo_reg_write(solo_dev, SOLO_P2M_CONFIG(p2m_id), config |
94 SOLO_P2M_DESC_MODE);
95 } else {
96
97 p2m_dev->desc_count = desc_cnt;
98 p2m_dev->desc_idx = 1;
99 p2m_dev->descs = desc;
100
101 solo_reg_write(solo_dev, SOLO_P2M_TAR_ADR(p2m_id),
102 desc[1].dma_addr);
103 solo_reg_write(solo_dev, SOLO_P2M_EXT_ADR(p2m_id),
104 desc[1].ext_addr);
105 solo_reg_write(solo_dev, SOLO_P2M_EXT_CFG(p2m_id),
106 desc[1].cfg);
107 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(p2m_id),
108 desc[1].ctrl);
109 }
110
111 timeout = wait_for_completion_timeout(&p2m_dev->completion,
112 solo_dev->p2m_jiffies);
113
114 if (WARN_ON_ONCE(p2m_dev->error))
115 ret = -EIO;
116 else if (timeout == 0) {
117 solo_dev->p2m_timeouts++;
118 ret = -EAGAIN;
119 }
120
121 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(p2m_id), 0);
122
123
124
125 if (desc_cnt > 1 && solo_dev->type != SOLO_DEV_6110 && config)
126 solo_reg_write(solo_dev, SOLO_P2M_CONFIG(p2m_id), config);
127
128 mutex_unlock(&p2m_dev->mutex);
129
130 return ret;
131}
132
133void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr,
134 dma_addr_t dma_addr, u32 ext_addr, u32 size,
135 int repeat, u32 ext_size)
136{
137 WARN_ON_ONCE(dma_addr & 0x03);
138 WARN_ON_ONCE(!size);
139
140 desc->cfg = SOLO_P2M_COPY_SIZE(size >> 2);
141 desc->ctrl = SOLO_P2M_BURST_SIZE(SOLO_P2M_BURST_256) |
142 (wr ? SOLO_P2M_WRITE : 0) | SOLO_P2M_TRANS_ON;
143
144 if (repeat) {
145 desc->cfg |= SOLO_P2M_EXT_INC(ext_size >> 2);
146 desc->ctrl |= SOLO_P2M_PCI_INC(size >> 2) |
147 SOLO_P2M_REPEAT(repeat);
148 }
149
150 desc->dma_addr = dma_addr;
151 desc->ext_addr = ext_addr;
152}
153
154int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr,
155 dma_addr_t dma_addr, u32 ext_addr, u32 size,
156 int repeat, u32 ext_size)
157{
158 struct solo_p2m_desc desc[2];
159
160 solo_p2m_fill_desc(&desc[1], wr, dma_addr, ext_addr, size, repeat,
161 ext_size);
162
163
164 return solo_p2m_dma_desc(solo_dev, desc, 0, 1);
165}
166
167void solo_p2m_isr(struct solo_dev *solo_dev, int id)
168{
169 struct solo_p2m_dev *p2m_dev = &solo_dev->p2m_dev[id];
170 struct solo_p2m_desc *desc;
171
172 if (p2m_dev->desc_count <= p2m_dev->desc_idx) {
173 complete(&p2m_dev->completion);
174 return;
175 }
176
177
178 p2m_dev->desc_idx++;
179 desc = &p2m_dev->descs[p2m_dev->desc_idx];
180
181 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(id), 0);
182 solo_reg_write(solo_dev, SOLO_P2M_TAR_ADR(id), desc->dma_addr);
183 solo_reg_write(solo_dev, SOLO_P2M_EXT_ADR(id), desc->ext_addr);
184 solo_reg_write(solo_dev, SOLO_P2M_EXT_CFG(id), desc->cfg);
185 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(id), desc->ctrl);
186}
187
188void solo_p2m_error_isr(struct solo_dev *solo_dev)
189{
190 unsigned int err = solo_reg_read(solo_dev, SOLO_PCI_ERR);
191 struct solo_p2m_dev *p2m_dev;
192 int i;
193
194 if (!(err & (SOLO_PCI_ERR_P2M | SOLO_PCI_ERR_P2M_DESC)))
195 return;
196
197 for (i = 0; i < SOLO_NR_P2M; i++) {
198 p2m_dev = &solo_dev->p2m_dev[i];
199 p2m_dev->error = 1;
200 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(i), 0);
201 complete(&p2m_dev->completion);
202 }
203}
204
205void solo_p2m_exit(struct solo_dev *solo_dev)
206{
207 int i;
208
209 for (i = 0; i < SOLO_NR_P2M; i++)
210 solo_irq_off(solo_dev, SOLO_IRQ_P2M(i));
211}
212
213static int solo_p2m_test(struct solo_dev *solo_dev, int base, int size)
214{
215 u32 *wr_buf;
216 u32 *rd_buf;
217 int i;
218 int ret = -EIO;
219 int order = get_order(size);
220
221 wr_buf = (u32 *)__get_free_pages(GFP_KERNEL, order);
222 if (wr_buf == NULL)
223 return -1;
224
225 rd_buf = (u32 *)__get_free_pages(GFP_KERNEL, order);
226 if (rd_buf == NULL) {
227 free_pages((unsigned long)wr_buf, order);
228 return -1;
229 }
230
231 for (i = 0; i < (size >> 3); i++)
232 *(wr_buf + i) = (i << 16) | (i + 1);
233
234 for (i = (size >> 3); i < (size >> 2); i++)
235 *(wr_buf + i) = ~((i << 16) | (i + 1));
236
237 memset(rd_buf, 0x55, size);
238
239 if (solo_p2m_dma(solo_dev, 1, wr_buf, base, size, 0, 0))
240 goto test_fail;
241
242 if (solo_p2m_dma(solo_dev, 0, rd_buf, base, size, 0, 0))
243 goto test_fail;
244
245 for (i = 0; i < (size >> 2); i++) {
246 if (*(wr_buf + i) != *(rd_buf + i))
247 goto test_fail;
248 }
249
250 ret = 0;
251
252test_fail:
253 free_pages((unsigned long)wr_buf, order);
254 free_pages((unsigned long)rd_buf, order);
255
256 return ret;
257}
258
259int solo_p2m_init(struct solo_dev *solo_dev)
260{
261 struct solo_p2m_dev *p2m_dev;
262 int i;
263
264 for (i = 0; i < SOLO_NR_P2M; i++) {
265 p2m_dev = &solo_dev->p2m_dev[i];
266
267 mutex_init(&p2m_dev->mutex);
268 init_completion(&p2m_dev->completion);
269
270 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(i), 0);
271 solo_reg_write(solo_dev, SOLO_P2M_CONFIG(i),
272 SOLO_P2M_CSC_16BIT_565 |
273 SOLO_P2M_DESC_INTR_OPT |
274 SOLO_P2M_DMA_INTERVAL(0) |
275 SOLO_P2M_PCI_MASTER_MODE);
276 solo_irq_on(solo_dev, SOLO_IRQ_P2M(i));
277 }
278
279
280 for (solo_dev->sdram_size = 0, i = 2; i >= 0; i--) {
281 solo_reg_write(solo_dev, SOLO_DMA_CTRL,
282 SOLO_DMA_CTRL_REFRESH_CYCLE(1) |
283 SOLO_DMA_CTRL_SDRAM_SIZE(i) |
284 SOLO_DMA_CTRL_SDRAM_CLK_INVERT |
285 SOLO_DMA_CTRL_READ_CLK_SELECT |
286 SOLO_DMA_CTRL_LATENCY(1));
287
288 solo_reg_write(solo_dev, SOLO_SYS_CFG, solo_dev->sys_config |
289 SOLO_SYS_CFG_RESET);
290 solo_reg_write(solo_dev, SOLO_SYS_CFG, solo_dev->sys_config);
291
292 switch (i) {
293 case 2:
294 if (solo_p2m_test(solo_dev, 0x07ff0000, 0x00010000) ||
295 solo_p2m_test(solo_dev, 0x05ff0000, 0x00010000))
296 continue;
297 break;
298
299 case 1:
300 if (solo_p2m_test(solo_dev, 0x03ff0000, 0x00010000))
301 continue;
302 break;
303
304 default:
305 if (solo_p2m_test(solo_dev, 0x01ff0000, 0x00010000))
306 continue;
307 }
308
309 solo_dev->sdram_size = (32 << 20) << i;
310 break;
311 }
312
313 if (!solo_dev->sdram_size) {
314 dev_err(&solo_dev->pdev->dev, "Error detecting SDRAM size\n");
315 return -EIO;
316 }
317
318 if (SOLO_SDRAM_END(solo_dev) > solo_dev->sdram_size) {
319 dev_err(&solo_dev->pdev->dev,
320 "SDRAM is not large enough (%u < %u)\n",
321 solo_dev->sdram_size, SOLO_SDRAM_END(solo_dev));
322 return -EIO;
323 }
324
325 return 0;
326}
327