1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35#include <linux/err.h>
36#include <linux/kernel.h>
37#include <linux/module.h>
38#include <linux/blkdev.h>
39#include <scsi/scsi_host.h>
40#include <linux/ata.h>
41#include <linux/libata.h>
42#include <linux/platform_device.h>
43#include <linux/delay.h>
44#include <linux/dmaengine.h>
45#include <linux/ktime.h>
46
47#include <linux/platform_data/dma-ep93xx.h>
48#include <linux/soc/cirrus/ep93xx.h>
49
50#define DRV_NAME "ep93xx-ide"
51#define DRV_VERSION "1.0"
52
53enum {
54
55 IDECTRL = 0x00,
56 IDECTRL_CS0N = (1 << 0),
57 IDECTRL_CS1N = (1 << 1),
58 IDECTRL_DIORN = (1 << 5),
59 IDECTRL_DIOWN = (1 << 6),
60 IDECTRL_INTRQ = (1 << 9),
61 IDECTRL_IORDY = (1 << 10),
62
63
64
65
66
67
68
69
70
71
72
73 IDECTRL_ADDR_CMD = 0 + 2,
74 IDECTRL_ADDR_DATA = (ATA_REG_DATA << 2) + 2,
75 IDECTRL_ADDR_ERROR = (ATA_REG_ERR << 2) + 2,
76 IDECTRL_ADDR_FEATURE = (ATA_REG_FEATURE << 2) + 2,
77 IDECTRL_ADDR_NSECT = (ATA_REG_NSECT << 2) + 2,
78 IDECTRL_ADDR_LBAL = (ATA_REG_LBAL << 2) + 2,
79 IDECTRL_ADDR_LBAM = (ATA_REG_LBAM << 2) + 2,
80 IDECTRL_ADDR_LBAH = (ATA_REG_LBAH << 2) + 2,
81 IDECTRL_ADDR_DEVICE = (ATA_REG_DEVICE << 2) + 2,
82 IDECTRL_ADDR_STATUS = (ATA_REG_STATUS << 2) + 2,
83 IDECTRL_ADDR_COMMAND = (ATA_REG_CMD << 2) + 2,
84 IDECTRL_ADDR_ALTSTATUS = (0x06 << 2) + 1,
85 IDECTRL_ADDR_CTL = (0x06 << 2) + 1,
86
87
88 IDECFG = 0x04,
89 IDECFG_IDEEN = (1 << 0),
90 IDECFG_PIO = (1 << 1),
91 IDECFG_MDMA = (1 << 2),
92 IDECFG_UDMA = (1 << 3),
93 IDECFG_MODE_SHIFT = 4,
94 IDECFG_MODE_MASK = (0xf << 4),
95 IDECFG_WST_SHIFT = 8,
96 IDECFG_WST_MASK = (0x3 << 8),
97
98
99 IDEMDMAOP = 0x08,
100
101
102 IDEUDMAOP = 0x0c,
103 IDEUDMAOP_UEN = (1 << 0),
104 IDEUDMAOP_RWOP = (1 << 1),
105
106
107 IDEDATAOUT = 0x10,
108 IDEDATAIN = 0x14,
109 IDEMDMADATAOUT = 0x18,
110 IDEMDMADATAIN = 0x1c,
111 IDEUDMADATAOUT = 0x20,
112 IDEUDMADATAIN = 0x24,
113
114
115 IDEUDMASTS = 0x28,
116 IDEUDMASTS_DMAIDE = (1 << 16),
117 IDEUDMASTS_INTIDE = (1 << 17),
118 IDEUDMASTS_SBUSY = (1 << 18),
119 IDEUDMASTS_NDO = (1 << 24),
120 IDEUDMASTS_NDI = (1 << 25),
121 IDEUDMASTS_N4X = (1 << 26),
122
123
124 IDEUDMADEBUG = 0x2c,
125};
126
127struct ep93xx_pata_data {
128 const struct platform_device *pdev;
129 void __iomem *ide_base;
130 struct ata_timing t;
131 bool iordy;
132
133 unsigned long udma_in_phys;
134 unsigned long udma_out_phys;
135
136 struct dma_chan *dma_rx_channel;
137 struct ep93xx_dma_data dma_rx_data;
138 struct dma_chan *dma_tx_channel;
139 struct ep93xx_dma_data dma_tx_data;
140};
141
142static void ep93xx_pata_clear_regs(void __iomem *base)
143{
144 writel(IDECTRL_CS0N | IDECTRL_CS1N | IDECTRL_DIORN |
145 IDECTRL_DIOWN, base + IDECTRL);
146
147 writel(0, base + IDECFG);
148 writel(0, base + IDEMDMAOP);
149 writel(0, base + IDEUDMAOP);
150 writel(0, base + IDEDATAOUT);
151 writel(0, base + IDEDATAIN);
152 writel(0, base + IDEMDMADATAOUT);
153 writel(0, base + IDEMDMADATAIN);
154 writel(0, base + IDEUDMADATAOUT);
155 writel(0, base + IDEUDMADATAIN);
156 writel(0, base + IDEUDMADEBUG);
157}
158
159static bool ep93xx_pata_check_iordy(void __iomem *base)
160{
161 return !!(readl(base + IDECTRL) & IDECTRL_IORDY);
162}
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178static int ep93xx_pata_get_wst(int pio_mode)
179{
180 int val;
181
182 if (pio_mode == 0)
183 val = 3;
184 else if (pio_mode < 3)
185 val = 2;
186 else
187 val = 1;
188
189 return val << IDECFG_WST_SHIFT;
190}
191
192static void ep93xx_pata_enable_pio(void __iomem *base, int pio_mode)
193{
194 writel(IDECFG_IDEEN | IDECFG_PIO |
195 ep93xx_pata_get_wst(pio_mode) |
196 (pio_mode << IDECFG_MODE_SHIFT), base + IDECFG);
197}
198
199
200
201
202
203
204
205
206static void ep93xx_pata_delay(unsigned long count)
207{
208 __asm__ volatile (
209 "0:\n"
210 "mov r0, r0\n"
211 "subs %0, %1, #1\n"
212 "bge 0b\n"
213 : "=r" (count)
214 : "0" (count)
215 );
216}
217
218static unsigned long ep93xx_pata_wait_for_iordy(void __iomem *base,
219 unsigned long t2)
220{
221
222
223
224
225
226
227
228
229
230 unsigned long start = (1250 + 35) / 25 - t2;
231 unsigned long counter = start;
232
233 while (!ep93xx_pata_check_iordy(base) && counter--)
234 ep93xx_pata_delay(1);
235 return start - counter;
236}
237
238
239static void ep93xx_pata_rw_begin(void __iomem *base, unsigned long addr,
240 unsigned long t1)
241{
242 writel(IDECTRL_DIOWN | IDECTRL_DIORN | addr, base + IDECTRL);
243 ep93xx_pata_delay(t1);
244}
245
246
247static void ep93xx_pata_rw_end(void __iomem *base, unsigned long addr,
248 bool iordy, unsigned long t0, unsigned long t2,
249 unsigned long t2i)
250{
251 ep93xx_pata_delay(t2);
252
253 if (iordy)
254 t2 += ep93xx_pata_wait_for_iordy(base, t2);
255 writel(IDECTRL_DIOWN | IDECTRL_DIORN | addr, base + IDECTRL);
256 if (t0 > t2 && t0 - t2 > t2i)
257 ep93xx_pata_delay(t0 - t2);
258 else
259 ep93xx_pata_delay(t2i);
260}
261
262static u16 ep93xx_pata_read(struct ep93xx_pata_data *drv_data,
263 unsigned long addr,
264 bool reg)
265{
266 void __iomem *base = drv_data->ide_base;
267 const struct ata_timing *t = &drv_data->t;
268 unsigned long t0 = reg ? t->cyc8b : t->cycle;
269 unsigned long t2 = reg ? t->act8b : t->active;
270 unsigned long t2i = reg ? t->rec8b : t->recover;
271
272 ep93xx_pata_rw_begin(base, addr, t->setup);
273 writel(IDECTRL_DIOWN | addr, base + IDECTRL);
274
275
276
277
278 ep93xx_pata_rw_end(base, addr, drv_data->iordy, t0, t2, t2i);
279 return readl(base + IDEDATAIN);
280}
281
282
283static u16 ep93xx_pata_read_reg(struct ep93xx_pata_data *drv_data,
284 unsigned long addr)
285{
286 return ep93xx_pata_read(drv_data, addr, true);
287}
288
289
290static u16 ep93xx_pata_read_data(struct ep93xx_pata_data *drv_data,
291 unsigned long addr)
292{
293 return ep93xx_pata_read(drv_data, addr, false);
294}
295
296static void ep93xx_pata_write(struct ep93xx_pata_data *drv_data,
297 u16 value, unsigned long addr,
298 bool reg)
299{
300 void __iomem *base = drv_data->ide_base;
301 const struct ata_timing *t = &drv_data->t;
302 unsigned long t0 = reg ? t->cyc8b : t->cycle;
303 unsigned long t2 = reg ? t->act8b : t->active;
304 unsigned long t2i = reg ? t->rec8b : t->recover;
305
306 ep93xx_pata_rw_begin(base, addr, t->setup);
307
308
309
310
311 writel(value, base + IDEDATAOUT);
312 writel(IDECTRL_DIORN | addr, base + IDECTRL);
313 ep93xx_pata_rw_end(base, addr, drv_data->iordy, t0, t2, t2i);
314}
315
316
317static void ep93xx_pata_write_reg(struct ep93xx_pata_data *drv_data,
318 u16 value, unsigned long addr)
319{
320 ep93xx_pata_write(drv_data, value, addr, true);
321}
322
323
324static void ep93xx_pata_write_data(struct ep93xx_pata_data *drv_data,
325 u16 value, unsigned long addr)
326{
327 ep93xx_pata_write(drv_data, value, addr, false);
328}
329
330static void ep93xx_pata_set_piomode(struct ata_port *ap,
331 struct ata_device *adev)
332{
333 struct ep93xx_pata_data *drv_data = ap->host->private_data;
334 struct ata_device *pair = ata_dev_pair(adev);
335
336
337
338
339
340
341
342 unsigned long T = 1000000 / (200 / 5);
343
344 ata_timing_compute(adev, adev->pio_mode, &drv_data->t, T, 0);
345 if (pair && pair->pio_mode) {
346 struct ata_timing t;
347 ata_timing_compute(pair, pair->pio_mode, &t, T, 0);
348 ata_timing_merge(&t, &drv_data->t, &drv_data->t,
349 ATA_TIMING_SETUP | ATA_TIMING_8BIT);
350 }
351 drv_data->iordy = ata_pio_need_iordy(adev);
352
353 ep93xx_pata_enable_pio(drv_data->ide_base,
354 adev->pio_mode - XFER_PIO_0);
355}
356
357
358static u8 ep93xx_pata_check_status(struct ata_port *ap)
359{
360 struct ep93xx_pata_data *drv_data = ap->host->private_data;
361
362 return ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_STATUS);
363}
364
365static u8 ep93xx_pata_check_altstatus(struct ata_port *ap)
366{
367 struct ep93xx_pata_data *drv_data = ap->host->private_data;
368
369 return ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_ALTSTATUS);
370}
371
372
373static void ep93xx_pata_tf_load(struct ata_port *ap,
374 const struct ata_taskfile *tf)
375{
376 struct ep93xx_pata_data *drv_data = ap->host->private_data;
377 unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR;
378
379 if (tf->ctl != ap->last_ctl) {
380 ep93xx_pata_write_reg(drv_data, tf->ctl, IDECTRL_ADDR_CTL);
381 ap->last_ctl = tf->ctl;
382 ata_wait_idle(ap);
383 }
384
385 if (is_addr && (tf->flags & ATA_TFLAG_LBA48)) {
386 ep93xx_pata_write_reg(drv_data, tf->hob_feature,
387 IDECTRL_ADDR_FEATURE);
388 ep93xx_pata_write_reg(drv_data, tf->hob_nsect,
389 IDECTRL_ADDR_NSECT);
390 ep93xx_pata_write_reg(drv_data, tf->hob_lbal,
391 IDECTRL_ADDR_LBAL);
392 ep93xx_pata_write_reg(drv_data, tf->hob_lbam,
393 IDECTRL_ADDR_LBAM);
394 ep93xx_pata_write_reg(drv_data, tf->hob_lbah,
395 IDECTRL_ADDR_LBAH);
396 }
397
398 if (is_addr) {
399 ep93xx_pata_write_reg(drv_data, tf->feature,
400 IDECTRL_ADDR_FEATURE);
401 ep93xx_pata_write_reg(drv_data, tf->nsect, IDECTRL_ADDR_NSECT);
402 ep93xx_pata_write_reg(drv_data, tf->lbal, IDECTRL_ADDR_LBAL);
403 ep93xx_pata_write_reg(drv_data, tf->lbam, IDECTRL_ADDR_LBAM);
404 ep93xx_pata_write_reg(drv_data, tf->lbah, IDECTRL_ADDR_LBAH);
405 }
406
407 if (tf->flags & ATA_TFLAG_DEVICE)
408 ep93xx_pata_write_reg(drv_data, tf->device,
409 IDECTRL_ADDR_DEVICE);
410
411 ata_wait_idle(ap);
412}
413
414
415static void ep93xx_pata_tf_read(struct ata_port *ap, struct ata_taskfile *tf)
416{
417 struct ep93xx_pata_data *drv_data = ap->host->private_data;
418
419 tf->command = ep93xx_pata_check_status(ap);
420 tf->feature = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_FEATURE);
421 tf->nsect = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_NSECT);
422 tf->lbal = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAL);
423 tf->lbam = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAM);
424 tf->lbah = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAH);
425 tf->device = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_DEVICE);
426
427 if (tf->flags & ATA_TFLAG_LBA48) {
428 ep93xx_pata_write_reg(drv_data, tf->ctl | ATA_HOB,
429 IDECTRL_ADDR_CTL);
430 tf->hob_feature = ep93xx_pata_read_reg(drv_data,
431 IDECTRL_ADDR_FEATURE);
432 tf->hob_nsect = ep93xx_pata_read_reg(drv_data,
433 IDECTRL_ADDR_NSECT);
434 tf->hob_lbal = ep93xx_pata_read_reg(drv_data,
435 IDECTRL_ADDR_LBAL);
436 tf->hob_lbam = ep93xx_pata_read_reg(drv_data,
437 IDECTRL_ADDR_LBAM);
438 tf->hob_lbah = ep93xx_pata_read_reg(drv_data,
439 IDECTRL_ADDR_LBAH);
440 ep93xx_pata_write_reg(drv_data, tf->ctl, IDECTRL_ADDR_CTL);
441 ap->last_ctl = tf->ctl;
442 }
443}
444
445
446static void ep93xx_pata_exec_command(struct ata_port *ap,
447 const struct ata_taskfile *tf)
448{
449 struct ep93xx_pata_data *drv_data = ap->host->private_data;
450
451 ep93xx_pata_write_reg(drv_data, tf->command,
452 IDECTRL_ADDR_COMMAND);
453 ata_sff_pause(ap);
454}
455
456
457static void ep93xx_pata_dev_select(struct ata_port *ap, unsigned int device)
458{
459 struct ep93xx_pata_data *drv_data = ap->host->private_data;
460 u8 tmp = ATA_DEVICE_OBS;
461
462 if (device != 0)
463 tmp |= ATA_DEV1;
464
465 ep93xx_pata_write_reg(drv_data, tmp, IDECTRL_ADDR_DEVICE);
466 ata_sff_pause(ap);
467}
468
469
470static void ep93xx_pata_set_devctl(struct ata_port *ap, u8 ctl)
471{
472 struct ep93xx_pata_data *drv_data = ap->host->private_data;
473
474 ep93xx_pata_write_reg(drv_data, ctl, IDECTRL_ADDR_CTL);
475}
476
477
478static unsigned int ep93xx_pata_data_xfer(struct ata_queued_cmd *qc,
479 unsigned char *buf,
480 unsigned int buflen, int rw)
481{
482 struct ata_port *ap = qc->dev->link->ap;
483 struct ep93xx_pata_data *drv_data = ap->host->private_data;
484 u16 *data = (u16 *)buf;
485 unsigned int words = buflen >> 1;
486
487
488 while (words--)
489 if (rw == READ)
490 *data++ = cpu_to_le16(
491 ep93xx_pata_read_data(
492 drv_data, IDECTRL_ADDR_DATA));
493 else
494 ep93xx_pata_write_data(drv_data, le16_to_cpu(*data++),
495 IDECTRL_ADDR_DATA);
496
497
498 if (unlikely(buflen & 0x01)) {
499 unsigned char pad[2] = { };
500
501 buf += buflen - 1;
502
503 if (rw == READ) {
504 *pad = cpu_to_le16(
505 ep93xx_pata_read_data(
506 drv_data, IDECTRL_ADDR_DATA));
507 *buf = pad[0];
508 } else {
509 pad[0] = *buf;
510 ep93xx_pata_write_data(drv_data, le16_to_cpu(*pad),
511 IDECTRL_ADDR_DATA);
512 }
513 words++;
514 }
515
516 return words << 1;
517}
518
519
520static bool ep93xx_pata_device_is_present(struct ata_port *ap,
521 unsigned int device)
522{
523 struct ep93xx_pata_data *drv_data = ap->host->private_data;
524 u8 nsect, lbal;
525
526 ap->ops->sff_dev_select(ap, device);
527
528 ep93xx_pata_write_reg(drv_data, 0x55, IDECTRL_ADDR_NSECT);
529 ep93xx_pata_write_reg(drv_data, 0xaa, IDECTRL_ADDR_LBAL);
530
531 ep93xx_pata_write_reg(drv_data, 0xaa, IDECTRL_ADDR_NSECT);
532 ep93xx_pata_write_reg(drv_data, 0x55, IDECTRL_ADDR_LBAL);
533
534 ep93xx_pata_write_reg(drv_data, 0x55, IDECTRL_ADDR_NSECT);
535 ep93xx_pata_write_reg(drv_data, 0xaa, IDECTRL_ADDR_LBAL);
536
537 nsect = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_NSECT);
538 lbal = ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_LBAL);
539
540 if ((nsect == 0x55) && (lbal == 0xaa))
541 return true;
542
543 return false;
544}
545
546
547static int ep93xx_pata_wait_after_reset(struct ata_link *link,
548 unsigned int devmask,
549 unsigned long deadline)
550{
551 struct ata_port *ap = link->ap;
552 struct ep93xx_pata_data *drv_data = ap->host->private_data;
553 unsigned int dev0 = devmask & (1 << 0);
554 unsigned int dev1 = devmask & (1 << 1);
555 int rc, ret = 0;
556
557 ata_msleep(ap, ATA_WAIT_AFTER_RESET);
558
559
560 rc = ata_sff_wait_ready(link, deadline);
561
562
563
564
565 if (rc)
566 return rc;
567
568
569
570
571
572 if (dev1) {
573 int i;
574
575 ap->ops->sff_dev_select(ap, 1);
576
577
578
579
580
581
582 for (i = 0; i < 2; i++) {
583 u8 nsect, lbal;
584
585 nsect = ep93xx_pata_read_reg(drv_data,
586 IDECTRL_ADDR_NSECT);
587 lbal = ep93xx_pata_read_reg(drv_data,
588 IDECTRL_ADDR_LBAL);
589 if (nsect == 1 && lbal == 1)
590 break;
591 msleep(50);
592 }
593
594 rc = ata_sff_wait_ready(link, deadline);
595 if (rc) {
596 if (rc != -ENODEV)
597 return rc;
598 ret = rc;
599 }
600 }
601
602 ap->ops->sff_dev_select(ap, 0);
603 if (dev1)
604 ap->ops->sff_dev_select(ap, 1);
605 if (dev0)
606 ap->ops->sff_dev_select(ap, 0);
607
608 return ret;
609}
610
611
612static int ep93xx_pata_bus_softreset(struct ata_port *ap, unsigned int devmask,
613 unsigned long deadline)
614{
615 struct ep93xx_pata_data *drv_data = ap->host->private_data;
616
617 ep93xx_pata_write_reg(drv_data, ap->ctl, IDECTRL_ADDR_CTL);
618 udelay(20);
619 ep93xx_pata_write_reg(drv_data, ap->ctl | ATA_SRST, IDECTRL_ADDR_CTL);
620 udelay(20);
621 ep93xx_pata_write_reg(drv_data, ap->ctl, IDECTRL_ADDR_CTL);
622 ap->last_ctl = ap->ctl;
623
624 return ep93xx_pata_wait_after_reset(&ap->link, devmask, deadline);
625}
626
627static void ep93xx_pata_release_dma(struct ep93xx_pata_data *drv_data)
628{
629 if (drv_data->dma_rx_channel) {
630 dma_release_channel(drv_data->dma_rx_channel);
631 drv_data->dma_rx_channel = NULL;
632 }
633 if (drv_data->dma_tx_channel) {
634 dma_release_channel(drv_data->dma_tx_channel);
635 drv_data->dma_tx_channel = NULL;
636 }
637}
638
639static bool ep93xx_pata_dma_filter(struct dma_chan *chan, void *filter_param)
640{
641 if (ep93xx_dma_chan_is_m2p(chan))
642 return false;
643
644 chan->private = filter_param;
645 return true;
646}
647
648static void ep93xx_pata_dma_init(struct ep93xx_pata_data *drv_data)
649{
650 const struct platform_device *pdev = drv_data->pdev;
651 dma_cap_mask_t mask;
652 struct dma_slave_config conf;
653
654 dma_cap_zero(mask);
655 dma_cap_set(DMA_SLAVE, mask);
656
657
658
659
660
661
662 drv_data->dma_rx_data.port = EP93XX_DMA_IDE;
663 drv_data->dma_rx_data.direction = DMA_DEV_TO_MEM;
664 drv_data->dma_rx_data.name = "ep93xx-pata-rx";
665 drv_data->dma_rx_channel = dma_request_channel(mask,
666 ep93xx_pata_dma_filter, &drv_data->dma_rx_data);
667 if (!drv_data->dma_rx_channel)
668 return;
669
670 drv_data->dma_tx_data.port = EP93XX_DMA_IDE;
671 drv_data->dma_tx_data.direction = DMA_MEM_TO_DEV;
672 drv_data->dma_tx_data.name = "ep93xx-pata-tx";
673 drv_data->dma_tx_channel = dma_request_channel(mask,
674 ep93xx_pata_dma_filter, &drv_data->dma_tx_data);
675 if (!drv_data->dma_tx_channel) {
676 dma_release_channel(drv_data->dma_rx_channel);
677 return;
678 }
679
680
681 memset(&conf, 0, sizeof(conf));
682 conf.direction = DMA_DEV_TO_MEM;
683 conf.src_addr = drv_data->udma_in_phys;
684 conf.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
685 if (dmaengine_slave_config(drv_data->dma_rx_channel, &conf)) {
686 dev_err(&pdev->dev, "failed to configure rx dma channel\n");
687 ep93xx_pata_release_dma(drv_data);
688 return;
689 }
690
691
692 memset(&conf, 0, sizeof(conf));
693 conf.direction = DMA_MEM_TO_DEV;
694 conf.dst_addr = drv_data->udma_out_phys;
695 conf.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
696 if (dmaengine_slave_config(drv_data->dma_tx_channel, &conf)) {
697 dev_err(&pdev->dev, "failed to configure tx dma channel\n");
698 ep93xx_pata_release_dma(drv_data);
699 }
700}
701
702static void ep93xx_pata_dma_start(struct ata_queued_cmd *qc)
703{
704 struct dma_async_tx_descriptor *txd;
705 struct ep93xx_pata_data *drv_data = qc->ap->host->private_data;
706 void __iomem *base = drv_data->ide_base;
707 struct ata_device *adev = qc->dev;
708 u32 v = qc->dma_dir == DMA_TO_DEVICE ? IDEUDMAOP_RWOP : 0;
709 struct dma_chan *channel = qc->dma_dir == DMA_TO_DEVICE
710 ? drv_data->dma_tx_channel : drv_data->dma_rx_channel;
711
712 txd = dmaengine_prep_slave_sg(channel, qc->sg, qc->n_elem, qc->dma_dir,
713 DMA_CTRL_ACK);
714 if (!txd) {
715 dev_err(qc->ap->dev, "failed to prepare slave for sg dma\n");
716 return;
717 }
718 txd->callback = NULL;
719 txd->callback_param = NULL;
720
721 if (dmaengine_submit(txd) < 0) {
722 dev_err(qc->ap->dev, "failed to submit dma transfer\n");
723 return;
724 }
725 dma_async_issue_pending(channel);
726
727
728
729
730
731
732
733
734 writel(v, base + IDEUDMAOP);
735 readl(base + IDEUDMAOP);
736 writel(v | IDEUDMAOP_UEN, base + IDEUDMAOP);
737
738 writel(IDECFG_IDEEN | IDECFG_UDMA |
739 ((adev->xfer_mode - XFER_UDMA_0) << IDECFG_MODE_SHIFT),
740 base + IDECFG);
741}
742
743static void ep93xx_pata_dma_stop(struct ata_queued_cmd *qc)
744{
745 struct ep93xx_pata_data *drv_data = qc->ap->host->private_data;
746 void __iomem *base = drv_data->ide_base;
747
748
749 dmaengine_terminate_all(drv_data->dma_rx_channel);
750 dmaengine_terminate_all(drv_data->dma_tx_channel);
751
752
753
754
755
756 writel(0, base + IDEUDMAOP);
757 writel(readl(base + IDECTRL) | IDECTRL_DIOWN | IDECTRL_DIORN |
758 IDECTRL_CS0N | IDECTRL_CS1N, base + IDECTRL);
759
760 ep93xx_pata_enable_pio(drv_data->ide_base,
761 qc->dev->pio_mode - XFER_PIO_0);
762
763 ata_sff_dma_pause(qc->ap);
764}
765
766static void ep93xx_pata_dma_setup(struct ata_queued_cmd *qc)
767{
768 qc->ap->ops->sff_exec_command(qc->ap, &qc->tf);
769}
770
771static u8 ep93xx_pata_dma_status(struct ata_port *ap)
772{
773 struct ep93xx_pata_data *drv_data = ap->host->private_data;
774 u32 val = readl(drv_data->ide_base + IDEUDMASTS);
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789 if (val & IDEUDMASTS_NDO || val & IDEUDMASTS_NDI ||
790 val & IDEUDMASTS_N4X || val & IDEUDMASTS_INTIDE)
791 return ATA_DMA_ERR;
792
793
794 if (readl(drv_data->ide_base + IDECTRL) & IDECTRL_INTRQ)
795 return ATA_DMA_INTR;
796
797 if (val & IDEUDMASTS_SBUSY || val & IDEUDMASTS_DMAIDE)
798 return ATA_DMA_ACTIVE;
799
800 return 0;
801}
802
803
804static int ep93xx_pata_softreset(struct ata_link *al, unsigned int *classes,
805 unsigned long deadline)
806{
807 struct ata_port *ap = al->ap;
808 unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS;
809 unsigned int devmask = 0;
810 int rc;
811 u8 err;
812
813
814 if (ep93xx_pata_device_is_present(ap, 0))
815 devmask |= (1 << 0);
816 if (slave_possible && ep93xx_pata_device_is_present(ap, 1))
817 devmask |= (1 << 1);
818
819
820 ap->ops->sff_dev_select(al->ap, 0);
821
822
823 rc = ep93xx_pata_bus_softreset(ap, devmask, deadline);
824
825 if (rc && (rc != -ENODEV || sata_scr_valid(al))) {
826 ata_link_err(al, "SRST failed (errno=%d)\n", rc);
827 return rc;
828 }
829
830
831 classes[0] = ata_sff_dev_classify(&al->device[0], devmask & (1 << 0),
832 &err);
833 if (slave_possible && err != 0x81)
834 classes[1] = ata_sff_dev_classify(&al->device[1],
835 devmask & (1 << 1), &err);
836
837 return 0;
838}
839
840
841static void ep93xx_pata_drain_fifo(struct ata_queued_cmd *qc)
842{
843 int count;
844 struct ata_port *ap;
845 struct ep93xx_pata_data *drv_data;
846
847
848 if (qc == NULL || qc->dma_dir == DMA_TO_DEVICE)
849 return;
850
851 ap = qc->ap;
852 drv_data = ap->host->private_data;
853
854 for (count = 0; (ap->ops->sff_check_status(ap) & ATA_DRQ)
855 && count < 65536; count += 2)
856 ep93xx_pata_read_reg(drv_data, IDECTRL_ADDR_DATA);
857
858
859 if (count)
860 ata_port_dbg(ap, "drained %d bytes to clear DRQ.\n", count);
861
862}
863
864static int ep93xx_pata_port_start(struct ata_port *ap)
865{
866 struct ep93xx_pata_data *drv_data = ap->host->private_data;
867
868
869
870
871
872 drv_data->t = *ata_timing_find_mode(XFER_PIO_0);
873 return 0;
874}
875
876static struct scsi_host_template ep93xx_pata_sht = {
877 ATA_BASE_SHT(DRV_NAME),
878
879 .sg_tablesize = 32,
880
881 .dma_boundary = 0x7fff,
882};
883
884static struct ata_port_operations ep93xx_pata_port_ops = {
885 .inherits = &ata_bmdma_port_ops,
886
887 .qc_prep = ata_noop_qc_prep,
888
889 .softreset = ep93xx_pata_softreset,
890 .hardreset = ATA_OP_NULL,
891
892 .sff_dev_select = ep93xx_pata_dev_select,
893 .sff_set_devctl = ep93xx_pata_set_devctl,
894 .sff_check_status = ep93xx_pata_check_status,
895 .sff_check_altstatus = ep93xx_pata_check_altstatus,
896 .sff_tf_load = ep93xx_pata_tf_load,
897 .sff_tf_read = ep93xx_pata_tf_read,
898 .sff_exec_command = ep93xx_pata_exec_command,
899 .sff_data_xfer = ep93xx_pata_data_xfer,
900 .sff_drain_fifo = ep93xx_pata_drain_fifo,
901 .sff_irq_clear = ATA_OP_NULL,
902
903 .set_piomode = ep93xx_pata_set_piomode,
904
905 .bmdma_setup = ep93xx_pata_dma_setup,
906 .bmdma_start = ep93xx_pata_dma_start,
907 .bmdma_stop = ep93xx_pata_dma_stop,
908 .bmdma_status = ep93xx_pata_dma_status,
909
910 .cable_detect = ata_cable_unknown,
911 .port_start = ep93xx_pata_port_start,
912};
913
914static int ep93xx_pata_probe(struct platform_device *pdev)
915{
916 struct ep93xx_pata_data *drv_data;
917 struct ata_host *host;
918 struct ata_port *ap;
919 int irq;
920 struct resource *mem_res;
921 void __iomem *ide_base;
922 int err;
923
924 err = ep93xx_ide_acquire_gpio(pdev);
925 if (err)
926 return err;
927
928
929 irq = platform_get_irq(pdev, 0);
930 if (irq < 0) {
931 err = irq;
932 goto err_rel_gpio;
933 }
934
935 mem_res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
936 ide_base = devm_ioremap_resource(&pdev->dev, mem_res);
937 if (IS_ERR(ide_base)) {
938 err = PTR_ERR(ide_base);
939 goto err_rel_gpio;
940 }
941
942 drv_data = devm_kzalloc(&pdev->dev, sizeof(*drv_data), GFP_KERNEL);
943 if (!drv_data) {
944 err = -ENXIO;
945 goto err_rel_gpio;
946 }
947
948 drv_data->pdev = pdev;
949 drv_data->ide_base = ide_base;
950 drv_data->udma_in_phys = mem_res->start + IDEUDMADATAIN;
951 drv_data->udma_out_phys = mem_res->start + IDEUDMADATAOUT;
952 ep93xx_pata_dma_init(drv_data);
953
954
955 host = ata_host_alloc(&pdev->dev, 1);
956 if (!host) {
957 err = -ENXIO;
958 goto err_rel_dma;
959 }
960
961 ep93xx_pata_clear_regs(ide_base);
962
963 host->private_data = drv_data;
964
965 ap = host->ports[0];
966 ap->dev = &pdev->dev;
967 ap->ops = &ep93xx_pata_port_ops;
968 ap->flags |= ATA_FLAG_SLAVE_POSS;
969 ap->pio_mask = ATA_PIO4;
970
971
972
973
974
975
976
977
978
979
980 if (drv_data->dma_rx_channel && drv_data->dma_tx_channel) {
981 int chip_rev = ep93xx_chip_revision();
982
983 if (chip_rev == EP93XX_CHIP_REV_E1)
984 ap->udma_mask = ATA_UDMA3;
985 else if (chip_rev == EP93XX_CHIP_REV_E2)
986 ap->udma_mask = ATA_UDMA4;
987 else
988 ap->udma_mask = ATA_UDMA2;
989 }
990
991
992 ep93xx_pata_enable_pio(ide_base, 0);
993
994 dev_info(&pdev->dev, "version " DRV_VERSION "\n");
995
996
997 err = ata_host_activate(host, irq, ata_bmdma_interrupt, 0,
998 &ep93xx_pata_sht);
999 if (err == 0)
1000 return 0;
1001
1002err_rel_dma:
1003 ep93xx_pata_release_dma(drv_data);
1004err_rel_gpio:
1005 ep93xx_ide_release_gpio(pdev);
1006 return err;
1007}
1008
1009static int ep93xx_pata_remove(struct platform_device *pdev)
1010{
1011 struct ata_host *host = platform_get_drvdata(pdev);
1012 struct ep93xx_pata_data *drv_data = host->private_data;
1013
1014 ata_host_detach(host);
1015 ep93xx_pata_release_dma(drv_data);
1016 ep93xx_pata_clear_regs(drv_data->ide_base);
1017 ep93xx_ide_release_gpio(pdev);
1018 return 0;
1019}
1020
1021static struct platform_driver ep93xx_pata_platform_driver = {
1022 .driver = {
1023 .name = DRV_NAME,
1024 },
1025 .probe = ep93xx_pata_probe,
1026 .remove = ep93xx_pata_remove,
1027};
1028
1029module_platform_driver(ep93xx_pata_platform_driver);
1030
1031MODULE_AUTHOR("Alessandro Zummo, Lennert Buytenhek, Joao Ramos, "
1032 "Bartlomiej Zolnierkiewicz, Rafal Prylowski");
1033MODULE_DESCRIPTION("low-level driver for cirrus ep93xx IDE controller");
1034MODULE_LICENSE("GPL");
1035MODULE_VERSION(DRV_VERSION);
1036MODULE_ALIAS("platform:pata_ep93xx");
1037