1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42#include <linux/module.h>
43#include <linux/kernel.h>
44#include <linux/types.h>
45#include <linux/skbuff.h>
46#include <linux/netdevice.h>
47#include <linux/ioport.h>
48#include <linux/delay.h>
49#include <linux/slab.h>
50#include <linux/init.h>
51#include <linux/rtnetlink.h>
52#include <linux/dma-mapping.h>
53
54#include <asm/io.h>
55#include <asm/dma.h>
56#include <asm/byteorder.h>
57
58#include <net/irda/irda.h>
59#include <net/irda/wrapper.h>
60#include <net/irda/irda_device.h>
61#include "w83977af.h"
62#include "w83977af_ir.h"
63
64#ifdef CONFIG_ARCH_NETWINDER
65#undef CONFIG_NETWINDER_TX_DMA_PROBLEMS
66#define CONFIG_NETWINDER_RX_DMA_PROBLEMS
67#endif
68#undef CONFIG_USE_INTERNAL_TIMER
69#define CONFIG_USE_W977_PNP
70#define PIO_MAX_SPEED 115200
71
72static char *driver_name = "w83977af_ir";
73static int qos_mtt_bits = 0x07;
74
75#define CHIP_IO_EXTENT 8
76
77static unsigned int io[] = { 0x180, ~0, ~0, ~0 };
78#ifdef CONFIG_ARCH_NETWINDER
79static unsigned int irq[] = { 6, 0, 0, 0 };
80#else
81static unsigned int irq[] = { 11, 0, 0, 0 };
82#endif
83static unsigned int dma[] = { 1, 0, 0, 0 };
84static unsigned int efbase[] = { W977_EFIO_BASE, W977_EFIO2_BASE };
85static unsigned int efio = W977_EFIO_BASE;
86
87static struct w83977af_ir *dev_self[] = { NULL, NULL, NULL, NULL};
88
89
90static int w83977af_open(int i, unsigned int iobase, unsigned int irq,
91 unsigned int dma);
92static int w83977af_close(struct w83977af_ir *self);
93static int w83977af_probe(int iobase, int irq, int dma);
94static int w83977af_dma_receive(struct w83977af_ir *self);
95static int w83977af_dma_receive_complete(struct w83977af_ir *self);
96static netdev_tx_t w83977af_hard_xmit(struct sk_buff *skb,
97 struct net_device *dev);
98static int w83977af_pio_write(int iobase, __u8 *buf, int len, int fifo_size);
99static void w83977af_dma_write(struct w83977af_ir *self, int iobase);
100static void w83977af_change_speed(struct w83977af_ir *self, __u32 speed);
101static int w83977af_is_receiving(struct w83977af_ir *self);
102
103static int w83977af_net_open(struct net_device *dev);
104static int w83977af_net_close(struct net_device *dev);
105static int w83977af_net_ioctl(struct net_device *dev, struct ifreq *rq, int cmd);
106
107
108
109
110
111
112
113static int __init w83977af_init(void)
114{
115 int i;
116
117 IRDA_DEBUG(0, "%s()\n", __func__ );
118
119 for (i=0; i < ARRAY_SIZE(dev_self) && io[i] < 2000; i++) {
120 if (w83977af_open(i, io[i], irq[i], dma[i]) == 0)
121 return 0;
122 }
123 return -ENODEV;
124}
125
126
127
128
129
130
131
132static void __exit w83977af_cleanup(void)
133{
134 int i;
135
136 IRDA_DEBUG(4, "%s()\n", __func__ );
137
138 for (i=0; i < ARRAY_SIZE(dev_self); i++) {
139 if (dev_self[i])
140 w83977af_close(dev_self[i]);
141 }
142}
143
144static const struct net_device_ops w83977_netdev_ops = {
145 .ndo_open = w83977af_net_open,
146 .ndo_stop = w83977af_net_close,
147 .ndo_start_xmit = w83977af_hard_xmit,
148 .ndo_do_ioctl = w83977af_net_ioctl,
149};
150
151
152
153
154
155
156
157static int w83977af_open(int i, unsigned int iobase, unsigned int irq,
158 unsigned int dma)
159{
160 struct net_device *dev;
161 struct w83977af_ir *self;
162 int err;
163
164 IRDA_DEBUG(0, "%s()\n", __func__ );
165
166
167 if (!request_region(iobase, CHIP_IO_EXTENT, driver_name)) {
168 IRDA_DEBUG(0, "%s(), can't get iobase of 0x%03x\n",
169 __func__ , iobase);
170 return -ENODEV;
171 }
172
173 if (w83977af_probe(iobase, irq, dma) == -1) {
174 err = -1;
175 goto err_out;
176 }
177
178
179
180 dev = alloc_irdadev(sizeof(struct w83977af_ir));
181 if (dev == NULL) {
182 printk( KERN_ERR "IrDA: Can't allocate memory for "
183 "IrDA control block!\n");
184 err = -ENOMEM;
185 goto err_out;
186 }
187
188 self = netdev_priv(dev);
189 spin_lock_init(&self->lock);
190
191
192
193 self->io.fir_base = iobase;
194 self->io.irq = irq;
195 self->io.fir_ext = CHIP_IO_EXTENT;
196 self->io.dma = dma;
197 self->io.fifo_size = 32;
198
199
200 irda_init_max_qos_capabilies(&self->qos);
201
202
203
204
205 self->qos.baud_rate.bits = IR_9600|IR_19200|IR_38400|IR_57600|
206 IR_115200|IR_576000|IR_1152000|(IR_4000000 << 8);
207
208
209 self->qos.min_turn_time.bits = qos_mtt_bits;
210 irda_qos_bits_to_value(&self->qos);
211
212
213 self->rx_buff.truesize = 14384;
214 self->tx_buff.truesize = 4000;
215
216
217 self->rx_buff.head =
218 dma_alloc_coherent(NULL, self->rx_buff.truesize,
219 &self->rx_buff_dma, GFP_KERNEL);
220 if (self->rx_buff.head == NULL) {
221 err = -ENOMEM;
222 goto err_out1;
223 }
224
225 memset(self->rx_buff.head, 0, self->rx_buff.truesize);
226
227 self->tx_buff.head =
228 dma_alloc_coherent(NULL, self->tx_buff.truesize,
229 &self->tx_buff_dma, GFP_KERNEL);
230 if (self->tx_buff.head == NULL) {
231 err = -ENOMEM;
232 goto err_out2;
233 }
234 memset(self->tx_buff.head, 0, self->tx_buff.truesize);
235
236 self->rx_buff.in_frame = FALSE;
237 self->rx_buff.state = OUTSIDE_FRAME;
238 self->tx_buff.data = self->tx_buff.head;
239 self->rx_buff.data = self->rx_buff.head;
240 self->netdev = dev;
241
242 dev->netdev_ops = &w83977_netdev_ops;
243
244 err = register_netdev(dev);
245 if (err) {
246 IRDA_ERROR("%s(), register_netdevice() failed!\n", __func__);
247 goto err_out3;
248 }
249 IRDA_MESSAGE("IrDA: Registered device %s\n", dev->name);
250
251
252 dev_self[i] = self;
253
254 return 0;
255err_out3:
256 dma_free_coherent(NULL, self->tx_buff.truesize,
257 self->tx_buff.head, self->tx_buff_dma);
258err_out2:
259 dma_free_coherent(NULL, self->rx_buff.truesize,
260 self->rx_buff.head, self->rx_buff_dma);
261err_out1:
262 free_netdev(dev);
263err_out:
264 release_region(iobase, CHIP_IO_EXTENT);
265 return err;
266}
267
268
269
270
271
272
273
274static int w83977af_close(struct w83977af_ir *self)
275{
276 int iobase;
277
278 IRDA_DEBUG(0, "%s()\n", __func__ );
279
280 iobase = self->io.fir_base;
281
282#ifdef CONFIG_USE_W977_PNP
283
284 w977_efm_enter(efio);
285
286 w977_select_device(W977_DEVICE_IR, efio);
287
288
289 w977_write_reg(0x30, 0x00, efio);
290
291 w977_efm_exit(efio);
292#endif
293
294
295 unregister_netdev(self->netdev);
296
297
298 IRDA_DEBUG(0 , "%s(), Releasing Region %03x\n",
299 __func__ , self->io.fir_base);
300 release_region(self->io.fir_base, self->io.fir_ext);
301
302 if (self->tx_buff.head)
303 dma_free_coherent(NULL, self->tx_buff.truesize,
304 self->tx_buff.head, self->tx_buff_dma);
305
306 if (self->rx_buff.head)
307 dma_free_coherent(NULL, self->rx_buff.truesize,
308 self->rx_buff.head, self->rx_buff_dma);
309
310 free_netdev(self->netdev);
311
312 return 0;
313}
314
315static int w83977af_probe(int iobase, int irq, int dma)
316{
317 int version;
318 int i;
319
320 for (i=0; i < 2; i++) {
321 IRDA_DEBUG( 0, "%s()\n", __func__ );
322#ifdef CONFIG_USE_W977_PNP
323
324 w977_efm_enter(efbase[i]);
325
326 w977_select_device(W977_DEVICE_IR, efbase[i]);
327
328
329 w977_write_reg(0x60, (iobase >> 8) & 0xff, efbase[i]);
330 w977_write_reg(0x61, (iobase) & 0xff, efbase[i]);
331
332 w977_write_reg(0x70, irq, efbase[i]);
333#ifdef CONFIG_ARCH_NETWINDER
334
335 w977_write_reg(0x74, dma+1, efbase[i]);
336#else
337 w977_write_reg(0x74, dma, efbase[i]);
338#endif
339 w977_write_reg(0x75, 0x04, efbase[i]);
340
341
342 w977_write_reg(0xf0, APEDCRC|ENBNKSEL, efbase[i]);
343
344
345 w977_write_reg(0x30, 0x01, efbase[i]);
346
347 w977_efm_exit(efbase[i]);
348#endif
349
350 switch_bank(iobase, SET2);
351 outb(iobase+2, 0x00);
352
353
354 switch_bank(iobase, SET0);
355 outb(HCR_EN_IRQ, iobase+HCR);
356
357
358 switch_bank(iobase, SET2);
359 outb(inb(iobase+ADCR1) | ADCR1_ADV_SL, iobase+ADCR1);
360
361
362 switch_bank(iobase, SET0);
363 outb(HCR_SIR, iobase+HCR);
364
365
366 switch_bank(iobase, SET3);
367 version = inb(iobase+AUID);
368
369
370 if (0x10 == (version & 0xf0)) {
371 efio = efbase[i];
372
373
374 switch_bank(iobase, SET2);
375 outb(ADCR2_RXFS32|ADCR2_TXFS32, iobase+ADCR2);
376
377
378 switch_bank(iobase, SET0);
379 outb(UFR_RXTL|UFR_TXTL|UFR_TXF_RST|UFR_RXF_RST|
380 UFR_EN_FIFO,iobase+UFR);
381
382
383 switch_bank(iobase, SET4);
384 outb(2048 & 0xff, iobase+6);
385 outb((2048 >> 8) & 0x1f, iobase+7);
386
387
388
389
390
391
392
393
394
395
396
397
398
399 switch_bank(iobase, SET7);
400 outb(0x40, iobase+7);
401
402 IRDA_MESSAGE("W83977AF (IR) driver loaded. "
403 "Version: 0x%02x\n", version);
404
405 return 0;
406 } else {
407
408 IRDA_DEBUG( 0, "%s(), Wrong chip version", __func__ );
409 }
410 }
411 return -1;
412}
413
414static void w83977af_change_speed(struct w83977af_ir *self, __u32 speed)
415{
416 int ir_mode = HCR_SIR;
417 int iobase;
418 __u8 set;
419
420 iobase = self->io.fir_base;
421
422
423 self->io.speed = speed;
424
425
426 set = inb(iobase+SSR);
427
428
429 switch_bank(iobase, SET0);
430 outb(0, iobase+ICR);
431
432
433 switch_bank(iobase, SET2);
434 outb(0x00, iobase+ABHL);
435
436 switch (speed) {
437 case 9600: outb(0x0c, iobase+ABLL); break;
438 case 19200: outb(0x06, iobase+ABLL); break;
439 case 38400: outb(0x03, iobase+ABLL); break;
440 case 57600: outb(0x02, iobase+ABLL); break;
441 case 115200: outb(0x01, iobase+ABLL); break;
442 case 576000:
443 ir_mode = HCR_MIR_576;
444 IRDA_DEBUG(0, "%s(), handling baud of 576000\n", __func__ );
445 break;
446 case 1152000:
447 ir_mode = HCR_MIR_1152;
448 IRDA_DEBUG(0, "%s(), handling baud of 1152000\n", __func__ );
449 break;
450 case 4000000:
451 ir_mode = HCR_FIR;
452 IRDA_DEBUG(0, "%s(), handling baud of 4000000\n", __func__ );
453 break;
454 default:
455 ir_mode = HCR_FIR;
456 IRDA_DEBUG(0, "%s(), unknown baud rate of %d\n", __func__ , speed);
457 break;
458 }
459
460
461 switch_bank(iobase, SET0);
462 outb(ir_mode, iobase+HCR);
463
464
465 switch_bank(iobase, SET2);
466 outb(ADCR2_RXFS32|ADCR2_TXFS32, iobase+ADCR2);
467
468
469 switch_bank(iobase, SET0);
470 outb(0x00, iobase+UFR);
471 outb(UFR_EN_FIFO, iobase+UFR);
472 outb(0xa7, iobase+UFR);
473
474 netif_wake_queue(self->netdev);
475
476
477 switch_bank(iobase, SET0);
478 if (speed > PIO_MAX_SPEED) {
479 outb(ICR_EFSFI, iobase+ICR);
480 w83977af_dma_receive(self);
481 } else
482 outb(ICR_ERBRI, iobase+ICR);
483
484
485 outb(set, iobase+SSR);
486}
487
488
489
490
491
492
493
494static netdev_tx_t w83977af_hard_xmit(struct sk_buff *skb,
495 struct net_device *dev)
496{
497 struct w83977af_ir *self;
498 __s32 speed;
499 int iobase;
500 __u8 set;
501 int mtt;
502
503 self = netdev_priv(dev);
504
505 iobase = self->io.fir_base;
506
507 IRDA_DEBUG(4, "%s(%ld), skb->len=%d\n", __func__ , jiffies,
508 (int) skb->len);
509
510
511 netif_stop_queue(dev);
512
513
514 speed = irda_get_next_speed(skb);
515 if ((speed != self->io.speed) && (speed != -1)) {
516
517 if (!skb->len) {
518 w83977af_change_speed(self, speed);
519 dev->trans_start = jiffies;
520 dev_kfree_skb(skb);
521 return NETDEV_TX_OK;
522 } else
523 self->new_speed = speed;
524 }
525
526
527 set = inb(iobase+SSR);
528
529
530 if (self->io.speed > PIO_MAX_SPEED) {
531 self->tx_buff.data = self->tx_buff.head;
532 skb_copy_from_linear_data(skb, self->tx_buff.data, skb->len);
533 self->tx_buff.len = skb->len;
534
535 mtt = irda_get_mtt(skb);
536#ifdef CONFIG_USE_INTERNAL_TIMER
537 if (mtt > 50) {
538
539 mtt /= 1000+1;
540
541
542 switch_bank(iobase, SET4);
543 outb(mtt & 0xff, iobase+TMRL);
544 outb((mtt >> 8) & 0x0f, iobase+TMRH);
545
546
547 outb(IR_MSL_EN_TMR, iobase+IR_MSL);
548 self->io.direction = IO_XMIT;
549
550
551 switch_bank(iobase, SET0);
552 outb(ICR_ETMRI, iobase+ICR);
553 } else {
554#endif
555 IRDA_DEBUG(4, "%s(%ld), mtt=%d\n", __func__ , jiffies, mtt);
556 if (mtt)
557 udelay(mtt);
558
559
560 switch_bank(iobase, SET0);
561 outb(ICR_EDMAI, iobase+ICR);
562 w83977af_dma_write(self, iobase);
563#ifdef CONFIG_USE_INTERNAL_TIMER
564 }
565#endif
566 } else {
567 self->tx_buff.data = self->tx_buff.head;
568 self->tx_buff.len = async_wrap_skb(skb, self->tx_buff.data,
569 self->tx_buff.truesize);
570
571
572 switch_bank(iobase, SET0);
573 outb(ICR_ETXTHI, iobase+ICR);
574 }
575 dev->trans_start = jiffies;
576 dev_kfree_skb(skb);
577
578
579 outb(set, iobase+SSR);
580
581 return NETDEV_TX_OK;
582}
583
584
585
586
587
588
589
590static void w83977af_dma_write(struct w83977af_ir *self, int iobase)
591{
592 __u8 set;
593#ifdef CONFIG_NETWINDER_TX_DMA_PROBLEMS
594 unsigned long flags;
595 __u8 hcr;
596#endif
597 IRDA_DEBUG(4, "%s(), len=%d\n", __func__ , self->tx_buff.len);
598
599
600 set = inb(iobase+SSR);
601
602
603 switch_bank(iobase, SET0);
604 outb(inb(iobase+HCR) & ~HCR_EN_DMA, iobase+HCR);
605
606
607 switch_bank(iobase, SET2);
608 outb(ADCR1_D_CHSW|ADCR1_ADV_SL, iobase+ADCR1);
609#ifdef CONFIG_NETWINDER_TX_DMA_PROBLEMS
610 spin_lock_irqsave(&self->lock, flags);
611
612 disable_dma(self->io.dma);
613 clear_dma_ff(self->io.dma);
614 set_dma_mode(self->io.dma, DMA_MODE_READ);
615 set_dma_addr(self->io.dma, self->tx_buff_dma);
616 set_dma_count(self->io.dma, self->tx_buff.len);
617#else
618 irda_setup_dma(self->io.dma, self->tx_buff_dma, self->tx_buff.len,
619 DMA_MODE_WRITE);
620#endif
621 self->io.direction = IO_XMIT;
622
623
624 switch_bank(iobase, SET0);
625#ifdef CONFIG_NETWINDER_TX_DMA_PROBLEMS
626 hcr = inb(iobase+HCR);
627 outb(hcr | HCR_EN_DMA, iobase+HCR);
628 enable_dma(self->io.dma);
629 spin_unlock_irqrestore(&self->lock, flags);
630#else
631 outb(inb(iobase+HCR) | HCR_EN_DMA | HCR_TX_WT, iobase+HCR);
632#endif
633
634
635 outb(set, iobase+SSR);
636}
637
638
639
640
641
642
643
644static int w83977af_pio_write(int iobase, __u8 *buf, int len, int fifo_size)
645{
646 int actual = 0;
647 __u8 set;
648
649 IRDA_DEBUG(4, "%s()\n", __func__ );
650
651
652 set = inb(iobase+SSR);
653
654 switch_bank(iobase, SET0);
655 if (!(inb_p(iobase+USR) & USR_TSRE)) {
656 IRDA_DEBUG(4,
657 "%s(), warning, FIFO not empty yet!\n", __func__ );
658
659 fifo_size -= 17;
660 IRDA_DEBUG(4, "%s(), %d bytes left in tx fifo\n",
661 __func__ , fifo_size);
662 }
663
664
665 while ((fifo_size-- > 0) && (actual < len)) {
666
667 outb(buf[actual++], iobase+TBR);
668 }
669
670 IRDA_DEBUG(4, "%s(), fifo_size %d ; %d sent of %d\n",
671 __func__ , fifo_size, actual, len);
672
673
674 outb(set, iobase+SSR);
675
676 return actual;
677}
678
679
680
681
682
683
684
685
686static void w83977af_dma_xmit_complete(struct w83977af_ir *self)
687{
688 int iobase;
689 __u8 set;
690
691 IRDA_DEBUG(4, "%s(%ld)\n", __func__ , jiffies);
692
693 IRDA_ASSERT(self != NULL, return;);
694
695 iobase = self->io.fir_base;
696
697
698 set = inb(iobase+SSR);
699
700
701 switch_bank(iobase, SET0);
702 outb(inb(iobase+HCR) & ~HCR_EN_DMA, iobase+HCR);
703
704
705 if (inb(iobase+AUDR) & AUDR_UNDR) {
706 IRDA_DEBUG(0, "%s(), Transmit underrun!\n", __func__ );
707
708 self->netdev->stats.tx_errors++;
709 self->netdev->stats.tx_fifo_errors++;
710
711
712 outb(AUDR_UNDR, iobase+AUDR);
713 } else
714 self->netdev->stats.tx_packets++;
715
716
717 if (self->new_speed) {
718 w83977af_change_speed(self, self->new_speed);
719 self->new_speed = 0;
720 }
721
722
723
724 netif_wake_queue(self->netdev);
725
726
727 outb(set, iobase+SSR);
728}
729
730
731
732
733
734
735
736
737static int w83977af_dma_receive(struct w83977af_ir *self)
738{
739 int iobase;
740 __u8 set;
741#ifdef CONFIG_NETWINDER_RX_DMA_PROBLEMS
742 unsigned long flags;
743 __u8 hcr;
744#endif
745 IRDA_ASSERT(self != NULL, return -1;);
746
747 IRDA_DEBUG(4, "%s\n", __func__ );
748
749 iobase= self->io.fir_base;
750
751
752 set = inb(iobase+SSR);
753
754
755 switch_bank(iobase, SET0);
756 outb(inb(iobase+HCR) & ~HCR_EN_DMA, iobase+HCR);
757
758
759 switch_bank(iobase, SET2);
760 outb((inb(iobase+ADCR1) & ~ADCR1_D_CHSW)|ADCR1_ADV_SL,
761 iobase+ADCR1);
762
763 self->io.direction = IO_RECV;
764 self->rx_buff.data = self->rx_buff.head;
765
766#ifdef CONFIG_NETWINDER_RX_DMA_PROBLEMS
767 spin_lock_irqsave(&self->lock, flags);
768
769 disable_dma(self->io.dma);
770 clear_dma_ff(self->io.dma);
771 set_dma_mode(self->io.dma, DMA_MODE_READ);
772 set_dma_addr(self->io.dma, self->rx_buff_dma);
773 set_dma_count(self->io.dma, self->rx_buff.truesize);
774#else
775 irda_setup_dma(self->io.dma, self->rx_buff_dma, self->rx_buff.truesize,
776 DMA_MODE_READ);
777#endif
778
779
780
781
782
783 switch_bank(iobase, SET0);
784 outb(UFR_RXTL|UFR_TXTL|UFR_RXF_RST|UFR_EN_FIFO, iobase+UFR);
785 self->st_fifo.len = self->st_fifo.tail = self->st_fifo.head = 0;
786
787
788 switch_bank(iobase, SET0);
789#ifdef CONFIG_NETWINDER_RX_DMA_PROBLEMS
790 hcr = inb(iobase+HCR);
791 outb(hcr | HCR_EN_DMA, iobase+HCR);
792 enable_dma(self->io.dma);
793 spin_unlock_irqrestore(&self->lock, flags);
794#else
795 outb(inb(iobase+HCR) | HCR_EN_DMA, iobase+HCR);
796#endif
797
798 outb(set, iobase+SSR);
799
800 return 0;
801}
802
803
804
805
806
807
808
809static int w83977af_dma_receive_complete(struct w83977af_ir *self)
810{
811 struct sk_buff *skb;
812 struct st_fifo *st_fifo;
813 int len;
814 int iobase;
815 __u8 set;
816 __u8 status;
817
818 IRDA_DEBUG(4, "%s\n", __func__ );
819
820 st_fifo = &self->st_fifo;
821
822 iobase = self->io.fir_base;
823
824
825 set = inb(iobase+SSR);
826
827 iobase = self->io.fir_base;
828
829
830 switch_bank(iobase, SET5);
831 while ((status = inb(iobase+FS_FO)) & FS_FO_FSFDR) {
832 st_fifo->entries[st_fifo->tail].status = status;
833
834 st_fifo->entries[st_fifo->tail].len = inb(iobase+RFLFL);
835 st_fifo->entries[st_fifo->tail].len |= inb(iobase+RFLFH) << 8;
836
837 st_fifo->tail++;
838 st_fifo->len++;
839 }
840
841 while (st_fifo->len) {
842
843 status = st_fifo->entries[st_fifo->head].status;
844 len = st_fifo->entries[st_fifo->head].len;
845 st_fifo->head++;
846 st_fifo->len--;
847
848
849 if (status & FS_FO_ERR_MSK) {
850 if (status & FS_FO_LST_FR) {
851
852 self->netdev->stats.rx_errors += len;
853 } else {
854
855 self->netdev->stats.rx_errors++;
856
857 self->rx_buff.data += len;
858
859 if (status & FS_FO_MX_LEX)
860 self->netdev->stats.rx_length_errors++;
861
862 if (status & FS_FO_PHY_ERR)
863 self->netdev->stats.rx_frame_errors++;
864
865 if (status & FS_FO_CRC_ERR)
866 self->netdev->stats.rx_crc_errors++;
867 }
868
869 if (status & FS_FO_RX_OV)
870 self->netdev->stats.rx_fifo_errors++;
871
872 if (status & FS_FO_FSF_OV)
873 self->netdev->stats.rx_fifo_errors++;
874
875 } else {
876
877 switch_bank(iobase, SET0);
878 if (inb(iobase+USR) & USR_RDR) {
879#ifdef CONFIG_USE_INTERNAL_TIMER
880
881 st_fifo->head--;
882 st_fifo->len++;
883 st_fifo->entries[st_fifo->head].status = status;
884 st_fifo->entries[st_fifo->head].len = len;
885
886
887 outb(set, iobase+SSR);
888
889 return FALSE;
890#else
891 udelay(80);
892#endif
893 }
894
895 skb = dev_alloc_skb(len+1);
896 if (skb == NULL) {
897 printk(KERN_INFO
898 "%s(), memory squeeze, dropping frame.\n", __func__);
899
900 outb(set, iobase+SSR);
901
902 return FALSE;
903 }
904
905
906 skb_reserve(skb, 1);
907
908
909 if (self->io.speed < 4000000) {
910 skb_put(skb, len-2);
911 skb_copy_to_linear_data(skb,
912 self->rx_buff.data,
913 len - 2);
914 } else {
915 skb_put(skb, len-4);
916 skb_copy_to_linear_data(skb,
917 self->rx_buff.data,
918 len - 4);
919 }
920
921
922 self->rx_buff.data += len;
923 self->netdev->stats.rx_packets++;
924
925 skb->dev = self->netdev;
926 skb_reset_mac_header(skb);
927 skb->protocol = htons(ETH_P_IRDA);
928 netif_rx(skb);
929 }
930 }
931
932 outb(set, iobase+SSR);
933
934 return TRUE;
935}
936
937
938
939
940
941
942
943static void w83977af_pio_receive(struct w83977af_ir *self)
944{
945 __u8 byte = 0x00;
946 int iobase;
947
948 IRDA_DEBUG(4, "%s()\n", __func__ );
949
950 IRDA_ASSERT(self != NULL, return;);
951
952 iobase = self->io.fir_base;
953
954
955 do {
956 byte = inb(iobase+RBR);
957 async_unwrap_char(self->netdev, &self->netdev->stats, &self->rx_buff,
958 byte);
959 } while (inb(iobase+USR) & USR_RDR);
960}
961
962
963
964
965
966
967
968static __u8 w83977af_sir_interrupt(struct w83977af_ir *self, int isr)
969{
970 int actual;
971 __u8 new_icr = 0;
972 __u8 set;
973 int iobase;
974
975 IRDA_DEBUG(4, "%s(), isr=%#x\n", __func__ , isr);
976
977 iobase = self->io.fir_base;
978
979 if (isr & ISR_TXTH_I) {
980
981 actual = w83977af_pio_write(self->io.fir_base,
982 self->tx_buff.data,
983 self->tx_buff.len,
984 self->io.fifo_size);
985
986 self->tx_buff.data += actual;
987 self->tx_buff.len -= actual;
988
989 self->io.direction = IO_XMIT;
990
991
992 if (self->tx_buff.len > 0) {
993 new_icr |= ICR_ETXTHI;
994 } else {
995 set = inb(iobase+SSR);
996 switch_bank(iobase, SET0);
997 outb(AUDR_SFEND, iobase+AUDR);
998 outb(set, iobase+SSR);
999
1000 self->netdev->stats.tx_packets++;
1001
1002
1003 netif_wake_queue(self->netdev);
1004 new_icr |= ICR_ETBREI;
1005 }
1006 }
1007
1008 if (isr & ISR_TXEMP_I) {
1009
1010 if (self->new_speed) {
1011 IRDA_DEBUG(2,
1012 "%s(), Changing speed!\n", __func__ );
1013 w83977af_change_speed(self, self->new_speed);
1014 self->new_speed = 0;
1015 }
1016
1017
1018 self->io.direction = IO_RECV;
1019 new_icr |= ICR_ERBRI;
1020 }
1021
1022
1023 if (isr & ISR_RXTH_I) {
1024 w83977af_pio_receive(self);
1025
1026
1027 new_icr |= ICR_ERBRI;
1028 }
1029 return new_icr;
1030}
1031
1032
1033
1034
1035
1036
1037
1038static __u8 w83977af_fir_interrupt(struct w83977af_ir *self, int isr)
1039{
1040 __u8 new_icr = 0;
1041 __u8 set;
1042 int iobase;
1043
1044 iobase = self->io.fir_base;
1045 set = inb(iobase+SSR);
1046
1047
1048 if (isr & (ISR_FEND_I|ISR_FSF_I)) {
1049 if (w83977af_dma_receive_complete(self)) {
1050
1051
1052 new_icr |= ICR_EFSFI;
1053 } else {
1054
1055
1056
1057 switch_bank(iobase, SET4);
1058 outb(0x01, iobase+TMRL);
1059 outb(0x00, iobase+TMRH);
1060
1061
1062 outb(IR_MSL_EN_TMR, iobase+IR_MSL);
1063
1064 new_icr |= ICR_ETMRI;
1065 }
1066 }
1067
1068 if (isr & ISR_TMR_I) {
1069
1070 switch_bank(iobase, SET4);
1071 outb(0, iobase+IR_MSL);
1072
1073
1074
1075
1076
1077
1078 if (self->io.direction == IO_XMIT) {
1079 w83977af_dma_write(self, iobase);
1080
1081 new_icr |= ICR_EDMAI;
1082 } else {
1083
1084 w83977af_dma_receive_complete(self);
1085
1086 new_icr |= ICR_EFSFI;
1087 }
1088 }
1089
1090 if (isr & ISR_DMA_I) {
1091 w83977af_dma_xmit_complete(self);
1092
1093
1094
1095
1096
1097
1098
1099
1100 w83977af_dma_receive(self);
1101 new_icr = ICR_EFSFI;
1102
1103 }
1104
1105
1106 outb(set, iobase+SSR);
1107
1108 return new_icr;
1109}
1110
1111
1112
1113
1114
1115
1116
1117static irqreturn_t w83977af_interrupt(int irq, void *dev_id)
1118{
1119 struct net_device *dev = dev_id;
1120 struct w83977af_ir *self;
1121 __u8 set, icr, isr;
1122 int iobase;
1123
1124 self = netdev_priv(dev);
1125
1126 iobase = self->io.fir_base;
1127
1128
1129 set = inb(iobase+SSR);
1130 switch_bank(iobase, SET0);
1131
1132 icr = inb(iobase+ICR);
1133 isr = inb(iobase+ISR) & icr;
1134
1135 outb(0, iobase+ICR);
1136
1137 if (isr) {
1138
1139 if (self->io.speed > PIO_MAX_SPEED )
1140 icr = w83977af_fir_interrupt(self, isr);
1141 else
1142 icr = w83977af_sir_interrupt(self, isr);
1143 }
1144
1145 outb(icr, iobase+ICR);
1146 outb(set, iobase+SSR);
1147 return IRQ_RETVAL(isr);
1148}
1149
1150
1151
1152
1153
1154
1155
1156static int w83977af_is_receiving(struct w83977af_ir *self)
1157{
1158 int status = FALSE;
1159 int iobase;
1160 __u8 set;
1161
1162 IRDA_ASSERT(self != NULL, return FALSE;);
1163
1164 if (self->io.speed > 115200) {
1165 iobase = self->io.fir_base;
1166
1167
1168 set = inb(iobase+SSR);
1169 switch_bank(iobase, SET2);
1170 if ((inb(iobase+RXFDTH) & 0x3f) != 0) {
1171
1172 status = TRUE;
1173 }
1174 outb(set, iobase+SSR);
1175 } else
1176 status = (self->rx_buff.state != OUTSIDE_FRAME);
1177
1178 return status;
1179}
1180
1181
1182
1183
1184
1185
1186
1187static int w83977af_net_open(struct net_device *dev)
1188{
1189 struct w83977af_ir *self;
1190 int iobase;
1191 char hwname[32];
1192 __u8 set;
1193
1194 IRDA_DEBUG(0, "%s()\n", __func__ );
1195
1196 IRDA_ASSERT(dev != NULL, return -1;);
1197 self = netdev_priv(dev);
1198
1199 IRDA_ASSERT(self != NULL, return 0;);
1200
1201 iobase = self->io.fir_base;
1202
1203 if (request_irq(self->io.irq, w83977af_interrupt, 0, dev->name,
1204 (void *) dev)) {
1205 return -EAGAIN;
1206 }
1207
1208
1209
1210
1211 if (request_dma(self->io.dma, dev->name)) {
1212 free_irq(self->io.irq, self);
1213 return -EAGAIN;
1214 }
1215
1216
1217 set = inb(iobase+SSR);
1218
1219
1220 switch_bank(iobase, SET0);
1221 if (self->io.speed > 115200) {
1222 outb(ICR_EFSFI, iobase+ICR);
1223 w83977af_dma_receive(self);
1224 } else
1225 outb(ICR_ERBRI, iobase+ICR);
1226
1227
1228 outb(set, iobase+SSR);
1229
1230
1231 netif_start_queue(dev);
1232
1233
1234 sprintf(hwname, "w83977af @ 0x%03x", self->io.fir_base);
1235
1236
1237
1238
1239
1240 self->irlap = irlap_open(dev, &self->qos, hwname);
1241
1242 return 0;
1243}
1244
1245
1246
1247
1248
1249
1250
1251static int w83977af_net_close(struct net_device *dev)
1252{
1253 struct w83977af_ir *self;
1254 int iobase;
1255 __u8 set;
1256
1257 IRDA_DEBUG(0, "%s()\n", __func__ );
1258
1259 IRDA_ASSERT(dev != NULL, return -1;);
1260
1261 self = netdev_priv(dev);
1262
1263 IRDA_ASSERT(self != NULL, return 0;);
1264
1265 iobase = self->io.fir_base;
1266
1267
1268 netif_stop_queue(dev);
1269
1270
1271 if (self->irlap)
1272 irlap_close(self->irlap);
1273 self->irlap = NULL;
1274
1275 disable_dma(self->io.dma);
1276
1277
1278 set = inb(iobase+SSR);
1279
1280
1281 switch_bank(iobase, SET0);
1282 outb(0, iobase+ICR);
1283
1284 free_irq(self->io.irq, dev);
1285 free_dma(self->io.dma);
1286
1287
1288 outb(set, iobase+SSR);
1289
1290 return 0;
1291}
1292
1293
1294
1295
1296
1297
1298
1299static int w83977af_net_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
1300{
1301 struct if_irda_req *irq = (struct if_irda_req *) rq;
1302 struct w83977af_ir *self;
1303 unsigned long flags;
1304 int ret = 0;
1305
1306 IRDA_ASSERT(dev != NULL, return -1;);
1307
1308 self = netdev_priv(dev);
1309
1310 IRDA_ASSERT(self != NULL, return -1;);
1311
1312 IRDA_DEBUG(2, "%s(), %s, (cmd=0x%X)\n", __func__ , dev->name, cmd);
1313
1314 spin_lock_irqsave(&self->lock, flags);
1315
1316 switch (cmd) {
1317 case SIOCSBANDWIDTH:
1318 if (!capable(CAP_NET_ADMIN)) {
1319 ret = -EPERM;
1320 goto out;
1321 }
1322 w83977af_change_speed(self, irq->ifr_baudrate);
1323 break;
1324 case SIOCSMEDIABUSY:
1325 if (!capable(CAP_NET_ADMIN)) {
1326 ret = -EPERM;
1327 goto out;
1328 }
1329 irda_device_set_media_busy(self->netdev, TRUE);
1330 break;
1331 case SIOCGRECEIVING:
1332 irq->ifr_receiving = w83977af_is_receiving(self);
1333 break;
1334 default:
1335 ret = -EOPNOTSUPP;
1336 }
1337out:
1338 spin_unlock_irqrestore(&self->lock, flags);
1339 return ret;
1340}
1341
1342MODULE_AUTHOR("Dag Brattli <dagb@cs.uit.no>");
1343MODULE_DESCRIPTION("Winbond W83977AF IrDA Device Driver");
1344MODULE_LICENSE("GPL");
1345
1346
1347module_param(qos_mtt_bits, int, 0);
1348MODULE_PARM_DESC(qos_mtt_bits, "Mimimum Turn Time");
1349module_param_array(io, int, NULL, 0);
1350MODULE_PARM_DESC(io, "Base I/O addresses");
1351module_param_array(irq, int, NULL, 0);
1352MODULE_PARM_DESC(irq, "IRQ lines");
1353
1354
1355
1356
1357
1358
1359
1360module_init(w83977af_init);
1361
1362
1363
1364
1365
1366
1367
1368module_exit(w83977af_cleanup);
1369