1
2
3
4
5
6
7
8
9
10
11
12
13
14
15#include <linux/crc32.h>
16#include <linux/delay.h>
17#include <linux/errno.h>
18#include <linux/netdevice.h>
19#include <linux/etherdevice.h>
20#include <linux/module.h>
21#include <linux/kernel.h>
22#include <linux/types.h>
23#include <linux/fcntl.h>
24#include <linux/interrupt.h>
25#include <linux/ioport.h>
26#include <linux/in.h>
27#include <linux/route.h>
28#include <linux/string.h>
29#include <linux/skbuff.h>
30#include <asm/irq.h>
31
32#include <linux/socket.h>
33#include <linux/bitops.h>
34
35#include <asm/io.h>
36#include <asm/dma.h>
37#include <asm/pgtable.h>
38#ifdef CONFIG_HP300
39#include <asm/blinken.h>
40#endif
41
42#include "7990.h"
43
44#define WRITERAP(lp, x) out_be16(lp->base + LANCE_RAP, (x))
45#define WRITERDP(lp, x) out_be16(lp->base + LANCE_RDP, (x))
46#define READRDP(lp) in_be16(lp->base + LANCE_RDP)
47
48#if IS_ENABLED(CONFIG_HPLANCE)
49#include "hplance.h"
50
51#undef WRITERAP
52#undef WRITERDP
53#undef READRDP
54
55#if IS_ENABLED(CONFIG_MVME147_NET)
56
57
58#define WRITERAP(lp, x) (lp->writerap(lp, x))
59#define WRITERDP(lp, x) (lp->writerdp(lp, x))
60#define READRDP(lp) (lp->readrdp(lp))
61
62#else
63
64
65static inline void WRITERAP(struct lance_private *lp, __u16 value)
66{
67 do {
68 out_be16(lp->base + HPLANCE_REGOFF + LANCE_RAP, value);
69 } while ((in_8(lp->base + HPLANCE_STATUS) & LE_ACK) == 0);
70}
71
72static inline void WRITERDP(struct lance_private *lp, __u16 value)
73{
74 do {
75 out_be16(lp->base + HPLANCE_REGOFF + LANCE_RDP, value);
76 } while ((in_8(lp->base + HPLANCE_STATUS) & LE_ACK) == 0);
77}
78
79static inline __u16 READRDP(struct lance_private *lp)
80{
81 __u16 value;
82 do {
83 value = in_be16(lp->base + HPLANCE_REGOFF + LANCE_RDP);
84 } while ((in_8(lp->base + HPLANCE_STATUS) & LE_ACK) == 0);
85 return value;
86}
87
88#endif
89#endif
90
91
92
93#ifdef UNDEF
94#define PRINT_RINGS() \
95do { \
96 int t; \
97 for (t = 0; t < RX_RING_SIZE; t++) { \
98 printk("R%d: @(%02X %04X) len %04X, mblen %04X, bits %02X\n", \
99 t, ib->brx_ring[t].rmd1_hadr, ib->brx_ring[t].rmd0, \
100 ib->brx_ring[t].length, \
101 ib->brx_ring[t].mblength, ib->brx_ring[t].rmd1_bits); \
102 } \
103 for (t = 0; t < TX_RING_SIZE; t++) { \
104 printk("T%d: @(%02X %04X) len %04X, misc %04X, bits %02X\n", \
105 t, ib->btx_ring[t].tmd1_hadr, ib->btx_ring[t].tmd0, \
106 ib->btx_ring[t].length, \
107 ib->btx_ring[t].misc, ib->btx_ring[t].tmd1_bits); \
108 } \
109} while (0)
110#else
111#define PRINT_RINGS()
112#endif
113
114
115static void load_csrs(struct lance_private *lp)
116{
117 volatile struct lance_init_block *aib = lp->lance_init_block;
118 int leptr;
119
120 leptr = LANCE_ADDR(aib);
121
122 WRITERAP(lp, LE_CSR1);
123 WRITERDP(lp, leptr & 0xFFFF);
124 WRITERAP(lp, LE_CSR2);
125 WRITERDP(lp, leptr >> 16);
126 WRITERAP(lp, LE_CSR3);
127 WRITERDP(lp, lp->busmaster_regval);
128
129
130 WRITERAP(lp, LE_CSR0);
131}
132
133
134#define DEBUG_IRING 0
135
136static void lance_init_ring(struct net_device *dev)
137{
138 struct lance_private *lp = netdev_priv(dev);
139 volatile struct lance_init_block *ib = lp->init_block;
140 volatile struct lance_init_block *aib;
141 int leptr;
142 int i;
143
144 aib = lp->lance_init_block;
145
146 lp->rx_new = lp->tx_new = 0;
147 lp->rx_old = lp->tx_old = 0;
148
149 ib->mode = LE_MO_PROM;
150
151
152
153
154
155
156
157
158
159
160
161#ifdef __BIG_ENDIAN
162 ib->phys_addr[0] = dev->dev_addr[1];
163 ib->phys_addr[1] = dev->dev_addr[0];
164 ib->phys_addr[2] = dev->dev_addr[3];
165 ib->phys_addr[3] = dev->dev_addr[2];
166 ib->phys_addr[4] = dev->dev_addr[5];
167 ib->phys_addr[5] = dev->dev_addr[4];
168#else
169 for (i = 0; i < 6; i++)
170 ib->phys_addr[i] = dev->dev_addr[i];
171#endif
172
173 if (DEBUG_IRING)
174 printk("TX rings:\n");
175
176 lp->tx_full = 0;
177
178 for (i = 0; i < (1 << lp->lance_log_tx_bufs); i++) {
179 leptr = LANCE_ADDR(&aib->tx_buf[i][0]);
180 ib->btx_ring[i].tmd0 = leptr;
181 ib->btx_ring[i].tmd1_hadr = leptr >> 16;
182 ib->btx_ring[i].tmd1_bits = 0;
183 ib->btx_ring[i].length = 0xf000;
184 ib->btx_ring[i].misc = 0;
185 if (DEBUG_IRING)
186 printk("%d: 0x%8.8x\n", i, leptr);
187 }
188
189
190 if (DEBUG_IRING)
191 printk("RX rings:\n");
192 for (i = 0; i < (1 << lp->lance_log_rx_bufs); i++) {
193 leptr = LANCE_ADDR(&aib->rx_buf[i][0]);
194
195 ib->brx_ring[i].rmd0 = leptr;
196 ib->brx_ring[i].rmd1_hadr = leptr >> 16;
197 ib->brx_ring[i].rmd1_bits = LE_R1_OWN;
198
199 ib->brx_ring[i].length = -RX_BUFF_SIZE | 0xf000;
200 ib->brx_ring[i].mblength = 0;
201 if (DEBUG_IRING)
202 printk("%d: 0x%8.8x\n", i, leptr);
203 }
204
205
206
207
208 leptr = LANCE_ADDR(&aib->brx_ring);
209 ib->rx_len = (lp->lance_log_rx_bufs << 13) | (leptr >> 16);
210 ib->rx_ptr = leptr;
211 if (DEBUG_IRING)
212 printk("RX ptr: %8.8x\n", leptr);
213
214
215 leptr = LANCE_ADDR(&aib->btx_ring);
216 ib->tx_len = (lp->lance_log_tx_bufs << 13) | (leptr >> 16);
217 ib->tx_ptr = leptr;
218 if (DEBUG_IRING)
219 printk("TX ptr: %8.8x\n", leptr);
220
221
222 ib->filter[0] = 0;
223 ib->filter[1] = 0;
224 PRINT_RINGS();
225}
226
227
228static int init_restart_lance(struct lance_private *lp)
229{
230 int i;
231
232 WRITERAP(lp, LE_CSR0);
233 WRITERDP(lp, LE_C0_INIT);
234
235
236
237
238 for (i = 0; (i < 100) && !(READRDP(lp) & (LE_C0_ERR | LE_C0_IDON)); i++)
239 barrier();
240 if ((i == 100) || (READRDP(lp) & LE_C0_ERR)) {
241 printk("LANCE unopened after %d ticks, csr0=%4.4x.\n", i, READRDP(lp));
242 return -1;
243 }
244
245
246 WRITERDP(lp, LE_C0_IDON);
247 WRITERDP(lp, LE_C0_INEA | LE_C0_STRT);
248
249 return 0;
250}
251
252static int lance_reset(struct net_device *dev)
253{
254 struct lance_private *lp = netdev_priv(dev);
255 int status;
256
257
258 WRITERAP(lp, LE_CSR0);
259 WRITERDP(lp, LE_C0_STOP);
260
261 load_csrs(lp);
262 lance_init_ring(dev);
263 netif_trans_update(dev);
264 status = init_restart_lance(lp);
265#ifdef DEBUG_DRIVER
266 printk("Lance restart=%d\n", status);
267#endif
268 return status;
269}
270
271static int lance_rx(struct net_device *dev)
272{
273 struct lance_private *lp = netdev_priv(dev);
274 volatile struct lance_init_block *ib = lp->init_block;
275 volatile struct lance_rx_desc *rd;
276 unsigned char bits;
277#ifdef TEST_HITS
278 int i;
279#endif
280
281#ifdef TEST_HITS
282 printk("[");
283 for (i = 0; i < RX_RING_SIZE; i++) {
284 if (i == lp->rx_new)
285 printk("%s",
286 ib->brx_ring[i].rmd1_bits & LE_R1_OWN ? "_" : "X");
287 else
288 printk("%s",
289 ib->brx_ring[i].rmd1_bits & LE_R1_OWN ? "." : "1");
290 }
291 printk("]");
292#endif
293#ifdef CONFIG_HP300
294 blinken_leds(0x40, 0);
295#endif
296 WRITERDP(lp, LE_C0_RINT | LE_C0_INEA);
297 for (rd = &ib->brx_ring[lp->rx_new];
298 !((bits = rd->rmd1_bits) & LE_R1_OWN);
299 rd = &ib->brx_ring[lp->rx_new]) {
300
301
302 if ((bits & LE_R1_POK) != LE_R1_POK) {
303 dev->stats.rx_over_errors++;
304 dev->stats.rx_errors++;
305 continue;
306 } else if (bits & LE_R1_ERR) {
307
308
309
310 if (bits & LE_R1_BUF)
311 dev->stats.rx_fifo_errors++;
312 if (bits & LE_R1_CRC)
313 dev->stats.rx_crc_errors++;
314 if (bits & LE_R1_OFL)
315 dev->stats.rx_over_errors++;
316 if (bits & LE_R1_FRA)
317 dev->stats.rx_frame_errors++;
318 if (bits & LE_R1_EOP)
319 dev->stats.rx_errors++;
320 } else {
321 int len = (rd->mblength & 0xfff) - 4;
322 struct sk_buff *skb = netdev_alloc_skb(dev, len + 2);
323
324 if (!skb) {
325 dev->stats.rx_dropped++;
326 rd->mblength = 0;
327 rd->rmd1_bits = LE_R1_OWN;
328 lp->rx_new = (lp->rx_new + 1) & lp->rx_ring_mod_mask;
329 return 0;
330 }
331
332 skb_reserve(skb, 2);
333 skb_put(skb, len);
334 skb_copy_to_linear_data(skb,
335 (unsigned char *)&(ib->rx_buf[lp->rx_new][0]),
336 len);
337 skb->protocol = eth_type_trans(skb, dev);
338 netif_rx(skb);
339 dev->stats.rx_packets++;
340 dev->stats.rx_bytes += len;
341 }
342
343
344 rd->mblength = 0;
345 rd->rmd1_bits = LE_R1_OWN;
346 lp->rx_new = (lp->rx_new + 1) & lp->rx_ring_mod_mask;
347 }
348 return 0;
349}
350
351static int lance_tx(struct net_device *dev)
352{
353 struct lance_private *lp = netdev_priv(dev);
354 volatile struct lance_init_block *ib = lp->init_block;
355 volatile struct lance_tx_desc *td;
356 int i, j;
357 int status;
358
359#ifdef CONFIG_HP300
360 blinken_leds(0x80, 0);
361#endif
362
363 WRITERDP(lp, LE_C0_TINT | LE_C0_INEA);
364
365
366 j = lp->tx_old;
367 for (i = j; i != lp->tx_new; i = j) {
368 td = &ib->btx_ring[i];
369
370
371 if (td->tmd1_bits & LE_T1_OWN)
372 break;
373
374 if (td->tmd1_bits & LE_T1_ERR) {
375 status = td->misc;
376
377 dev->stats.tx_errors++;
378 if (status & LE_T3_RTY)
379 dev->stats.tx_aborted_errors++;
380 if (status & LE_T3_LCOL)
381 dev->stats.tx_window_errors++;
382
383 if (status & LE_T3_CLOS) {
384 dev->stats.tx_carrier_errors++;
385 if (lp->auto_select) {
386 lp->tpe = 1 - lp->tpe;
387 printk("%s: Carrier Lost, trying %s\n",
388 dev->name,
389 lp->tpe ? "TPE" : "AUI");
390
391 WRITERAP(lp, LE_CSR0);
392 WRITERDP(lp, LE_C0_STOP);
393 lance_init_ring(dev);
394 load_csrs(lp);
395 init_restart_lance(lp);
396 return 0;
397 }
398 }
399
400
401
402 if (status & (LE_T3_BUF|LE_T3_UFL)) {
403 dev->stats.tx_fifo_errors++;
404
405 printk("%s: Tx: ERR_BUF|ERR_UFL, restarting\n",
406 dev->name);
407
408 WRITERAP(lp, LE_CSR0);
409 WRITERDP(lp, LE_C0_STOP);
410 lance_init_ring(dev);
411 load_csrs(lp);
412 init_restart_lance(lp);
413 return 0;
414 }
415 } else if ((td->tmd1_bits & LE_T1_POK) == LE_T1_POK) {
416
417
418
419 td->tmd1_bits &= ~(LE_T1_POK);
420
421
422 if (td->tmd1_bits & LE_T1_EONE)
423 dev->stats.collisions++;
424
425
426 if (td->tmd1_bits & LE_T1_EMORE)
427 dev->stats.collisions += 2;
428
429 dev->stats.tx_packets++;
430 }
431
432 j = (j + 1) & lp->tx_ring_mod_mask;
433 }
434 lp->tx_old = j;
435 WRITERDP(lp, LE_C0_TINT | LE_C0_INEA);
436 return 0;
437}
438
439static irqreturn_t
440lance_interrupt(int irq, void *dev_id)
441{
442 struct net_device *dev = (struct net_device *)dev_id;
443 struct lance_private *lp = netdev_priv(dev);
444 int csr0;
445
446 spin_lock(&lp->devlock);
447
448 WRITERAP(lp, LE_CSR0);
449 csr0 = READRDP(lp);
450
451 PRINT_RINGS();
452
453 if (!(csr0 & LE_C0_INTR)) {
454 spin_unlock(&lp->devlock);
455 return IRQ_NONE;
456 }
457
458
459 WRITERDP(lp, csr0 & ~(LE_C0_INEA|LE_C0_TDMD|LE_C0_STOP|LE_C0_STRT|LE_C0_INIT));
460
461 if ((csr0 & LE_C0_ERR)) {
462
463 WRITERDP(lp, LE_C0_BABL|LE_C0_ERR|LE_C0_MISS|LE_C0_INEA);
464 }
465
466 if (csr0 & LE_C0_RINT)
467 lance_rx(dev);
468
469 if (csr0 & LE_C0_TINT)
470 lance_tx(dev);
471
472
473 if (csr0 & LE_C0_BABL)
474 dev->stats.tx_errors++;
475 if (csr0 & LE_C0_MISS)
476 dev->stats.rx_errors++;
477 if (csr0 & LE_C0_MERR) {
478 printk("%s: Bus master arbitration failure, status %4.4x.\n",
479 dev->name, csr0);
480
481 WRITERDP(lp, LE_C0_STRT);
482 }
483
484 if (lp->tx_full && netif_queue_stopped(dev) && (TX_BUFFS_AVAIL >= 0)) {
485 lp->tx_full = 0;
486 netif_wake_queue(dev);
487 }
488
489 WRITERAP(lp, LE_CSR0);
490 WRITERDP(lp, LE_C0_BABL|LE_C0_CERR|LE_C0_MISS|LE_C0_MERR|LE_C0_IDON|LE_C0_INEA);
491
492 spin_unlock(&lp->devlock);
493 return IRQ_HANDLED;
494}
495
496int lance_open(struct net_device *dev)
497{
498 struct lance_private *lp = netdev_priv(dev);
499 int res;
500
501
502 if (request_irq(lp->irq, lance_interrupt, IRQF_SHARED, lp->name, dev))
503 return -EAGAIN;
504
505 res = lance_reset(dev);
506 spin_lock_init(&lp->devlock);
507 netif_start_queue(dev);
508
509 return res;
510}
511EXPORT_SYMBOL_GPL(lance_open);
512
513int lance_close(struct net_device *dev)
514{
515 struct lance_private *lp = netdev_priv(dev);
516
517 netif_stop_queue(dev);
518
519
520 WRITERAP(lp, LE_CSR0);
521 WRITERDP(lp, LE_C0_STOP);
522
523 free_irq(lp->irq, dev);
524
525 return 0;
526}
527EXPORT_SYMBOL_GPL(lance_close);
528
529void lance_tx_timeout(struct net_device *dev)
530{
531 printk("lance_tx_timeout\n");
532 lance_reset(dev);
533 netif_trans_update(dev);
534 netif_wake_queue(dev);
535}
536EXPORT_SYMBOL_GPL(lance_tx_timeout);
537
538int lance_start_xmit(struct sk_buff *skb, struct net_device *dev)
539{
540 struct lance_private *lp = netdev_priv(dev);
541 volatile struct lance_init_block *ib = lp->init_block;
542 int entry, skblen, len;
543 static int outs;
544 unsigned long flags;
545
546 netif_stop_queue(dev);
547
548 if (!TX_BUFFS_AVAIL) {
549 dev_consume_skb_any(skb);
550 return NETDEV_TX_OK;
551 }
552
553 skblen = skb->len;
554
555#ifdef DEBUG_DRIVER
556
557 {
558 int i;
559
560 for (i = 0; i < 64; i++) {
561 if ((i % 16) == 0)
562 printk("\n");
563 printk("%2.2x ", skb->data[i]);
564 }
565 }
566#endif
567 len = (skblen <= ETH_ZLEN) ? ETH_ZLEN : skblen;
568 entry = lp->tx_new & lp->tx_ring_mod_mask;
569 ib->btx_ring[entry].length = (-len) | 0xf000;
570 ib->btx_ring[entry].misc = 0;
571
572 if (skb->len < ETH_ZLEN)
573 memset((void *)&ib->tx_buf[entry][0], 0, ETH_ZLEN);
574 skb_copy_from_linear_data(skb, (void *)&ib->tx_buf[entry][0], skblen);
575
576
577 ib->btx_ring[entry].tmd1_bits = (LE_T1_POK|LE_T1_OWN);
578 lp->tx_new = (lp->tx_new + 1) & lp->tx_ring_mod_mask;
579
580 outs++;
581
582 WRITERDP(lp, LE_C0_INEA | LE_C0_TDMD);
583 dev_consume_skb_any(skb);
584
585 spin_lock_irqsave(&lp->devlock, flags);
586 if (TX_BUFFS_AVAIL)
587 netif_start_queue(dev);
588 else
589 lp->tx_full = 1;
590 spin_unlock_irqrestore(&lp->devlock, flags);
591
592 return NETDEV_TX_OK;
593}
594EXPORT_SYMBOL_GPL(lance_start_xmit);
595
596
597static void lance_load_multicast(struct net_device *dev)
598{
599 struct lance_private *lp = netdev_priv(dev);
600 volatile struct lance_init_block *ib = lp->init_block;
601 volatile u16 *mcast_table = (u16 *)&ib->filter;
602 struct netdev_hw_addr *ha;
603 u32 crc;
604
605
606 if (dev->flags & IFF_ALLMULTI) {
607 ib->filter[0] = 0xffffffff;
608 ib->filter[1] = 0xffffffff;
609 return;
610 }
611
612 ib->filter[0] = 0;
613 ib->filter[1] = 0;
614
615
616 netdev_for_each_mc_addr(ha, dev) {
617 crc = ether_crc_le(6, ha->addr);
618 crc = crc >> 26;
619 mcast_table[crc >> 4] |= 1 << (crc & 0xf);
620 }
621}
622
623
624void lance_set_multicast(struct net_device *dev)
625{
626 struct lance_private *lp = netdev_priv(dev);
627 volatile struct lance_init_block *ib = lp->init_block;
628 int stopped;
629
630 stopped = netif_queue_stopped(dev);
631 if (!stopped)
632 netif_stop_queue(dev);
633
634 while (lp->tx_old != lp->tx_new)
635 schedule();
636
637 WRITERAP(lp, LE_CSR0);
638 WRITERDP(lp, LE_C0_STOP);
639 lance_init_ring(dev);
640
641 if (dev->flags & IFF_PROMISC) {
642 ib->mode |= LE_MO_PROM;
643 } else {
644 ib->mode &= ~LE_MO_PROM;
645 lance_load_multicast(dev);
646 }
647 load_csrs(lp);
648 init_restart_lance(lp);
649
650 if (!stopped)
651 netif_start_queue(dev);
652}
653EXPORT_SYMBOL_GPL(lance_set_multicast);
654
655#ifdef CONFIG_NET_POLL_CONTROLLER
656void lance_poll(struct net_device *dev)
657{
658 struct lance_private *lp = netdev_priv(dev);
659
660 spin_lock(&lp->devlock);
661 WRITERAP(lp, LE_CSR0);
662 WRITERDP(lp, LE_C0_STRT);
663 spin_unlock(&lp->devlock);
664 lance_interrupt(dev->irq, dev);
665}
666EXPORT_SYMBOL_GPL(lance_poll);
667#endif
668
669MODULE_LICENSE("GPL");
670