1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22#include "common.h"
23#include "net.h"
24#include "malloc.h"
25#include "asm/errno.h"
26#include "asm/io.h"
27#include "asm/immap_qe.h"
28#include "qe.h"
29#include "uccf.h"
30#include "uec.h"
31#include "uec_phy.h"
32#include "miiphy.h"
33#include <phy.h>
34
35
36#ifndef CONFIG_UTBIPAR_INIT_TBIPA
37#define CONFIG_UTBIPAR_INIT_TBIPA 0x1F
38#endif
39
40static uec_info_t uec_info[] = {
41#ifdef CONFIG_UEC_ETH1
42 STD_UEC_INFO(1),
43#endif
44#ifdef CONFIG_UEC_ETH2
45 STD_UEC_INFO(2),
46#endif
47#ifdef CONFIG_UEC_ETH3
48 STD_UEC_INFO(3),
49#endif
50#ifdef CONFIG_UEC_ETH4
51 STD_UEC_INFO(4),
52#endif
53#ifdef CONFIG_UEC_ETH5
54 STD_UEC_INFO(5),
55#endif
56#ifdef CONFIG_UEC_ETH6
57 STD_UEC_INFO(6),
58#endif
59#ifdef CONFIG_UEC_ETH7
60 STD_UEC_INFO(7),
61#endif
62#ifdef CONFIG_UEC_ETH8
63 STD_UEC_INFO(8),
64#endif
65};
66
67#define MAXCONTROLLERS (8)
68
69static struct eth_device *devlist[MAXCONTROLLERS];
70
71static int uec_mac_enable(uec_private_t *uec, comm_dir_e mode)
72{
73 uec_t *uec_regs;
74 u32 maccfg1;
75
76 if (!uec) {
77 printf("%s: uec not initial\n", __FUNCTION__);
78 return -EINVAL;
79 }
80 uec_regs = uec->uec_regs;
81
82 maccfg1 = in_be32(&uec_regs->maccfg1);
83
84 if (mode & COMM_DIR_TX) {
85 maccfg1 |= MACCFG1_ENABLE_TX;
86 out_be32(&uec_regs->maccfg1, maccfg1);
87 uec->mac_tx_enabled = 1;
88 }
89
90 if (mode & COMM_DIR_RX) {
91 maccfg1 |= MACCFG1_ENABLE_RX;
92 out_be32(&uec_regs->maccfg1, maccfg1);
93 uec->mac_rx_enabled = 1;
94 }
95
96 return 0;
97}
98
99static int uec_mac_disable(uec_private_t *uec, comm_dir_e mode)
100{
101 uec_t *uec_regs;
102 u32 maccfg1;
103
104 if (!uec) {
105 printf("%s: uec not initial\n", __FUNCTION__);
106 return -EINVAL;
107 }
108 uec_regs = uec->uec_regs;
109
110 maccfg1 = in_be32(&uec_regs->maccfg1);
111
112 if (mode & COMM_DIR_TX) {
113 maccfg1 &= ~MACCFG1_ENABLE_TX;
114 out_be32(&uec_regs->maccfg1, maccfg1);
115 uec->mac_tx_enabled = 0;
116 }
117
118 if (mode & COMM_DIR_RX) {
119 maccfg1 &= ~MACCFG1_ENABLE_RX;
120 out_be32(&uec_regs->maccfg1, maccfg1);
121 uec->mac_rx_enabled = 0;
122 }
123
124 return 0;
125}
126
127static int uec_graceful_stop_tx(uec_private_t *uec)
128{
129 ucc_fast_t *uf_regs;
130 u32 cecr_subblock;
131 u32 ucce;
132
133 if (!uec || !uec->uccf) {
134 printf("%s: No handle passed.\n", __FUNCTION__);
135 return -EINVAL;
136 }
137
138 uf_regs = uec->uccf->uf_regs;
139
140
141 out_be32(&uf_regs->ucce, UCCE_GRA);
142
143
144 cecr_subblock =
145 ucc_fast_get_qe_cr_subblock(uec->uec_info->uf_info.ucc_num);
146 qe_issue_cmd(QE_GRACEFUL_STOP_TX, cecr_subblock,
147 (u8)QE_CR_PROTOCOL_ETHERNET, 0);
148
149
150 do {
151 ucce = in_be32(&uf_regs->ucce);
152 } while (! (ucce & UCCE_GRA));
153
154 uec->grace_stopped_tx = 1;
155
156 return 0;
157}
158
159static int uec_graceful_stop_rx(uec_private_t *uec)
160{
161 u32 cecr_subblock;
162 u8 ack;
163
164 if (!uec) {
165 printf("%s: No handle passed.\n", __FUNCTION__);
166 return -EINVAL;
167 }
168
169 if (!uec->p_rx_glbl_pram) {
170 printf("%s: No init rx global parameter\n", __FUNCTION__);
171 return -EINVAL;
172 }
173
174
175 ack = uec->p_rx_glbl_pram->rxgstpack;
176 ack &= ~GRACEFUL_STOP_ACKNOWLEDGE_RX;
177 uec->p_rx_glbl_pram->rxgstpack = ack;
178
179
180 do {
181
182 cecr_subblock =
183 ucc_fast_get_qe_cr_subblock(uec->uec_info->uf_info.ucc_num);
184 qe_issue_cmd(QE_GRACEFUL_STOP_RX, cecr_subblock,
185 (u8)QE_CR_PROTOCOL_ETHERNET, 0);
186 ack = uec->p_rx_glbl_pram->rxgstpack;
187 } while (! (ack & GRACEFUL_STOP_ACKNOWLEDGE_RX ));
188
189 uec->grace_stopped_rx = 1;
190
191 return 0;
192}
193
194static int uec_restart_tx(uec_private_t *uec)
195{
196 u32 cecr_subblock;
197
198 if (!uec || !uec->uec_info) {
199 printf("%s: No handle passed.\n", __FUNCTION__);
200 return -EINVAL;
201 }
202
203 cecr_subblock =
204 ucc_fast_get_qe_cr_subblock(uec->uec_info->uf_info.ucc_num);
205 qe_issue_cmd(QE_RESTART_TX, cecr_subblock,
206 (u8)QE_CR_PROTOCOL_ETHERNET, 0);
207
208 uec->grace_stopped_tx = 0;
209
210 return 0;
211}
212
213static int uec_restart_rx(uec_private_t *uec)
214{
215 u32 cecr_subblock;
216
217 if (!uec || !uec->uec_info) {
218 printf("%s: No handle passed.\n", __FUNCTION__);
219 return -EINVAL;
220 }
221
222 cecr_subblock =
223 ucc_fast_get_qe_cr_subblock(uec->uec_info->uf_info.ucc_num);
224 qe_issue_cmd(QE_RESTART_RX, cecr_subblock,
225 (u8)QE_CR_PROTOCOL_ETHERNET, 0);
226
227 uec->grace_stopped_rx = 0;
228
229 return 0;
230}
231
232static int uec_open(uec_private_t *uec, comm_dir_e mode)
233{
234 ucc_fast_private_t *uccf;
235
236 if (!uec || !uec->uccf) {
237 printf("%s: No handle passed.\n", __FUNCTION__);
238 return -EINVAL;
239 }
240 uccf = uec->uccf;
241
242
243 if (uec->uec_info->uf_info.ucc_num >= UCC_MAX_NUM) {
244 printf("%s: ucc_num out of range.\n", __FUNCTION__);
245 return -EINVAL;
246 }
247
248
249 uec_mac_enable(uec, mode);
250
251
252 ucc_fast_enable(uccf, mode);
253
254
255 if ((mode & COMM_DIR_TX) && uec->grace_stopped_tx) {
256 uec_restart_tx(uec);
257 }
258 if ((mode & COMM_DIR_RX) && uec->grace_stopped_rx) {
259 uec_restart_rx(uec);
260 }
261
262 return 0;
263}
264
265static int uec_stop(uec_private_t *uec, comm_dir_e mode)
266{
267 if (!uec || !uec->uccf) {
268 printf("%s: No handle passed.\n", __FUNCTION__);
269 return -EINVAL;
270 }
271
272
273 if (uec->uec_info->uf_info.ucc_num >= UCC_MAX_NUM) {
274 printf("%s: ucc_num out of range.\n", __FUNCTION__);
275 return -EINVAL;
276 }
277
278 if ((mode & COMM_DIR_TX) && !uec->grace_stopped_tx) {
279 uec_graceful_stop_tx(uec);
280 }
281
282 if ((mode & COMM_DIR_RX) && !uec->grace_stopped_rx) {
283 uec_graceful_stop_rx(uec);
284 }
285
286
287 ucc_fast_disable(uec->uccf, mode);
288
289
290 uec_mac_disable(uec, mode);
291
292 return 0;
293}
294
295static int uec_set_mac_duplex(uec_private_t *uec, int duplex)
296{
297 uec_t *uec_regs;
298 u32 maccfg2;
299
300 if (!uec) {
301 printf("%s: uec not initial\n", __FUNCTION__);
302 return -EINVAL;
303 }
304 uec_regs = uec->uec_regs;
305
306 if (duplex == DUPLEX_HALF) {
307 maccfg2 = in_be32(&uec_regs->maccfg2);
308 maccfg2 &= ~MACCFG2_FDX;
309 out_be32(&uec_regs->maccfg2, maccfg2);
310 }
311
312 if (duplex == DUPLEX_FULL) {
313 maccfg2 = in_be32(&uec_regs->maccfg2);
314 maccfg2 |= MACCFG2_FDX;
315 out_be32(&uec_regs->maccfg2, maccfg2);
316 }
317
318 return 0;
319}
320
321static int uec_set_mac_if_mode(uec_private_t *uec,
322 phy_interface_t if_mode, int speed)
323{
324 phy_interface_t enet_if_mode;
325 uec_t *uec_regs;
326 u32 upsmr;
327 u32 maccfg2;
328
329 if (!uec) {
330 printf("%s: uec not initial\n", __FUNCTION__);
331 return -EINVAL;
332 }
333
334 uec_regs = uec->uec_regs;
335 enet_if_mode = if_mode;
336
337 maccfg2 = in_be32(&uec_regs->maccfg2);
338 maccfg2 &= ~MACCFG2_INTERFACE_MODE_MASK;
339
340 upsmr = in_be32(&uec->uccf->uf_regs->upsmr);
341 upsmr &= ~(UPSMR_RPM | UPSMR_TBIM | UPSMR_R10M | UPSMR_RMM);
342
343 switch (speed) {
344 case SPEED_10:
345 maccfg2 |= MACCFG2_INTERFACE_MODE_NIBBLE;
346 switch (enet_if_mode) {
347 case PHY_INTERFACE_MODE_MII:
348 break;
349 case PHY_INTERFACE_MODE_RGMII:
350 upsmr |= (UPSMR_RPM | UPSMR_R10M);
351 break;
352 case PHY_INTERFACE_MODE_RMII:
353 upsmr |= (UPSMR_R10M | UPSMR_RMM);
354 break;
355 default:
356 return -EINVAL;
357 break;
358 }
359 break;
360 case SPEED_100:
361 maccfg2 |= MACCFG2_INTERFACE_MODE_NIBBLE;
362 switch (enet_if_mode) {
363 case PHY_INTERFACE_MODE_MII:
364 break;
365 case PHY_INTERFACE_MODE_RGMII:
366 upsmr |= UPSMR_RPM;
367 break;
368 case PHY_INTERFACE_MODE_RMII:
369 upsmr |= UPSMR_RMM;
370 break;
371 default:
372 return -EINVAL;
373 break;
374 }
375 break;
376 case SPEED_1000:
377 maccfg2 |= MACCFG2_INTERFACE_MODE_BYTE;
378 switch (enet_if_mode) {
379 case PHY_INTERFACE_MODE_GMII:
380 break;
381 case PHY_INTERFACE_MODE_TBI:
382 upsmr |= UPSMR_TBIM;
383 break;
384 case PHY_INTERFACE_MODE_RTBI:
385 upsmr |= (UPSMR_RPM | UPSMR_TBIM);
386 break;
387 case PHY_INTERFACE_MODE_RGMII_RXID:
388 case PHY_INTERFACE_MODE_RGMII_TXID:
389 case PHY_INTERFACE_MODE_RGMII_ID:
390 case PHY_INTERFACE_MODE_RGMII:
391 upsmr |= UPSMR_RPM;
392 break;
393 case PHY_INTERFACE_MODE_SGMII:
394 upsmr |= UPSMR_SGMM;
395 break;
396 default:
397 return -EINVAL;
398 break;
399 }
400 break;
401 default:
402 return -EINVAL;
403 break;
404 }
405
406 out_be32(&uec_regs->maccfg2, maccfg2);
407 out_be32(&uec->uccf->uf_regs->upsmr, upsmr);
408
409 return 0;
410}
411
412static int init_mii_management_configuration(uec_mii_t *uec_mii_regs)
413{
414 uint timeout = 0x1000;
415 u32 miimcfg = 0;
416
417 miimcfg = in_be32(&uec_mii_regs->miimcfg);
418 miimcfg |= MIIMCFG_MNGMNT_CLC_DIV_INIT_VALUE;
419 out_be32(&uec_mii_regs->miimcfg, miimcfg);
420
421
422 while ((in_be32(&uec_mii_regs->miimcfg) & MIIMIND_BUSY) && timeout--);
423 if (timeout <= 0) {
424 printf("%s: The MII Bus is stuck!", __FUNCTION__);
425 return -ETIMEDOUT;
426 }
427
428 return 0;
429}
430
431static int init_phy(struct eth_device *dev)
432{
433 uec_private_t *uec;
434 uec_mii_t *umii_regs;
435 struct uec_mii_info *mii_info;
436 struct phy_info *curphy;
437 int err;
438
439 uec = (uec_private_t *)dev->priv;
440 umii_regs = uec->uec_mii_regs;
441
442 uec->oldlink = 0;
443 uec->oldspeed = 0;
444 uec->oldduplex = -1;
445
446 mii_info = malloc(sizeof(*mii_info));
447 if (!mii_info) {
448 printf("%s: Could not allocate mii_info", dev->name);
449 return -ENOMEM;
450 }
451 memset(mii_info, 0, sizeof(*mii_info));
452
453 if (uec->uec_info->uf_info.eth_type == GIGA_ETH) {
454 mii_info->speed = SPEED_1000;
455 } else {
456 mii_info->speed = SPEED_100;
457 }
458
459 mii_info->duplex = DUPLEX_FULL;
460 mii_info->pause = 0;
461 mii_info->link = 1;
462
463 mii_info->advertising = (ADVERTISED_10baseT_Half |
464 ADVERTISED_10baseT_Full |
465 ADVERTISED_100baseT_Half |
466 ADVERTISED_100baseT_Full |
467 ADVERTISED_1000baseT_Full);
468 mii_info->autoneg = 1;
469 mii_info->mii_id = uec->uec_info->phy_address;
470 mii_info->dev = dev;
471
472 mii_info->mdio_read = &uec_read_phy_reg;
473 mii_info->mdio_write = &uec_write_phy_reg;
474
475 uec->mii_info = mii_info;
476
477 qe_set_mii_clk_src(uec->uec_info->uf_info.ucc_num);
478
479 if (init_mii_management_configuration(umii_regs)) {
480 printf("%s: The MII Bus is stuck!", dev->name);
481 err = -1;
482 goto bus_fail;
483 }
484
485
486 curphy = uec_get_phy_info(uec->mii_info);
487 if (!curphy) {
488 printf("%s: No PHY found", dev->name);
489 err = -1;
490 goto no_phy;
491 }
492
493 mii_info->phyinfo = curphy;
494
495
496 if (curphy->init) {
497 err = curphy->init(uec->mii_info);
498 if (err)
499 goto phy_init_fail;
500 }
501
502 return 0;
503
504phy_init_fail:
505no_phy:
506bus_fail:
507 free(mii_info);
508 return err;
509}
510
511static void adjust_link(struct eth_device *dev)
512{
513 uec_private_t *uec = (uec_private_t *)dev->priv;
514 struct uec_mii_info *mii_info = uec->mii_info;
515
516 extern void change_phy_interface_mode(struct eth_device *dev,
517 phy_interface_t mode, int speed);
518
519 if (mii_info->link) {
520
521
522 if (mii_info->duplex != uec->oldduplex) {
523 if (!(mii_info->duplex)) {
524 uec_set_mac_duplex(uec, DUPLEX_HALF);
525 printf("%s: Half Duplex\n", dev->name);
526 } else {
527 uec_set_mac_duplex(uec, DUPLEX_FULL);
528 printf("%s: Full Duplex\n", dev->name);
529 }
530 uec->oldduplex = mii_info->duplex;
531 }
532
533 if (mii_info->speed != uec->oldspeed) {
534 phy_interface_t mode =
535 uec->uec_info->enet_interface_type;
536 if (uec->uec_info->uf_info.eth_type == GIGA_ETH) {
537 switch (mii_info->speed) {
538 case SPEED_1000:
539 break;
540 case SPEED_100:
541 printf ("switching to rgmii 100\n");
542 mode = PHY_INTERFACE_MODE_RGMII;
543 break;
544 case SPEED_10:
545 printf ("switching to rgmii 10\n");
546 mode = PHY_INTERFACE_MODE_RGMII;
547 break;
548 default:
549 printf("%s: Ack,Speed(%d)is illegal\n",
550 dev->name, mii_info->speed);
551 break;
552 }
553 }
554
555
556 change_phy_interface_mode(dev, mode, mii_info->speed);
557
558 uec_set_mac_if_mode(uec, mode, mii_info->speed);
559
560 printf("%s: Speed %dBT\n", dev->name, mii_info->speed);
561 uec->oldspeed = mii_info->speed;
562 }
563
564 if (!uec->oldlink) {
565 printf("%s: Link is up\n", dev->name);
566 uec->oldlink = 1;
567 }
568
569 } else {
570 if (uec->oldlink) {
571 printf("%s: Link is down\n", dev->name);
572 uec->oldlink = 0;
573 uec->oldspeed = 0;
574 uec->oldduplex = -1;
575 }
576 }
577}
578
579static void phy_change(struct eth_device *dev)
580{
581 uec_private_t *uec = (uec_private_t *)dev->priv;
582
583#if defined(CONFIG_P1012) || defined(CONFIG_P1021) || defined(CONFIG_P1025)
584 ccsr_gur_t *gur = (void *)(CONFIG_SYS_MPC85xx_GUTS_ADDR);
585
586
587 setbits_be32(&gur->pmuxcr, MPC85xx_PMUXCR_QE9);
588 setbits_be32(&gur->pmuxcr, MPC85xx_PMUXCR_QE12);
589#endif
590
591
592 uec->mii_info->phyinfo->read_status(uec->mii_info);
593
594#if defined(CONFIG_P1012) || defined(CONFIG_P1021) || defined(CONFIG_P1025)
595
596
597
598
599 clrbits_be32(&gur->pmuxcr, MPC85xx_PMUXCR_QE12);
600#endif
601
602
603 adjust_link(dev);
604}
605
606#if defined(CONFIG_MII) || defined(CONFIG_CMD_MII)
607
608
609
610
611
612
613
614static int uec_miiphy_find_dev_by_name(const char *devname)
615{
616 int i;
617
618 for (i = 0; i < MAXCONTROLLERS; i++) {
619 if (strncmp(devname, devlist[i]->name, strlen(devname)) == 0) {
620 break;
621 }
622 }
623
624
625 if (i == MAXCONTROLLERS) {
626 debug ("%s: device %s not found in devlist\n", __FUNCTION__, devname);
627 i = -1;
628 }
629
630 return i;
631}
632
633
634
635
636
637
638
639static int uec_miiphy_read(const char *devname, unsigned char addr,
640 unsigned char reg, unsigned short *value)
641{
642 int devindex = 0;
643
644 if (devname == NULL || value == NULL) {
645 debug("%s: NULL pointer given\n", __FUNCTION__);
646 } else {
647 devindex = uec_miiphy_find_dev_by_name(devname);
648 if (devindex >= 0) {
649 *value = uec_read_phy_reg(devlist[devindex], addr, reg);
650 }
651 }
652 return 0;
653}
654
655
656
657
658
659
660
661static int uec_miiphy_write(const char *devname, unsigned char addr,
662 unsigned char reg, unsigned short value)
663{
664 int devindex = 0;
665
666 if (devname == NULL) {
667 debug("%s: NULL pointer given\n", __FUNCTION__);
668 } else {
669 devindex = uec_miiphy_find_dev_by_name(devname);
670 if (devindex >= 0) {
671 uec_write_phy_reg(devlist[devindex], addr, reg, value);
672 }
673 }
674 return 0;
675}
676#endif
677
678static int uec_set_mac_address(uec_private_t *uec, u8 *mac_addr)
679{
680 uec_t *uec_regs;
681 u32 mac_addr1;
682 u32 mac_addr2;
683
684 if (!uec) {
685 printf("%s: uec not initial\n", __FUNCTION__);
686 return -EINVAL;
687 }
688
689 uec_regs = uec->uec_regs;
690
691
692
693
694
695 mac_addr1 = (mac_addr[5] << 24) | (mac_addr[4] << 16) | \
696 (mac_addr[3] << 8) | (mac_addr[2]);
697 out_be32(&uec_regs->macstnaddr1, mac_addr1);
698
699 mac_addr2 = ((mac_addr[1] << 24) | (mac_addr[0] << 16)) & 0xffff0000;
700 out_be32(&uec_regs->macstnaddr2, mac_addr2);
701
702 return 0;
703}
704
705static int uec_convert_threads_num(uec_num_of_threads_e threads_num,
706 int *threads_num_ret)
707{
708 int num_threads_numerica;
709
710 switch (threads_num) {
711 case UEC_NUM_OF_THREADS_1:
712 num_threads_numerica = 1;
713 break;
714 case UEC_NUM_OF_THREADS_2:
715 num_threads_numerica = 2;
716 break;
717 case UEC_NUM_OF_THREADS_4:
718 num_threads_numerica = 4;
719 break;
720 case UEC_NUM_OF_THREADS_6:
721 num_threads_numerica = 6;
722 break;
723 case UEC_NUM_OF_THREADS_8:
724 num_threads_numerica = 8;
725 break;
726 default:
727 printf("%s: Bad number of threads value.",
728 __FUNCTION__);
729 return -EINVAL;
730 }
731
732 *threads_num_ret = num_threads_numerica;
733
734 return 0;
735}
736
737static void uec_init_tx_parameter(uec_private_t *uec, int num_threads_tx)
738{
739 uec_info_t *uec_info;
740 u32 end_bd;
741 u8 bmrx = 0;
742 int i;
743
744 uec_info = uec->uec_info;
745
746
747 uec->tx_glbl_pram_offset = qe_muram_alloc(
748 sizeof(uec_tx_global_pram_t),
749 UEC_TX_GLOBAL_PRAM_ALIGNMENT);
750 uec->p_tx_glbl_pram = (uec_tx_global_pram_t *)
751 qe_muram_addr(uec->tx_glbl_pram_offset);
752
753
754 memset(uec->p_tx_glbl_pram, 0, sizeof(uec_tx_global_pram_t));
755
756
757
758
759 out_be16(&uec->p_tx_glbl_pram->temoder, TEMODER_INIT_VALUE);
760
761
762 uec->send_q_mem_reg_offset = qe_muram_alloc(
763 sizeof(uec_send_queue_qd_t),
764 UEC_SEND_QUEUE_QUEUE_DESCRIPTOR_ALIGNMENT);
765 uec->p_send_q_mem_reg = (uec_send_queue_mem_region_t *)
766 qe_muram_addr(uec->send_q_mem_reg_offset);
767 out_be32(&uec->p_tx_glbl_pram->sqptr, uec->send_q_mem_reg_offset);
768
769
770 end_bd = (u32)uec->p_tx_bd_ring + (uec_info->tx_bd_ring_len - 1)
771 * SIZEOFBD;
772 out_be32(&uec->p_send_q_mem_reg->sqqd[0].bd_ring_base,
773 (u32)(uec->p_tx_bd_ring));
774 out_be32(&uec->p_send_q_mem_reg->sqqd[0].last_bd_completed_address,
775 end_bd);
776
777
778 out_be32(&uec->p_tx_glbl_pram->schedulerbasepointer, 0);
779
780
781 out_be32(&uec->p_tx_glbl_pram->txrmonbaseptr, 0);
782
783
784 bmrx = BMR_INIT_VALUE;
785 out_be32(&uec->p_tx_glbl_pram->tstate, ((u32)(bmrx) << BMR_SHIFT));
786
787
788 for (i = 0; i < MAX_IPH_OFFSET_ENTRY; i++) {
789 out_8(&uec->p_tx_glbl_pram->iphoffset[i], 0);
790 }
791
792
793 for (i = 0; i < UEC_TX_VTAG_TABLE_ENTRY_MAX; i++) {
794 out_be32(&uec->p_tx_glbl_pram->vtagtable[i], 0);
795 }
796
797
798 uec->thread_dat_tx_offset = qe_muram_alloc(
799 num_threads_tx * sizeof(uec_thread_data_tx_t) +
800 32 *(num_threads_tx == 1), UEC_THREAD_DATA_ALIGNMENT);
801
802 uec->p_thread_data_tx = (uec_thread_data_tx_t *)
803 qe_muram_addr(uec->thread_dat_tx_offset);
804 out_be32(&uec->p_tx_glbl_pram->tqptr, uec->thread_dat_tx_offset);
805}
806
807static void uec_init_rx_parameter(uec_private_t *uec, int num_threads_rx)
808{
809 u8 bmrx = 0;
810 int i;
811 uec_82xx_address_filtering_pram_t *p_af_pram;
812
813
814 uec->rx_glbl_pram_offset = qe_muram_alloc(
815 sizeof(uec_rx_global_pram_t), UEC_RX_GLOBAL_PRAM_ALIGNMENT);
816 uec->p_rx_glbl_pram = (uec_rx_global_pram_t *)
817 qe_muram_addr(uec->rx_glbl_pram_offset);
818
819
820 memset(uec->p_rx_glbl_pram, 0, sizeof(uec_rx_global_pram_t));
821
822
823
824
825
826
827
828
829 out_be32(&uec->p_rx_glbl_pram->remoder, REMODER_INIT_VALUE);
830
831
832 uec->thread_dat_rx_offset = qe_muram_alloc(
833 num_threads_rx * sizeof(uec_thread_data_rx_t),
834 UEC_THREAD_DATA_ALIGNMENT);
835 uec->p_thread_data_rx = (uec_thread_data_rx_t *)
836 qe_muram_addr(uec->thread_dat_rx_offset);
837 out_be32(&uec->p_rx_glbl_pram->rqptr, uec->thread_dat_rx_offset);
838
839
840 out_be16(&uec->p_rx_glbl_pram->typeorlen, 3072);
841
842
843 out_be32(&uec->p_rx_glbl_pram->rxrmonbaseptr, 0);
844
845
846 out_be32(&uec->p_rx_glbl_pram->intcoalescingptr, 0);
847
848
849 bmrx = BMR_INIT_VALUE;
850 out_8(&uec->p_rx_glbl_pram->rstate, bmrx);
851
852
853 out_be16(&uec->p_rx_glbl_pram->mrblr, MAX_RXBUF_LEN);
854
855
856 uec->rx_bd_qs_tbl_offset = qe_muram_alloc(
857 sizeof(uec_rx_bd_queues_entry_t) + \
858 sizeof(uec_rx_prefetched_bds_t),
859 UEC_RX_BD_QUEUES_ALIGNMENT);
860 uec->p_rx_bd_qs_tbl = (uec_rx_bd_queues_entry_t *)
861 qe_muram_addr(uec->rx_bd_qs_tbl_offset);
862
863
864 memset(uec->p_rx_bd_qs_tbl, 0, sizeof(uec_rx_bd_queues_entry_t) + \
865 sizeof(uec_rx_prefetched_bds_t));
866 out_be32(&uec->p_rx_glbl_pram->rbdqptr, uec->rx_bd_qs_tbl_offset);
867 out_be32(&uec->p_rx_bd_qs_tbl->externalbdbaseptr,
868 (u32)uec->p_rx_bd_ring);
869
870
871 out_be16(&uec->p_rx_glbl_pram->mflr, MAX_FRAME_LEN);
872
873 out_be16(&uec->p_rx_glbl_pram->minflr, MIN_FRAME_LEN);
874
875 out_be16(&uec->p_rx_glbl_pram->maxd1, MAX_DMA1_LEN);
876
877 out_be16(&uec->p_rx_glbl_pram->maxd2, MAX_DMA2_LEN);
878
879 out_be32(&uec->p_rx_glbl_pram->ecamptr, 0);
880
881 out_be32(&uec->p_rx_glbl_pram->l2qt, 0);
882
883 for (i = 0; i < 8; i++) {
884 out_be32(&uec->p_rx_glbl_pram->l3qt[i], 0);
885 }
886
887
888 out_be16(&uec->p_rx_glbl_pram->vlantype, 0x8100);
889
890 out_be16(&uec->p_rx_glbl_pram->vlantci, 0);
891
892
893 p_af_pram = (uec_82xx_address_filtering_pram_t *) \
894 uec->p_rx_glbl_pram->addressfiltering;
895
896 p_af_pram->iaddr_h = 0;
897 p_af_pram->iaddr_l = 0;
898 p_af_pram->gaddr_h = 0;
899 p_af_pram->gaddr_l = 0;
900}
901
902static int uec_issue_init_enet_rxtx_cmd(uec_private_t *uec,
903 int thread_tx, int thread_rx)
904{
905 uec_init_cmd_pram_t *p_init_enet_param;
906 u32 init_enet_param_offset;
907 uec_info_t *uec_info;
908 int i;
909 int snum;
910 u32 init_enet_offset;
911 u32 entry_val;
912 u32 command;
913 u32 cecr_subblock;
914
915 uec_info = uec->uec_info;
916
917
918 uec->init_enet_param_offset = qe_muram_alloc(
919 sizeof(uec_init_cmd_pram_t), 4);
920 init_enet_param_offset = uec->init_enet_param_offset;
921 uec->p_init_enet_param = (uec_init_cmd_pram_t *)
922 qe_muram_addr(uec->init_enet_param_offset);
923
924
925 memset((void *)uec->p_init_enet_param, 0, sizeof(uec_init_cmd_pram_t));
926
927
928 p_init_enet_param = uec->p_init_enet_param;
929 p_init_enet_param->resinit0 = ENET_INIT_PARAM_MAGIC_RES_INIT0;
930 p_init_enet_param->resinit1 = ENET_INIT_PARAM_MAGIC_RES_INIT1;
931 p_init_enet_param->resinit2 = ENET_INIT_PARAM_MAGIC_RES_INIT2;
932 p_init_enet_param->resinit3 = ENET_INIT_PARAM_MAGIC_RES_INIT3;
933 p_init_enet_param->resinit4 = ENET_INIT_PARAM_MAGIC_RES_INIT4;
934 p_init_enet_param->largestexternallookupkeysize = 0;
935
936 p_init_enet_param->rgftgfrxglobal |= ((u32)uec_info->num_threads_rx)
937 << ENET_INIT_PARAM_RGF_SHIFT;
938 p_init_enet_param->rgftgfrxglobal |= ((u32)uec_info->num_threads_tx)
939 << ENET_INIT_PARAM_TGF_SHIFT;
940
941
942 p_init_enet_param->rgftgfrxglobal |= uec->rx_glbl_pram_offset |
943 (u32)uec_info->risc_rx;
944
945
946 for (i = 0; i < (thread_rx + 1); i++) {
947 if ((snum = qe_get_snum()) < 0) {
948 printf("%s can not get snum\n", __FUNCTION__);
949 return -ENOMEM;
950 }
951
952 if (i==0) {
953 init_enet_offset = 0;
954 } else {
955 init_enet_offset = qe_muram_alloc(
956 sizeof(uec_thread_rx_pram_t),
957 UEC_THREAD_RX_PRAM_ALIGNMENT);
958 }
959
960 entry_val = ((u32)snum << ENET_INIT_PARAM_SNUM_SHIFT) |
961 init_enet_offset | (u32)uec_info->risc_rx;
962 p_init_enet_param->rxthread[i] = entry_val;
963 }
964
965
966 p_init_enet_param->txglobal = uec->tx_glbl_pram_offset |
967 (u32)uec_info->risc_tx;
968
969
970 for (i = 0; i < thread_tx; i++) {
971 if ((snum = qe_get_snum()) < 0) {
972 printf("%s can not get snum\n", __FUNCTION__);
973 return -ENOMEM;
974 }
975
976 init_enet_offset = qe_muram_alloc(sizeof(uec_thread_tx_pram_t),
977 UEC_THREAD_TX_PRAM_ALIGNMENT);
978
979 entry_val = ((u32)snum << ENET_INIT_PARAM_SNUM_SHIFT) |
980 init_enet_offset | (u32)uec_info->risc_tx;
981 p_init_enet_param->txthread[i] = entry_val;
982 }
983
984 __asm__ __volatile__("sync");
985
986
987 command = QE_INIT_TX_RX;
988 cecr_subblock = ucc_fast_get_qe_cr_subblock(
989 uec->uec_info->uf_info.ucc_num);
990 qe_issue_cmd(command, cecr_subblock, (u8) QE_CR_PROTOCOL_ETHERNET,
991 init_enet_param_offset);
992
993 return 0;
994}
995
996static int uec_startup(uec_private_t *uec)
997{
998 uec_info_t *uec_info;
999 ucc_fast_info_t *uf_info;
1000 ucc_fast_private_t *uccf;
1001 ucc_fast_t *uf_regs;
1002 uec_t *uec_regs;
1003 int num_threads_tx;
1004 int num_threads_rx;
1005 u32 utbipar;
1006 u32 length;
1007 u32 align;
1008 qe_bd_t *bd;
1009 u8 *buf;
1010 int i;
1011
1012 if (!uec || !uec->uec_info) {
1013 printf("%s: uec or uec_info not initial\n", __FUNCTION__);
1014 return -EINVAL;
1015 }
1016
1017 uec_info = uec->uec_info;
1018 uf_info = &(uec_info->uf_info);
1019
1020
1021 if ((uec_info->rx_bd_ring_len < UEC_RX_BD_RING_SIZE_MIN) || \
1022 (uec_info->rx_bd_ring_len % UEC_RX_BD_RING_SIZE_ALIGNMENT)) {
1023 printf("%s: Rx BD ring len must be multiple of 4, and > 8.\n",
1024 __FUNCTION__);
1025 return -EINVAL;
1026 }
1027
1028
1029 if (uec_info->tx_bd_ring_len < UEC_TX_BD_RING_SIZE_MIN) {
1030 printf("%s: Tx BD ring length must not be smaller than 2.\n",
1031 __FUNCTION__);
1032 return -EINVAL;
1033 }
1034
1035
1036 if ((MAX_RXBUF_LEN == 0) || (MAX_RXBUF_LEN % UEC_MRBLR_ALIGNMENT)) {
1037 printf("%s: max rx buffer length must be mutliple of 128.\n",
1038 __FUNCTION__);
1039 return -EINVAL;
1040 }
1041
1042
1043 uec->grace_stopped_rx = 1;
1044 uec->grace_stopped_tx = 1;
1045
1046
1047 if (ucc_fast_init(uf_info, &uccf)) {
1048 printf("%s: failed to init ucc fast\n", __FUNCTION__);
1049 return -ENOMEM;
1050 }
1051
1052
1053 uec->uccf = uccf;
1054
1055
1056 if (uec_convert_threads_num(uec_info->num_threads_tx,
1057 &num_threads_tx)) {
1058 return -EINVAL;
1059 }
1060
1061
1062 if (uec_convert_threads_num(uec_info->num_threads_rx,
1063 &num_threads_rx)) {
1064 return -EINVAL;
1065 }
1066
1067 uf_regs = uccf->uf_regs;
1068
1069
1070 uec_regs = (uec_t *)(&uf_regs->ucc_eth);
1071
1072
1073 uec->uec_regs = uec_regs;
1074
1075
1076 out_be32(&uec->uccf->uf_regs->upsmr, UPSMR_INIT_VALUE);
1077
1078
1079 out_be32(&uec_regs->maccfg1, MACCFG1_INIT_VALUE);
1080
1081
1082 out_be32(&uec_regs->maccfg2, MACCFG2_INIT_VALUE);
1083
1084
1085 uec_set_mac_if_mode(uec, uec_info->enet_interface_type, uec_info->speed);
1086
1087
1088#ifndef CONFIG_eTSEC_MDIO_BUS
1089 uec->uec_mii_regs = (uec_mii_t *)(&uec_regs->miimcfg);
1090#else
1091 uec->uec_mii_regs = (uec_mii_t *) CONFIG_MIIM_ADDRESS;
1092#endif
1093
1094
1095 qe_set_mii_clk_src(uec_info->uf_info.ucc_num);
1096
1097
1098 utbipar = in_be32(&uec_regs->utbipar);
1099 utbipar &= ~UTBIPAR_PHY_ADDRESS_MASK;
1100
1101
1102
1103
1104 utbipar |= CONFIG_UTBIPAR_INIT_TBIPA << UTBIPAR_PHY_ADDRESS_SHIFT;
1105 out_be32(&uec_regs->utbipar, utbipar);
1106
1107
1108 if ((uec->uec_info->enet_interface_type == PHY_INTERFACE_MODE_SGMII) &&
1109 (uec->uec_info->speed == SPEED_1000)) {
1110 uec_write_phy_reg(uec->dev, uec_regs->utbipar,
1111 ENET_TBI_MII_ANA, TBIANA_SETTINGS);
1112
1113 uec_write_phy_reg(uec->dev, uec_regs->utbipar,
1114 ENET_TBI_MII_TBICON, TBICON_CLK_SELECT);
1115
1116 uec_write_phy_reg(uec->dev, uec_regs->utbipar,
1117 ENET_TBI_MII_CR, TBICR_SETTINGS);
1118 }
1119
1120
1121 length = ((uec_info->tx_bd_ring_len * SIZEOFBD) /
1122 UEC_TX_BD_RING_SIZE_MEMORY_ALIGNMENT) *
1123 UEC_TX_BD_RING_SIZE_MEMORY_ALIGNMENT;
1124 if ((uec_info->tx_bd_ring_len * SIZEOFBD) %
1125 UEC_TX_BD_RING_SIZE_MEMORY_ALIGNMENT) {
1126 length += UEC_TX_BD_RING_SIZE_MEMORY_ALIGNMENT;
1127 }
1128
1129 align = UEC_TX_BD_RING_ALIGNMENT;
1130 uec->tx_bd_ring_offset = (u32)malloc((u32)(length + align));
1131 if (uec->tx_bd_ring_offset != 0) {
1132 uec->p_tx_bd_ring = (u8 *)((uec->tx_bd_ring_offset + align)
1133 & ~(align - 1));
1134 }
1135
1136
1137 memset((void *)(uec->tx_bd_ring_offset), 0, length + align);
1138
1139
1140 length = uec_info->rx_bd_ring_len * SIZEOFBD;
1141 align = UEC_RX_BD_RING_ALIGNMENT;
1142 uec->rx_bd_ring_offset = (u32)(malloc((u32)(length + align)));
1143 if (uec->rx_bd_ring_offset != 0) {
1144 uec->p_rx_bd_ring = (u8 *)((uec->rx_bd_ring_offset + align)
1145 & ~(align - 1));
1146 }
1147
1148
1149 memset((void *)(uec->rx_bd_ring_offset), 0, length + align);
1150
1151
1152 length = uec_info->rx_bd_ring_len * MAX_RXBUF_LEN;
1153 align = UEC_RX_DATA_BUF_ALIGNMENT;
1154 uec->rx_buf_offset = (u32)malloc(length + align);
1155 if (uec->rx_buf_offset != 0) {
1156 uec->p_rx_buf = (u8 *)((uec->rx_buf_offset + align)
1157 & ~(align - 1));
1158 }
1159
1160
1161 memset((void *)(uec->rx_buf_offset), 0, length + align);
1162
1163
1164 bd = (qe_bd_t *)uec->p_tx_bd_ring;
1165 uec->txBd = bd;
1166
1167 for (i = 0; i < uec_info->tx_bd_ring_len; i++) {
1168 BD_DATA_CLEAR(bd);
1169 BD_STATUS_SET(bd, 0);
1170 BD_LENGTH_SET(bd, 0);
1171 bd ++;
1172 }
1173 BD_STATUS_SET((--bd), TxBD_WRAP);
1174
1175
1176 bd = (qe_bd_t *)uec->p_rx_bd_ring;
1177 uec->rxBd = bd;
1178 buf = uec->p_rx_buf;
1179 for (i = 0; i < uec_info->rx_bd_ring_len; i++) {
1180 BD_DATA_SET(bd, buf);
1181 BD_LENGTH_SET(bd, 0);
1182 BD_STATUS_SET(bd, RxBD_EMPTY);
1183 buf += MAX_RXBUF_LEN;
1184 bd ++;
1185 }
1186 BD_STATUS_SET((--bd), RxBD_WRAP | RxBD_EMPTY);
1187
1188
1189 uec_init_tx_parameter(uec, num_threads_tx);
1190
1191
1192 uec_init_rx_parameter(uec, num_threads_rx);
1193
1194
1195 if (uec_issue_init_enet_rxtx_cmd(uec, num_threads_tx,
1196 num_threads_rx)) {
1197 printf("%s issue init enet cmd failed\n", __FUNCTION__);
1198 return -ENOMEM;
1199 }
1200
1201 return 0;
1202}
1203
1204static int uec_init(struct eth_device* dev, bd_t *bd)
1205{
1206 uec_private_t *uec;
1207 int err, i;
1208 struct phy_info *curphy;
1209#if defined(CONFIG_P1012) || defined(CONFIG_P1021) || defined(CONFIG_P1025)
1210 ccsr_gur_t *gur = (void *)(CONFIG_SYS_MPC85xx_GUTS_ADDR);
1211#endif
1212
1213 uec = (uec_private_t *)dev->priv;
1214
1215 if (uec->the_first_run == 0) {
1216#if defined(CONFIG_P1012) || defined(CONFIG_P1021) || defined(CONFIG_P1025)
1217
1218 setbits_be32(&gur->pmuxcr, MPC85xx_PMUXCR_QE9);
1219 setbits_be32(&gur->pmuxcr, MPC85xx_PMUXCR_QE12);
1220#endif
1221
1222 err = init_phy(dev);
1223 if (err) {
1224 printf("%s: Cannot initialize PHY, aborting.\n",
1225 dev->name);
1226 return err;
1227 }
1228
1229 curphy = uec->mii_info->phyinfo;
1230
1231 if (curphy->config_aneg) {
1232 err = curphy->config_aneg(uec->mii_info);
1233 if (err) {
1234 printf("%s: Can't negotiate PHY\n", dev->name);
1235 return err;
1236 }
1237 }
1238
1239
1240 i = 50;
1241 do {
1242 err = curphy->read_status(uec->mii_info);
1243 if (!(((i-- > 0) && !uec->mii_info->link) || err))
1244 break;
1245 udelay(100000);
1246 } while (1);
1247
1248#if defined(CONFIG_P1012) || defined(CONFIG_P1021) || defined(CONFIG_P1025)
1249
1250 clrbits_be32(&gur->pmuxcr, MPC85xx_PMUXCR_QE12);
1251#endif
1252
1253 if (err || i <= 0)
1254 printf("warning: %s: timeout on PHY link\n", dev->name);
1255
1256 adjust_link(dev);
1257 uec->the_first_run = 1;
1258 }
1259
1260
1261 if (dev->enetaddr[0] & 0x01) {
1262 printf("%s: MacAddress is multcast address\n",
1263 __FUNCTION__);
1264 return -1;
1265 }
1266 uec_set_mac_address(uec, dev->enetaddr);
1267
1268
1269 err = uec_open(uec, COMM_DIR_RX_AND_TX);
1270 if (err) {
1271 printf("%s: cannot enable UEC device\n", dev->name);
1272 return -1;
1273 }
1274
1275 phy_change(dev);
1276
1277 return (uec->mii_info->link ? 0 : -1);
1278}
1279
1280static void uec_halt(struct eth_device* dev)
1281{
1282 uec_private_t *uec = (uec_private_t *)dev->priv;
1283 uec_stop(uec, COMM_DIR_RX_AND_TX);
1284}
1285
1286static int uec_send(struct eth_device *dev, void *buf, int len)
1287{
1288 uec_private_t *uec;
1289 ucc_fast_private_t *uccf;
1290 volatile qe_bd_t *bd;
1291 u16 status;
1292 int i;
1293 int result = 0;
1294
1295 uec = (uec_private_t *)dev->priv;
1296 uccf = uec->uccf;
1297 bd = uec->txBd;
1298
1299
1300 for (i = 0; bd->status & TxBD_READY; i++) {
1301 if (i > 0x100000) {
1302 printf("%s: tx buffer not ready\n", dev->name);
1303 return result;
1304 }
1305 }
1306
1307
1308 BD_DATA_SET(bd, buf);
1309 BD_LENGTH_SET(bd, len);
1310 status = bd->status;
1311 status &= BD_WRAP;
1312 status |= (TxBD_READY | TxBD_LAST);
1313 BD_STATUS_SET(bd, status);
1314
1315
1316 ucc_fast_transmit_on_demand(uccf);
1317
1318
1319 for (i = 0; bd->status & TxBD_READY; i++) {
1320 if (i > 0x100000) {
1321 printf("%s: tx error\n", dev->name);
1322 return result;
1323 }
1324 }
1325
1326
1327 BD_ADVANCE(bd, status, uec->p_tx_bd_ring);
1328 uec->txBd = bd;
1329 result = 1;
1330
1331 return result;
1332}
1333
1334static int uec_recv(struct eth_device* dev)
1335{
1336 uec_private_t *uec = dev->priv;
1337 volatile qe_bd_t *bd;
1338 u16 status;
1339 u16 len;
1340 u8 *data;
1341
1342 bd = uec->rxBd;
1343 status = bd->status;
1344
1345 while (!(status & RxBD_EMPTY)) {
1346 if (!(status & RxBD_ERROR)) {
1347 data = BD_DATA(bd);
1348 len = BD_LENGTH(bd);
1349 NetReceive(data, len);
1350 } else {
1351 printf("%s: Rx error\n", dev->name);
1352 }
1353 status &= BD_CLEAN;
1354 BD_LENGTH_SET(bd, 0);
1355 BD_STATUS_SET(bd, status | RxBD_EMPTY);
1356 BD_ADVANCE(bd, status, uec->p_rx_bd_ring);
1357 status = bd->status;
1358 }
1359 uec->rxBd = bd;
1360
1361 return 1;
1362}
1363
1364int uec_initialize(bd_t *bis, uec_info_t *uec_info)
1365{
1366 struct eth_device *dev;
1367 int i;
1368 uec_private_t *uec;
1369 int err;
1370
1371 dev = (struct eth_device *)malloc(sizeof(struct eth_device));
1372 if (!dev)
1373 return 0;
1374 memset(dev, 0, sizeof(struct eth_device));
1375
1376
1377 uec = (uec_private_t *)malloc(sizeof(uec_private_t));
1378 if (!uec) {
1379 return -ENOMEM;
1380 }
1381 memset(uec, 0, sizeof(uec_private_t));
1382
1383
1384#if (MAX_QE_RISC == 4)
1385 uec_info->risc_tx = QE_RISC_ALLOCATION_FOUR_RISCS;
1386 uec_info->risc_rx = QE_RISC_ALLOCATION_FOUR_RISCS;
1387#endif
1388
1389 devlist[uec_info->uf_info.ucc_num] = dev;
1390
1391 uec->uec_info = uec_info;
1392 uec->dev = dev;
1393
1394 sprintf(dev->name, "UEC%d", uec_info->uf_info.ucc_num);
1395 dev->iobase = 0;
1396 dev->priv = (void *)uec;
1397 dev->init = uec_init;
1398 dev->halt = uec_halt;
1399 dev->send = uec_send;
1400 dev->recv = uec_recv;
1401
1402
1403 for (i = 0; i < 6; i++)
1404 dev->enetaddr[i] = 0;
1405
1406 eth_register(dev);
1407
1408 err = uec_startup(uec);
1409 if (err) {
1410 printf("%s: Cannot configure net device, aborting.",dev->name);
1411 return err;
1412 }
1413
1414#if defined(CONFIG_MII) || defined(CONFIG_CMD_MII)
1415 miiphy_register(dev->name, uec_miiphy_read, uec_miiphy_write);
1416#endif
1417
1418 return 1;
1419}
1420
1421int uec_eth_init(bd_t *bis, uec_info_t *uecs, int num)
1422{
1423 int i;
1424
1425 for (i = 0; i < num; i++)
1426 uec_initialize(bis, &uecs[i]);
1427
1428 return 0;
1429}
1430
1431int uec_standard_init(bd_t *bis)
1432{
1433 return uec_eth_init(bis, uec_info, ARRAY_SIZE(uec_info));
1434}
1435