1
2#include <linux/delay.h>
3
4#include "nitrox_dev.h"
5#include "nitrox_csr.h"
6#include "nitrox_hal.h"
7
8#define PLL_REF_CLK 50
9#define MAX_CSR_RETRIES 10
10
11
12
13
14
15static void emu_enable_cores(struct nitrox_device *ndev)
16{
17 union emu_se_enable emu_se;
18 union emu_ae_enable emu_ae;
19 int i;
20
21
22 emu_ae.value = 0;
23 emu_ae.s.enable = 0xfffff;
24
25
26 emu_se.value = 0;
27 emu_se.s.enable = 0xffff;
28
29
30 for (i = 0; i < NR_CLUSTERS; i++) {
31 nitrox_write_csr(ndev, EMU_AE_ENABLEX(i), emu_ae.value);
32 nitrox_write_csr(ndev, EMU_SE_ENABLEX(i), emu_se.value);
33 }
34}
35
36
37
38
39
40void nitrox_config_emu_unit(struct nitrox_device *ndev)
41{
42 union emu_wd_int_ena_w1s emu_wd_int;
43 union emu_ge_int_ena_w1s emu_ge_int;
44 u64 offset;
45 int i;
46
47
48 emu_enable_cores(ndev);
49
50
51 emu_ge_int.value = 0;
52 emu_ge_int.s.se_ge = 0xffff;
53 emu_ge_int.s.ae_ge = 0xfffff;
54 emu_wd_int.value = 0;
55 emu_wd_int.s.se_wd = 1;
56
57 for (i = 0; i < NR_CLUSTERS; i++) {
58 offset = EMU_WD_INT_ENA_W1SX(i);
59 nitrox_write_csr(ndev, offset, emu_wd_int.value);
60 offset = EMU_GE_INT_ENA_W1SX(i);
61 nitrox_write_csr(ndev, offset, emu_ge_int.value);
62 }
63}
64
65static void reset_pkt_input_ring(struct nitrox_device *ndev, int ring)
66{
67 union nps_pkt_in_instr_ctl pkt_in_ctl;
68 union nps_pkt_in_done_cnts pkt_in_cnts;
69 int max_retries = MAX_CSR_RETRIES;
70 u64 offset;
71
72
73 offset = NPS_PKT_IN_INSTR_CTLX(ring);
74 pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
75 pkt_in_ctl.s.enb = 0;
76 nitrox_write_csr(ndev, offset, pkt_in_ctl.value);
77
78
79 usleep_range(100, 150);
80 do {
81 pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
82 if (!pkt_in_ctl.s.enb)
83 break;
84 udelay(50);
85 } while (max_retries--);
86
87
88 offset = NPS_PKT_IN_DONE_CNTSX(ring);
89 pkt_in_cnts.value = nitrox_read_csr(ndev, offset);
90 nitrox_write_csr(ndev, offset, pkt_in_cnts.value);
91 usleep_range(50, 100);
92}
93
94void enable_pkt_input_ring(struct nitrox_device *ndev, int ring)
95{
96 union nps_pkt_in_instr_ctl pkt_in_ctl;
97 int max_retries = MAX_CSR_RETRIES;
98 u64 offset;
99
100
101 offset = NPS_PKT_IN_INSTR_CTLX(ring);
102 pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
103 pkt_in_ctl.s.is64b = 1;
104 pkt_in_ctl.s.enb = 1;
105 nitrox_write_csr(ndev, offset, pkt_in_ctl.value);
106
107
108 do {
109 pkt_in_ctl.value = nitrox_read_csr(ndev, offset);
110 if (pkt_in_ctl.s.enb)
111 break;
112 udelay(50);
113 } while (max_retries--);
114}
115
116
117
118
119
120void nitrox_config_pkt_input_rings(struct nitrox_device *ndev)
121{
122 int i;
123
124 for (i = 0; i < ndev->nr_queues; i++) {
125 struct nitrox_cmdq *cmdq = &ndev->pkt_inq[i];
126 union nps_pkt_in_instr_rsize pkt_in_rsize;
127 union nps_pkt_in_instr_baoff_dbell pkt_in_dbell;
128 u64 offset;
129
130 reset_pkt_input_ring(ndev, i);
131
132
133
134
135
136
137 offset = NPS_PKT_IN_INSTR_BADDRX(i);
138 nitrox_write_csr(ndev, offset, cmdq->dma);
139
140
141 offset = NPS_PKT_IN_INSTR_RSIZEX(i);
142 pkt_in_rsize.value = 0;
143 pkt_in_rsize.s.rsize = ndev->qlen;
144 nitrox_write_csr(ndev, offset, pkt_in_rsize.value);
145
146
147 offset = NPS_PKT_IN_INT_LEVELSX(i);
148 nitrox_write_csr(ndev, offset, 0xffffffff);
149
150
151 offset = NPS_PKT_IN_INSTR_BAOFF_DBELLX(i);
152 pkt_in_dbell.value = 0;
153 pkt_in_dbell.s.dbell = 0xffffffff;
154 nitrox_write_csr(ndev, offset, pkt_in_dbell.value);
155
156
157 enable_pkt_input_ring(ndev, i);
158 }
159}
160
161static void reset_pkt_solicit_port(struct nitrox_device *ndev, int port)
162{
163 union nps_pkt_slc_ctl pkt_slc_ctl;
164 union nps_pkt_slc_cnts pkt_slc_cnts;
165 int max_retries = MAX_CSR_RETRIES;
166 u64 offset;
167
168
169 offset = NPS_PKT_SLC_CTLX(port);
170 pkt_slc_ctl.value = nitrox_read_csr(ndev, offset);
171 pkt_slc_ctl.s.enb = 0;
172 nitrox_write_csr(ndev, offset, pkt_slc_ctl.value);
173
174
175 usleep_range(100, 150);
176
177 do {
178 pkt_slc_ctl.value = nitrox_read_csr(ndev, offset);
179 if (!pkt_slc_ctl.s.enb)
180 break;
181 udelay(50);
182 } while (max_retries--);
183
184
185 offset = NPS_PKT_SLC_CNTSX(port);
186 pkt_slc_cnts.value = nitrox_read_csr(ndev, offset);
187 nitrox_write_csr(ndev, offset, pkt_slc_cnts.value);
188 usleep_range(50, 100);
189}
190
191void enable_pkt_solicit_port(struct nitrox_device *ndev, int port)
192{
193 union nps_pkt_slc_ctl pkt_slc_ctl;
194 int max_retries = MAX_CSR_RETRIES;
195 u64 offset;
196
197 offset = NPS_PKT_SLC_CTLX(port);
198 pkt_slc_ctl.value = 0;
199 pkt_slc_ctl.s.enb = 1;
200
201
202
203
204 pkt_slc_ctl.s.z = 1;
205
206 pkt_slc_ctl.s.rh = 1;
207 nitrox_write_csr(ndev, offset, pkt_slc_ctl.value);
208
209
210 do {
211 pkt_slc_ctl.value = nitrox_read_csr(ndev, offset);
212 if (pkt_slc_ctl.s.enb)
213 break;
214 udelay(50);
215 } while (max_retries--);
216}
217
218static void config_pkt_solicit_port(struct nitrox_device *ndev, int port)
219{
220 union nps_pkt_slc_int_levels pkt_slc_int;
221 u64 offset;
222
223 reset_pkt_solicit_port(ndev, port);
224
225
226 offset = NPS_PKT_SLC_INT_LEVELSX(port);
227 pkt_slc_int.value = 0;
228
229 pkt_slc_int.s.timet = 0x3fffff;
230 nitrox_write_csr(ndev, offset, pkt_slc_int.value);
231
232
233 enable_pkt_solicit_port(ndev, port);
234}
235
236void nitrox_config_pkt_solicit_ports(struct nitrox_device *ndev)
237{
238 int i;
239
240 for (i = 0; i < ndev->nr_queues; i++)
241 config_pkt_solicit_port(ndev, i);
242}
243
244
245
246
247
248
249
250static void enable_nps_core_interrupts(struct nitrox_device *ndev)
251{
252 union nps_core_int_ena_w1s core_int;
253
254
255 core_int.value = 0;
256 core_int.s.host_wr_err = 1;
257 core_int.s.host_wr_timeout = 1;
258 core_int.s.exec_wr_timeout = 1;
259 core_int.s.npco_dma_malform = 1;
260 core_int.s.host_nps_wr_err = 1;
261 nitrox_write_csr(ndev, NPS_CORE_INT_ENA_W1S, core_int.value);
262}
263
264void nitrox_config_nps_core_unit(struct nitrox_device *ndev)
265{
266 union nps_core_gbl_vfcfg core_gbl_vfcfg;
267
268
269 nitrox_write_csr(ndev, NPS_CORE_CONTROL, 1ULL);
270
271
272 core_gbl_vfcfg.value = 0;
273 core_gbl_vfcfg.s.ilk_disable = 1;
274 core_gbl_vfcfg.s.cfg = __NDEV_MODE_PF;
275 nitrox_write_csr(ndev, NPS_CORE_GBL_VFCFG, core_gbl_vfcfg.value);
276
277
278 enable_nps_core_interrupts(ndev);
279}
280
281
282
283
284
285
286
287static void enable_nps_pkt_interrupts(struct nitrox_device *ndev)
288{
289
290 nitrox_write_csr(ndev, NPS_PKT_IN_RERR_LO_ENA_W1S, (~0ULL));
291 nitrox_write_csr(ndev, NPS_PKT_IN_RERR_HI_ENA_W1S, (~0ULL));
292 nitrox_write_csr(ndev, NPS_PKT_IN_ERR_TYPE_ENA_W1S, (~0ULL));
293
294 nitrox_write_csr(ndev, NPS_PKT_SLC_RERR_HI_ENA_W1S, (~0ULL));
295 nitrox_write_csr(ndev, NPS_PKT_SLC_RERR_LO_ENA_W1S, (~0ULL));
296 nitrox_write_csr(ndev, NPS_PKT_SLC_ERR_TYPE_ENA_W1S, (~0uLL));
297}
298
299void nitrox_config_nps_pkt_unit(struct nitrox_device *ndev)
300{
301
302 nitrox_config_pkt_input_rings(ndev);
303 nitrox_config_pkt_solicit_ports(ndev);
304
305
306 enable_nps_pkt_interrupts(ndev);
307}
308
309static void reset_aqm_ring(struct nitrox_device *ndev, int ring)
310{
311 union aqmq_en aqmq_en_reg;
312 union aqmq_activity_stat activity_stat;
313 union aqmq_cmp_cnt cmp_cnt;
314 int max_retries = MAX_CSR_RETRIES;
315 u64 offset;
316
317
318 offset = AQMQ_ENX(ring);
319 aqmq_en_reg.value = 0;
320 aqmq_en_reg.queue_enable = 0;
321 nitrox_write_csr(ndev, offset, aqmq_en_reg.value);
322
323
324 usleep_range(100, 150);
325 offset = AQMQ_ACTIVITY_STATX(ring);
326 do {
327 activity_stat.value = nitrox_read_csr(ndev, offset);
328 if (!activity_stat.queue_active)
329 break;
330 udelay(50);
331 } while (max_retries--);
332
333
334 offset = AQMQ_CMP_CNTX(ring);
335 cmp_cnt.value = nitrox_read_csr(ndev, offset);
336 nitrox_write_csr(ndev, offset, cmp_cnt.value);
337 usleep_range(50, 100);
338}
339
340void enable_aqm_ring(struct nitrox_device *ndev, int ring)
341{
342 union aqmq_en aqmq_en_reg;
343 u64 offset;
344
345 offset = AQMQ_ENX(ring);
346 aqmq_en_reg.value = 0;
347 aqmq_en_reg.queue_enable = 1;
348 nitrox_write_csr(ndev, offset, aqmq_en_reg.value);
349 usleep_range(50, 100);
350}
351
352void nitrox_config_aqm_rings(struct nitrox_device *ndev)
353{
354 int ring;
355
356 for (ring = 0; ring < ndev->nr_queues; ring++) {
357 struct nitrox_cmdq *cmdq = ndev->aqmq[ring];
358 union aqmq_drbl drbl;
359 union aqmq_qsz qsize;
360 union aqmq_cmp_thr cmp_thr;
361 u64 offset;
362
363
364 reset_aqm_ring(ndev, ring);
365
366
367 offset = AQMQ_DRBLX(ring);
368 drbl.value = 0;
369 drbl.dbell_count = 0xFFFFFFFF;
370 nitrox_write_csr(ndev, offset, drbl.value);
371
372
373
374
375 offset = AQMQ_NXT_CMDX(ring);
376 nitrox_write_csr(ndev, offset, 0ULL);
377
378
379 offset = AQMQ_BADRX(ring);
380 nitrox_write_csr(ndev, offset, cmdq->dma);
381
382
383 offset = AQMQ_QSZX(ring);
384 qsize.value = 0;
385 qsize.host_queue_size = ndev->qlen;
386 nitrox_write_csr(ndev, offset, qsize.value);
387
388
389 offset = AQMQ_CMP_THRX(ring);
390 cmp_thr.value = 0;
391 cmp_thr.commands_completed_threshold = 1;
392 nitrox_write_csr(ndev, offset, cmp_thr.value);
393
394
395 enable_aqm_ring(ndev, ring);
396 }
397}
398
399static void enable_aqm_interrupts(struct nitrox_device *ndev)
400{
401
402 nitrox_write_csr(ndev, AQM_DBELL_OVF_LO_ENA_W1S, (~0ULL));
403 nitrox_write_csr(ndev, AQM_DBELL_OVF_HI_ENA_W1S, (~0ULL));
404 nitrox_write_csr(ndev, AQM_DMA_RD_ERR_LO_ENA_W1S, (~0ULL));
405 nitrox_write_csr(ndev, AQM_DMA_RD_ERR_HI_ENA_W1S, (~0ULL));
406 nitrox_write_csr(ndev, AQM_EXEC_NA_LO_ENA_W1S, (~0ULL));
407 nitrox_write_csr(ndev, AQM_EXEC_NA_HI_ENA_W1S, (~0ULL));
408 nitrox_write_csr(ndev, AQM_EXEC_ERR_LO_ENA_W1S, (~0ULL));
409 nitrox_write_csr(ndev, AQM_EXEC_ERR_HI_ENA_W1S, (~0ULL));
410}
411
412void nitrox_config_aqm_unit(struct nitrox_device *ndev)
413{
414
415 nitrox_config_aqm_rings(ndev);
416
417
418 enable_aqm_interrupts(ndev);
419}
420
421void nitrox_config_pom_unit(struct nitrox_device *ndev)
422{
423 union pom_int_ena_w1s pom_int;
424 int i;
425
426
427 pom_int.value = 0;
428 pom_int.s.illegal_dport = 1;
429 nitrox_write_csr(ndev, POM_INT_ENA_W1S, pom_int.value);
430
431
432 for (i = 0; i < ndev->hw.se_cores; i++)
433 nitrox_write_csr(ndev, POM_PERF_CTL, BIT_ULL(i));
434}
435
436
437
438
439
440void nitrox_config_rand_unit(struct nitrox_device *ndev)
441{
442 union efl_rnm_ctl_status efl_rnm_ctl;
443 u64 offset;
444
445 offset = EFL_RNM_CTL_STATUS;
446 efl_rnm_ctl.value = nitrox_read_csr(ndev, offset);
447 efl_rnm_ctl.s.ent_en = 1;
448 efl_rnm_ctl.s.rng_en = 1;
449 nitrox_write_csr(ndev, offset, efl_rnm_ctl.value);
450}
451
452void nitrox_config_efl_unit(struct nitrox_device *ndev)
453{
454 int i;
455
456 for (i = 0; i < NR_CLUSTERS; i++) {
457 union efl_core_int_ena_w1s efl_core_int;
458 u64 offset;
459
460
461 offset = EFL_CORE_INT_ENA_W1SX(i);
462 efl_core_int.value = 0;
463 efl_core_int.s.len_ovr = 1;
464 efl_core_int.s.d_left = 1;
465 efl_core_int.s.epci_decode_err = 1;
466 nitrox_write_csr(ndev, offset, efl_core_int.value);
467
468 offset = EFL_CORE_VF_ERR_INT0_ENA_W1SX(i);
469 nitrox_write_csr(ndev, offset, (~0ULL));
470 offset = EFL_CORE_VF_ERR_INT1_ENA_W1SX(i);
471 nitrox_write_csr(ndev, offset, (~0ULL));
472 }
473}
474
475void nitrox_config_bmi_unit(struct nitrox_device *ndev)
476{
477 union bmi_ctl bmi_ctl;
478 union bmi_int_ena_w1s bmi_int_ena;
479 u64 offset;
480
481
482 offset = BMI_CTL;
483 bmi_ctl.value = nitrox_read_csr(ndev, offset);
484 bmi_ctl.s.max_pkt_len = 0xff;
485 bmi_ctl.s.nps_free_thrsh = 0xff;
486 bmi_ctl.s.nps_hdrq_thrsh = 0x7a;
487 nitrox_write_csr(ndev, offset, bmi_ctl.value);
488
489
490 offset = BMI_INT_ENA_W1S;
491 bmi_int_ena.value = 0;
492 bmi_int_ena.s.max_len_err_nps = 1;
493 bmi_int_ena.s.pkt_rcv_err_nps = 1;
494 bmi_int_ena.s.fpf_undrrn = 1;
495 nitrox_write_csr(ndev, offset, bmi_int_ena.value);
496}
497
498void nitrox_config_bmo_unit(struct nitrox_device *ndev)
499{
500 union bmo_ctl2 bmo_ctl2;
501 u64 offset;
502
503
504 offset = BMO_CTL2;
505 bmo_ctl2.value = nitrox_read_csr(ndev, offset);
506 bmo_ctl2.s.nps_slc_buf_thrsh = 0xff;
507 nitrox_write_csr(ndev, offset, bmo_ctl2.value);
508}
509
510void invalidate_lbc(struct nitrox_device *ndev)
511{
512 union lbc_inval_ctl lbc_ctl;
513 union lbc_inval_status lbc_stat;
514 int max_retries = MAX_CSR_RETRIES;
515 u64 offset;
516
517
518 offset = LBC_INVAL_CTL;
519 lbc_ctl.value = nitrox_read_csr(ndev, offset);
520 lbc_ctl.s.cam_inval_start = 1;
521 nitrox_write_csr(ndev, offset, lbc_ctl.value);
522
523 offset = LBC_INVAL_STATUS;
524 do {
525 lbc_stat.value = nitrox_read_csr(ndev, offset);
526 if (lbc_stat.s.done)
527 break;
528 udelay(50);
529 } while (max_retries--);
530}
531
532void nitrox_config_lbc_unit(struct nitrox_device *ndev)
533{
534 union lbc_int_ena_w1s lbc_int_ena;
535 u64 offset;
536
537 invalidate_lbc(ndev);
538
539
540 offset = LBC_INT_ENA_W1S;
541 lbc_int_ena.value = 0;
542 lbc_int_ena.s.dma_rd_err = 1;
543 lbc_int_ena.s.over_fetch_err = 1;
544 lbc_int_ena.s.cam_inval_abort = 1;
545 lbc_int_ena.s.cam_hard_err = 1;
546 nitrox_write_csr(ndev, offset, lbc_int_ena.value);
547
548 offset = LBC_PLM_VF1_64_INT_ENA_W1S;
549 nitrox_write_csr(ndev, offset, (~0ULL));
550 offset = LBC_PLM_VF65_128_INT_ENA_W1S;
551 nitrox_write_csr(ndev, offset, (~0ULL));
552
553 offset = LBC_ELM_VF1_64_INT_ENA_W1S;
554 nitrox_write_csr(ndev, offset, (~0ULL));
555 offset = LBC_ELM_VF65_128_INT_ENA_W1S;
556 nitrox_write_csr(ndev, offset, (~0ULL));
557}
558
559void config_nps_core_vfcfg_mode(struct nitrox_device *ndev, enum vf_mode mode)
560{
561 union nps_core_gbl_vfcfg vfcfg;
562
563 vfcfg.value = nitrox_read_csr(ndev, NPS_CORE_GBL_VFCFG);
564 vfcfg.s.cfg = mode & 0x7;
565
566 nitrox_write_csr(ndev, NPS_CORE_GBL_VFCFG, vfcfg.value);
567}
568
569static const char *get_core_option(u8 se_cores, u8 ae_cores)
570{
571 const char *option = "";
572
573 if (ae_cores == AE_MAX_CORES) {
574 switch (se_cores) {
575 case SE_MAX_CORES:
576 option = "60";
577 break;
578 case 40:
579 option = "60s";
580 break;
581 }
582 } else if (ae_cores == (AE_MAX_CORES / 2)) {
583 option = "30";
584 } else {
585 option = "60i";
586 }
587
588 return option;
589}
590
591static const char *get_feature_option(u8 zip_cores, int core_freq)
592{
593 if (zip_cores == 0)
594 return "";
595 else if (zip_cores < ZIP_MAX_CORES)
596 return "-C15";
597
598 if (core_freq >= 850)
599 return "-C45";
600 else if (core_freq >= 750)
601 return "-C35";
602 else if (core_freq >= 550)
603 return "-C25";
604
605 return "";
606}
607
608void nitrox_get_hwinfo(struct nitrox_device *ndev)
609{
610 union emu_fuse_map emu_fuse;
611 union rst_boot rst_boot;
612 union fus_dat1 fus_dat1;
613 unsigned char name[IFNAMSIZ * 2] = {};
614 int i, dead_cores;
615 u64 offset;
616
617
618 offset = RST_BOOT;
619 rst_boot.value = nitrox_read_csr(ndev, offset);
620 ndev->hw.freq = (rst_boot.pnr_mul + 3) * PLL_REF_CLK;
621
622 for (i = 0; i < NR_CLUSTERS; i++) {
623 offset = EMU_FUSE_MAPX(i);
624 emu_fuse.value = nitrox_read_csr(ndev, offset);
625 if (emu_fuse.s.valid) {
626 dead_cores = hweight32(emu_fuse.s.ae_fuse);
627 ndev->hw.ae_cores += AE_CORES_PER_CLUSTER - dead_cores;
628 dead_cores = hweight16(emu_fuse.s.se_fuse);
629 ndev->hw.se_cores += SE_CORES_PER_CLUSTER - dead_cores;
630 }
631 }
632
633 offset = FUS_DAT1;
634 fus_dat1.value = nitrox_read_csr(ndev, offset);
635 if (!fus_dat1.nozip) {
636 dead_cores = hweight8(fus_dat1.zip_info);
637 ndev->hw.zip_cores = ZIP_MAX_CORES - dead_cores;
638 }
639
640
641
642
643 snprintf(name, sizeof(name), "CNN55%s-%3dBG676%s-1.%u",
644 get_core_option(ndev->hw.se_cores, ndev->hw.ae_cores),
645 ndev->hw.freq,
646 get_feature_option(ndev->hw.zip_cores, ndev->hw.freq),
647 ndev->hw.revision_id);
648
649
650 strncpy(ndev->hw.partname, name, sizeof(ndev->hw.partname));
651}
652
653void enable_pf2vf_mbox_interrupts(struct nitrox_device *ndev)
654{
655 u64 value = ~0ULL;
656 u64 reg_addr;
657
658
659 reg_addr = NPS_PKT_MBOX_INT_LO_ENA_W1S;
660 nitrox_write_csr(ndev, reg_addr, value);
661
662
663 reg_addr = NPS_PKT_MBOX_INT_HI_ENA_W1S;
664 nitrox_write_csr(ndev, reg_addr, value);
665}
666
667void disable_pf2vf_mbox_interrupts(struct nitrox_device *ndev)
668{
669 u64 value = ~0ULL;
670 u64 reg_addr;
671
672
673 reg_addr = NPS_PKT_MBOX_INT_LO_ENA_W1C;
674 nitrox_write_csr(ndev, reg_addr, value);
675
676
677 reg_addr = NPS_PKT_MBOX_INT_HI_ENA_W1C;
678 nitrox_write_csr(ndev, reg_addr, value);
679}
680