1
2
3
4
5
6
7
8
9
10
11
12#include <linux/delay.h>
13#include <linux/export.h>
14#include <linux/gfp.h>
15#include <linux/kernel.h>
16#include <linux/pci.h>
17#include <linux/string.h>
18
19#include <asm/pci-bridge.h>
20#include <asm/ppc-pci.h>
21
22static int eeh_pe_aux_size = 0;
23static LIST_HEAD(eeh_phb_pe);
24
25
26
27
28
29
30
31void eeh_set_pe_aux_size(int size)
32{
33 if (size < 0)
34 return;
35
36 eeh_pe_aux_size = size;
37}
38
39
40
41
42
43
44
45
46static struct eeh_pe *eeh_pe_alloc(struct pci_controller *phb, int type)
47{
48 struct eeh_pe *pe;
49 size_t alloc_size;
50
51 alloc_size = sizeof(struct eeh_pe);
52 if (eeh_pe_aux_size) {
53 alloc_size = ALIGN(alloc_size, cache_line_size());
54 alloc_size += eeh_pe_aux_size;
55 }
56
57
58 pe = kzalloc(alloc_size, GFP_KERNEL);
59 if (!pe) return NULL;
60
61
62 pe->type = type;
63 pe->phb = phb;
64 INIT_LIST_HEAD(&pe->child_list);
65 INIT_LIST_HEAD(&pe->edevs);
66
67 pe->data = (void *)pe + ALIGN(sizeof(struct eeh_pe),
68 cache_line_size());
69 return pe;
70}
71
72
73
74
75
76
77
78
79int eeh_phb_pe_create(struct pci_controller *phb)
80{
81 struct eeh_pe *pe;
82
83
84 pe = eeh_pe_alloc(phb, EEH_PE_PHB);
85 if (!pe) {
86 pr_err("%s: out of memory!\n", __func__);
87 return -ENOMEM;
88 }
89
90
91 list_add_tail(&pe->child, &eeh_phb_pe);
92
93 pr_debug("EEH: Add PE for PHB#%x\n", phb->global_number);
94
95 return 0;
96}
97
98
99
100
101
102
103
104
105
106int eeh_wait_state(struct eeh_pe *pe, int max_wait)
107{
108 int ret;
109 int mwait;
110
111
112
113
114
115
116
117
118
119#define EEH_STATE_MIN_WAIT_TIME (1000)
120#define EEH_STATE_MAX_WAIT_TIME (300 * 1000)
121
122 while (1) {
123 ret = eeh_ops->get_state(pe, &mwait);
124
125 if (ret != EEH_STATE_UNAVAILABLE)
126 return ret;
127
128 if (max_wait <= 0) {
129 pr_warn("%s: Timeout when getting PE's state (%d)\n",
130 __func__, max_wait);
131 return EEH_STATE_NOT_SUPPORT;
132 }
133
134 if (mwait < EEH_STATE_MIN_WAIT_TIME) {
135 pr_warn("%s: Firmware returned bad wait value %d\n",
136 __func__, mwait);
137 mwait = EEH_STATE_MIN_WAIT_TIME;
138 } else if (mwait > EEH_STATE_MAX_WAIT_TIME) {
139 pr_warn("%s: Firmware returned too long wait value %d\n",
140 __func__, mwait);
141 mwait = EEH_STATE_MAX_WAIT_TIME;
142 }
143
144 msleep(min(mwait, max_wait));
145 max_wait -= mwait;
146 }
147}
148
149
150
151
152
153
154
155
156
157struct eeh_pe *eeh_phb_pe_get(struct pci_controller *phb)
158{
159 struct eeh_pe *pe;
160
161 list_for_each_entry(pe, &eeh_phb_pe, child) {
162
163
164
165
166
167 if ((pe->type & EEH_PE_PHB) && pe->phb == phb)
168 return pe;
169 }
170
171 return NULL;
172}
173
174
175
176
177
178
179
180
181
182struct eeh_pe *eeh_pe_next(struct eeh_pe *pe, struct eeh_pe *root)
183{
184 struct list_head *next = pe->child_list.next;
185
186 if (next == &pe->child_list) {
187 while (1) {
188 if (pe == root)
189 return NULL;
190 next = pe->child.next;
191 if (next != &pe->parent->child_list)
192 break;
193 pe = pe->parent;
194 }
195 }
196
197 return list_entry(next, struct eeh_pe, child);
198}
199
200
201
202
203
204
205
206
207
208
209
210
211void *eeh_pe_traverse(struct eeh_pe *root,
212 eeh_pe_traverse_func fn, void *flag)
213{
214 struct eeh_pe *pe;
215 void *ret;
216
217 eeh_for_each_pe(root, pe) {
218 ret = fn(pe, flag);
219 if (ret) return ret;
220 }
221
222 return NULL;
223}
224
225
226
227
228
229
230
231
232
233
234void eeh_pe_dev_traverse(struct eeh_pe *root,
235 eeh_edev_traverse_func fn, void *flag)
236{
237 struct eeh_pe *pe;
238 struct eeh_dev *edev, *tmp;
239
240 if (!root) {
241 pr_warn("%s: Invalid PE %p\n",
242 __func__, root);
243 return;
244 }
245
246
247 eeh_for_each_pe(root, pe)
248 eeh_pe_for_each_dev(pe, edev, tmp)
249 fn(edev, flag);
250}
251
252
253
254
255
256
257
258
259
260
261
262struct eeh_pe_get_flag {
263 int pe_no;
264 int config_addr;
265};
266
267static void *__eeh_pe_get(struct eeh_pe *pe, void *flag)
268{
269 struct eeh_pe_get_flag *tmp = (struct eeh_pe_get_flag *) flag;
270
271
272 if (pe->type & EEH_PE_PHB)
273 return NULL;
274
275
276
277
278
279 if (eeh_has_flag(EEH_VALID_PE_ZERO)) {
280 if (tmp->pe_no == pe->addr)
281 return pe;
282 } else {
283 if (tmp->pe_no &&
284 (tmp->pe_no == pe->addr))
285 return pe;
286 }
287
288
289 if (tmp->config_addr &&
290 (tmp->config_addr == pe->config_addr))
291 return pe;
292
293 return NULL;
294}
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309struct eeh_pe *eeh_pe_get(struct pci_controller *phb,
310 int pe_no, int config_addr)
311{
312 struct eeh_pe *root = eeh_phb_pe_get(phb);
313 struct eeh_pe_get_flag tmp = { pe_no, config_addr };
314 struct eeh_pe *pe;
315
316 pe = eeh_pe_traverse(root, __eeh_pe_get, &tmp);
317
318 return pe;
319}
320
321
322
323
324
325
326
327
328
329static struct eeh_pe *eeh_pe_get_parent(struct eeh_dev *edev)
330{
331 struct eeh_dev *parent;
332 struct pci_dn *pdn = eeh_dev_to_pdn(edev);
333
334
335
336
337
338
339 if (edev->physfn)
340 pdn = pci_get_pdn(edev->physfn);
341 else
342 pdn = pdn ? pdn->parent : NULL;
343 while (pdn) {
344
345 parent = pdn_to_eeh_dev(pdn);
346 if (!parent)
347 return NULL;
348
349 if (parent->pe)
350 return parent->pe;
351
352 pdn = pdn->parent;
353 }
354
355 return NULL;
356}
357
358
359
360
361
362
363
364
365
366
367int eeh_add_to_parent_pe(struct eeh_dev *edev)
368{
369 struct eeh_pe *pe, *parent;
370 struct pci_dn *pdn = eeh_dev_to_pdn(edev);
371 int config_addr = (pdn->busno << 8) | (pdn->devfn);
372
373
374 if (!eeh_has_flag(EEH_VALID_PE_ZERO) && !edev->pe_config_addr) {
375 eeh_edev_err(edev, "PE#0 is invalid for this PHB!\n");
376 return -EINVAL;
377 }
378
379
380
381
382
383
384
385 pe = eeh_pe_get(pdn->phb, edev->pe_config_addr, config_addr);
386 if (pe) {
387 if (pe->type & EEH_PE_INVALID) {
388 list_add_tail(&edev->entry, &pe->edevs);
389 edev->pe = pe;
390
391
392
393
394 parent = pe;
395 while (parent) {
396 if (!(parent->type & EEH_PE_INVALID))
397 break;
398 parent->type &= ~EEH_PE_INVALID;
399 parent = parent->parent;
400 }
401
402 eeh_edev_dbg(edev,
403 "Added to device PE (parent: PE#%x)\n",
404 pe->parent->addr);
405 } else {
406
407 pe->type = EEH_PE_BUS;
408 edev->pe = pe;
409
410
411 list_add_tail(&edev->entry, &pe->edevs);
412 eeh_edev_dbg(edev, "Added to bus PE\n");
413 }
414 return 0;
415 }
416
417
418 if (edev->physfn)
419 pe = eeh_pe_alloc(pdn->phb, EEH_PE_VF);
420 else
421 pe = eeh_pe_alloc(pdn->phb, EEH_PE_DEVICE);
422 if (!pe) {
423 pr_err("%s: out of memory!\n", __func__);
424 return -ENOMEM;
425 }
426 pe->addr = edev->pe_config_addr;
427 pe->config_addr = config_addr;
428
429
430
431
432
433
434
435 parent = eeh_pe_get_parent(edev);
436 if (!parent) {
437 parent = eeh_phb_pe_get(pdn->phb);
438 if (!parent) {
439 pr_err("%s: No PHB PE is found (PHB Domain=%d)\n",
440 __func__, pdn->phb->global_number);
441 edev->pe = NULL;
442 kfree(pe);
443 return -EEXIST;
444 }
445 }
446 pe->parent = parent;
447
448
449
450
451
452 list_add_tail(&pe->child, &parent->child_list);
453 list_add_tail(&edev->entry, &pe->edevs);
454 edev->pe = pe;
455 eeh_edev_dbg(edev, "Added to device PE (parent: PE#%x)\n",
456 pe->parent->addr);
457
458 return 0;
459}
460
461
462
463
464
465
466
467
468
469
470int eeh_rmv_from_parent_pe(struct eeh_dev *edev)
471{
472 struct eeh_pe *pe, *parent, *child;
473 bool keep, recover;
474 int cnt;
475
476 pe = eeh_dev_to_pe(edev);
477 if (!pe) {
478 eeh_edev_dbg(edev, "No PE found for device.\n");
479 return -EEXIST;
480 }
481
482
483 edev->pe = NULL;
484 list_del(&edev->entry);
485
486
487
488
489
490
491
492 while (1) {
493 parent = pe->parent;
494
495
496 if (pe->type & EEH_PE_PHB)
497 break;
498
499
500
501
502
503
504 keep = !!(pe->state & EEH_PE_KEEP);
505 recover = !!(pe->state & EEH_PE_RECOVERING);
506 WARN_ON(keep && !recover);
507
508 if (!keep && !recover) {
509 if (list_empty(&pe->edevs) &&
510 list_empty(&pe->child_list)) {
511 list_del(&pe->child);
512 kfree(pe);
513 } else {
514 break;
515 }
516 } else {
517
518
519
520
521
522
523
524
525
526 if (list_empty(&pe->edevs)) {
527 cnt = 0;
528 list_for_each_entry(child, &pe->child_list, child) {
529 if (!(child->type & EEH_PE_INVALID)) {
530 cnt++;
531 break;
532 }
533 }
534
535 if (!cnt)
536 pe->type |= EEH_PE_INVALID;
537 else
538 break;
539 }
540 }
541
542 pe = parent;
543 }
544
545 return 0;
546}
547
548
549
550
551
552
553
554
555
556
557void eeh_pe_update_time_stamp(struct eeh_pe *pe)
558{
559 time64_t tstamp;
560
561 if (!pe) return;
562
563 if (pe->freeze_count <= 0) {
564 pe->freeze_count = 0;
565 pe->tstamp = ktime_get_seconds();
566 } else {
567 tstamp = ktime_get_seconds();
568 if (tstamp - pe->tstamp > 3600) {
569 pe->tstamp = tstamp;
570 pe->freeze_count = 0;
571 }
572 }
573}
574
575
576
577
578
579
580
581
582
583void eeh_pe_state_mark(struct eeh_pe *root, int state)
584{
585 struct eeh_pe *pe;
586
587 eeh_for_each_pe(root, pe)
588 if (!(pe->state & EEH_PE_REMOVED))
589 pe->state |= state;
590}
591EXPORT_SYMBOL_GPL(eeh_pe_state_mark);
592
593
594
595
596
597
598
599
600
601void eeh_pe_mark_isolated(struct eeh_pe *root)
602{
603 struct eeh_pe *pe;
604 struct eeh_dev *edev;
605 struct pci_dev *pdev;
606
607 eeh_pe_state_mark(root, EEH_PE_ISOLATED);
608 eeh_for_each_pe(root, pe) {
609 list_for_each_entry(edev, &pe->edevs, entry) {
610 pdev = eeh_dev_to_pci_dev(edev);
611 if (pdev)
612 pdev->error_state = pci_channel_io_frozen;
613 }
614
615 if (pe->state & EEH_PE_CFG_RESTRICTED)
616 pe->state |= EEH_PE_CFG_BLOCKED;
617 }
618}
619EXPORT_SYMBOL_GPL(eeh_pe_mark_isolated);
620
621static void __eeh_pe_dev_mode_mark(struct eeh_dev *edev, void *flag)
622{
623 int mode = *((int *)flag);
624
625 edev->mode |= mode;
626}
627
628
629
630
631
632
633
634void eeh_pe_dev_mode_mark(struct eeh_pe *pe, int mode)
635{
636 eeh_pe_dev_traverse(pe, __eeh_pe_dev_mode_mark, &mode);
637}
638
639
640
641
642
643
644
645
646
647
648
649void eeh_pe_state_clear(struct eeh_pe *root, int state, bool include_passed)
650{
651 struct eeh_pe *pe;
652 struct eeh_dev *edev, *tmp;
653 struct pci_dev *pdev;
654
655 eeh_for_each_pe(root, pe) {
656
657 if (pe->state & EEH_PE_REMOVED)
658 continue;
659
660 if (!include_passed && eeh_pe_passed(pe))
661 continue;
662
663 pe->state &= ~state;
664
665
666
667
668
669
670 if (!(state & EEH_PE_ISOLATED))
671 continue;
672
673 pe->check_count = 0;
674 eeh_pe_for_each_dev(pe, edev, tmp) {
675 pdev = eeh_dev_to_pci_dev(edev);
676 if (!pdev)
677 continue;
678
679 pdev->error_state = pci_channel_io_normal;
680 }
681
682
683 if (pe->state & EEH_PE_CFG_RESTRICTED)
684 pe->state &= ~EEH_PE_CFG_BLOCKED;
685 }
686}
687
688
689
690
691
692
693
694
695
696
697
698
699static void eeh_bridge_check_link(struct eeh_dev *edev)
700{
701 struct pci_dn *pdn = eeh_dev_to_pdn(edev);
702 int cap;
703 uint32_t val;
704 int timeout = 0;
705
706
707
708
709
710 if (!(edev->mode & (EEH_DEV_ROOT_PORT | EEH_DEV_DS_PORT)))
711 return;
712
713 eeh_edev_dbg(edev, "Checking PCIe link...\n");
714
715
716 cap = edev->pcie_cap;
717 eeh_ops->read_config(pdn, cap + PCI_EXP_SLTSTA, 2, &val);
718 if (!(val & PCI_EXP_SLTSTA_PDS)) {
719 eeh_edev_dbg(edev, "No card in the slot (0x%04x) !\n", val);
720 return;
721 }
722
723
724 eeh_ops->read_config(pdn, cap + PCI_EXP_SLTCAP, 2, &val);
725 if (val & PCI_EXP_SLTCAP_PCP) {
726 eeh_ops->read_config(pdn, cap + PCI_EXP_SLTCTL, 2, &val);
727 if (val & PCI_EXP_SLTCTL_PCC) {
728 eeh_edev_dbg(edev, "In power-off state, power it on ...\n");
729 val &= ~(PCI_EXP_SLTCTL_PCC | PCI_EXP_SLTCTL_PIC);
730 val |= (0x0100 & PCI_EXP_SLTCTL_PIC);
731 eeh_ops->write_config(pdn, cap + PCI_EXP_SLTCTL, 2, val);
732 msleep(2 * 1000);
733 }
734 }
735
736
737 eeh_ops->read_config(pdn, cap + PCI_EXP_LNKCTL, 2, &val);
738 val &= ~PCI_EXP_LNKCTL_LD;
739 eeh_ops->write_config(pdn, cap + PCI_EXP_LNKCTL, 2, val);
740
741
742 eeh_ops->read_config(pdn, cap + PCI_EXP_LNKCAP, 4, &val);
743 if (!(val & PCI_EXP_LNKCAP_DLLLARC)) {
744 eeh_edev_dbg(edev, "No link reporting capability (0x%08x) \n", val);
745 msleep(1000);
746 return;
747 }
748
749
750 timeout = 0;
751 while (timeout < 5000) {
752 msleep(20);
753 timeout += 20;
754
755 eeh_ops->read_config(pdn, cap + PCI_EXP_LNKSTA, 2, &val);
756 if (val & PCI_EXP_LNKSTA_DLLLA)
757 break;
758 }
759
760 if (val & PCI_EXP_LNKSTA_DLLLA)
761 eeh_edev_dbg(edev, "Link up (%s)\n",
762 (val & PCI_EXP_LNKSTA_CLS_2_5GB) ? "2.5GB" : "5GB");
763 else
764 eeh_edev_dbg(edev, "Link not ready (0x%04x)\n", val);
765}
766
767#define BYTE_SWAP(OFF) (8*((OFF)/4)+3-(OFF))
768#define SAVED_BYTE(OFF) (((u8 *)(edev->config_space))[BYTE_SWAP(OFF)])
769
770static void eeh_restore_bridge_bars(struct eeh_dev *edev)
771{
772 struct pci_dn *pdn = eeh_dev_to_pdn(edev);
773 int i;
774
775
776
777
778
779 for (i = 4; i < 13; i++)
780 eeh_ops->write_config(pdn, i*4, 4, edev->config_space[i]);
781
782 eeh_ops->write_config(pdn, 14*4, 4, edev->config_space[14]);
783
784
785 eeh_ops->write_config(pdn, PCI_CACHE_LINE_SIZE, 1,
786 SAVED_BYTE(PCI_CACHE_LINE_SIZE));
787 eeh_ops->write_config(pdn, PCI_LATENCY_TIMER, 1,
788 SAVED_BYTE(PCI_LATENCY_TIMER));
789
790 eeh_ops->write_config(pdn, 15*4, 4, edev->config_space[15]);
791
792
793 eeh_ops->write_config(pdn, PCI_COMMAND, 4, edev->config_space[1] |
794 PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
795
796
797 eeh_bridge_check_link(edev);
798}
799
800static void eeh_restore_device_bars(struct eeh_dev *edev)
801{
802 struct pci_dn *pdn = eeh_dev_to_pdn(edev);
803 int i;
804 u32 cmd;
805
806 for (i = 4; i < 10; i++)
807 eeh_ops->write_config(pdn, i*4, 4, edev->config_space[i]);
808
809 eeh_ops->write_config(pdn, 12*4, 4, edev->config_space[12]);
810
811 eeh_ops->write_config(pdn, PCI_CACHE_LINE_SIZE, 1,
812 SAVED_BYTE(PCI_CACHE_LINE_SIZE));
813 eeh_ops->write_config(pdn, PCI_LATENCY_TIMER, 1,
814 SAVED_BYTE(PCI_LATENCY_TIMER));
815
816
817 eeh_ops->write_config(pdn, 15*4, 4, edev->config_space[15]);
818
819
820
821
822
823 eeh_ops->read_config(pdn, PCI_COMMAND, 4, &cmd);
824 if (edev->config_space[1] & PCI_COMMAND_PARITY)
825 cmd |= PCI_COMMAND_PARITY;
826 else
827 cmd &= ~PCI_COMMAND_PARITY;
828 if (edev->config_space[1] & PCI_COMMAND_SERR)
829 cmd |= PCI_COMMAND_SERR;
830 else
831 cmd &= ~PCI_COMMAND_SERR;
832 eeh_ops->write_config(pdn, PCI_COMMAND, 4, cmd);
833}
834
835
836
837
838
839
840
841
842
843
844static void eeh_restore_one_device_bars(struct eeh_dev *edev, void *flag)
845{
846 struct pci_dn *pdn = eeh_dev_to_pdn(edev);
847
848
849 if (edev->mode & EEH_DEV_BRIDGE)
850 eeh_restore_bridge_bars(edev);
851 else
852 eeh_restore_device_bars(edev);
853
854 if (eeh_ops->restore_config && pdn)
855 eeh_ops->restore_config(pdn);
856}
857
858
859
860
861
862
863
864
865void eeh_pe_restore_bars(struct eeh_pe *pe)
866{
867
868
869
870
871 eeh_pe_dev_traverse(pe, eeh_restore_one_device_bars, NULL);
872}
873
874
875
876
877
878
879
880
881
882
883const char *eeh_pe_loc_get(struct eeh_pe *pe)
884{
885 struct pci_bus *bus = eeh_pe_bus_get(pe);
886 struct device_node *dn;
887 const char *loc = NULL;
888
889 while (bus) {
890 dn = pci_bus_to_OF_node(bus);
891 if (!dn) {
892 bus = bus->parent;
893 continue;
894 }
895
896 if (pci_is_root_bus(bus))
897 loc = of_get_property(dn, "ibm,io-base-loc-code", NULL);
898 else
899 loc = of_get_property(dn, "ibm,slot-location-code",
900 NULL);
901
902 if (loc)
903 return loc;
904
905 bus = bus->parent;
906 }
907
908 return "N/A";
909}
910
911
912
913
914
915
916
917
918
919
920
921struct pci_bus *eeh_pe_bus_get(struct eeh_pe *pe)
922{
923 struct eeh_dev *edev;
924 struct pci_dev *pdev;
925
926 if (pe->type & EEH_PE_PHB)
927 return pe->phb->bus;
928
929
930 if (pe->state & EEH_PE_PRI_BUS)
931 return pe->bus;
932
933
934 edev = list_first_entry_or_null(&pe->edevs, struct eeh_dev, entry);
935 pdev = eeh_dev_to_pci_dev(edev);
936 if (pdev)
937 return pdev->bus;
938
939 return NULL;
940}
941