1
2
3
4
5
6
7
8
9
10
11
12#include <linux/module.h>
13#include <linux/pci.h>
14#include <linux/init.h>
15#include <linux/agp_backend.h>
16#include <linux/mmzone.h>
17#include <asm/page.h>
18#include <asm/e820/api.h>
19#include <asm/amd_nb.h>
20#include <asm/gart.h>
21#include "agp.h"
22
23
24#define NVIDIA_X86_64_0_APBASE 0x10
25#define NVIDIA_X86_64_1_APBASE1 0x50
26#define NVIDIA_X86_64_1_APLIMIT1 0x54
27#define NVIDIA_X86_64_1_APSIZE 0xa8
28#define NVIDIA_X86_64_1_APBASE2 0xd8
29#define NVIDIA_X86_64_1_APLIMIT2 0xdc
30
31
32#define ULI_X86_64_BASE_ADDR 0x10
33#define ULI_X86_64_HTT_FEA_REG 0x50
34#define ULI_X86_64_ENU_SCR_REG 0x54
35
36static struct resource *aperture_resource;
37static bool __initdata agp_try_unsupported = 1;
38static int agp_bridges_found;
39
40static void amd64_tlbflush(struct agp_memory *temp)
41{
42 amd_flush_garts();
43}
44
45static int amd64_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
46{
47 int i, j, num_entries;
48 long long tmp;
49 int mask_type;
50 struct agp_bridge_data *bridge = mem->bridge;
51 u32 pte;
52
53 num_entries = agp_num_entries();
54
55 if (type != mem->type)
56 return -EINVAL;
57 mask_type = bridge->driver->agp_type_to_mask_type(bridge, type);
58 if (mask_type != 0)
59 return -EINVAL;
60
61
62
63
64 if (((unsigned long)pg_start + mem->page_count) > num_entries)
65 return -EINVAL;
66
67 j = pg_start;
68
69
70 while (j < (pg_start + mem->page_count)) {
71 if (!PGE_EMPTY(agp_bridge, readl(agp_bridge->gatt_table+j)))
72 return -EBUSY;
73 j++;
74 }
75
76 if (!mem->is_flushed) {
77 global_cache_flush();
78 mem->is_flushed = true;
79 }
80
81 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
82 tmp = agp_bridge->driver->mask_memory(agp_bridge,
83 page_to_phys(mem->pages[i]),
84 mask_type);
85
86 BUG_ON(tmp & 0xffffff0000000ffcULL);
87 pte = (tmp & 0x000000ff00000000ULL) >> 28;
88 pte |=(tmp & 0x00000000fffff000ULL);
89 pte |= GPTE_VALID | GPTE_COHERENT;
90
91 writel(pte, agp_bridge->gatt_table+j);
92 readl(agp_bridge->gatt_table+j);
93 }
94 amd64_tlbflush(mem);
95 return 0;
96}
97
98
99
100
101
102
103static struct aper_size_info_32 amd64_aperture_sizes[7] =
104{
105 {32, 8192, 3+(sizeof(long)/8), 0 },
106 {64, 16384, 4+(sizeof(long)/8), 1<<1 },
107 {128, 32768, 5+(sizeof(long)/8), 1<<2 },
108 {256, 65536, 6+(sizeof(long)/8), 1<<1 | 1<<2 },
109 {512, 131072, 7+(sizeof(long)/8), 1<<3 },
110 {1024, 262144, 8+(sizeof(long)/8), 1<<1 | 1<<3},
111 {2048, 524288, 9+(sizeof(long)/8), 1<<2 | 1<<3}
112};
113
114
115
116
117
118
119
120
121static int amd64_fetch_size(void)
122{
123 struct pci_dev *dev;
124 int i;
125 u32 temp;
126 struct aper_size_info_32 *values;
127
128 dev = node_to_amd_nb(0)->misc;
129 if (dev==NULL)
130 return 0;
131
132 pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &temp);
133 temp = (temp & 0xe);
134 values = A_SIZE_32(amd64_aperture_sizes);
135
136 for (i = 0; i < agp_bridge->driver->num_aperture_sizes; i++) {
137 if (temp == values[i].size_value) {
138 agp_bridge->previous_size =
139 agp_bridge->current_size = (void *) (values + i);
140
141 agp_bridge->aperture_size_idx = i;
142 return values[i].size;
143 }
144 }
145 return 0;
146}
147
148
149
150
151
152static u64 amd64_configure(struct pci_dev *hammer, u64 gatt_table)
153{
154 u64 aperturebase;
155 u32 tmp;
156 u64 aper_base;
157
158
159 pci_read_config_dword(hammer, AMD64_GARTAPERTUREBASE, &tmp);
160 aperturebase = (u64)tmp << 25;
161 aper_base = (aperturebase & PCI_BASE_ADDRESS_MEM_MASK);
162
163 enable_gart_translation(hammer, gatt_table);
164
165 return aper_base;
166}
167
168
169static const struct aper_size_info_32 amd_8151_sizes[7] =
170{
171 {2048, 524288, 9, 0x00000000 },
172 {1024, 262144, 8, 0x00000400 },
173 {512, 131072, 7, 0x00000600 },
174 {256, 65536, 6, 0x00000700 },
175 {128, 32768, 5, 0x00000720 },
176 {64, 16384, 4, 0x00000730 },
177 {32, 8192, 3, 0x00000738 }
178};
179
180static int amd_8151_configure(void)
181{
182 unsigned long gatt_bus = virt_to_phys(agp_bridge->gatt_table_real);
183 int i;
184
185 if (!amd_nb_has_feature(AMD_NB_GART))
186 return 0;
187
188
189 for (i = 0; i < amd_nb_num(); i++) {
190 agp_bridge->gart_bus_addr =
191 amd64_configure(node_to_amd_nb(i)->misc, gatt_bus);
192 }
193 amd_flush_garts();
194 return 0;
195}
196
197
198static void amd64_cleanup(void)
199{
200 u32 tmp;
201 int i;
202
203 if (!amd_nb_has_feature(AMD_NB_GART))
204 return;
205
206 for (i = 0; i < amd_nb_num(); i++) {
207 struct pci_dev *dev = node_to_amd_nb(i)->misc;
208
209 pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &tmp);
210 tmp &= ~GARTEN;
211 pci_write_config_dword(dev, AMD64_GARTAPERTURECTL, tmp);
212 }
213}
214
215
216static const struct agp_bridge_driver amd_8151_driver = {
217 .owner = THIS_MODULE,
218 .aperture_sizes = amd_8151_sizes,
219 .size_type = U32_APER_SIZE,
220 .num_aperture_sizes = 7,
221 .needs_scratch_page = true,
222 .configure = amd_8151_configure,
223 .fetch_size = amd64_fetch_size,
224 .cleanup = amd64_cleanup,
225 .tlb_flush = amd64_tlbflush,
226 .mask_memory = agp_generic_mask_memory,
227 .masks = NULL,
228 .agp_enable = agp_generic_enable,
229 .cache_flush = global_cache_flush,
230 .create_gatt_table = agp_generic_create_gatt_table,
231 .free_gatt_table = agp_generic_free_gatt_table,
232 .insert_memory = amd64_insert_memory,
233 .remove_memory = agp_generic_remove_memory,
234 .alloc_by_type = agp_generic_alloc_by_type,
235 .free_by_type = agp_generic_free_by_type,
236 .agp_alloc_page = agp_generic_alloc_page,
237 .agp_alloc_pages = agp_generic_alloc_pages,
238 .agp_destroy_page = agp_generic_destroy_page,
239 .agp_destroy_pages = agp_generic_destroy_pages,
240 .agp_type_to_mask_type = agp_generic_type_to_mask_type,
241};
242
243
244static int agp_aperture_valid(u64 aper, u32 size)
245{
246 if (!aperture_valid(aper, size, 32*1024*1024))
247 return 0;
248
249
250
251
252
253
254 if (!aperture_resource &&
255 !(aperture_resource = request_mem_region(aper, size, "aperture"))) {
256 printk(KERN_ERR PFX "Aperture conflicts with PCI mapping.\n");
257 return 0;
258 }
259 return 1;
260}
261
262
263
264
265
266
267
268
269
270
271static int fix_northbridge(struct pci_dev *nb, struct pci_dev *agp, u16 cap)
272{
273 u64 aper, nb_aper;
274 int order = 0;
275 u32 nb_order, nb_base;
276 u16 apsize;
277
278 pci_read_config_dword(nb, AMD64_GARTAPERTURECTL, &nb_order);
279 nb_order = (nb_order >> 1) & 7;
280 pci_read_config_dword(nb, AMD64_GARTAPERTUREBASE, &nb_base);
281 nb_aper = (u64)nb_base << 25;
282
283
284
285 pci_read_config_word(agp, cap+0x14, &apsize);
286 if (apsize == 0xffff) {
287 if (agp_aperture_valid(nb_aper, (32*1024*1024)<<nb_order))
288 return 0;
289 return -1;
290 }
291
292 apsize &= 0xfff;
293
294 if (apsize & 0xff)
295 apsize |= 0xf00;
296 order = 7 - hweight16(apsize);
297
298 aper = pci_bus_address(agp, AGP_APERTURE_BAR);
299
300
301
302
303
304 if (order >=0 && aper + (32ULL<<(20 + order)) > 0x100000000ULL) {
305 dev_info(&agp->dev, "aperture size %u MB is not right, using settings from NB\n",
306 32 << order);
307 order = nb_order;
308 }
309
310 if (nb_order >= order) {
311 if (agp_aperture_valid(nb_aper, (32*1024*1024)<<nb_order))
312 return 0;
313 }
314
315 dev_info(&agp->dev, "aperture from AGP @ %Lx size %u MB\n",
316 aper, 32 << order);
317 if (order < 0 || !agp_aperture_valid(aper, (32*1024*1024)<<order))
318 return -1;
319
320 gart_set_size_and_enable(nb, order);
321 pci_write_config_dword(nb, AMD64_GARTAPERTUREBASE, aper >> 25);
322
323 return 0;
324}
325
326static int cache_nbs(struct pci_dev *pdev, u32 cap_ptr)
327{
328 int i;
329
330 if (amd_cache_northbridges() < 0)
331 return -ENODEV;
332
333 if (!amd_nb_has_feature(AMD_NB_GART))
334 return -ENODEV;
335
336 i = 0;
337 for (i = 0; i < amd_nb_num(); i++) {
338 struct pci_dev *dev = node_to_amd_nb(i)->misc;
339 if (fix_northbridge(dev, pdev, cap_ptr) < 0) {
340 dev_err(&dev->dev, "no usable aperture found\n");
341#ifdef __x86_64__
342
343 dev_err(&dev->dev, "consider rebooting with iommu=memaper=2 to get a good aperture\n");
344#endif
345 return -1;
346 }
347 }
348 return 0;
349}
350
351
352static void amd8151_init(struct pci_dev *pdev, struct agp_bridge_data *bridge)
353{
354 char *revstring;
355
356 switch (pdev->revision) {
357 case 0x01: revstring="A0"; break;
358 case 0x02: revstring="A1"; break;
359 case 0x11: revstring="B0"; break;
360 case 0x12: revstring="B1"; break;
361 case 0x13: revstring="B2"; break;
362 case 0x14: revstring="B3"; break;
363 default: revstring="??"; break;
364 }
365
366 dev_info(&pdev->dev, "AMD 8151 AGP Bridge rev %s\n", revstring);
367
368
369
370
371
372 if (pdev->revision < 0x13) {
373 dev_info(&pdev->dev, "correcting AGP revision (reports 3.5, is really 3.0)\n");
374 bridge->major_version = 3;
375 bridge->minor_version = 0;
376 }
377}
378
379
380static const struct aper_size_info_32 uli_sizes[7] =
381{
382 {256, 65536, 6, 10},
383 {128, 32768, 5, 9},
384 {64, 16384, 4, 8},
385 {32, 8192, 3, 7},
386 {16, 4096, 2, 6},
387 {8, 2048, 1, 4},
388 {4, 1024, 0, 3}
389};
390static int uli_agp_init(struct pci_dev *pdev)
391{
392 u32 httfea,baseaddr,enuscr;
393 struct pci_dev *dev1;
394 int i, ret;
395 unsigned size = amd64_fetch_size();
396
397 dev_info(&pdev->dev, "setting up ULi AGP\n");
398 dev1 = pci_get_slot (pdev->bus,PCI_DEVFN(0,0));
399 if (dev1 == NULL) {
400 dev_info(&pdev->dev, "can't find ULi secondary device\n");
401 return -ENODEV;
402 }
403
404 for (i = 0; i < ARRAY_SIZE(uli_sizes); i++)
405 if (uli_sizes[i].size == size)
406 break;
407
408 if (i == ARRAY_SIZE(uli_sizes)) {
409 dev_info(&pdev->dev, "no ULi size found for %d\n", size);
410 ret = -ENODEV;
411 goto put;
412 }
413
414
415 pci_read_config_dword (node_to_amd_nb(0)->misc, AMD64_GARTAPERTUREBASE,
416 &httfea);
417
418
419 if ((httfea & 0x7fff) >> (32 - 25)) {
420 ret = -ENODEV;
421 goto put;
422 }
423
424 httfea = (httfea& 0x7fff) << 25;
425
426 pci_read_config_dword(pdev, ULI_X86_64_BASE_ADDR, &baseaddr);
427 baseaddr&= ~PCI_BASE_ADDRESS_MEM_MASK;
428 baseaddr|= httfea;
429 pci_write_config_dword(pdev, ULI_X86_64_BASE_ADDR, baseaddr);
430
431 enuscr= httfea+ (size * 1024 * 1024) - 1;
432 pci_write_config_dword(dev1, ULI_X86_64_HTT_FEA_REG, httfea);
433 pci_write_config_dword(dev1, ULI_X86_64_ENU_SCR_REG, enuscr);
434 ret = 0;
435put:
436 pci_dev_put(dev1);
437 return ret;
438}
439
440
441static const struct aper_size_info_32 nforce3_sizes[5] =
442{
443 {512, 131072, 7, 0x00000000 },
444 {256, 65536, 6, 0x00000008 },
445 {128, 32768, 5, 0x0000000C },
446 {64, 16384, 4, 0x0000000E },
447 {32, 8192, 3, 0x0000000F }
448};
449
450
451
452static int nforce3_agp_init(struct pci_dev *pdev)
453{
454 u32 tmp, apbase, apbar, aplimit;
455 struct pci_dev *dev1;
456 int i, ret;
457 unsigned size = amd64_fetch_size();
458
459 dev_info(&pdev->dev, "setting up Nforce3 AGP\n");
460
461 dev1 = pci_get_slot(pdev->bus, PCI_DEVFN(11, 0));
462 if (dev1 == NULL) {
463 dev_info(&pdev->dev, "can't find Nforce3 secondary device\n");
464 return -ENODEV;
465 }
466
467 for (i = 0; i < ARRAY_SIZE(nforce3_sizes); i++)
468 if (nforce3_sizes[i].size == size)
469 break;
470
471 if (i == ARRAY_SIZE(nforce3_sizes)) {
472 dev_info(&pdev->dev, "no NForce3 size found for %d\n", size);
473 ret = -ENODEV;
474 goto put;
475 }
476
477 pci_read_config_dword(dev1, NVIDIA_X86_64_1_APSIZE, &tmp);
478 tmp &= ~(0xf);
479 tmp |= nforce3_sizes[i].size_value;
480 pci_write_config_dword(dev1, NVIDIA_X86_64_1_APSIZE, tmp);
481
482
483 pci_read_config_dword (node_to_amd_nb(0)->misc, AMD64_GARTAPERTUREBASE,
484 &apbase);
485
486
487 if ( (apbase & 0x7fff) >> (32 - 25) ) {
488 dev_info(&pdev->dev, "aperture base > 4G\n");
489 ret = -ENODEV;
490 goto put;
491 }
492
493 apbase = (apbase & 0x7fff) << 25;
494
495 pci_read_config_dword(pdev, NVIDIA_X86_64_0_APBASE, &apbar);
496 apbar &= ~PCI_BASE_ADDRESS_MEM_MASK;
497 apbar |= apbase;
498 pci_write_config_dword(pdev, NVIDIA_X86_64_0_APBASE, apbar);
499
500 aplimit = apbase + (size * 1024 * 1024) - 1;
501 pci_write_config_dword(dev1, NVIDIA_X86_64_1_APBASE1, apbase);
502 pci_write_config_dword(dev1, NVIDIA_X86_64_1_APLIMIT1, aplimit);
503 pci_write_config_dword(dev1, NVIDIA_X86_64_1_APBASE2, apbase);
504 pci_write_config_dword(dev1, NVIDIA_X86_64_1_APLIMIT2, aplimit);
505
506 ret = 0;
507put:
508 pci_dev_put(dev1);
509
510 return ret;
511}
512
513static int agp_amd64_probe(struct pci_dev *pdev,
514 const struct pci_device_id *ent)
515{
516 struct agp_bridge_data *bridge;
517 u8 cap_ptr;
518 int err;
519
520
521 if (agp_bridges_found)
522 return -ENODEV;
523
524 cap_ptr = pci_find_capability(pdev, PCI_CAP_ID_AGP);
525 if (!cap_ptr)
526 return -ENODEV;
527
528
529
530 bridge = agp_alloc_bridge();
531 if (!bridge)
532 return -ENOMEM;
533
534 if (pdev->vendor == PCI_VENDOR_ID_AMD &&
535 pdev->device == PCI_DEVICE_ID_AMD_8151_0) {
536 amd8151_init(pdev, bridge);
537 } else {
538 dev_info(&pdev->dev, "AGP bridge [%04x/%04x]\n",
539 pdev->vendor, pdev->device);
540 }
541
542 bridge->driver = &amd_8151_driver;
543 bridge->dev = pdev;
544 bridge->capndx = cap_ptr;
545
546
547 pci_read_config_dword(pdev, bridge->capndx+PCI_AGP_STATUS, &bridge->mode);
548
549 if (cache_nbs(pdev, cap_ptr) == -1) {
550 agp_put_bridge(bridge);
551 return -ENODEV;
552 }
553
554 if (pdev->vendor == PCI_VENDOR_ID_NVIDIA) {
555 int ret = nforce3_agp_init(pdev);
556 if (ret) {
557 agp_put_bridge(bridge);
558 return ret;
559 }
560 }
561
562 if (pdev->vendor == PCI_VENDOR_ID_AL) {
563 int ret = uli_agp_init(pdev);
564 if (ret) {
565 agp_put_bridge(bridge);
566 return ret;
567 }
568 }
569
570 pci_set_drvdata(pdev, bridge);
571 err = agp_add_bridge(bridge);
572 if (err < 0)
573 return err;
574
575 agp_bridges_found++;
576 return 0;
577}
578
579static void agp_amd64_remove(struct pci_dev *pdev)
580{
581 struct agp_bridge_data *bridge = pci_get_drvdata(pdev);
582
583 release_mem_region(virt_to_phys(bridge->gatt_table_real),
584 amd64_aperture_sizes[bridge->aperture_size_idx].size);
585 agp_remove_bridge(bridge);
586 agp_put_bridge(bridge);
587
588 agp_bridges_found--;
589}
590
591#ifdef CONFIG_PM
592
593static int agp_amd64_suspend(struct pci_dev *pdev, pm_message_t state)
594{
595 pci_save_state(pdev);
596 pci_set_power_state(pdev, pci_choose_state(pdev, state));
597
598 return 0;
599}
600
601static int agp_amd64_resume(struct pci_dev *pdev)
602{
603 pci_set_power_state(pdev, PCI_D0);
604 pci_restore_state(pdev);
605
606 if (pdev->vendor == PCI_VENDOR_ID_NVIDIA)
607 nforce3_agp_init(pdev);
608
609 return amd_8151_configure();
610}
611
612#endif
613
614static const struct pci_device_id agp_amd64_pci_table[] = {
615 {
616 .class = (PCI_CLASS_BRIDGE_HOST << 8),
617 .class_mask = ~0,
618 .vendor = PCI_VENDOR_ID_AMD,
619 .device = PCI_DEVICE_ID_AMD_8151_0,
620 .subvendor = PCI_ANY_ID,
621 .subdevice = PCI_ANY_ID,
622 },
623
624 {
625 .class = (PCI_CLASS_BRIDGE_HOST << 8),
626 .class_mask = ~0,
627 .vendor = PCI_VENDOR_ID_AL,
628 .device = PCI_DEVICE_ID_AL_M1689,
629 .subvendor = PCI_ANY_ID,
630 .subdevice = PCI_ANY_ID,
631 },
632
633 {
634 .class = (PCI_CLASS_BRIDGE_HOST << 8),
635 .class_mask = ~0,
636 .vendor = PCI_VENDOR_ID_VIA,
637 .device = PCI_DEVICE_ID_VIA_K8T800PRO_0,
638 .subvendor = PCI_ANY_ID,
639 .subdevice = PCI_ANY_ID,
640 },
641
642 {
643 .class = (PCI_CLASS_BRIDGE_HOST << 8),
644 .class_mask = ~0,
645 .vendor = PCI_VENDOR_ID_VIA,
646 .device = PCI_DEVICE_ID_VIA_8385_0,
647 .subvendor = PCI_ANY_ID,
648 .subdevice = PCI_ANY_ID,
649 },
650
651 {
652 .class = (PCI_CLASS_BRIDGE_HOST << 8),
653 .class_mask = ~0,
654 .vendor = PCI_VENDOR_ID_VIA,
655 .device = PCI_DEVICE_ID_VIA_8380_0,
656 .subvendor = PCI_ANY_ID,
657 .subdevice = PCI_ANY_ID,
658 },
659
660 {
661 .class = (PCI_CLASS_BRIDGE_HOST << 8),
662 .class_mask = ~0,
663 .vendor = PCI_VENDOR_ID_VIA,
664 .device = PCI_DEVICE_ID_VIA_VT3336,
665 .subvendor = PCI_ANY_ID,
666 .subdevice = PCI_ANY_ID,
667 },
668
669 {
670 .class = (PCI_CLASS_BRIDGE_HOST << 8),
671 .class_mask = ~0,
672 .vendor = PCI_VENDOR_ID_VIA,
673 .device = PCI_DEVICE_ID_VIA_3238_0,
674 .subvendor = PCI_ANY_ID,
675 .subdevice = PCI_ANY_ID,
676 },
677
678 {
679 .class = (PCI_CLASS_BRIDGE_HOST << 8),
680 .class_mask = ~0,
681 .vendor = PCI_VENDOR_ID_VIA,
682 .device = PCI_DEVICE_ID_VIA_838X_1,
683 .subvendor = PCI_ANY_ID,
684 .subdevice = PCI_ANY_ID,
685 },
686
687 {
688 .class = (PCI_CLASS_BRIDGE_HOST << 8),
689 .class_mask = ~0,
690 .vendor = PCI_VENDOR_ID_NVIDIA,
691 .device = PCI_DEVICE_ID_NVIDIA_NFORCE3,
692 .subvendor = PCI_ANY_ID,
693 .subdevice = PCI_ANY_ID,
694 },
695 {
696 .class = (PCI_CLASS_BRIDGE_HOST << 8),
697 .class_mask = ~0,
698 .vendor = PCI_VENDOR_ID_NVIDIA,
699 .device = PCI_DEVICE_ID_NVIDIA_NFORCE3S,
700 .subvendor = PCI_ANY_ID,
701 .subdevice = PCI_ANY_ID,
702 },
703
704 {
705 .class = (PCI_CLASS_BRIDGE_HOST << 8),
706 .class_mask = ~0,
707 .vendor = PCI_VENDOR_ID_SI,
708 .device = PCI_DEVICE_ID_SI_755,
709 .subvendor = PCI_ANY_ID,
710 .subdevice = PCI_ANY_ID,
711 },
712
713 {
714 .class = (PCI_CLASS_BRIDGE_HOST << 8),
715 .class_mask = ~0,
716 .vendor = PCI_VENDOR_ID_SI,
717 .device = PCI_DEVICE_ID_SI_760,
718 .subvendor = PCI_ANY_ID,
719 .subdevice = PCI_ANY_ID,
720 },
721
722 {
723 .class = (PCI_CLASS_BRIDGE_HOST << 8),
724 .class_mask = ~0,
725 .vendor = PCI_VENDOR_ID_AL,
726 .device = 0x1695,
727 .subvendor = PCI_ANY_ID,
728 .subdevice = PCI_ANY_ID,
729 },
730
731 { }
732};
733
734MODULE_DEVICE_TABLE(pci, agp_amd64_pci_table);
735
736static const struct pci_device_id agp_amd64_pci_promisc_table[] = {
737 { PCI_DEVICE_CLASS(0, 0) },
738 { }
739};
740
741static struct pci_driver agp_amd64_pci_driver = {
742 .name = "agpgart-amd64",
743 .id_table = agp_amd64_pci_table,
744 .probe = agp_amd64_probe,
745 .remove = agp_amd64_remove,
746#ifdef CONFIG_PM
747 .suspend = agp_amd64_suspend,
748 .resume = agp_amd64_resume,
749#endif
750};
751
752
753
754int __init agp_amd64_init(void)
755{
756 int err = 0;
757
758 if (agp_off)
759 return -EINVAL;
760
761 err = pci_register_driver(&agp_amd64_pci_driver);
762 if (err < 0)
763 return err;
764
765 if (agp_bridges_found == 0) {
766 if (!agp_try_unsupported && !agp_try_unsupported_boot) {
767 printk(KERN_INFO PFX "No supported AGP bridge found.\n");
768#ifdef MODULE
769 printk(KERN_INFO PFX "You can try agp_try_unsupported=1\n");
770#else
771 printk(KERN_INFO PFX "You can boot with agp=try_unsupported\n");
772#endif
773 pci_unregister_driver(&agp_amd64_pci_driver);
774 return -ENODEV;
775 }
776
777
778 if (!amd_nb_num()) {
779 pci_unregister_driver(&agp_amd64_pci_driver);
780 return -ENODEV;
781 }
782
783
784 agp_amd64_pci_driver.id_table = agp_amd64_pci_promisc_table;
785 err = driver_attach(&agp_amd64_pci_driver.driver);
786 if (err == 0 && agp_bridges_found == 0) {
787 pci_unregister_driver(&agp_amd64_pci_driver);
788 err = -ENODEV;
789 }
790 }
791 return err;
792}
793
794static int __init agp_amd64_mod_init(void)
795{
796#ifndef MODULE
797 if (gart_iommu_aperture)
798 return agp_bridges_found ? 0 : -ENODEV;
799#endif
800 return agp_amd64_init();
801}
802
803static void __exit agp_amd64_cleanup(void)
804{
805#ifndef MODULE
806 if (gart_iommu_aperture)
807 return;
808#endif
809 if (aperture_resource)
810 release_resource(aperture_resource);
811 pci_unregister_driver(&agp_amd64_pci_driver);
812}
813
814module_init(agp_amd64_mod_init);
815module_exit(agp_amd64_cleanup);
816
817MODULE_AUTHOR("Dave Jones, Andi Kleen");
818module_param(agp_try_unsupported, bool, 0);
819MODULE_LICENSE("GPL");
820