1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30#include <linux/pci.h>
31#include <linux/interrupt.h>
32#include <linux/slab.h>
33#include <linux/i2o.h>
34#include <linux/module.h>
35#include "core.h"
36
37#define OSM_DESCRIPTION "I2O-subsystem"
38
39
40static struct pci_device_id __devinitdata i2o_pci_ids[] = {
41 {PCI_DEVICE_CLASS(PCI_CLASS_INTELLIGENT_I2O << 8, 0xffff00)},
42 {PCI_DEVICE(PCI_VENDOR_ID_DPT, 0xa511)},
43 {.vendor = PCI_VENDOR_ID_INTEL,.device = 0x1962,
44 .subvendor = PCI_VENDOR_ID_PROMISE,.subdevice = PCI_ANY_ID},
45 {0}
46};
47
48
49
50
51
52
53
54
55static void i2o_pci_free(struct i2o_controller *c)
56{
57 struct device *dev;
58
59 dev = &c->pdev->dev;
60
61 i2o_dma_free(dev, &c->out_queue);
62 i2o_dma_free(dev, &c->status_block);
63 kfree(c->lct);
64 i2o_dma_free(dev, &c->dlct);
65 i2o_dma_free(dev, &c->hrt);
66 i2o_dma_free(dev, &c->status);
67
68 if (c->raptor && c->in_queue.virt)
69 iounmap(c->in_queue.virt);
70
71 if (c->base.virt)
72 iounmap(c->base.virt);
73
74 pci_release_regions(c->pdev);
75}
76
77
78
79
80
81
82
83
84
85
86
87static int __devinit i2o_pci_alloc(struct i2o_controller *c)
88{
89 struct pci_dev *pdev = c->pdev;
90 struct device *dev = &pdev->dev;
91 int i;
92
93 if (pci_request_regions(pdev, OSM_DESCRIPTION)) {
94 printk(KERN_ERR "%s: device already claimed\n", c->name);
95 return -ENODEV;
96 }
97
98 for (i = 0; i < 6; i++) {
99
100 if (!(pci_resource_flags(pdev, i) & IORESOURCE_IO)) {
101 if (!c->base.phys) {
102 c->base.phys = pci_resource_start(pdev, i);
103 c->base.len = pci_resource_len(pdev, i);
104
105
106
107
108
109 if (pdev->device == 0xa501) {
110 if (pdev->subsystem_device >= 0xc032 &&
111 pdev->subsystem_device <= 0xc03b) {
112 if (c->base.len > 0x400000)
113 c->base.len = 0x400000;
114 } else {
115 if (c->base.len > 0x100000)
116 c->base.len = 0x100000;
117 }
118 }
119 if (!c->raptor)
120 break;
121 } else {
122 c->in_queue.phys = pci_resource_start(pdev, i);
123 c->in_queue.len = pci_resource_len(pdev, i);
124 break;
125 }
126 }
127 }
128
129 if (i == 6) {
130 printk(KERN_ERR "%s: I2O controller has no memory regions"
131 " defined.\n", c->name);
132 i2o_pci_free(c);
133 return -EINVAL;
134 }
135
136
137 if (c->raptor) {
138 printk(KERN_INFO "%s: PCI I2O controller\n", c->name);
139 printk(KERN_INFO " BAR0 at 0x%08lX size=%ld\n",
140 (unsigned long)c->base.phys, (unsigned long)c->base.len);
141 printk(KERN_INFO " BAR1 at 0x%08lX size=%ld\n",
142 (unsigned long)c->in_queue.phys,
143 (unsigned long)c->in_queue.len);
144 } else
145 printk(KERN_INFO "%s: PCI I2O controller at %08lX size=%ld\n",
146 c->name, (unsigned long)c->base.phys,
147 (unsigned long)c->base.len);
148
149 c->base.virt = ioremap_nocache(c->base.phys, c->base.len);
150 if (!c->base.virt) {
151 printk(KERN_ERR "%s: Unable to map controller.\n", c->name);
152 i2o_pci_free(c);
153 return -ENOMEM;
154 }
155
156 if (c->raptor) {
157 c->in_queue.virt =
158 ioremap_nocache(c->in_queue.phys, c->in_queue.len);
159 if (!c->in_queue.virt) {
160 printk(KERN_ERR "%s: Unable to map controller.\n",
161 c->name);
162 i2o_pci_free(c);
163 return -ENOMEM;
164 }
165 } else
166 c->in_queue = c->base;
167
168 c->irq_status = c->base.virt + I2O_IRQ_STATUS;
169 c->irq_mask = c->base.virt + I2O_IRQ_MASK;
170 c->in_port = c->base.virt + I2O_IN_PORT;
171 c->out_port = c->base.virt + I2O_OUT_PORT;
172
173
174 if (pdev->vendor == PCI_VENDOR_ID_MOTOROLA && pdev->device == 0x18c0) {
175
176 if (be32_to_cpu(readl(c->base.virt + 0x10000)) & 0x10000000) {
177 printk(KERN_INFO "%s: MPC82XX needs CPU running to "
178 "service I2O.\n", c->name);
179 i2o_pci_free(c);
180 return -ENODEV;
181 } else {
182 c->irq_status += I2O_MOTOROLA_PORT_OFFSET;
183 c->irq_mask += I2O_MOTOROLA_PORT_OFFSET;
184 c->in_port += I2O_MOTOROLA_PORT_OFFSET;
185 c->out_port += I2O_MOTOROLA_PORT_OFFSET;
186 printk(KERN_INFO "%s: MPC82XX workarounds activated.\n",
187 c->name);
188 }
189 }
190
191 if (i2o_dma_alloc(dev, &c->status, 8)) {
192 i2o_pci_free(c);
193 return -ENOMEM;
194 }
195
196 if (i2o_dma_alloc(dev, &c->hrt, sizeof(i2o_hrt))) {
197 i2o_pci_free(c);
198 return -ENOMEM;
199 }
200
201 if (i2o_dma_alloc(dev, &c->dlct, 8192)) {
202 i2o_pci_free(c);
203 return -ENOMEM;
204 }
205
206 if (i2o_dma_alloc(dev, &c->status_block, sizeof(i2o_status_block))) {
207 i2o_pci_free(c);
208 return -ENOMEM;
209 }
210
211 if (i2o_dma_alloc(dev, &c->out_queue,
212 I2O_MAX_OUTBOUND_MSG_FRAMES * I2O_OUTBOUND_MSG_FRAME_SIZE *
213 sizeof(u32))) {
214 i2o_pci_free(c);
215 return -ENOMEM;
216 }
217
218 pci_set_drvdata(pdev, c);
219
220 return 0;
221}
222
223
224
225
226
227
228
229
230
231static irqreturn_t i2o_pci_interrupt(int irq, void *dev_id)
232{
233 struct i2o_controller *c = dev_id;
234 u32 m;
235 irqreturn_t rc = IRQ_NONE;
236
237 while (readl(c->irq_status) & I2O_IRQ_OUTBOUND_POST) {
238 m = readl(c->out_port);
239 if (m == I2O_QUEUE_EMPTY) {
240
241
242
243
244 m = readl(c->out_port);
245 if (unlikely(m == I2O_QUEUE_EMPTY))
246 break;
247 }
248
249
250 if (i2o_driver_dispatch(c, m))
251
252 i2o_flush_reply(c, m);
253
254 rc = IRQ_HANDLED;
255 }
256
257 return rc;
258}
259
260
261
262
263
264
265
266
267
268
269static int i2o_pci_irq_enable(struct i2o_controller *c)
270{
271 struct pci_dev *pdev = c->pdev;
272 int rc;
273
274 writel(0xffffffff, c->irq_mask);
275
276 if (pdev->irq) {
277 rc = request_irq(pdev->irq, i2o_pci_interrupt, IRQF_SHARED,
278 c->name, c);
279 if (rc < 0) {
280 printk(KERN_ERR "%s: unable to allocate interrupt %d."
281 "\n", c->name, pdev->irq);
282 return rc;
283 }
284 }
285
286 writel(0x00000000, c->irq_mask);
287
288 printk(KERN_INFO "%s: Installed at IRQ %d\n", c->name, pdev->irq);
289
290 return 0;
291}
292
293
294
295
296
297
298
299static void i2o_pci_irq_disable(struct i2o_controller *c)
300{
301 writel(0xffffffff, c->irq_mask);
302
303 if (c->pdev->irq > 0)
304 free_irq(c->pdev->irq, c);
305}
306
307
308
309
310
311
312
313
314
315
316
317
318static int __devinit i2o_pci_probe(struct pci_dev *pdev,
319 const struct pci_device_id *id)
320{
321 struct i2o_controller *c;
322 int rc;
323 struct pci_dev *i960 = NULL;
324
325 printk(KERN_INFO "i2o: Checking for PCI I2O controllers...\n");
326
327 if ((pdev->class & 0xff) > 1) {
328 printk(KERN_WARNING "i2o: %s does not support I2O 1.5 "
329 "(skipping).\n", pci_name(pdev));
330 return -ENODEV;
331 }
332
333 if ((rc = pci_enable_device(pdev))) {
334 printk(KERN_WARNING "i2o: couldn't enable device %s\n",
335 pci_name(pdev));
336 return rc;
337 }
338
339 if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
340 printk(KERN_WARNING "i2o: no suitable DMA found for %s\n",
341 pci_name(pdev));
342 rc = -ENODEV;
343 goto disable;
344 }
345
346 pci_set_master(pdev);
347
348 c = i2o_iop_alloc();
349 if (IS_ERR(c)) {
350 printk(KERN_ERR "i2o: couldn't allocate memory for %s\n",
351 pci_name(pdev));
352 rc = PTR_ERR(c);
353 goto disable;
354 } else
355 printk(KERN_INFO "%s: controller found (%s)\n", c->name,
356 pci_name(pdev));
357
358 c->pdev = pdev;
359 c->device.parent = &pdev->dev;
360
361
362 if (pdev->vendor == PCI_VENDOR_ID_NCR && pdev->device == 0x0630) {
363 c->short_req = 1;
364 printk(KERN_INFO "%s: Symbios FC920 workarounds activated.\n",
365 c->name);
366 }
367
368 if (pdev->subsystem_vendor == PCI_VENDOR_ID_PROMISE) {
369
370
371
372
373 i960 = pci_get_slot(c->pdev->bus,
374 PCI_DEVFN(PCI_SLOT(c->pdev->devfn), 0));
375
376 if (i960) {
377 pci_write_config_word(i960, 0x42, 0);
378 pci_dev_put(i960);
379 }
380
381 c->promise = 1;
382 c->limit_sectors = 1;
383 }
384
385 if (pdev->subsystem_vendor == PCI_VENDOR_ID_DPT)
386 c->adaptec = 1;
387
388
389 if (pdev->vendor == PCI_VENDOR_ID_DPT) {
390 c->no_quiesce = 1;
391 if (pdev->device == 0xa511)
392 c->raptor = 1;
393
394 if (pdev->subsystem_device == 0xc05a) {
395 c->limit_sectors = 1;
396 printk(KERN_INFO
397 "%s: limit sectors per request to %d\n", c->name,
398 I2O_MAX_SECTORS_LIMITED);
399 }
400#ifdef CONFIG_I2O_EXT_ADAPTEC_DMA64
401 if (sizeof(dma_addr_t) > 4) {
402 if (pci_set_dma_mask(pdev, DMA_BIT_MASK(64)))
403 printk(KERN_INFO "%s: 64-bit DMA unavailable\n",
404 c->name);
405 else {
406 c->pae_support = 1;
407 printk(KERN_INFO "%s: using 64-bit DMA\n",
408 c->name);
409 }
410 }
411#endif
412 }
413
414 if ((rc = i2o_pci_alloc(c))) {
415 printk(KERN_ERR "%s: DMA / IO allocation for I2O controller "
416 "failed\n", c->name);
417 goto free_controller;
418 }
419
420 if (i2o_pci_irq_enable(c)) {
421 printk(KERN_ERR "%s: unable to enable interrupts for I2O "
422 "controller\n", c->name);
423 goto free_pci;
424 }
425
426 if ((rc = i2o_iop_add(c)))
427 goto uninstall;
428
429 if (i960)
430 pci_write_config_word(i960, 0x42, 0x03ff);
431
432 return 0;
433
434 uninstall:
435 i2o_pci_irq_disable(c);
436
437 free_pci:
438 i2o_pci_free(c);
439
440 free_controller:
441 i2o_iop_free(c);
442
443 disable:
444 pci_disable_device(pdev);
445
446 return rc;
447}
448
449
450
451
452
453
454
455
456static void __devexit i2o_pci_remove(struct pci_dev *pdev)
457{
458 struct i2o_controller *c;
459 c = pci_get_drvdata(pdev);
460
461 i2o_iop_remove(c);
462 i2o_pci_irq_disable(c);
463 i2o_pci_free(c);
464
465 pci_disable_device(pdev);
466
467 printk(KERN_INFO "%s: Controller removed.\n", c->name);
468
469 put_device(&c->device);
470};
471
472
473static struct pci_driver i2o_pci_driver = {
474 .name = "PCI_I2O",
475 .id_table = i2o_pci_ids,
476 .probe = i2o_pci_probe,
477 .remove = __devexit_p(i2o_pci_remove),
478};
479
480
481
482
483
484
485int __init i2o_pci_init(void)
486{
487 return pci_register_driver(&i2o_pci_driver);
488};
489
490
491
492
493void __exit i2o_pci_exit(void)
494{
495 pci_unregister_driver(&i2o_pci_driver);
496};
497
498MODULE_DEVICE_TABLE(pci, i2o_pci_ids);
499