1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16#include <linux/module.h>
17#include <linux/i2o.h>
18#include <linux/delay.h>
19#include <linux/string.h>
20#include <linux/slab.h>
21#include "core.h"
22
23
24
25
26
27
28
29
30
31
32
33
34
35static inline int i2o_device_issue_claim(struct i2o_device *dev, u32 cmd,
36 u32 type)
37{
38 struct i2o_message *msg;
39
40 msg = i2o_msg_get_wait(dev->iop, I2O_TIMEOUT_MESSAGE_GET);
41 if (IS_ERR(msg))
42 return PTR_ERR(msg);
43
44 msg->u.head[0] = cpu_to_le32(FIVE_WORD_MSG_SIZE | SGL_OFFSET_0);
45 msg->u.head[1] =
46 cpu_to_le32(cmd << 24 | HOST_TID << 12 | dev->lct_data.tid);
47 msg->body[0] = cpu_to_le32(type);
48
49 return i2o_msg_post_wait(dev->iop, msg, 60);
50}
51
52
53
54
55
56
57
58
59
60int i2o_device_claim(struct i2o_device *dev)
61{
62 int rc = 0;
63
64 mutex_lock(&dev->lock);
65
66 rc = i2o_device_issue_claim(dev, I2O_CMD_UTIL_CLAIM, I2O_CLAIM_PRIMARY);
67 if (!rc)
68 pr_debug("i2o: claim of device %d succeeded\n",
69 dev->lct_data.tid);
70 else
71 pr_debug("i2o: claim of device %d failed %d\n",
72 dev->lct_data.tid, rc);
73
74 mutex_unlock(&dev->lock);
75
76 return rc;
77}
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92int i2o_device_claim_release(struct i2o_device *dev)
93{
94 int tries;
95 int rc = 0;
96
97 mutex_lock(&dev->lock);
98
99
100
101
102
103 for (tries = 0; tries < 10; tries++) {
104 rc = i2o_device_issue_claim(dev, I2O_CMD_UTIL_RELEASE,
105 I2O_CLAIM_PRIMARY);
106 if (!rc)
107 break;
108
109 ssleep(1);
110 }
111
112 if (!rc)
113 pr_debug("i2o: claim release of device %d succeeded\n",
114 dev->lct_data.tid);
115 else
116 pr_debug("i2o: claim release of device %d failed %d\n",
117 dev->lct_data.tid, rc);
118
119 mutex_unlock(&dev->lock);
120
121 return rc;
122}
123
124
125
126
127
128
129
130
131static void i2o_device_release(struct device *dev)
132{
133 struct i2o_device *i2o_dev = to_i2o_device(dev);
134
135 pr_debug("i2o: device %s released\n", dev_name(dev));
136
137 kfree(i2o_dev);
138}
139
140
141
142
143
144
145
146
147
148static ssize_t class_id_show(struct device *dev, struct device_attribute *attr,
149 char *buf)
150{
151 struct i2o_device *i2o_dev = to_i2o_device(dev);
152
153 sprintf(buf, "0x%03x\n", i2o_dev->lct_data.class_id);
154 return strlen(buf) + 1;
155}
156static DEVICE_ATTR_RO(class_id);
157
158
159
160
161
162
163
164
165
166static ssize_t tid_show(struct device *dev, struct device_attribute *attr,
167 char *buf)
168{
169 struct i2o_device *i2o_dev = to_i2o_device(dev);
170
171 sprintf(buf, "0x%03x\n", i2o_dev->lct_data.tid);
172 return strlen(buf) + 1;
173}
174static DEVICE_ATTR_RO(tid);
175
176
177static struct attribute *i2o_device_attrs[] = {
178 &dev_attr_class_id.attr,
179 &dev_attr_tid.attr,
180 NULL,
181};
182
183static const struct attribute_group i2o_device_group = {
184 .attrs = i2o_device_attrs,
185};
186
187const struct attribute_group *i2o_device_groups[] = {
188 &i2o_device_group,
189 NULL,
190};
191
192
193
194
195
196
197
198
199
200static struct i2o_device *i2o_device_alloc(void)
201{
202 struct i2o_device *dev;
203
204 dev = kzalloc(sizeof(*dev), GFP_KERNEL);
205 if (!dev)
206 return ERR_PTR(-ENOMEM);
207
208 INIT_LIST_HEAD(&dev->list);
209 mutex_init(&dev->lock);
210
211 dev->device.bus = &i2o_bus_type;
212 dev->device.release = &i2o_device_release;
213
214 return dev;
215}
216
217
218
219
220
221
222
223
224
225
226
227static int i2o_device_add(struct i2o_controller *c, i2o_lct_entry *entry)
228{
229 struct i2o_device *i2o_dev, *tmp;
230 int rc;
231
232 i2o_dev = i2o_device_alloc();
233 if (IS_ERR(i2o_dev)) {
234 printk(KERN_ERR "i2o: unable to allocate i2o device\n");
235 return PTR_ERR(i2o_dev);
236 }
237
238 i2o_dev->lct_data = *entry;
239
240 dev_set_name(&i2o_dev->device, "%d:%03x", c->unit,
241 i2o_dev->lct_data.tid);
242
243 i2o_dev->iop = c;
244 i2o_dev->device.parent = &c->device;
245
246 rc = device_register(&i2o_dev->device);
247 if (rc)
248 goto err;
249
250 list_add_tail(&i2o_dev->list, &c->devices);
251
252
253 tmp = i2o_iop_find_device(i2o_dev->iop, i2o_dev->lct_data.user_tid);
254 if (tmp && (tmp != i2o_dev)) {
255 rc = sysfs_create_link(&i2o_dev->device.kobj,
256 &tmp->device.kobj, "user");
257 if (rc)
258 goto unreg_dev;
259 }
260
261
262 list_for_each_entry(tmp, &c->devices, list)
263 if ((tmp->lct_data.user_tid == i2o_dev->lct_data.tid)
264 && (tmp != i2o_dev)) {
265 rc = sysfs_create_link(&tmp->device.kobj,
266 &i2o_dev->device.kobj, "user");
267 if (rc)
268 goto rmlink1;
269 }
270
271
272 tmp = i2o_iop_find_device(i2o_dev->iop, i2o_dev->lct_data.parent_tid);
273 if (tmp && (tmp != i2o_dev)) {
274 rc = sysfs_create_link(&i2o_dev->device.kobj,
275 &tmp->device.kobj, "parent");
276 if (rc)
277 goto rmlink1;
278 }
279
280
281 list_for_each_entry(tmp, &c->devices, list)
282 if ((tmp->lct_data.parent_tid == i2o_dev->lct_data.tid)
283 && (tmp != i2o_dev)) {
284 rc = sysfs_create_link(&tmp->device.kobj,
285 &i2o_dev->device.kobj, "parent");
286 if (rc)
287 goto rmlink2;
288 }
289
290 i2o_driver_notify_device_add_all(i2o_dev);
291
292 pr_debug("i2o: device %s added\n", dev_name(&i2o_dev->device));
293
294 return 0;
295
296rmlink2:
297
298
299
300
301 list_for_each_entry(tmp, &c->devices, list) {
302 if (tmp->lct_data.parent_tid == i2o_dev->lct_data.tid)
303 sysfs_remove_link(&tmp->device.kobj, "parent");
304 }
305 sysfs_remove_link(&i2o_dev->device.kobj, "parent");
306rmlink1:
307 list_for_each_entry(tmp, &c->devices, list)
308 if (tmp->lct_data.user_tid == i2o_dev->lct_data.tid)
309 sysfs_remove_link(&tmp->device.kobj, "user");
310 sysfs_remove_link(&i2o_dev->device.kobj, "user");
311unreg_dev:
312 list_del(&i2o_dev->list);
313 device_unregister(&i2o_dev->device);
314err:
315 kfree(i2o_dev);
316 return rc;
317}
318
319
320
321
322
323
324
325
326
327void i2o_device_remove(struct i2o_device *i2o_dev)
328{
329 struct i2o_device *tmp;
330 struct i2o_controller *c = i2o_dev->iop;
331
332 i2o_driver_notify_device_remove_all(i2o_dev);
333
334 sysfs_remove_link(&i2o_dev->device.kobj, "parent");
335 sysfs_remove_link(&i2o_dev->device.kobj, "user");
336
337 list_for_each_entry(tmp, &c->devices, list) {
338 if (tmp->lct_data.parent_tid == i2o_dev->lct_data.tid)
339 sysfs_remove_link(&tmp->device.kobj, "parent");
340 if (tmp->lct_data.user_tid == i2o_dev->lct_data.tid)
341 sysfs_remove_link(&tmp->device.kobj, "user");
342 }
343 list_del(&i2o_dev->list);
344
345 device_unregister(&i2o_dev->device);
346}
347
348
349
350
351
352
353
354
355
356
357
358int i2o_device_parse_lct(struct i2o_controller *c)
359{
360 struct i2o_device *dev, *tmp;
361 i2o_lct *lct;
362 u32 *dlct = c->dlct.virt;
363 int max = 0, i = 0;
364 u16 table_size;
365 u32 buf;
366
367 mutex_lock(&c->lct_lock);
368
369 kfree(c->lct);
370
371 buf = le32_to_cpu(*dlct++);
372 table_size = buf & 0xffff;
373
374 lct = c->lct = kmalloc(table_size * 4, GFP_KERNEL);
375 if (!lct) {
376 mutex_unlock(&c->lct_lock);
377 return -ENOMEM;
378 }
379
380 lct->lct_ver = buf >> 28;
381 lct->boot_tid = buf >> 16 & 0xfff;
382 lct->table_size = table_size;
383 lct->change_ind = le32_to_cpu(*dlct++);
384 lct->iop_flags = le32_to_cpu(*dlct++);
385
386 table_size -= 3;
387
388 pr_debug("%s: LCT has %d entries (LCT size: %d)\n", c->name, max,
389 lct->table_size);
390
391 while (table_size > 0) {
392 i2o_lct_entry *entry = &lct->lct_entry[max];
393 int found = 0;
394
395 buf = le32_to_cpu(*dlct++);
396 entry->entry_size = buf & 0xffff;
397 entry->tid = buf >> 16 & 0xfff;
398
399 entry->change_ind = le32_to_cpu(*dlct++);
400 entry->device_flags = le32_to_cpu(*dlct++);
401
402 buf = le32_to_cpu(*dlct++);
403 entry->class_id = buf & 0xfff;
404 entry->version = buf >> 12 & 0xf;
405 entry->vendor_id = buf >> 16;
406
407 entry->sub_class = le32_to_cpu(*dlct++);
408
409 buf = le32_to_cpu(*dlct++);
410 entry->user_tid = buf & 0xfff;
411 entry->parent_tid = buf >> 12 & 0xfff;
412 entry->bios_info = buf >> 24;
413
414 memcpy(&entry->identity_tag, dlct, 8);
415 dlct += 2;
416
417 entry->event_capabilities = le32_to_cpu(*dlct++);
418
419
420 list_for_each_entry_safe(dev, tmp, &c->devices, list) {
421 if (entry->tid == dev->lct_data.tid) {
422 found = 1;
423 break;
424 }
425 }
426
427 if (!found)
428 i2o_device_add(c, entry);
429
430 table_size -= 9;
431 max++;
432 }
433
434
435 list_for_each_entry_safe(dev, tmp, &c->devices, list) {
436 int found = 0;
437
438 for (i = 0; i < max; i++) {
439 if (lct->lct_entry[i].tid == dev->lct_data.tid) {
440 found = 1;
441 break;
442 }
443 }
444
445 if (!found)
446 i2o_device_remove(dev);
447 }
448
449 mutex_unlock(&c->lct_lock);
450
451 return 0;
452}
453
454
455
456
457
458
459
460
461
462
463
464
465
466int i2o_parm_issue(struct i2o_device *i2o_dev, int cmd, void *oplist,
467 int oplen, void *reslist, int reslen)
468{
469 struct i2o_message *msg;
470 int i = 0;
471 int rc;
472 struct i2o_dma res;
473 struct i2o_controller *c = i2o_dev->iop;
474 struct device *dev = &c->pdev->dev;
475
476 res.virt = NULL;
477
478 if (i2o_dma_alloc(dev, &res, reslen))
479 return -ENOMEM;
480
481 msg = i2o_msg_get_wait(c, I2O_TIMEOUT_MESSAGE_GET);
482 if (IS_ERR(msg)) {
483 i2o_dma_free(dev, &res);
484 return PTR_ERR(msg);
485 }
486
487 i = 0;
488 msg->u.head[1] =
489 cpu_to_le32(cmd << 24 | HOST_TID << 12 | i2o_dev->lct_data.tid);
490 msg->body[i++] = cpu_to_le32(0x00000000);
491 msg->body[i++] = cpu_to_le32(0x4C000000 | oplen);
492 memcpy(&msg->body[i], oplist, oplen);
493 i += (oplen / 4 + (oplen % 4 ? 1 : 0));
494 msg->body[i++] = cpu_to_le32(0xD0000000 | res.len);
495 msg->body[i++] = cpu_to_le32(res.phys);
496
497 msg->u.head[0] =
498 cpu_to_le32(I2O_MESSAGE_SIZE(i + sizeof(struct i2o_message) / 4) |
499 SGL_OFFSET_5);
500
501 rc = i2o_msg_post_wait_mem(c, msg, 10, &res);
502
503
504 if (rc == -ETIMEDOUT)
505 return rc;
506
507 memcpy(reslist, res.virt, res.len);
508 i2o_dma_free(dev, &res);
509
510 return rc;
511}
512
513
514
515
516int i2o_parm_field_get(struct i2o_device *i2o_dev, int group, int field,
517 void *buf, int buflen)
518{
519 u32 opblk[] = { cpu_to_le32(0x00000001),
520 cpu_to_le32((u16) group << 16 | I2O_PARAMS_FIELD_GET),
521 cpu_to_le32((s16) field << 16 | 0x00000001)
522 };
523 u8 *resblk;
524 int rc;
525
526 resblk = kmalloc(buflen + 8, GFP_KERNEL);
527 if (!resblk)
528 return -ENOMEM;
529
530 rc = i2o_parm_issue(i2o_dev, I2O_CMD_UTIL_PARAMS_GET, opblk,
531 sizeof(opblk), resblk, buflen + 8);
532
533 memcpy(buf, resblk + 8, buflen);
534
535 kfree(resblk);
536
537 return rc;
538}
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556int i2o_parm_table_get(struct i2o_device *dev, int oper, int group,
557 int fieldcount, void *ibuf, int ibuflen, void *resblk,
558 int reslen)
559{
560 u16 *opblk;
561 int size;
562
563 size = 10 + ibuflen;
564 if (size % 4)
565 size += 4 - size % 4;
566
567 opblk = kmalloc(size, GFP_KERNEL);
568 if (opblk == NULL) {
569 printk(KERN_ERR "i2o: no memory for query buffer.\n");
570 return -ENOMEM;
571 }
572
573 opblk[0] = 1;
574 opblk[1] = 0;
575 opblk[2] = oper;
576 opblk[3] = group;
577 opblk[4] = fieldcount;
578 memcpy(opblk + 5, ibuf, ibuflen);
579
580 size = i2o_parm_issue(dev, I2O_CMD_UTIL_PARAMS_GET, opblk,
581 size, resblk, reslen);
582
583 kfree(opblk);
584 if (size > reslen)
585 return reslen;
586
587 return size;
588}
589
590EXPORT_SYMBOL(i2o_device_claim);
591EXPORT_SYMBOL(i2o_device_claim_release);
592EXPORT_SYMBOL(i2o_parm_field_get);
593EXPORT_SYMBOL(i2o_parm_table_get);
594EXPORT_SYMBOL(i2o_parm_issue);
595