1
2
3
4
5
6
7
8#include <linux/init.h>
9#include <linux/fs.h>
10#include <linux/kdev_t.h>
11#include <linux/slab.h>
12#include <linux/string.h>
13
14#include <linux/major.h>
15#include <linux/errno.h>
16#include <linux/module.h>
17#include <linux/seq_file.h>
18
19#include <linux/kobject.h>
20#include <linux/kobj_map.h>
21#include <linux/cdev.h>
22#include <linux/mutex.h>
23#include <linux/backing-dev.h>
24#include <linux/tty.h>
25
26#include "internal.h"
27
28static struct kobj_map *cdev_map;
29
30static DEFINE_MUTEX(chrdevs_lock);
31
32#define CHRDEV_MAJOR_HASH_SIZE 255
33
34static struct char_device_struct {
35 struct char_device_struct *next;
36 unsigned int major;
37 unsigned int baseminor;
38 int minorct;
39 char name[64];
40 struct cdev *cdev;
41} *chrdevs[CHRDEV_MAJOR_HASH_SIZE];
42
43
44static inline int major_to_index(unsigned major)
45{
46 return major % CHRDEV_MAJOR_HASH_SIZE;
47}
48
49#ifdef CONFIG_PROC_FS
50
51void chrdev_show(struct seq_file *f, off_t offset)
52{
53 struct char_device_struct *cd;
54
55 mutex_lock(&chrdevs_lock);
56 for (cd = chrdevs[major_to_index(offset)]; cd; cd = cd->next) {
57 if (cd->major == offset)
58 seq_printf(f, "%3d %s\n", cd->major, cd->name);
59 }
60 mutex_unlock(&chrdevs_lock);
61}
62
63#endif
64
65static int find_dynamic_major(void)
66{
67 int i;
68 struct char_device_struct *cd;
69
70 for (i = ARRAY_SIZE(chrdevs)-1; i >= CHRDEV_MAJOR_DYN_END; i--) {
71 if (chrdevs[i] == NULL)
72 return i;
73 }
74
75 for (i = CHRDEV_MAJOR_DYN_EXT_START;
76 i >= CHRDEV_MAJOR_DYN_EXT_END; i--) {
77 for (cd = chrdevs[major_to_index(i)]; cd; cd = cd->next)
78 if (cd->major == i)
79 break;
80
81 if (cd == NULL)
82 return i;
83 }
84
85 return -EBUSY;
86}
87
88
89
90
91
92
93
94
95
96
97
98
99static struct char_device_struct *
100__register_chrdev_region(unsigned int major, unsigned int baseminor,
101 int minorct, const char *name)
102{
103 struct char_device_struct *cd, **cp;
104 int ret = 0;
105 int i;
106
107 cd = kzalloc(sizeof(struct char_device_struct), GFP_KERNEL);
108 if (cd == NULL)
109 return ERR_PTR(-ENOMEM);
110
111 mutex_lock(&chrdevs_lock);
112
113 if (major == 0) {
114 ret = find_dynamic_major();
115 if (ret < 0) {
116 pr_err("CHRDEV \"%s\" dynamic allocation region is full\n",
117 name);
118 goto out;
119 }
120 major = ret;
121 }
122
123 if (major >= CHRDEV_MAJOR_MAX) {
124 pr_err("CHRDEV \"%s\" major requested (%u) is greater than the maximum (%u)\n",
125 name, major, CHRDEV_MAJOR_MAX-1);
126 ret = -EINVAL;
127 goto out;
128 }
129
130 cd->major = major;
131 cd->baseminor = baseminor;
132 cd->minorct = minorct;
133 strlcpy(cd->name, name, sizeof(cd->name));
134
135 i = major_to_index(major);
136
137 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next)
138 if ((*cp)->major > major ||
139 ((*cp)->major == major &&
140 (((*cp)->baseminor >= baseminor) ||
141 ((*cp)->baseminor + (*cp)->minorct > baseminor))))
142 break;
143
144
145 if (*cp && (*cp)->major == major) {
146 int old_min = (*cp)->baseminor;
147 int old_max = (*cp)->baseminor + (*cp)->minorct - 1;
148 int new_min = baseminor;
149 int new_max = baseminor + minorct - 1;
150
151
152 if (new_max >= old_min && new_max <= old_max) {
153 ret = -EBUSY;
154 goto out;
155 }
156
157
158 if (new_min <= old_max && new_min >= old_min) {
159 ret = -EBUSY;
160 goto out;
161 }
162 }
163
164 cd->next = *cp;
165 *cp = cd;
166 mutex_unlock(&chrdevs_lock);
167 return cd;
168out:
169 mutex_unlock(&chrdevs_lock);
170 kfree(cd);
171 return ERR_PTR(ret);
172}
173
174static struct char_device_struct *
175__unregister_chrdev_region(unsigned major, unsigned baseminor, int minorct)
176{
177 struct char_device_struct *cd = NULL, **cp;
178 int i = major_to_index(major);
179
180 mutex_lock(&chrdevs_lock);
181 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next)
182 if ((*cp)->major == major &&
183 (*cp)->baseminor == baseminor &&
184 (*cp)->minorct == minorct)
185 break;
186 if (*cp) {
187 cd = *cp;
188 *cp = cd->next;
189 }
190 mutex_unlock(&chrdevs_lock);
191 return cd;
192}
193
194
195
196
197
198
199
200
201
202
203int register_chrdev_region(dev_t from, unsigned count, const char *name)
204{
205 struct char_device_struct *cd;
206 dev_t to = from + count;
207 dev_t n, next;
208
209 for (n = from; n < to; n = next) {
210 next = MKDEV(MAJOR(n)+1, 0);
211 if (next > to)
212 next = to;
213 cd = __register_chrdev_region(MAJOR(n), MINOR(n),
214 next - n, name);
215 if (IS_ERR(cd))
216 goto fail;
217 }
218 return 0;
219fail:
220 to = n;
221 for (n = from; n < to; n = next) {
222 next = MKDEV(MAJOR(n)+1, 0);
223 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n));
224 }
225 return PTR_ERR(cd);
226}
227
228
229
230
231
232
233
234
235
236
237
238
239int alloc_chrdev_region(dev_t *dev, unsigned baseminor, unsigned count,
240 const char *name)
241{
242 struct char_device_struct *cd;
243 cd = __register_chrdev_region(0, baseminor, count, name);
244 if (IS_ERR(cd))
245 return PTR_ERR(cd);
246 *dev = MKDEV(cd->major, cd->baseminor);
247 return 0;
248}
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271int __register_chrdev(unsigned int major, unsigned int baseminor,
272 unsigned int count, const char *name,
273 const struct file_operations *fops)
274{
275 struct char_device_struct *cd;
276 struct cdev *cdev;
277 int err = -ENOMEM;
278
279 cd = __register_chrdev_region(major, baseminor, count, name);
280 if (IS_ERR(cd))
281 return PTR_ERR(cd);
282
283 cdev = cdev_alloc();
284 if (!cdev)
285 goto out2;
286
287 cdev->owner = fops->owner;
288 cdev->ops = fops;
289 kobject_set_name(&cdev->kobj, "%s", name);
290
291 err = cdev_add(cdev, MKDEV(cd->major, baseminor), count);
292 if (err)
293 goto out;
294
295 cd->cdev = cdev;
296
297 return major ? 0 : cd->major;
298out:
299 kobject_put(&cdev->kobj);
300out2:
301 kfree(__unregister_chrdev_region(cd->major, baseminor, count));
302 return err;
303}
304
305
306
307
308
309
310
311
312
313
314void unregister_chrdev_region(dev_t from, unsigned count)
315{
316 dev_t to = from + count;
317 dev_t n, next;
318
319 for (n = from; n < to; n = next) {
320 next = MKDEV(MAJOR(n)+1, 0);
321 if (next > to)
322 next = to;
323 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n));
324 }
325}
326
327
328
329
330
331
332
333
334
335
336
337
338void __unregister_chrdev(unsigned int major, unsigned int baseminor,
339 unsigned int count, const char *name)
340{
341 struct char_device_struct *cd;
342
343 cd = __unregister_chrdev_region(major, baseminor, count);
344 if (cd && cd->cdev)
345 cdev_del(cd->cdev);
346 kfree(cd);
347}
348
349static DEFINE_SPINLOCK(cdev_lock);
350
351static struct kobject *cdev_get(struct cdev *p)
352{
353 struct module *owner = p->owner;
354 struct kobject *kobj;
355
356 if (owner && !try_module_get(owner))
357 return NULL;
358 kobj = kobject_get(&p->kobj);
359 if (!kobj)
360 module_put(owner);
361 return kobj;
362}
363
364void cdev_put(struct cdev *p)
365{
366 if (p) {
367 struct module *owner = p->owner;
368 kobject_put(&p->kobj);
369 module_put(owner);
370 }
371}
372
373
374
375
376static int chrdev_open(struct inode *inode, struct file *filp)
377{
378 const struct file_operations *fops;
379 struct cdev *p;
380 struct cdev *new = NULL;
381 int ret = 0;
382
383 spin_lock(&cdev_lock);
384 p = inode->i_cdev;
385 if (!p) {
386 struct kobject *kobj;
387 int idx;
388 spin_unlock(&cdev_lock);
389 kobj = kobj_lookup(cdev_map, inode->i_rdev, &idx);
390 if (!kobj)
391 return -ENXIO;
392 new = container_of(kobj, struct cdev, kobj);
393 spin_lock(&cdev_lock);
394
395
396 p = inode->i_cdev;
397 if (!p) {
398 inode->i_cdev = p = new;
399 list_add(&inode->i_devices, &p->list);
400 new = NULL;
401 } else if (!cdev_get(p))
402 ret = -ENXIO;
403 } else if (!cdev_get(p))
404 ret = -ENXIO;
405 spin_unlock(&cdev_lock);
406 cdev_put(new);
407 if (ret)
408 return ret;
409
410 ret = -ENXIO;
411 fops = fops_get(p->ops);
412 if (!fops)
413 goto out_cdev_put;
414
415 replace_fops(filp, fops);
416 if (filp->f_op->open) {
417 ret = filp->f_op->open(inode, filp);
418 if (ret)
419 goto out_cdev_put;
420 }
421
422 return 0;
423
424 out_cdev_put:
425 cdev_put(p);
426 return ret;
427}
428
429void cd_forget(struct inode *inode)
430{
431 spin_lock(&cdev_lock);
432 list_del_init(&inode->i_devices);
433 inode->i_cdev = NULL;
434 inode->i_mapping = &inode->i_data;
435 spin_unlock(&cdev_lock);
436}
437
438static void cdev_purge(struct cdev *cdev)
439{
440 spin_lock(&cdev_lock);
441 while (!list_empty(&cdev->list)) {
442 struct inode *inode;
443 inode = container_of(cdev->list.next, struct inode, i_devices);
444 list_del_init(&inode->i_devices);
445 inode->i_cdev = NULL;
446 }
447 spin_unlock(&cdev_lock);
448}
449
450
451
452
453
454
455const struct file_operations def_chr_fops = {
456 .open = chrdev_open,
457 .llseek = noop_llseek,
458};
459
460static struct kobject *exact_match(dev_t dev, int *part, void *data)
461{
462 struct cdev *p = data;
463 return &p->kobj;
464}
465
466static int exact_lock(dev_t dev, void *data)
467{
468 struct cdev *p = data;
469 return cdev_get(p) ? 0 : -1;
470}
471
472
473
474
475
476
477
478
479
480
481
482int cdev_add(struct cdev *p, dev_t dev, unsigned count)
483{
484 int error;
485
486 p->dev = dev;
487 p->count = count;
488
489 error = kobj_map(cdev_map, dev, count, NULL,
490 exact_match, exact_lock, p);
491 if (error)
492 return error;
493
494 kobject_get(p->kobj.parent);
495
496 return 0;
497}
498
499
500
501
502
503
504
505
506
507
508void cdev_set_parent(struct cdev *p, struct kobject *kobj)
509{
510 WARN_ON(!kobj->state_initialized);
511 p->kobj.parent = kobj;
512}
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537int cdev_device_add(struct cdev *cdev, struct device *dev)
538{
539 int rc = 0;
540
541 if (dev->devt) {
542 cdev_set_parent(cdev, &dev->kobj);
543
544 rc = cdev_add(cdev, dev->devt, 1);
545 if (rc)
546 return rc;
547 }
548
549 rc = device_add(dev);
550 if (rc)
551 cdev_del(cdev);
552
553 return rc;
554}
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571void cdev_device_del(struct cdev *cdev, struct device *dev)
572{
573 device_del(dev);
574 if (dev->devt)
575 cdev_del(cdev);
576}
577
578static void cdev_unmap(dev_t dev, unsigned count)
579{
580 kobj_unmap(cdev_map, dev, count);
581}
582
583
584
585
586
587
588
589
590
591
592
593
594void cdev_del(struct cdev *p)
595{
596 cdev_unmap(p->dev, p->count);
597 kobject_put(&p->kobj);
598}
599
600
601static void cdev_default_release(struct kobject *kobj)
602{
603 struct cdev *p = container_of(kobj, struct cdev, kobj);
604 struct kobject *parent = kobj->parent;
605
606 cdev_purge(p);
607 kobject_put(parent);
608}
609
610static void cdev_dynamic_release(struct kobject *kobj)
611{
612 struct cdev *p = container_of(kobj, struct cdev, kobj);
613 struct kobject *parent = kobj->parent;
614
615 cdev_purge(p);
616 kfree(p);
617 kobject_put(parent);
618}
619
620static struct kobj_type ktype_cdev_default = {
621 .release = cdev_default_release,
622};
623
624static struct kobj_type ktype_cdev_dynamic = {
625 .release = cdev_dynamic_release,
626};
627
628
629
630
631
632
633struct cdev *cdev_alloc(void)
634{
635 struct cdev *p = kzalloc(sizeof(struct cdev), GFP_KERNEL);
636 if (p) {
637 INIT_LIST_HEAD(&p->list);
638 kobject_init(&p->kobj, &ktype_cdev_dynamic);
639 }
640 return p;
641}
642
643
644
645
646
647
648
649
650
651void cdev_init(struct cdev *cdev, const struct file_operations *fops)
652{
653 memset(cdev, 0, sizeof *cdev);
654 INIT_LIST_HEAD(&cdev->list);
655 kobject_init(&cdev->kobj, &ktype_cdev_default);
656 cdev->ops = fops;
657}
658
659static struct kobject *base_probe(dev_t dev, int *part, void *data)
660{
661 if (request_module("char-major-%d-%d", MAJOR(dev), MINOR(dev)) > 0)
662
663 request_module("char-major-%d", MAJOR(dev));
664 return NULL;
665}
666
667void __init chrdev_init(void)
668{
669 cdev_map = kobj_map_init(base_probe, &chrdevs_lock);
670}
671
672
673
674EXPORT_SYMBOL(register_chrdev_region);
675EXPORT_SYMBOL(unregister_chrdev_region);
676EXPORT_SYMBOL(alloc_chrdev_region);
677EXPORT_SYMBOL(cdev_init);
678EXPORT_SYMBOL(cdev_alloc);
679EXPORT_SYMBOL(cdev_del);
680EXPORT_SYMBOL(cdev_add);
681EXPORT_SYMBOL(cdev_set_parent);
682EXPORT_SYMBOL(cdev_device_add);
683EXPORT_SYMBOL(cdev_device_del);
684EXPORT_SYMBOL(__register_chrdev);
685EXPORT_SYMBOL(__unregister_chrdev);
686