1
2
3
4
5
6
7
8
9
10
11
12
13
14#include <linux/kernel.h>
15#include <linux/export.h>
16#include <linux/module.h>
17#include <linux/err.h>
18#include <linux/device.h>
19#include <linux/slab.h>
20#include <linux/of.h>
21#include <linux/phy/phy.h>
22#include <linux/idr.h>
23#include <linux/pm_runtime.h>
24
25static struct class *phy_class;
26static DEFINE_MUTEX(phy_provider_mutex);
27static LIST_HEAD(phy_provider_list);
28static DEFINE_IDA(phy_ida);
29
30static void devm_phy_release(struct device *dev, void *res)
31{
32 struct phy *phy = *(struct phy **)res;
33
34 phy_put(phy);
35}
36
37static void devm_phy_provider_release(struct device *dev, void *res)
38{
39 struct phy_provider *phy_provider = *(struct phy_provider **)res;
40
41 of_phy_provider_unregister(phy_provider);
42}
43
44static void devm_phy_consume(struct device *dev, void *res)
45{
46 struct phy *phy = *(struct phy **)res;
47
48 phy_destroy(phy);
49}
50
51static int devm_phy_match(struct device *dev, void *res, void *match_data)
52{
53 return res == match_data;
54}
55
56static struct phy *phy_lookup(struct device *device, const char *port)
57{
58 unsigned int count;
59 struct phy *phy;
60 struct device *dev;
61 struct phy_consumer *consumers;
62 struct class_dev_iter iter;
63
64 class_dev_iter_init(&iter, phy_class, NULL, NULL);
65 while ((dev = class_dev_iter_next(&iter))) {
66 phy = to_phy(dev);
67
68 if (!phy->init_data)
69 continue;
70 count = phy->init_data->num_consumers;
71 consumers = phy->init_data->consumers;
72 while (count--) {
73 if (!strcmp(consumers->dev_name, dev_name(device)) &&
74 !strcmp(consumers->port, port)) {
75 class_dev_iter_exit(&iter);
76 return phy;
77 }
78 consumers++;
79 }
80 }
81
82 class_dev_iter_exit(&iter);
83 return ERR_PTR(-ENODEV);
84}
85
86static struct phy_provider *of_phy_provider_lookup(struct device_node *node)
87{
88 struct phy_provider *phy_provider;
89
90 list_for_each_entry(phy_provider, &phy_provider_list, list) {
91 if (phy_provider->dev->of_node == node)
92 return phy_provider;
93 }
94
95 return ERR_PTR(-EPROBE_DEFER);
96}
97
98int phy_pm_runtime_get(struct phy *phy)
99{
100 int ret;
101
102 if (!pm_runtime_enabled(&phy->dev))
103 return -ENOTSUPP;
104
105 ret = pm_runtime_get(&phy->dev);
106 if (ret < 0 && ret != -EINPROGRESS)
107 pm_runtime_put_noidle(&phy->dev);
108
109 return ret;
110}
111EXPORT_SYMBOL_GPL(phy_pm_runtime_get);
112
113int phy_pm_runtime_get_sync(struct phy *phy)
114{
115 int ret;
116
117 if (!pm_runtime_enabled(&phy->dev))
118 return -ENOTSUPP;
119
120 ret = pm_runtime_get_sync(&phy->dev);
121 if (ret < 0)
122 pm_runtime_put_sync(&phy->dev);
123
124 return ret;
125}
126EXPORT_SYMBOL_GPL(phy_pm_runtime_get_sync);
127
128int phy_pm_runtime_put(struct phy *phy)
129{
130 if (!pm_runtime_enabled(&phy->dev))
131 return -ENOTSUPP;
132
133 return pm_runtime_put(&phy->dev);
134}
135EXPORT_SYMBOL_GPL(phy_pm_runtime_put);
136
137int phy_pm_runtime_put_sync(struct phy *phy)
138{
139 if (!pm_runtime_enabled(&phy->dev))
140 return -ENOTSUPP;
141
142 return pm_runtime_put_sync(&phy->dev);
143}
144EXPORT_SYMBOL_GPL(phy_pm_runtime_put_sync);
145
146void phy_pm_runtime_allow(struct phy *phy)
147{
148 if (!pm_runtime_enabled(&phy->dev))
149 return;
150
151 pm_runtime_allow(&phy->dev);
152}
153EXPORT_SYMBOL_GPL(phy_pm_runtime_allow);
154
155void phy_pm_runtime_forbid(struct phy *phy)
156{
157 if (!pm_runtime_enabled(&phy->dev))
158 return;
159
160 pm_runtime_forbid(&phy->dev);
161}
162EXPORT_SYMBOL_GPL(phy_pm_runtime_forbid);
163
164int phy_init(struct phy *phy)
165{
166 int ret;
167
168 if (!phy)
169 return 0;
170
171 ret = phy_pm_runtime_get_sync(phy);
172 if (ret < 0 && ret != -ENOTSUPP)
173 return ret;
174
175 mutex_lock(&phy->mutex);
176 if (phy->init_count == 0 && phy->ops->init) {
177 ret = phy->ops->init(phy);
178 if (ret < 0) {
179 dev_err(&phy->dev, "phy init failed --> %d\n", ret);
180 goto out;
181 }
182 } else {
183 ret = 0;
184 }
185 ++phy->init_count;
186
187out:
188 mutex_unlock(&phy->mutex);
189 phy_pm_runtime_put(phy);
190 return ret;
191}
192EXPORT_SYMBOL_GPL(phy_init);
193
194int phy_exit(struct phy *phy)
195{
196 int ret;
197
198 if (!phy)
199 return 0;
200
201 ret = phy_pm_runtime_get_sync(phy);
202 if (ret < 0 && ret != -ENOTSUPP)
203 return ret;
204
205 mutex_lock(&phy->mutex);
206 if (phy->init_count == 1 && phy->ops->exit) {
207 ret = phy->ops->exit(phy);
208 if (ret < 0) {
209 dev_err(&phy->dev, "phy exit failed --> %d\n", ret);
210 goto out;
211 }
212 }
213 --phy->init_count;
214
215out:
216 mutex_unlock(&phy->mutex);
217 phy_pm_runtime_put(phy);
218 return ret;
219}
220EXPORT_SYMBOL_GPL(phy_exit);
221
222int phy_power_on(struct phy *phy)
223{
224 int ret;
225
226 if (!phy)
227 return 0;
228
229 ret = phy_pm_runtime_get_sync(phy);
230 if (ret < 0 && ret != -ENOTSUPP)
231 return ret;
232
233 mutex_lock(&phy->mutex);
234 if (phy->power_count == 0 && phy->ops->power_on) {
235 ret = phy->ops->power_on(phy);
236 if (ret < 0) {
237 dev_err(&phy->dev, "phy poweron failed --> %d\n", ret);
238 goto out;
239 }
240 } else {
241 ret = 0;
242 }
243 ++phy->power_count;
244 mutex_unlock(&phy->mutex);
245 return 0;
246
247out:
248 mutex_unlock(&phy->mutex);
249 phy_pm_runtime_put_sync(phy);
250
251 return ret;
252}
253EXPORT_SYMBOL_GPL(phy_power_on);
254
255int phy_power_off(struct phy *phy)
256{
257 int ret;
258
259 if (!phy)
260 return 0;
261
262 mutex_lock(&phy->mutex);
263 if (phy->power_count == 1 && phy->ops->power_off) {
264 ret = phy->ops->power_off(phy);
265 if (ret < 0) {
266 dev_err(&phy->dev, "phy poweroff failed --> %d\n", ret);
267 mutex_unlock(&phy->mutex);
268 return ret;
269 }
270 }
271 --phy->power_count;
272 mutex_unlock(&phy->mutex);
273 phy_pm_runtime_put(phy);
274
275 return 0;
276}
277EXPORT_SYMBOL_GPL(phy_power_off);
278
279
280
281
282
283
284
285
286
287
288
289
290static struct phy *_of_phy_get(struct device_node *np, int index)
291{
292 int ret;
293 struct phy_provider *phy_provider;
294 struct phy *phy = NULL;
295 struct of_phandle_args args;
296
297 ret = of_parse_phandle_with_args(np, "phys", "#phy-cells",
298 index, &args);
299 if (ret)
300 return ERR_PTR(-ENODEV);
301
302 mutex_lock(&phy_provider_mutex);
303 phy_provider = of_phy_provider_lookup(args.np);
304 if (IS_ERR(phy_provider) || !try_module_get(phy_provider->owner)) {
305 phy = ERR_PTR(-EPROBE_DEFER);
306 goto err0;
307 }
308
309 phy = phy_provider->of_xlate(phy_provider->dev, &args);
310 module_put(phy_provider->owner);
311
312err0:
313 mutex_unlock(&phy_provider_mutex);
314 of_node_put(args.np);
315
316 return phy;
317}
318
319
320
321
322
323
324
325
326
327
328struct phy *of_phy_get(struct device_node *np, const char *con_id)
329{
330 struct phy *phy = NULL;
331 int index = 0;
332
333 if (con_id)
334 index = of_property_match_string(np, "phy-names", con_id);
335
336 phy = _of_phy_get(np, index);
337 if (IS_ERR(phy))
338 return phy;
339
340 if (!try_module_get(phy->ops->owner))
341 return ERR_PTR(-EPROBE_DEFER);
342
343 get_device(&phy->dev);
344
345 return phy;
346}
347EXPORT_SYMBOL_GPL(of_phy_get);
348
349
350
351
352
353
354
355void phy_put(struct phy *phy)
356{
357 if (!phy || IS_ERR(phy))
358 return;
359
360 module_put(phy->ops->owner);
361 put_device(&phy->dev);
362}
363EXPORT_SYMBOL_GPL(phy_put);
364
365
366
367
368
369
370
371
372
373void devm_phy_put(struct device *dev, struct phy *phy)
374{
375 int r;
376
377 if (!phy)
378 return;
379
380 r = devres_destroy(dev, devm_phy_release, devm_phy_match, phy);
381 dev_WARN_ONCE(dev, r, "couldn't find PHY resource\n");
382}
383EXPORT_SYMBOL_GPL(devm_phy_put);
384
385
386
387
388
389
390
391
392
393
394
395struct phy *of_phy_simple_xlate(struct device *dev, struct of_phandle_args
396 *args)
397{
398 struct phy *phy;
399 struct class_dev_iter iter;
400 struct device_node *node = dev->of_node;
401
402 class_dev_iter_init(&iter, phy_class, NULL, NULL);
403 while ((dev = class_dev_iter_next(&iter))) {
404 phy = to_phy(dev);
405 if (node != phy->dev.of_node)
406 continue;
407
408 class_dev_iter_exit(&iter);
409 return phy;
410 }
411
412 class_dev_iter_exit(&iter);
413 return ERR_PTR(-ENODEV);
414}
415EXPORT_SYMBOL_GPL(of_phy_simple_xlate);
416
417
418
419
420
421
422
423
424
425
426
427struct phy *phy_get(struct device *dev, const char *string)
428{
429 int index = 0;
430 struct phy *phy;
431
432 if (string == NULL) {
433 dev_WARN(dev, "missing string\n");
434 return ERR_PTR(-EINVAL);
435 }
436
437 if (dev->of_node) {
438 index = of_property_match_string(dev->of_node, "phy-names",
439 string);
440 phy = _of_phy_get(dev->of_node, index);
441 } else {
442 phy = phy_lookup(dev, string);
443 }
444 if (IS_ERR(phy))
445 return phy;
446
447 if (!try_module_get(phy->ops->owner))
448 return ERR_PTR(-EPROBE_DEFER);
449
450 get_device(&phy->dev);
451
452 return phy;
453}
454EXPORT_SYMBOL_GPL(phy_get);
455
456
457
458
459
460
461
462
463
464
465
466struct phy *phy_optional_get(struct device *dev, const char *string)
467{
468 struct phy *phy = phy_get(dev, string);
469
470 if (PTR_ERR(phy) == -ENODEV)
471 phy = NULL;
472
473 return phy;
474}
475EXPORT_SYMBOL_GPL(phy_optional_get);
476
477
478
479
480
481
482
483
484
485
486
487struct phy *devm_phy_get(struct device *dev, const char *string)
488{
489 struct phy **ptr, *phy;
490
491 ptr = devres_alloc(devm_phy_release, sizeof(*ptr), GFP_KERNEL);
492 if (!ptr)
493 return ERR_PTR(-ENOMEM);
494
495 phy = phy_get(dev, string);
496 if (!IS_ERR(phy)) {
497 *ptr = phy;
498 devres_add(dev, ptr);
499 } else {
500 devres_free(ptr);
501 }
502
503 return phy;
504}
505EXPORT_SYMBOL_GPL(devm_phy_get);
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520struct phy *devm_phy_optional_get(struct device *dev, const char *string)
521{
522 struct phy *phy = devm_phy_get(dev, string);
523
524 if (PTR_ERR(phy) == -ENODEV)
525 phy = NULL;
526
527 return phy;
528}
529EXPORT_SYMBOL_GPL(devm_phy_optional_get);
530
531
532
533
534
535
536
537
538
539
540
541struct phy *devm_of_phy_get(struct device *dev, struct device_node *np,
542 const char *con_id)
543{
544 struct phy **ptr, *phy;
545
546 ptr = devres_alloc(devm_phy_release, sizeof(*ptr), GFP_KERNEL);
547 if (!ptr)
548 return ERR_PTR(-ENOMEM);
549
550 phy = of_phy_get(np, con_id);
551 if (!IS_ERR(phy)) {
552 *ptr = phy;
553 devres_add(dev, ptr);
554 } else {
555 devres_free(ptr);
556 }
557
558 return phy;
559}
560EXPORT_SYMBOL_GPL(devm_of_phy_get);
561
562
563
564
565
566
567
568
569
570struct phy *phy_create(struct device *dev, const struct phy_ops *ops,
571 struct phy_init_data *init_data)
572{
573 int ret;
574 int id;
575 struct phy *phy;
576
577 if (WARN_ON(!dev))
578 return ERR_PTR(-EINVAL);
579
580 phy = kzalloc(sizeof(*phy), GFP_KERNEL);
581 if (!phy)
582 return ERR_PTR(-ENOMEM);
583
584 id = ida_simple_get(&phy_ida, 0, 0, GFP_KERNEL);
585 if (id < 0) {
586 dev_err(dev, "unable to get id\n");
587 ret = id;
588 goto free_phy;
589 }
590
591 device_initialize(&phy->dev);
592 mutex_init(&phy->mutex);
593
594 phy->dev.class = phy_class;
595 phy->dev.parent = dev;
596 phy->dev.of_node = dev->of_node;
597 phy->id = id;
598 phy->ops = ops;
599 phy->init_data = init_data;
600
601 ret = dev_set_name(&phy->dev, "phy-%s.%d", dev_name(dev), id);
602 if (ret)
603 goto put_dev;
604
605 ret = device_add(&phy->dev);
606 if (ret)
607 goto put_dev;
608
609 if (pm_runtime_enabled(dev)) {
610 pm_runtime_enable(&phy->dev);
611 pm_runtime_no_callbacks(&phy->dev);
612 }
613
614 return phy;
615
616put_dev:
617 put_device(&phy->dev);
618 ida_remove(&phy_ida, phy->id);
619free_phy:
620 kfree(phy);
621 return ERR_PTR(ret);
622}
623EXPORT_SYMBOL_GPL(phy_create);
624
625
626
627
628
629
630
631
632
633
634
635
636struct phy *devm_phy_create(struct device *dev, const struct phy_ops *ops,
637 struct phy_init_data *init_data)
638{
639 struct phy **ptr, *phy;
640
641 ptr = devres_alloc(devm_phy_consume, sizeof(*ptr), GFP_KERNEL);
642 if (!ptr)
643 return ERR_PTR(-ENOMEM);
644
645 phy = phy_create(dev, ops, init_data);
646 if (!IS_ERR(phy)) {
647 *ptr = phy;
648 devres_add(dev, ptr);
649 } else {
650 devres_free(ptr);
651 }
652
653 return phy;
654}
655EXPORT_SYMBOL_GPL(devm_phy_create);
656
657
658
659
660
661
662
663void phy_destroy(struct phy *phy)
664{
665 pm_runtime_disable(&phy->dev);
666 device_unregister(&phy->dev);
667}
668EXPORT_SYMBOL_GPL(phy_destroy);
669
670
671
672
673
674
675
676
677
678void devm_phy_destroy(struct device *dev, struct phy *phy)
679{
680 int r;
681
682 r = devres_destroy(dev, devm_phy_consume, devm_phy_match, phy);
683 dev_WARN_ONCE(dev, r, "couldn't find PHY resource\n");
684}
685EXPORT_SYMBOL_GPL(devm_phy_destroy);
686
687
688
689
690
691
692
693
694
695
696
697struct phy_provider *__of_phy_provider_register(struct device *dev,
698 struct module *owner, struct phy * (*of_xlate)(struct device *dev,
699 struct of_phandle_args *args))
700{
701 struct phy_provider *phy_provider;
702
703 phy_provider = kzalloc(sizeof(*phy_provider), GFP_KERNEL);
704 if (!phy_provider)
705 return ERR_PTR(-ENOMEM);
706
707 phy_provider->dev = dev;
708 phy_provider->owner = owner;
709 phy_provider->of_xlate = of_xlate;
710
711 mutex_lock(&phy_provider_mutex);
712 list_add_tail(&phy_provider->list, &phy_provider_list);
713 mutex_unlock(&phy_provider_mutex);
714
715 return phy_provider;
716}
717EXPORT_SYMBOL_GPL(__of_phy_provider_register);
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732struct phy_provider *__devm_of_phy_provider_register(struct device *dev,
733 struct module *owner, struct phy * (*of_xlate)(struct device *dev,
734 struct of_phandle_args *args))
735{
736 struct phy_provider **ptr, *phy_provider;
737
738 ptr = devres_alloc(devm_phy_provider_release, sizeof(*ptr), GFP_KERNEL);
739 if (!ptr)
740 return ERR_PTR(-ENOMEM);
741
742 phy_provider = __of_phy_provider_register(dev, owner, of_xlate);
743 if (!IS_ERR(phy_provider)) {
744 *ptr = phy_provider;
745 devres_add(dev, ptr);
746 } else {
747 devres_free(ptr);
748 }
749
750 return phy_provider;
751}
752EXPORT_SYMBOL_GPL(__devm_of_phy_provider_register);
753
754
755
756
757
758
759
760void of_phy_provider_unregister(struct phy_provider *phy_provider)
761{
762 if (IS_ERR(phy_provider))
763 return;
764
765 mutex_lock(&phy_provider_mutex);
766 list_del(&phy_provider->list);
767 kfree(phy_provider);
768 mutex_unlock(&phy_provider_mutex);
769}
770EXPORT_SYMBOL_GPL(of_phy_provider_unregister);
771
772
773
774
775
776
777
778
779void devm_of_phy_provider_unregister(struct device *dev,
780 struct phy_provider *phy_provider) {
781 int r;
782
783 r = devres_destroy(dev, devm_phy_provider_release, devm_phy_match,
784 phy_provider);
785 dev_WARN_ONCE(dev, r, "couldn't find PHY provider device resource\n");
786}
787EXPORT_SYMBOL_GPL(devm_of_phy_provider_unregister);
788
789
790
791
792
793
794
795
796static void phy_release(struct device *dev)
797{
798 struct phy *phy;
799
800 phy = to_phy(dev);
801 dev_vdbg(dev, "releasing '%s'\n", dev_name(dev));
802 ida_remove(&phy_ida, phy->id);
803 kfree(phy);
804}
805
806static int __init phy_core_init(void)
807{
808 phy_class = class_create(THIS_MODULE, "phy");
809 if (IS_ERR(phy_class)) {
810 pr_err("failed to create phy class --> %ld\n",
811 PTR_ERR(phy_class));
812 return PTR_ERR(phy_class);
813 }
814
815 phy_class->dev_release = phy_release;
816
817 return 0;
818}
819module_init(phy_core_init);
820
821static void __exit phy_core_exit(void)
822{
823 class_destroy(phy_class);
824}
825module_exit(phy_core_exit);
826
827MODULE_DESCRIPTION("Generic PHY Framework");
828MODULE_AUTHOR("Kishon Vijay Abraham I <kishon@ti.com>");
829MODULE_LICENSE("GPL v2");
830