1
2
3
4
5
6
7
8#include <linux/clk.h>
9#include <linux/clk-provider.h>
10#include <linux/device.h>
11#include <linux/export.h>
12#include <linux/of.h>
13#include <linux/slab.h>
14
15static int __must_check of_clk_bulk_get(struct device_node *np, int num_clks,
16 struct clk_bulk_data *clks)
17{
18 int ret;
19 int i;
20
21 for (i = 0; i < num_clks; i++)
22 clks[i].clk = NULL;
23
24 for (i = 0; i < num_clks; i++) {
25 clks[i].clk = of_clk_get(np, i);
26 if (IS_ERR(clks[i].clk)) {
27 ret = PTR_ERR(clks[i].clk);
28 pr_err("%pOF: Failed to get clk index: %d ret: %d\n",
29 np, i, ret);
30 clks[i].clk = NULL;
31 goto err;
32 }
33 }
34
35 return 0;
36
37err:
38 clk_bulk_put(i, clks);
39
40 return ret;
41}
42
43static int __must_check of_clk_bulk_get_all(struct device_node *np,
44 struct clk_bulk_data **clks)
45{
46 struct clk_bulk_data *clk_bulk;
47 int num_clks;
48 int ret;
49
50 num_clks = of_clk_get_parent_count(np);
51 if (!num_clks)
52 return 0;
53
54 clk_bulk = kmalloc_array(num_clks, sizeof(*clk_bulk), GFP_KERNEL);
55 if (!clk_bulk)
56 return -ENOMEM;
57
58 ret = of_clk_bulk_get(np, num_clks, clk_bulk);
59 if (ret) {
60 kfree(clk_bulk);
61 return ret;
62 }
63
64 *clks = clk_bulk;
65
66 return num_clks;
67}
68
69void clk_bulk_put(int num_clks, struct clk_bulk_data *clks)
70{
71 while (--num_clks >= 0) {
72 clk_put(clks[num_clks].clk);
73 clks[num_clks].clk = NULL;
74 }
75}
76EXPORT_SYMBOL_GPL(clk_bulk_put);
77
78int __must_check clk_bulk_get(struct device *dev, int num_clks,
79 struct clk_bulk_data *clks)
80{
81 int ret;
82 int i;
83
84 for (i = 0; i < num_clks; i++)
85 clks[i].clk = NULL;
86
87 for (i = 0; i < num_clks; i++) {
88 clks[i].clk = clk_get(dev, clks[i].id);
89 if (IS_ERR(clks[i].clk)) {
90 ret = PTR_ERR(clks[i].clk);
91 if (ret != -EPROBE_DEFER)
92 dev_err(dev, "Failed to get clk '%s': %d\n",
93 clks[i].id, ret);
94 clks[i].clk = NULL;
95 goto err;
96 }
97 }
98
99 return 0;
100
101err:
102 clk_bulk_put(i, clks);
103
104 return ret;
105}
106EXPORT_SYMBOL(clk_bulk_get);
107
108void clk_bulk_put_all(int num_clks, struct clk_bulk_data *clks)
109{
110 if (IS_ERR_OR_NULL(clks))
111 return;
112
113 clk_bulk_put(num_clks, clks);
114
115 kfree(clks);
116}
117EXPORT_SYMBOL(clk_bulk_put_all);
118
119int __must_check clk_bulk_get_all(struct device *dev,
120 struct clk_bulk_data **clks)
121{
122 struct device_node *np = dev_of_node(dev);
123
124 if (!np)
125 return 0;
126
127 return of_clk_bulk_get_all(np, clks);
128}
129EXPORT_SYMBOL(clk_bulk_get_all);
130
131#ifdef CONFIG_HAVE_CLK_PREPARE
132
133
134
135
136
137
138
139
140
141void clk_bulk_unprepare(int num_clks, const struct clk_bulk_data *clks)
142{
143 while (--num_clks >= 0)
144 clk_unprepare(clks[num_clks].clk);
145}
146EXPORT_SYMBOL_GPL(clk_bulk_unprepare);
147
148
149
150
151
152
153
154
155
156int __must_check clk_bulk_prepare(int num_clks,
157 const struct clk_bulk_data *clks)
158{
159 int ret;
160 int i;
161
162 for (i = 0; i < num_clks; i++) {
163 ret = clk_prepare(clks[i].clk);
164 if (ret) {
165 pr_err("Failed to prepare clk '%s': %d\n",
166 clks[i].id, ret);
167 goto err;
168 }
169 }
170
171 return 0;
172
173err:
174 clk_bulk_unprepare(i, clks);
175
176 return ret;
177}
178EXPORT_SYMBOL_GPL(clk_bulk_prepare);
179
180#endif
181
182
183
184
185
186
187
188
189
190
191void clk_bulk_disable(int num_clks, const struct clk_bulk_data *clks)
192{
193
194 while (--num_clks >= 0)
195 clk_disable(clks[num_clks].clk);
196}
197EXPORT_SYMBOL_GPL(clk_bulk_disable);
198
199
200
201
202
203
204
205
206
207int __must_check clk_bulk_enable(int num_clks, const struct clk_bulk_data *clks)
208{
209 int ret;
210 int i;
211
212 for (i = 0; i < num_clks; i++) {
213 ret = clk_enable(clks[i].clk);
214 if (ret) {
215 pr_err("Failed to enable clk '%s': %d\n",
216 clks[i].id, ret);
217 goto err;
218 }
219 }
220
221 return 0;
222
223err:
224 clk_bulk_disable(i, clks);
225
226 return ret;
227}
228EXPORT_SYMBOL_GPL(clk_bulk_enable);
229