1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41#define DEBUG_SUBSYSTEM S_LOV
42
43
44#include "../include/obd_class.h"
45
46#include "lov_cl_internal.h"
47#include "lov_internal.h"
48
49struct kmem_cache *lov_lock_kmem;
50struct kmem_cache *lov_object_kmem;
51struct kmem_cache *lov_thread_kmem;
52struct kmem_cache *lov_session_kmem;
53struct kmem_cache *lov_req_kmem;
54
55struct kmem_cache *lovsub_lock_kmem;
56struct kmem_cache *lovsub_object_kmem;
57struct kmem_cache *lovsub_req_kmem;
58
59struct kmem_cache *lov_lock_link_kmem;
60
61
62static struct lock_class_key cl_lov_device_mutex_class;
63
64struct lu_kmem_descr lov_caches[] = {
65 {
66 .ckd_cache = &lov_lock_kmem,
67 .ckd_name = "lov_lock_kmem",
68 .ckd_size = sizeof(struct lov_lock)
69 },
70 {
71 .ckd_cache = &lov_object_kmem,
72 .ckd_name = "lov_object_kmem",
73 .ckd_size = sizeof(struct lov_object)
74 },
75 {
76 .ckd_cache = &lov_thread_kmem,
77 .ckd_name = "lov_thread_kmem",
78 .ckd_size = sizeof(struct lov_thread_info)
79 },
80 {
81 .ckd_cache = &lov_session_kmem,
82 .ckd_name = "lov_session_kmem",
83 .ckd_size = sizeof(struct lov_session)
84 },
85 {
86 .ckd_cache = &lov_req_kmem,
87 .ckd_name = "lov_req_kmem",
88 .ckd_size = sizeof(struct lov_req)
89 },
90 {
91 .ckd_cache = &lovsub_lock_kmem,
92 .ckd_name = "lovsub_lock_kmem",
93 .ckd_size = sizeof(struct lovsub_lock)
94 },
95 {
96 .ckd_cache = &lovsub_object_kmem,
97 .ckd_name = "lovsub_object_kmem",
98 .ckd_size = sizeof(struct lovsub_object)
99 },
100 {
101 .ckd_cache = &lovsub_req_kmem,
102 .ckd_name = "lovsub_req_kmem",
103 .ckd_size = sizeof(struct lovsub_req)
104 },
105 {
106 .ckd_cache = &lov_lock_link_kmem,
107 .ckd_name = "lov_lock_link_kmem",
108 .ckd_size = sizeof(struct lov_lock_link)
109 },
110 {
111 .ckd_cache = NULL
112 }
113};
114
115
116
117
118
119
120
121static void lov_req_completion(const struct lu_env *env,
122 const struct cl_req_slice *slice, int ioret)
123{
124 struct lov_req *lr;
125
126 lr = cl2lov_req(slice);
127 kmem_cache_free(lov_req_kmem, lr);
128}
129
130static const struct cl_req_operations lov_req_ops = {
131 .cro_completion = lov_req_completion
132};
133
134
135
136
137
138
139
140static void *lov_key_init(const struct lu_context *ctx,
141 struct lu_context_key *key)
142{
143 struct lov_thread_info *info;
144
145 info = kmem_cache_alloc(lov_thread_kmem, GFP_NOFS | __GFP_ZERO);
146 if (info != NULL)
147 INIT_LIST_HEAD(&info->lti_closure.clc_list);
148 else
149 info = ERR_PTR(-ENOMEM);
150 return info;
151}
152
153static void lov_key_fini(const struct lu_context *ctx,
154 struct lu_context_key *key, void *data)
155{
156 struct lov_thread_info *info = data;
157
158 LINVRNT(list_empty(&info->lti_closure.clc_list));
159 kmem_cache_free(lov_thread_kmem, info);
160}
161
162struct lu_context_key lov_key = {
163 .lct_tags = LCT_CL_THREAD,
164 .lct_init = lov_key_init,
165 .lct_fini = lov_key_fini
166};
167
168static void *lov_session_key_init(const struct lu_context *ctx,
169 struct lu_context_key *key)
170{
171 struct lov_session *info;
172
173 info = kmem_cache_alloc(lov_session_kmem, GFP_NOFS | __GFP_ZERO);
174 if (info == NULL)
175 info = ERR_PTR(-ENOMEM);
176 return info;
177}
178
179static void lov_session_key_fini(const struct lu_context *ctx,
180 struct lu_context_key *key, void *data)
181{
182 struct lov_session *info = data;
183
184 kmem_cache_free(lov_session_kmem, info);
185}
186
187struct lu_context_key lov_session_key = {
188 .lct_tags = LCT_SESSION,
189 .lct_init = lov_session_key_init,
190 .lct_fini = lov_session_key_fini
191};
192
193
194LU_TYPE_INIT_FINI(lov, &lov_key, &lov_session_key);
195
196static struct lu_device *lov_device_fini(const struct lu_env *env,
197 struct lu_device *d)
198{
199 int i;
200 struct lov_device *ld = lu2lov_dev(d);
201
202 LASSERT(ld->ld_lov != NULL);
203 if (ld->ld_target == NULL)
204 return NULL;
205
206 lov_foreach_target(ld, i) {
207 struct lovsub_device *lsd;
208
209 lsd = ld->ld_target[i];
210 if (lsd != NULL) {
211 cl_stack_fini(env, lovsub2cl_dev(lsd));
212 ld->ld_target[i] = NULL;
213 }
214 }
215 return NULL;
216}
217
218static int lov_device_init(const struct lu_env *env, struct lu_device *d,
219 const char *name, struct lu_device *next)
220{
221 struct lov_device *ld = lu2lov_dev(d);
222 int i;
223 int rc = 0;
224
225 LASSERT(d->ld_site != NULL);
226 if (ld->ld_target == NULL)
227 return rc;
228
229 lov_foreach_target(ld, i) {
230 struct lovsub_device *lsd;
231 struct cl_device *cl;
232 struct lov_tgt_desc *desc;
233
234 desc = ld->ld_lov->lov_tgts[i];
235 if (desc == NULL)
236 continue;
237
238 cl = cl_type_setup(env, d->ld_site, &lovsub_device_type,
239 desc->ltd_obd->obd_lu_dev);
240 if (IS_ERR(cl)) {
241 rc = PTR_ERR(cl);
242 break;
243 }
244 lsd = cl2lovsub_dev(cl);
245 lsd->acid_idx = i;
246 lsd->acid_super = ld;
247 ld->ld_target[i] = lsd;
248 }
249
250 if (rc)
251 lov_device_fini(env, d);
252 else
253 ld->ld_flags |= LOV_DEV_INITIALIZED;
254
255 return rc;
256}
257
258static int lov_req_init(const struct lu_env *env, struct cl_device *dev,
259 struct cl_req *req)
260{
261 struct lov_req *lr;
262 int result;
263
264 lr = kmem_cache_alloc(lov_req_kmem, GFP_NOFS | __GFP_ZERO);
265 if (lr != NULL) {
266 cl_req_slice_add(req, &lr->lr_cl, dev, &lov_req_ops);
267 result = 0;
268 } else
269 result = -ENOMEM;
270 return result;
271}
272
273static const struct cl_device_operations lov_cl_ops = {
274 .cdo_req_init = lov_req_init
275};
276
277static void lov_emerg_free(struct lov_device_emerg **emrg, int nr)
278{
279 int i;
280
281 for (i = 0; i < nr; ++i) {
282 struct lov_device_emerg *em;
283
284 em = emrg[i];
285 if (em != NULL) {
286 LASSERT(em->emrg_page_list.pl_nr == 0);
287 if (em->emrg_env != NULL)
288 cl_env_put(em->emrg_env, &em->emrg_refcheck);
289 kfree(em);
290 }
291 }
292 kfree(emrg);
293}
294
295static struct lu_device *lov_device_free(const struct lu_env *env,
296 struct lu_device *d)
297{
298 struct lov_device *ld = lu2lov_dev(d);
299 const int nr = ld->ld_target_nr;
300
301 cl_device_fini(lu2cl_dev(d));
302 kfree(ld->ld_target);
303 if (ld->ld_emrg != NULL)
304 lov_emerg_free(ld->ld_emrg, nr);
305 kfree(ld);
306 return NULL;
307}
308
309static void lov_cl_del_target(const struct lu_env *env, struct lu_device *dev,
310 __u32 index)
311{
312 struct lov_device *ld = lu2lov_dev(dev);
313
314 if (ld->ld_target[index] != NULL) {
315 cl_stack_fini(env, lovsub2cl_dev(ld->ld_target[index]));
316 ld->ld_target[index] = NULL;
317 }
318}
319
320static struct lov_device_emerg **lov_emerg_alloc(int nr)
321{
322 struct lov_device_emerg **emerg;
323 int i;
324 int result;
325
326 emerg = kcalloc(nr, sizeof(emerg[0]), GFP_NOFS);
327 if (emerg == NULL)
328 return ERR_PTR(-ENOMEM);
329 for (result = i = 0; i < nr && result == 0; i++) {
330 struct lov_device_emerg *em;
331
332 em = kzalloc(sizeof(*em), GFP_NOFS);
333 if (em != NULL) {
334 emerg[i] = em;
335 cl_page_list_init(&em->emrg_page_list);
336 em->emrg_env = cl_env_alloc(&em->emrg_refcheck,
337 LCT_REMEMBER|LCT_NOREF);
338 if (!IS_ERR(em->emrg_env))
339 em->emrg_env->le_ctx.lc_cookie = 0x2;
340 else {
341 result = PTR_ERR(em->emrg_env);
342 em->emrg_env = NULL;
343 }
344 } else
345 result = -ENOMEM;
346 }
347 if (result != 0) {
348 lov_emerg_free(emerg, nr);
349 emerg = ERR_PTR(result);
350 }
351 return emerg;
352}
353
354static int lov_expand_targets(const struct lu_env *env, struct lov_device *dev)
355{
356 int result;
357 __u32 tgt_size;
358 __u32 sub_size;
359
360 result = 0;
361 tgt_size = dev->ld_lov->lov_tgt_size;
362 sub_size = dev->ld_target_nr;
363 if (sub_size < tgt_size) {
364 struct lovsub_device **newd;
365 struct lov_device_emerg **emerg;
366 const size_t sz = sizeof(newd[0]);
367
368 emerg = lov_emerg_alloc(tgt_size);
369 if (IS_ERR(emerg))
370 return PTR_ERR(emerg);
371
372 newd = kcalloc(tgt_size, sz, GFP_NOFS);
373 if (newd != NULL) {
374 mutex_lock(&dev->ld_mutex);
375 if (sub_size > 0) {
376 memcpy(newd, dev->ld_target, sub_size * sz);
377 kfree(dev->ld_target);
378 }
379 dev->ld_target = newd;
380 dev->ld_target_nr = tgt_size;
381
382 if (dev->ld_emrg != NULL)
383 lov_emerg_free(dev->ld_emrg, sub_size);
384 dev->ld_emrg = emerg;
385 mutex_unlock(&dev->ld_mutex);
386 } else {
387 lov_emerg_free(emerg, tgt_size);
388 result = -ENOMEM;
389 }
390 }
391 return result;
392}
393
394static int lov_cl_add_target(const struct lu_env *env, struct lu_device *dev,
395 __u32 index)
396{
397 struct obd_device *obd = dev->ld_obd;
398 struct lov_device *ld = lu2lov_dev(dev);
399 struct lov_tgt_desc *tgt;
400 struct lovsub_device *lsd;
401 struct cl_device *cl;
402 int rc;
403
404 obd_getref(obd);
405
406 tgt = obd->u.lov.lov_tgts[index];
407 LASSERT(tgt != NULL);
408 LASSERT(tgt->ltd_obd != NULL);
409
410 if (!tgt->ltd_obd->obd_set_up) {
411 CERROR("Target %s not set up\n", obd_uuid2str(&tgt->ltd_uuid));
412 return -EINVAL;
413 }
414
415 rc = lov_expand_targets(env, ld);
416 if (rc == 0 && ld->ld_flags & LOV_DEV_INITIALIZED) {
417 LASSERT(dev->ld_site != NULL);
418
419 cl = cl_type_setup(env, dev->ld_site, &lovsub_device_type,
420 tgt->ltd_obd->obd_lu_dev);
421 if (!IS_ERR(cl)) {
422 lsd = cl2lovsub_dev(cl);
423 lsd->acid_idx = index;
424 lsd->acid_super = ld;
425 ld->ld_target[index] = lsd;
426 } else {
427 CERROR("add failed (%d), deleting %s\n", rc,
428 obd_uuid2str(&tgt->ltd_uuid));
429 lov_cl_del_target(env, dev, index);
430 rc = PTR_ERR(cl);
431 }
432 }
433 obd_putref(obd);
434 return rc;
435}
436
437static int lov_process_config(const struct lu_env *env,
438 struct lu_device *d, struct lustre_cfg *cfg)
439{
440 struct obd_device *obd = d->ld_obd;
441 int cmd;
442 int rc;
443 int gen;
444 __u32 index;
445
446 obd_getref(obd);
447
448 cmd = cfg->lcfg_command;
449 rc = lov_process_config_base(d->ld_obd, cfg, &index, &gen);
450 if (rc == 0) {
451 switch (cmd) {
452 case LCFG_LOV_ADD_OBD:
453 case LCFG_LOV_ADD_INA:
454 rc = lov_cl_add_target(env, d, index);
455 if (rc != 0)
456 lov_del_target(d->ld_obd, index, NULL, 0);
457 break;
458 case LCFG_LOV_DEL_OBD:
459 lov_cl_del_target(env, d, index);
460 break;
461 }
462 }
463 obd_putref(obd);
464 return rc;
465}
466
467static const struct lu_device_operations lov_lu_ops = {
468 .ldo_object_alloc = lov_object_alloc,
469 .ldo_process_config = lov_process_config,
470};
471
472static struct lu_device *lov_device_alloc(const struct lu_env *env,
473 struct lu_device_type *t,
474 struct lustre_cfg *cfg)
475{
476 struct lu_device *d;
477 struct lov_device *ld;
478 struct obd_device *obd;
479 int rc;
480
481 ld = kzalloc(sizeof(*ld), GFP_NOFS);
482 if (!ld)
483 return ERR_PTR(-ENOMEM);
484
485 cl_device_init(&ld->ld_cl, t);
486 d = lov2lu_dev(ld);
487 d->ld_ops = &lov_lu_ops;
488 ld->ld_cl.cd_ops = &lov_cl_ops;
489
490 mutex_init(&ld->ld_mutex);
491 lockdep_set_class(&ld->ld_mutex, &cl_lov_device_mutex_class);
492
493
494 obd = class_name2obd(lustre_cfg_string(cfg, 0));
495 LASSERT(obd != NULL);
496 rc = lov_setup(obd, cfg);
497 if (rc) {
498 lov_device_free(env, d);
499 return ERR_PTR(rc);
500 }
501
502 ld->ld_lov = &obd->u.lov;
503 return d;
504}
505
506static const struct lu_device_type_operations lov_device_type_ops = {
507 .ldto_init = lov_type_init,
508 .ldto_fini = lov_type_fini,
509
510 .ldto_start = lov_type_start,
511 .ldto_stop = lov_type_stop,
512
513 .ldto_device_alloc = lov_device_alloc,
514 .ldto_device_free = lov_device_free,
515
516 .ldto_device_init = lov_device_init,
517 .ldto_device_fini = lov_device_fini
518};
519
520struct lu_device_type lov_device_type = {
521 .ldt_tags = LU_DEVICE_CL,
522 .ldt_name = LUSTRE_LOV_NAME,
523 .ldt_ops = &lov_device_type_ops,
524 .ldt_ctx_tags = LCT_CL_THREAD
525};
526EXPORT_SYMBOL(lov_device_type);
527
528
529