1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37#include <linux/module.h>
38#include <linux/fs.h>
39#include <linux/pagemap.h>
40#include <linux/types.h>
41#include <linux/slab.h>
42#include <linux/highmem.h>
43#include <linux/init.h>
44#include <linux/string.h>
45#include <linux/backing-dev.h>
46
47#include <asm/uaccess.h>
48
49
50#include "cluster/nodemanager.h"
51#include "cluster/heartbeat.h"
52#include "cluster/tcp.h"
53
54#include "dlmapi.h"
55
56#include "userdlm.h"
57
58#include "dlmfsver.h"
59
60#define MLOG_MASK_PREFIX ML_DLMFS
61#include "cluster/masklog.h"
62
63static const struct super_operations dlmfs_ops;
64static const struct file_operations dlmfs_file_operations;
65static const struct inode_operations dlmfs_dir_inode_operations;
66static const struct inode_operations dlmfs_root_inode_operations;
67static const struct inode_operations dlmfs_file_inode_operations;
68static struct kmem_cache *dlmfs_inode_cache;
69
70struct workqueue_struct *user_dlm_worker;
71
72
73
74
75
76
77
78
79
80
81static int dlmfs_decode_open_flags(int open_flags,
82 int *level,
83 int *flags)
84{
85 if (open_flags & (O_WRONLY|O_RDWR))
86 *level = LKM_EXMODE;
87 else
88 *level = LKM_PRMODE;
89
90 *flags = 0;
91 if (open_flags & O_NONBLOCK)
92 *flags |= LKM_NOQUEUE;
93
94 return 0;
95}
96
97static int dlmfs_file_open(struct inode *inode,
98 struct file *file)
99{
100 int status, level, flags;
101 struct dlmfs_filp_private *fp = NULL;
102 struct dlmfs_inode_private *ip;
103
104 if (S_ISDIR(inode->i_mode))
105 BUG();
106
107 mlog(0, "open called on inode %lu, flags 0x%x\n", inode->i_ino,
108 file->f_flags);
109
110 status = dlmfs_decode_open_flags(file->f_flags, &level, &flags);
111 if (status < 0)
112 goto bail;
113
114
115
116 file->f_flags &= ~O_APPEND;
117
118 fp = kmalloc(sizeof(*fp), GFP_NOFS);
119 if (!fp) {
120 status = -ENOMEM;
121 goto bail;
122 }
123 fp->fp_lock_level = level;
124
125 ip = DLMFS_I(inode);
126
127 status = user_dlm_cluster_lock(&ip->ip_lockres, level, flags);
128 if (status < 0) {
129
130
131
132
133 if (flags & LKM_NOQUEUE && status == -EAGAIN)
134 status = -ETXTBSY;
135 kfree(fp);
136 goto bail;
137 }
138
139 file->private_data = fp;
140bail:
141 return status;
142}
143
144static int dlmfs_file_release(struct inode *inode,
145 struct file *file)
146{
147 int level, status;
148 struct dlmfs_inode_private *ip = DLMFS_I(inode);
149 struct dlmfs_filp_private *fp =
150 (struct dlmfs_filp_private *) file->private_data;
151
152 if (S_ISDIR(inode->i_mode))
153 BUG();
154
155 mlog(0, "close called on inode %lu\n", inode->i_ino);
156
157 status = 0;
158 if (fp) {
159 level = fp->fp_lock_level;
160 if (level != LKM_IVMODE)
161 user_dlm_cluster_unlock(&ip->ip_lockres, level);
162
163 kfree(fp);
164 file->private_data = NULL;
165 }
166
167 return 0;
168}
169
170static ssize_t dlmfs_file_read(struct file *filp,
171 char __user *buf,
172 size_t count,
173 loff_t *ppos)
174{
175 int bytes_left;
176 ssize_t readlen;
177 char *lvb_buf;
178 struct inode *inode = filp->f_path.dentry->d_inode;
179
180 mlog(0, "inode %lu, count = %zu, *ppos = %llu\n",
181 inode->i_ino, count, *ppos);
182
183 if (*ppos >= i_size_read(inode))
184 return 0;
185
186 if (!count)
187 return 0;
188
189 if (!access_ok(VERIFY_WRITE, buf, count))
190 return -EFAULT;
191
192
193 if ((count + *ppos) > i_size_read(inode))
194 readlen = i_size_read(inode) - *ppos;
195 else
196 readlen = count - *ppos;
197
198 lvb_buf = kmalloc(readlen, GFP_NOFS);
199 if (!lvb_buf)
200 return -ENOMEM;
201
202 user_dlm_read_lvb(inode, lvb_buf, readlen);
203 bytes_left = __copy_to_user(buf, lvb_buf, readlen);
204 readlen -= bytes_left;
205
206 kfree(lvb_buf);
207
208 *ppos = *ppos + readlen;
209
210 mlog(0, "read %zd bytes\n", readlen);
211 return readlen;
212}
213
214static ssize_t dlmfs_file_write(struct file *filp,
215 const char __user *buf,
216 size_t count,
217 loff_t *ppos)
218{
219 int bytes_left;
220 ssize_t writelen;
221 char *lvb_buf;
222 struct inode *inode = filp->f_path.dentry->d_inode;
223
224 mlog(0, "inode %lu, count = %zu, *ppos = %llu\n",
225 inode->i_ino, count, *ppos);
226
227 if (*ppos >= i_size_read(inode))
228 return -ENOSPC;
229
230 if (!count)
231 return 0;
232
233 if (!access_ok(VERIFY_READ, buf, count))
234 return -EFAULT;
235
236
237 if ((count + *ppos) > i_size_read(inode))
238 writelen = i_size_read(inode) - *ppos;
239 else
240 writelen = count - *ppos;
241
242 lvb_buf = kmalloc(writelen, GFP_NOFS);
243 if (!lvb_buf)
244 return -ENOMEM;
245
246 bytes_left = copy_from_user(lvb_buf, buf, writelen);
247 writelen -= bytes_left;
248 if (writelen)
249 user_dlm_write_lvb(inode, lvb_buf, writelen);
250
251 kfree(lvb_buf);
252
253 *ppos = *ppos + writelen;
254 mlog(0, "wrote %zd bytes\n", writelen);
255 return writelen;
256}
257
258static void dlmfs_init_once(struct kmem_cache *cachep,
259 void *foo)
260{
261 struct dlmfs_inode_private *ip =
262 (struct dlmfs_inode_private *) foo;
263
264 ip->ip_dlm = NULL;
265 ip->ip_parent = NULL;
266
267 inode_init_once(&ip->ip_vfs_inode);
268}
269
270static struct inode *dlmfs_alloc_inode(struct super_block *sb)
271{
272 struct dlmfs_inode_private *ip;
273
274 ip = kmem_cache_alloc(dlmfs_inode_cache, GFP_NOFS);
275 if (!ip)
276 return NULL;
277
278 return &ip->ip_vfs_inode;
279}
280
281static void dlmfs_destroy_inode(struct inode *inode)
282{
283 kmem_cache_free(dlmfs_inode_cache, DLMFS_I(inode));
284}
285
286static void dlmfs_clear_inode(struct inode *inode)
287{
288 int status;
289 struct dlmfs_inode_private *ip;
290
291 if (!inode)
292 return;
293
294 mlog(0, "inode %lu\n", inode->i_ino);
295
296 ip = DLMFS_I(inode);
297
298 if (S_ISREG(inode->i_mode)) {
299 status = user_dlm_destroy_lock(&ip->ip_lockres);
300 if (status < 0)
301 mlog_errno(status);
302 iput(ip->ip_parent);
303 goto clear_fields;
304 }
305
306 mlog(0, "we're a directory, ip->ip_dlm = 0x%p\n", ip->ip_dlm);
307
308
309 if (ip->ip_dlm)
310 user_dlm_unregister_context(ip->ip_dlm);
311clear_fields:
312 ip->ip_parent = NULL;
313 ip->ip_dlm = NULL;
314}
315
316static struct backing_dev_info dlmfs_backing_dev_info = {
317 .ra_pages = 0,
318 .capabilities = BDI_CAP_NO_ACCT_DIRTY | BDI_CAP_NO_WRITEBACK,
319};
320
321static struct inode *dlmfs_get_root_inode(struct super_block *sb)
322{
323 struct inode *inode = new_inode(sb);
324 int mode = S_IFDIR | 0755;
325 struct dlmfs_inode_private *ip;
326
327 if (inode) {
328 ip = DLMFS_I(inode);
329
330 inode->i_mode = mode;
331 inode->i_uid = current->fsuid;
332 inode->i_gid = current->fsgid;
333 inode->i_blocks = 0;
334 inode->i_mapping->backing_dev_info = &dlmfs_backing_dev_info;
335 inode->i_atime = inode->i_mtime = inode->i_ctime = CURRENT_TIME;
336 inc_nlink(inode);
337
338 inode->i_fop = &simple_dir_operations;
339 inode->i_op = &dlmfs_root_inode_operations;
340 }
341
342 return inode;
343}
344
345static struct inode *dlmfs_get_inode(struct inode *parent,
346 struct dentry *dentry,
347 int mode)
348{
349 struct super_block *sb = parent->i_sb;
350 struct inode * inode = new_inode(sb);
351 struct dlmfs_inode_private *ip;
352
353 if (!inode)
354 return NULL;
355
356 inode->i_mode = mode;
357 inode->i_uid = current->fsuid;
358 inode->i_gid = current->fsgid;
359 inode->i_blocks = 0;
360 inode->i_mapping->backing_dev_info = &dlmfs_backing_dev_info;
361 inode->i_atime = inode->i_mtime = inode->i_ctime = CURRENT_TIME;
362
363 ip = DLMFS_I(inode);
364 ip->ip_dlm = DLMFS_I(parent)->ip_dlm;
365
366 switch (mode & S_IFMT) {
367 default:
368
369
370 BUG();
371 break;
372 case S_IFREG:
373 inode->i_op = &dlmfs_file_inode_operations;
374 inode->i_fop = &dlmfs_file_operations;
375
376 i_size_write(inode, DLM_LVB_LEN);
377
378 user_dlm_lock_res_init(&ip->ip_lockres, dentry);
379
380
381
382
383
384 ip->ip_parent = igrab(parent);
385 BUG_ON(!ip->ip_parent);
386 break;
387 case S_IFDIR:
388 inode->i_op = &dlmfs_dir_inode_operations;
389 inode->i_fop = &simple_dir_operations;
390
391
392
393 inc_nlink(inode);
394 break;
395 }
396
397 if (parent->i_mode & S_ISGID) {
398 inode->i_gid = parent->i_gid;
399 if (S_ISDIR(mode))
400 inode->i_mode |= S_ISGID;
401 }
402
403 return inode;
404}
405
406
407
408
409
410static int dlmfs_mkdir(struct inode * dir,
411 struct dentry * dentry,
412 int mode)
413{
414 int status;
415 struct inode *inode = NULL;
416 struct qstr *domain = &dentry->d_name;
417 struct dlmfs_inode_private *ip;
418 struct dlm_ctxt *dlm;
419
420 mlog(0, "mkdir %.*s\n", domain->len, domain->name);
421
422
423 if (domain->len >= O2NM_MAX_NAME_LEN) {
424 status = -EINVAL;
425 mlog(ML_ERROR, "invalid domain name for directory.\n");
426 goto bail;
427 }
428
429 inode = dlmfs_get_inode(dir, dentry, mode | S_IFDIR);
430 if (!inode) {
431 status = -ENOMEM;
432 mlog_errno(status);
433 goto bail;
434 }
435
436 ip = DLMFS_I(inode);
437
438 dlm = user_dlm_register_context(domain);
439 if (IS_ERR(dlm)) {
440 status = PTR_ERR(dlm);
441 mlog(ML_ERROR, "Error %d could not register domain \"%.*s\"\n",
442 status, domain->len, domain->name);
443 goto bail;
444 }
445 ip->ip_dlm = dlm;
446
447 inc_nlink(dir);
448 d_instantiate(dentry, inode);
449 dget(dentry);
450
451 status = 0;
452bail:
453 if (status < 0)
454 iput(inode);
455 return status;
456}
457
458static int dlmfs_create(struct inode *dir,
459 struct dentry *dentry,
460 int mode,
461 struct nameidata *nd)
462{
463 int status = 0;
464 struct inode *inode;
465 struct qstr *name = &dentry->d_name;
466
467 mlog(0, "create %.*s\n", name->len, name->name);
468
469
470
471 if (name->len >= USER_DLM_LOCK_ID_MAX_LEN ||
472 name->name[0] == '$') {
473 status = -EINVAL;
474 mlog(ML_ERROR, "invalid lock name, %.*s\n", name->len,
475 name->name);
476 goto bail;
477 }
478
479 inode = dlmfs_get_inode(dir, dentry, mode | S_IFREG);
480 if (!inode) {
481 status = -ENOMEM;
482 mlog_errno(status);
483 goto bail;
484 }
485
486 d_instantiate(dentry, inode);
487 dget(dentry);
488bail:
489 return status;
490}
491
492static int dlmfs_unlink(struct inode *dir,
493 struct dentry *dentry)
494{
495 int status;
496 struct inode *inode = dentry->d_inode;
497
498 mlog(0, "unlink inode %lu\n", inode->i_ino);
499
500
501
502 status = user_dlm_destroy_lock(&DLMFS_I(inode)->ip_lockres);
503 if (status < 0) {
504 mlog(ML_ERROR, "unlink %.*s, error %d from destroy\n",
505 dentry->d_name.len, dentry->d_name.name, status);
506 goto bail;
507 }
508 status = simple_unlink(dir, dentry);
509bail:
510 return status;
511}
512
513static int dlmfs_fill_super(struct super_block * sb,
514 void * data,
515 int silent)
516{
517 struct inode * inode;
518 struct dentry * root;
519
520 sb->s_maxbytes = MAX_LFS_FILESIZE;
521 sb->s_blocksize = PAGE_CACHE_SIZE;
522 sb->s_blocksize_bits = PAGE_CACHE_SHIFT;
523 sb->s_magic = DLMFS_MAGIC;
524 sb->s_op = &dlmfs_ops;
525 inode = dlmfs_get_root_inode(sb);
526 if (!inode)
527 return -ENOMEM;
528
529 root = d_alloc_root(inode);
530 if (!root) {
531 iput(inode);
532 return -ENOMEM;
533 }
534 sb->s_root = root;
535 return 0;
536}
537
538static const struct file_operations dlmfs_file_operations = {
539 .open = dlmfs_file_open,
540 .release = dlmfs_file_release,
541 .read = dlmfs_file_read,
542 .write = dlmfs_file_write,
543};
544
545static const struct inode_operations dlmfs_dir_inode_operations = {
546 .create = dlmfs_create,
547 .lookup = simple_lookup,
548 .unlink = dlmfs_unlink,
549};
550
551
552static const struct inode_operations dlmfs_root_inode_operations = {
553 .lookup = simple_lookup,
554 .mkdir = dlmfs_mkdir,
555 .rmdir = simple_rmdir,
556};
557
558static const struct super_operations dlmfs_ops = {
559 .statfs = simple_statfs,
560 .alloc_inode = dlmfs_alloc_inode,
561 .destroy_inode = dlmfs_destroy_inode,
562 .clear_inode = dlmfs_clear_inode,
563 .drop_inode = generic_delete_inode,
564};
565
566static const struct inode_operations dlmfs_file_inode_operations = {
567 .getattr = simple_getattr,
568};
569
570static int dlmfs_get_sb(struct file_system_type *fs_type,
571 int flags, const char *dev_name, void *data, struct vfsmount *mnt)
572{
573 return get_sb_nodev(fs_type, flags, data, dlmfs_fill_super, mnt);
574}
575
576static struct file_system_type dlmfs_fs_type = {
577 .owner = THIS_MODULE,
578 .name = "ocfs2_dlmfs",
579 .get_sb = dlmfs_get_sb,
580 .kill_sb = kill_litter_super,
581};
582
583static int __init init_dlmfs_fs(void)
584{
585 int status;
586 int cleanup_inode = 0, cleanup_worker = 0;
587
588 dlmfs_print_version();
589
590 status = bdi_init(&dlmfs_backing_dev_info);
591 if (status)
592 return status;
593
594 dlmfs_inode_cache = kmem_cache_create("dlmfs_inode_cache",
595 sizeof(struct dlmfs_inode_private),
596 0, (SLAB_HWCACHE_ALIGN|SLAB_RECLAIM_ACCOUNT|
597 SLAB_MEM_SPREAD),
598 dlmfs_init_once);
599 if (!dlmfs_inode_cache)
600 goto bail;
601 cleanup_inode = 1;
602
603 user_dlm_worker = create_singlethread_workqueue("user_dlm");
604 if (!user_dlm_worker) {
605 status = -ENOMEM;
606 goto bail;
607 }
608 cleanup_worker = 1;
609
610 status = register_filesystem(&dlmfs_fs_type);
611bail:
612 if (status) {
613 if (cleanup_inode)
614 kmem_cache_destroy(dlmfs_inode_cache);
615 if (cleanup_worker)
616 destroy_workqueue(user_dlm_worker);
617 bdi_destroy(&dlmfs_backing_dev_info);
618 } else
619 printk("OCFS2 User DLM kernel interface loaded\n");
620 return status;
621}
622
623static void __exit exit_dlmfs_fs(void)
624{
625 unregister_filesystem(&dlmfs_fs_type);
626
627 flush_workqueue(user_dlm_worker);
628 destroy_workqueue(user_dlm_worker);
629
630 kmem_cache_destroy(dlmfs_inode_cache);
631
632 bdi_destroy(&dlmfs_backing_dev_info);
633}
634
635MODULE_AUTHOR("Oracle");
636MODULE_LICENSE("GPL");
637
638module_init(init_dlmfs_fs)
639module_exit(exit_dlmfs_fs)
640