1
2
3
4
5
6
7
8
9
10
11
12
13
14
15#ifndef _MD_MD_H
16#define _MD_MD_H
17
18#include <linux/blkdev.h>
19#include <linux/backing-dev.h>
20#include <linux/badblocks.h>
21#include <linux/kobject.h>
22#include <linux/list.h>
23#include <linux/mm.h>
24#include <linux/mutex.h>
25#include <linux/timer.h>
26#include <linux/wait.h>
27#include <linux/workqueue.h>
28#include "md-cluster.h"
29
30#define MaxSector (~(sector_t)0)
31
32
33
34
35struct md_rdev {
36 struct list_head same_set;
37
38 sector_t sectors;
39 struct mddev *mddev;
40 int last_events;
41
42
43
44
45
46
47 struct block_device *meta_bdev;
48 struct block_device *bdev;
49
50 struct page *sb_page, *bb_page;
51 int sb_loaded;
52 __u64 sb_events;
53 sector_t data_offset;
54 sector_t new_data_offset;
55 sector_t sb_start;
56 int sb_size;
57 int preferred_minor;
58
59 struct kobject kobj;
60
61
62
63
64
65
66
67
68
69
70
71
72 unsigned long flags;
73 wait_queue_head_t blocked_wait;
74
75 int desc_nr;
76 int raid_disk;
77 int new_raid_disk;
78
79
80 int saved_raid_disk;
81
82
83
84 union {
85 sector_t recovery_offset;
86
87
88
89 sector_t journal_tail;
90
91
92
93 };
94
95 atomic_t nr_pending;
96
97
98
99 atomic_t read_errors;
100
101
102 struct timespec last_read_error;
103
104
105 atomic_t corrected_errors;
106
107
108
109 struct work_struct del_work;
110
111 struct kernfs_node *sysfs_state;
112
113
114 struct badblocks badblocks;
115};
116enum flag_bits {
117 Faulty,
118 In_sync,
119 Bitmap_sync,
120
121
122 WriteMostly,
123 AutoDetected,
124 Blocked,
125
126
127
128 WriteErrorSeen,
129
130
131 FaultRecorded,
132
133
134
135
136
137 BlockedBadBlocks,
138
139
140
141
142
143
144
145
146
147 WantReplacement,
148
149
150
151
152 Replacement,
153
154
155
156 Candidate,
157
158
159
160 Journal,
161
162
163
164
165 ClusterRemove,
166};
167
168static inline int is_badblock(struct md_rdev *rdev, sector_t s, int sectors,
169 sector_t *first_bad, int *bad_sectors)
170{
171 if (unlikely(rdev->badblocks.count)) {
172 int rv = badblocks_check(&rdev->badblocks, rdev->data_offset + s,
173 sectors,
174 first_bad, bad_sectors);
175 if (rv)
176 *first_bad -= rdev->data_offset;
177 return rv;
178 }
179 return 0;
180}
181extern int rdev_set_badblocks(struct md_rdev *rdev, sector_t s, int sectors,
182 int is_new);
183extern int rdev_clear_badblocks(struct md_rdev *rdev, sector_t s, int sectors,
184 int is_new);
185struct md_cluster_info;
186
187struct mddev {
188 void *private;
189 struct md_personality *pers;
190 dev_t unit;
191 int md_minor;
192 struct list_head disks;
193 unsigned long flags;
194#define MD_CHANGE_DEVS 0
195#define MD_CHANGE_CLEAN 1
196#define MD_CHANGE_PENDING 2
197#define MD_UPDATE_SB_FLAGS (1 | 2 | 4)
198#define MD_ARRAY_FIRST_USE 3
199#define MD_STILL_CLOSED 4
200
201
202#define MD_JOURNAL_CLEAN 5
203#define MD_HAS_JOURNAL 6
204#define MD_RELOAD_SB 7
205
206
207
208 int suspended;
209 atomic_t active_io;
210 int ro;
211 int sysfs_active;
212
213
214
215 struct gendisk *gendisk;
216
217 struct kobject kobj;
218 int hold_active;
219#define UNTIL_IOCTL 1
220#define UNTIL_STOP 2
221
222
223 int major_version,
224 minor_version,
225 patch_version;
226 int persistent;
227 int external;
228
229 char metadata_type[17];
230 int chunk_sectors;
231 time64_t ctime, utime;
232 int level, layout;
233 char clevel[16];
234 int raid_disks;
235 int max_disks;
236 sector_t dev_sectors;
237
238 sector_t array_sectors;
239 int external_size;
240
241 __u64 events;
242
243
244
245
246
247 int can_decrease_events;
248
249 char uuid[16];
250
251
252
253
254
255
256 sector_t reshape_position;
257 int delta_disks, new_level, new_layout;
258 int new_chunk_sectors;
259 int reshape_backwards;
260
261 struct md_thread *thread;
262 struct md_thread *sync_thread;
263
264
265
266
267
268
269
270 char *last_sync_action;
271 sector_t curr_resync;
272
273
274
275
276
277
278 sector_t curr_resync_completed;
279 unsigned long resync_mark;
280 sector_t resync_mark_cnt;
281 sector_t curr_mark_cnt;
282
283 sector_t resync_max_sectors;
284
285 atomic64_t resync_mismatches;
286
287
288
289
290 sector_t suspend_lo;
291 sector_t suspend_hi;
292
293 int sync_speed_min;
294 int sync_speed_max;
295
296
297 int parallel_resync;
298
299 int ok_start_degraded;
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314#define MD_RECOVERY_RUNNING 0
315#define MD_RECOVERY_SYNC 1
316#define MD_RECOVERY_RECOVER 2
317#define MD_RECOVERY_INTR 3
318#define MD_RECOVERY_DONE 4
319#define MD_RECOVERY_NEEDED 5
320#define MD_RECOVERY_REQUESTED 6
321#define MD_RECOVERY_CHECK 7
322#define MD_RECOVERY_RESHAPE 8
323#define MD_RECOVERY_FROZEN 9
324#define MD_RECOVERY_ERROR 10
325
326 unsigned long recovery;
327
328
329
330
331
332 int recovery_disabled;
333
334 int in_sync;
335
336
337
338
339
340
341
342
343
344 struct mutex open_mutex;
345 struct mutex reconfig_mutex;
346 atomic_t active;
347 atomic_t openers;
348
349 int changed;
350
351 int degraded;
352
353
354
355 atomic_t recovery_active;
356 wait_queue_head_t recovery_wait;
357 sector_t recovery_cp;
358 sector_t resync_min;
359
360 sector_t resync_max;
361
362
363 struct kernfs_node *sysfs_state;
364
365
366 struct kernfs_node *sysfs_action;
367
368 struct work_struct del_work;
369
370
371
372
373
374
375
376
377
378
379
380
381 spinlock_t lock;
382 wait_queue_head_t sb_wait;
383 atomic_t pending_writes;
384
385 unsigned int safemode;
386
387
388 unsigned int safemode_delay;
389 struct timer_list safemode_timer;
390 atomic_t writes_pending;
391 struct request_queue *queue;
392
393 struct bitmap *bitmap;
394 struct {
395 struct file *file;
396 loff_t offset;
397
398
399
400
401
402 unsigned long space;
403 loff_t default_offset;
404
405
406
407 unsigned long default_space;
408
409 struct mutex mutex;
410 unsigned long chunksize;
411 unsigned long daemon_sleep;
412 unsigned long max_write_behind;
413 int external;
414 int nodes;
415 char cluster_name[64];
416 } bitmap_info;
417
418 atomic_t max_corr_read_errors;
419 struct list_head all_mddevs;
420
421 struct attribute_group *to_remove;
422
423 struct bio_set *bio_set;
424
425
426
427
428
429 struct bio *flush_bio;
430 atomic_t flush_pending;
431 struct work_struct flush_work;
432 struct work_struct event_work;
433 void (*sync_super)(struct mddev *mddev, struct md_rdev *rdev);
434 struct md_cluster_info *cluster_info;
435 unsigned int good_device_nr;
436};
437
438static inline int __must_check mddev_lock(struct mddev *mddev)
439{
440 return mutex_lock_interruptible(&mddev->reconfig_mutex);
441}
442
443
444
445
446static inline void mddev_lock_nointr(struct mddev *mddev)
447{
448 mutex_lock(&mddev->reconfig_mutex);
449}
450
451static inline int mddev_is_locked(struct mddev *mddev)
452{
453 return mutex_is_locked(&mddev->reconfig_mutex);
454}
455
456static inline int mddev_trylock(struct mddev *mddev)
457{
458 return mutex_trylock(&mddev->reconfig_mutex);
459}
460extern void mddev_unlock(struct mddev *mddev);
461
462static inline void md_sync_acct(struct block_device *bdev, unsigned long nr_sectors)
463{
464 atomic_add(nr_sectors, &bdev->bd_contains->bd_disk->sync_io);
465}
466
467struct md_personality
468{
469 char *name;
470 int level;
471 struct list_head list;
472 struct module *owner;
473 void (*make_request)(struct mddev *mddev, struct bio *bio);
474 int (*run)(struct mddev *mddev);
475 void (*free)(struct mddev *mddev, void *priv);
476 void (*status)(struct seq_file *seq, struct mddev *mddev);
477
478
479
480 void (*error_handler)(struct mddev *mddev, struct md_rdev *rdev);
481 int (*hot_add_disk) (struct mddev *mddev, struct md_rdev *rdev);
482 int (*hot_remove_disk) (struct mddev *mddev, struct md_rdev *rdev);
483 int (*spare_active) (struct mddev *mddev);
484 sector_t (*sync_request)(struct mddev *mddev, sector_t sector_nr, int *skipped);
485 int (*resize) (struct mddev *mddev, sector_t sectors);
486 sector_t (*size) (struct mddev *mddev, sector_t sectors, int raid_disks);
487 int (*check_reshape) (struct mddev *mddev);
488 int (*start_reshape) (struct mddev *mddev);
489 void (*finish_reshape) (struct mddev *mddev);
490
491
492
493
494
495 void (*quiesce) (struct mddev *mddev, int state);
496
497
498
499
500
501
502
503
504
505 void *(*takeover) (struct mddev *mddev);
506
507
508 int (*congested)(struct mddev *mddev, int bits);
509};
510
511struct md_sysfs_entry {
512 struct attribute attr;
513 ssize_t (*show)(struct mddev *, char *);
514 ssize_t (*store)(struct mddev *, const char *, size_t);
515};
516extern struct attribute_group md_bitmap_group;
517
518static inline struct kernfs_node *sysfs_get_dirent_safe(struct kernfs_node *sd, char *name)
519{
520 if (sd)
521 return sysfs_get_dirent(sd, name);
522 return sd;
523}
524static inline void sysfs_notify_dirent_safe(struct kernfs_node *sd)
525{
526 if (sd)
527 sysfs_notify_dirent(sd);
528}
529
530static inline char * mdname (struct mddev * mddev)
531{
532 return mddev->gendisk ? mddev->gendisk->disk_name : "mdX";
533}
534
535static inline int sysfs_link_rdev(struct mddev *mddev, struct md_rdev *rdev)
536{
537 char nm[20];
538 if (!test_bit(Replacement, &rdev->flags) &&
539 !test_bit(Journal, &rdev->flags) &&
540 mddev->kobj.sd) {
541 sprintf(nm, "rd%d", rdev->raid_disk);
542 return sysfs_create_link(&mddev->kobj, &rdev->kobj, nm);
543 } else
544 return 0;
545}
546
547static inline void sysfs_unlink_rdev(struct mddev *mddev, struct md_rdev *rdev)
548{
549 char nm[20];
550 if (!test_bit(Replacement, &rdev->flags) &&
551 !test_bit(Journal, &rdev->flags) &&
552 mddev->kobj.sd) {
553 sprintf(nm, "rd%d", rdev->raid_disk);
554 sysfs_remove_link(&mddev->kobj, nm);
555 }
556}
557
558
559
560
561
562#define rdev_for_each_list(rdev, tmp, head) \
563 list_for_each_entry_safe(rdev, tmp, head, same_set)
564
565
566
567
568#define rdev_for_each(rdev, mddev) \
569 list_for_each_entry(rdev, &((mddev)->disks), same_set)
570
571#define rdev_for_each_safe(rdev, tmp, mddev) \
572 list_for_each_entry_safe(rdev, tmp, &((mddev)->disks), same_set)
573
574#define rdev_for_each_rcu(rdev, mddev) \
575 list_for_each_entry_rcu(rdev, &((mddev)->disks), same_set)
576
577struct md_thread {
578 void (*run) (struct md_thread *thread);
579 struct mddev *mddev;
580 wait_queue_head_t wqueue;
581 unsigned long flags;
582 struct task_struct *tsk;
583 unsigned long timeout;
584 void *private;
585};
586
587#define THREAD_WAKEUP 0
588
589static inline void safe_put_page(struct page *p)
590{
591 if (p) put_page(p);
592}
593
594extern int register_md_personality(struct md_personality *p);
595extern int unregister_md_personality(struct md_personality *p);
596extern int register_md_cluster_operations(struct md_cluster_operations *ops,
597 struct module *module);
598extern int unregister_md_cluster_operations(void);
599extern int md_setup_cluster(struct mddev *mddev, int nodes);
600extern void md_cluster_stop(struct mddev *mddev);
601extern struct md_thread *md_register_thread(
602 void (*run)(struct md_thread *thread),
603 struct mddev *mddev,
604 const char *name);
605extern void md_unregister_thread(struct md_thread **threadp);
606extern void md_wakeup_thread(struct md_thread *thread);
607extern void md_check_recovery(struct mddev *mddev);
608extern void md_reap_sync_thread(struct mddev *mddev);
609extern void md_write_start(struct mddev *mddev, struct bio *bi);
610extern void md_write_end(struct mddev *mddev);
611extern void md_done_sync(struct mddev *mddev, int blocks, int ok);
612extern void md_error(struct mddev *mddev, struct md_rdev *rdev);
613extern void md_finish_reshape(struct mddev *mddev);
614
615extern int mddev_congested(struct mddev *mddev, int bits);
616extern void md_flush_request(struct mddev *mddev, struct bio *bio);
617extern void md_super_write(struct mddev *mddev, struct md_rdev *rdev,
618 sector_t sector, int size, struct page *page);
619extern void md_super_wait(struct mddev *mddev);
620extern int sync_page_io(struct md_rdev *rdev, sector_t sector, int size,
621 struct page *page, int rw, bool metadata_op);
622extern void md_do_sync(struct md_thread *thread);
623extern void md_new_event(struct mddev *mddev);
624extern int md_allow_write(struct mddev *mddev);
625extern void md_wait_for_blocked_rdev(struct md_rdev *rdev, struct mddev *mddev);
626extern void md_set_array_sectors(struct mddev *mddev, sector_t array_sectors);
627extern int md_check_no_bitmap(struct mddev *mddev);
628extern int md_integrity_register(struct mddev *mddev);
629extern int md_integrity_add_rdev(struct md_rdev *rdev, struct mddev *mddev);
630extern int strict_strtoul_scaled(const char *cp, unsigned long *res, int scale);
631
632extern void mddev_init(struct mddev *mddev);
633extern int md_run(struct mddev *mddev);
634extern void md_stop(struct mddev *mddev);
635extern void md_stop_writes(struct mddev *mddev);
636extern int md_rdev_init(struct md_rdev *rdev);
637extern void md_rdev_clear(struct md_rdev *rdev);
638
639extern void mddev_suspend(struct mddev *mddev);
640extern void mddev_resume(struct mddev *mddev);
641extern struct bio *bio_clone_mddev(struct bio *bio, gfp_t gfp_mask,
642 struct mddev *mddev);
643extern struct bio *bio_alloc_mddev(gfp_t gfp_mask, int nr_iovecs,
644 struct mddev *mddev);
645
646extern void md_unplug(struct blk_plug_cb *cb, bool from_schedule);
647extern void md_reload_sb(struct mddev *mddev, int raid_disk);
648extern void md_update_sb(struct mddev *mddev, int force);
649extern void md_kick_rdev_from_array(struct md_rdev * rdev);
650struct md_rdev *md_find_rdev_nr_rcu(struct mddev *mddev, int nr);
651static inline int mddev_check_plugged(struct mddev *mddev)
652{
653 return !!blk_check_plugged(md_unplug, mddev,
654 sizeof(struct blk_plug_cb));
655}
656
657static inline void rdev_dec_pending(struct md_rdev *rdev, struct mddev *mddev)
658{
659 int faulty = test_bit(Faulty, &rdev->flags);
660 if (atomic_dec_and_test(&rdev->nr_pending) && faulty) {
661 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery);
662 md_wakeup_thread(mddev->thread);
663 }
664}
665
666extern struct md_cluster_operations *md_cluster_ops;
667static inline int mddev_is_clustered(struct mddev *mddev)
668{
669 return mddev->cluster_info && mddev->bitmap_info.nodes > 1;
670}
671#endif
672