linux/fs/btrfs/extent-io-tree.h
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0 */
   2
   3#ifndef BTRFS_EXTENT_IO_TREE_H
   4#define BTRFS_EXTENT_IO_TREE_H
   5
   6struct extent_changeset;
   7struct io_failure_record;
   8
   9/* Bits for the extent state */
  10#define EXTENT_DIRTY            (1U << 0)
  11#define EXTENT_UPTODATE         (1U << 1)
  12#define EXTENT_LOCKED           (1U << 2)
  13#define EXTENT_NEW              (1U << 3)
  14#define EXTENT_DELALLOC         (1U << 4)
  15#define EXTENT_DEFRAG           (1U << 5)
  16#define EXTENT_BOUNDARY         (1U << 6)
  17#define EXTENT_NODATASUM        (1U << 7)
  18#define EXTENT_CLEAR_META_RESV  (1U << 8)
  19#define EXTENT_NEED_WAIT        (1U << 9)
  20#define EXTENT_DAMAGED          (1U << 10)
  21#define EXTENT_NORESERVE        (1U << 11)
  22#define EXTENT_QGROUP_RESERVED  (1U << 12)
  23#define EXTENT_CLEAR_DATA_RESV  (1U << 13)
  24/*
  25 * Must be cleared only during ordered extent completion or on error paths if we
  26 * did not manage to submit bios and create the ordered extents for the range.
  27 * Should not be cleared during page release and page invalidation (if there is
  28 * an ordered extent in flight), that is left for the ordered extent completion.
  29 */
  30#define EXTENT_DELALLOC_NEW     (1U << 14)
  31/*
  32 * When an ordered extent successfully completes for a region marked as a new
  33 * delalloc range, use this flag when clearing a new delalloc range to indicate
  34 * that the VFS' inode number of bytes should be incremented and the inode's new
  35 * delalloc bytes decremented, in an atomic way to prevent races with stat(2).
  36 */
  37#define EXTENT_ADD_INODE_BYTES  (1U << 15)
  38#define EXTENT_DO_ACCOUNTING    (EXTENT_CLEAR_META_RESV | \
  39                                 EXTENT_CLEAR_DATA_RESV)
  40#define EXTENT_CTLBITS          (EXTENT_DO_ACCOUNTING | \
  41                                 EXTENT_ADD_INODE_BYTES)
  42
  43/*
  44 * Redefined bits above which are used only in the device allocation tree,
  45 * shouldn't be using EXTENT_LOCKED / EXTENT_BOUNDARY / EXTENT_CLEAR_META_RESV
  46 * / EXTENT_CLEAR_DATA_RESV because they have special meaning to the bit
  47 * manipulation functions
  48 */
  49#define CHUNK_ALLOCATED                         EXTENT_DIRTY
  50#define CHUNK_TRIMMED                           EXTENT_DEFRAG
  51#define CHUNK_STATE_MASK                        (CHUNK_ALLOCATED |              \
  52                                                 CHUNK_TRIMMED)
  53
  54enum {
  55        IO_TREE_FS_PINNED_EXTENTS,
  56        IO_TREE_FS_EXCLUDED_EXTENTS,
  57        IO_TREE_BTREE_INODE_IO,
  58        IO_TREE_INODE_IO,
  59        IO_TREE_INODE_IO_FAILURE,
  60        IO_TREE_RELOC_BLOCKS,
  61        IO_TREE_TRANS_DIRTY_PAGES,
  62        IO_TREE_ROOT_DIRTY_LOG_PAGES,
  63        IO_TREE_INODE_FILE_EXTENT,
  64        IO_TREE_LOG_CSUM_RANGE,
  65        IO_TREE_SELFTEST,
  66        IO_TREE_DEVICE_ALLOC_STATE,
  67};
  68
  69struct extent_io_tree {
  70        struct rb_root state;
  71        struct btrfs_fs_info *fs_info;
  72        void *private_data;
  73        u64 dirty_bytes;
  74        bool track_uptodate;
  75
  76        /* Who owns this io tree, should be one of IO_TREE_* */
  77        u8 owner;
  78
  79        spinlock_t lock;
  80};
  81
  82struct extent_state {
  83        u64 start;
  84        u64 end; /* inclusive */
  85        struct rb_node rb_node;
  86
  87        /* ADD NEW ELEMENTS AFTER THIS */
  88        wait_queue_head_t wq;
  89        refcount_t refs;
  90        u32 state;
  91
  92        struct io_failure_record *failrec;
  93
  94#ifdef CONFIG_BTRFS_DEBUG
  95        struct list_head leak_list;
  96#endif
  97};
  98
  99int __init extent_state_cache_init(void);
 100void __cold extent_state_cache_exit(void);
 101
 102void extent_io_tree_init(struct btrfs_fs_info *fs_info,
 103                         struct extent_io_tree *tree, unsigned int owner,
 104                         void *private_data);
 105void extent_io_tree_release(struct extent_io_tree *tree);
 106
 107int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
 108                     struct extent_state **cached);
 109
 110static inline int lock_extent(struct extent_io_tree *tree, u64 start, u64 end)
 111{
 112        return lock_extent_bits(tree, start, end, NULL);
 113}
 114
 115int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
 116
 117int __init extent_io_init(void);
 118void __cold extent_io_exit(void);
 119
 120u64 count_range_bits(struct extent_io_tree *tree,
 121                     u64 *start, u64 search_end,
 122                     u64 max_bytes, u32 bits, int contig);
 123
 124void free_extent_state(struct extent_state *state);
 125int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end,
 126                   u32 bits, int filled, struct extent_state *cached_state);
 127int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
 128                             u32 bits, struct extent_changeset *changeset);
 129int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
 130                     u32 bits, int wake, int delete,
 131                     struct extent_state **cached);
 132int __clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
 133                     u32 bits, int wake, int delete,
 134                     struct extent_state **cached, gfp_t mask,
 135                     struct extent_changeset *changeset);
 136
 137static inline int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end)
 138{
 139        return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL);
 140}
 141
 142static inline int unlock_extent_cached(struct extent_io_tree *tree, u64 start,
 143                u64 end, struct extent_state **cached)
 144{
 145        return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached,
 146                                GFP_NOFS, NULL);
 147}
 148
 149static inline int unlock_extent_cached_atomic(struct extent_io_tree *tree,
 150                u64 start, u64 end, struct extent_state **cached)
 151{
 152        return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached,
 153                                GFP_ATOMIC, NULL);
 154}
 155
 156static inline int clear_extent_bits(struct extent_io_tree *tree, u64 start,
 157                                    u64 end, u32 bits)
 158{
 159        int wake = 0;
 160
 161        if (bits & EXTENT_LOCKED)
 162                wake = 1;
 163
 164        return clear_extent_bit(tree, start, end, bits, wake, 0, NULL);
 165}
 166
 167int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
 168                           u32 bits, struct extent_changeset *changeset);
 169int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
 170                   u32 bits, unsigned exclusive_bits, u64 *failed_start,
 171                   struct extent_state **cached_state, gfp_t mask,
 172                   struct extent_changeset *changeset);
 173int set_extent_bits_nowait(struct extent_io_tree *tree, u64 start, u64 end,
 174                           u32 bits);
 175
 176static inline int set_extent_bits(struct extent_io_tree *tree, u64 start,
 177                u64 end, u32 bits)
 178{
 179        return set_extent_bit(tree, start, end, bits, 0, NULL, NULL, GFP_NOFS,
 180                              NULL);
 181}
 182
 183static inline int clear_extent_uptodate(struct extent_io_tree *tree, u64 start,
 184                u64 end, struct extent_state **cached_state)
 185{
 186        return __clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0,
 187                                cached_state, GFP_NOFS, NULL);
 188}
 189
 190static inline int set_extent_dirty(struct extent_io_tree *tree, u64 start,
 191                u64 end, gfp_t mask)
 192{
 193        return set_extent_bit(tree, start, end, EXTENT_DIRTY, 0, NULL, NULL,
 194                              mask, NULL);
 195}
 196
 197static inline int clear_extent_dirty(struct extent_io_tree *tree, u64 start,
 198                                     u64 end, struct extent_state **cached)
 199{
 200        return clear_extent_bit(tree, start, end,
 201                                EXTENT_DIRTY | EXTENT_DELALLOC |
 202                                EXTENT_DO_ACCOUNTING, 0, 0, cached);
 203}
 204
 205int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
 206                       u32 bits, u32 clear_bits,
 207                       struct extent_state **cached_state);
 208
 209static inline int set_extent_delalloc(struct extent_io_tree *tree, u64 start,
 210                                      u64 end, u32 extra_bits,
 211                                      struct extent_state **cached_state)
 212{
 213        return set_extent_bit(tree, start, end,
 214                              EXTENT_DELALLOC | EXTENT_UPTODATE | extra_bits,
 215                              0, NULL, cached_state, GFP_NOFS, NULL);
 216}
 217
 218static inline int set_extent_defrag(struct extent_io_tree *tree, u64 start,
 219                u64 end, struct extent_state **cached_state)
 220{
 221        return set_extent_bit(tree, start, end,
 222                              EXTENT_DELALLOC | EXTENT_UPTODATE | EXTENT_DEFRAG,
 223                              0, NULL, cached_state, GFP_NOFS, NULL);
 224}
 225
 226static inline int set_extent_new(struct extent_io_tree *tree, u64 start,
 227                u64 end)
 228{
 229        return set_extent_bit(tree, start, end, EXTENT_NEW, 0, NULL, NULL,
 230                              GFP_NOFS, NULL);
 231}
 232
 233static inline int set_extent_uptodate(struct extent_io_tree *tree, u64 start,
 234                u64 end, struct extent_state **cached_state, gfp_t mask)
 235{
 236        return set_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, NULL,
 237                              cached_state, mask, NULL);
 238}
 239
 240int find_first_extent_bit(struct extent_io_tree *tree, u64 start,
 241                          u64 *start_ret, u64 *end_ret, u32 bits,
 242                          struct extent_state **cached_state);
 243void find_first_clear_extent_bit(struct extent_io_tree *tree, u64 start,
 244                                 u64 *start_ret, u64 *end_ret, u32 bits);
 245int find_contiguous_extent_bit(struct extent_io_tree *tree, u64 start,
 246                               u64 *start_ret, u64 *end_ret, u32 bits);
 247int extent_invalidatepage(struct extent_io_tree *tree,
 248                          struct page *page, unsigned long offset);
 249bool btrfs_find_delalloc_range(struct extent_io_tree *tree, u64 *start,
 250                               u64 *end, u64 max_bytes,
 251                               struct extent_state **cached_state);
 252
 253/* This should be reworked in the future and put elsewhere. */
 254struct io_failure_record *get_state_failrec(struct extent_io_tree *tree, u64 start);
 255int set_state_failrec(struct extent_io_tree *tree, u64 start,
 256                      struct io_failure_record *failrec);
 257void btrfs_free_io_failure_record(struct btrfs_inode *inode, u64 start,
 258                u64 end);
 259int free_io_failure(struct extent_io_tree *failure_tree,
 260                    struct extent_io_tree *io_tree,
 261                    struct io_failure_record *rec);
 262int clean_io_failure(struct btrfs_fs_info *fs_info,
 263                     struct extent_io_tree *failure_tree,
 264                     struct extent_io_tree *io_tree, u64 start,
 265                     struct page *page, u64 ino, unsigned int pg_offset);
 266
 267#endif /* BTRFS_EXTENT_IO_TREE_H */
 268