linux/include/linux/ceph/osdmap.h
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0 */
   2#ifndef _FS_CEPH_OSDMAP_H
   3#define _FS_CEPH_OSDMAP_H
   4
   5#include <linux/rbtree.h>
   6#include <linux/ceph/types.h>
   7#include <linux/ceph/decode.h>
   8#include <linux/crush/crush.h>
   9
  10/*
  11 * The osd map describes the current membership of the osd cluster and
  12 * specifies the mapping of objects to placement groups and placement
  13 * groups to (sets of) osds.  That is, it completely specifies the
  14 * (desired) distribution of all data objects in the system at some
  15 * point in time.
  16 *
  17 * Each map version is identified by an epoch, which increases monotonically.
  18 *
  19 * The map can be updated either via an incremental map (diff) describing
  20 * the change between two successive epochs, or as a fully encoded map.
  21 */
  22struct ceph_pg {
  23        uint64_t pool;
  24        uint32_t seed;
  25};
  26
  27#define CEPH_SPG_NOSHARD        -1
  28
  29struct ceph_spg {
  30        struct ceph_pg pgid;
  31        s8 shard;
  32};
  33
  34int ceph_pg_compare(const struct ceph_pg *lhs, const struct ceph_pg *rhs);
  35int ceph_spg_compare(const struct ceph_spg *lhs, const struct ceph_spg *rhs);
  36
  37#define CEPH_POOL_FLAG_HASHPSPOOL       (1ULL << 0) /* hash pg seed and pool id
  38                                                       together */
  39#define CEPH_POOL_FLAG_FULL             (1ULL << 1) /* pool is full */
  40#define CEPH_POOL_FLAG_FULL_QUOTA       (1ULL << 10) /* pool ran out of quota,
  41                                                        will set FULL too */
  42#define CEPH_POOL_FLAG_NEARFULL         (1ULL << 11) /* pool is nearfull */
  43
  44struct ceph_pg_pool_info {
  45        struct rb_node node;
  46        s64 id;
  47        u8 type; /* CEPH_POOL_TYPE_* */
  48        u8 size;
  49        u8 min_size;
  50        u8 crush_ruleset;
  51        u8 object_hash;
  52        u32 last_force_request_resend;
  53        u32 pg_num, pgp_num;
  54        int pg_num_mask, pgp_num_mask;
  55        s64 read_tier;
  56        s64 write_tier; /* wins for read+write ops */
  57        u64 flags; /* CEPH_POOL_FLAG_* */
  58        char *name;
  59
  60        bool was_full;  /* for handle_one_map() */
  61};
  62
  63static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool)
  64{
  65        switch (pool->type) {
  66        case CEPH_POOL_TYPE_REP:
  67                return true;
  68        case CEPH_POOL_TYPE_EC:
  69                return false;
  70        default:
  71                BUG();
  72        }
  73}
  74
  75struct ceph_object_locator {
  76        s64 pool;
  77        struct ceph_string *pool_ns;
  78};
  79
  80static inline void ceph_oloc_init(struct ceph_object_locator *oloc)
  81{
  82        oloc->pool = -1;
  83        oloc->pool_ns = NULL;
  84}
  85
  86static inline bool ceph_oloc_empty(const struct ceph_object_locator *oloc)
  87{
  88        return oloc->pool == -1;
  89}
  90
  91void ceph_oloc_copy(struct ceph_object_locator *dest,
  92                    const struct ceph_object_locator *src);
  93void ceph_oloc_destroy(struct ceph_object_locator *oloc);
  94
  95/*
  96 * 51-char inline_name is long enough for all cephfs and all but one
  97 * rbd requests: <imgname> in "<imgname>.rbd"/"rbd_id.<imgname>" can be
  98 * arbitrarily long (~PAGE_SIZE).  It's done once during rbd map; all
  99 * other rbd requests fit into inline_name.
 100 *
 101 * Makes ceph_object_id 64 bytes on 64-bit.
 102 */
 103#define CEPH_OID_INLINE_LEN 52
 104
 105/*
 106 * Both inline and external buffers have space for a NUL-terminator,
 107 * which is carried around.  It's not required though - RADOS object
 108 * names don't have to be NUL-terminated and may contain NULs.
 109 */
 110struct ceph_object_id {
 111        char *name;
 112        char inline_name[CEPH_OID_INLINE_LEN];
 113        int name_len;
 114};
 115
 116#define __CEPH_OID_INITIALIZER(oid) { .name = (oid).inline_name }
 117
 118#define CEPH_DEFINE_OID_ONSTACK(oid)                            \
 119        struct ceph_object_id oid = __CEPH_OID_INITIALIZER(oid)
 120
 121static inline void ceph_oid_init(struct ceph_object_id *oid)
 122{
 123        *oid = (struct ceph_object_id) __CEPH_OID_INITIALIZER(*oid);
 124}
 125
 126static inline bool ceph_oid_empty(const struct ceph_object_id *oid)
 127{
 128        return oid->name == oid->inline_name && !oid->name_len;
 129}
 130
 131void ceph_oid_copy(struct ceph_object_id *dest,
 132                   const struct ceph_object_id *src);
 133__printf(2, 3)
 134void ceph_oid_printf(struct ceph_object_id *oid, const char *fmt, ...);
 135__printf(3, 4)
 136int ceph_oid_aprintf(struct ceph_object_id *oid, gfp_t gfp,
 137                     const char *fmt, ...);
 138void ceph_oid_destroy(struct ceph_object_id *oid);
 139
 140struct workspace_manager {
 141        struct list_head idle_ws;
 142        spinlock_t ws_lock;
 143        /* Number of free workspaces */
 144        int free_ws;
 145        /* Total number of allocated workspaces */
 146        atomic_t total_ws;
 147        /* Waiters for a free workspace */
 148        wait_queue_head_t ws_wait;
 149};
 150
 151struct ceph_pg_mapping {
 152        struct rb_node node;
 153        struct ceph_pg pgid;
 154
 155        union {
 156                struct {
 157                        int len;
 158                        int osds[];
 159                } pg_temp, pg_upmap;
 160                struct {
 161                        int osd;
 162                } primary_temp;
 163                struct {
 164                        int len;
 165                        int from_to[][2];
 166                } pg_upmap_items;
 167        };
 168};
 169
 170struct ceph_osdmap {
 171        struct ceph_fsid fsid;
 172        u32 epoch;
 173        struct ceph_timespec created, modified;
 174
 175        u32 flags;         /* CEPH_OSDMAP_* */
 176
 177        u32 max_osd;       /* size of osd_state, _offload, _addr arrays */
 178        u32 *osd_state;    /* CEPH_OSD_* */
 179        u32 *osd_weight;   /* 0 = failed, 0x10000 = 100% normal */
 180        struct ceph_entity_addr *osd_addr;
 181
 182        struct rb_root pg_temp;
 183        struct rb_root primary_temp;
 184
 185        /* remap (post-CRUSH, pre-up) */
 186        struct rb_root pg_upmap;        /* PG := raw set */
 187        struct rb_root pg_upmap_items;  /* from -> to within raw set */
 188
 189        u32 *osd_primary_affinity;
 190
 191        struct rb_root pg_pools;
 192        u32 pool_max;
 193
 194        /* the CRUSH map specifies the mapping of placement groups to
 195         * the list of osds that store+replicate them. */
 196        struct crush_map *crush;
 197
 198        struct workspace_manager crush_wsm;
 199};
 200
 201static inline bool ceph_osd_exists(struct ceph_osdmap *map, int osd)
 202{
 203        return osd >= 0 && osd < map->max_osd &&
 204               (map->osd_state[osd] & CEPH_OSD_EXISTS);
 205}
 206
 207static inline bool ceph_osd_is_up(struct ceph_osdmap *map, int osd)
 208{
 209        return ceph_osd_exists(map, osd) &&
 210               (map->osd_state[osd] & CEPH_OSD_UP);
 211}
 212
 213static inline bool ceph_osd_is_down(struct ceph_osdmap *map, int osd)
 214{
 215        return !ceph_osd_is_up(map, osd);
 216}
 217
 218char *ceph_osdmap_state_str(char *str, int len, u32 state);
 219extern u32 ceph_get_primary_affinity(struct ceph_osdmap *map, int osd);
 220
 221static inline struct ceph_entity_addr *ceph_osd_addr(struct ceph_osdmap *map,
 222                                                     int osd)
 223{
 224        if (osd >= map->max_osd)
 225                return NULL;
 226        return &map->osd_addr[osd];
 227}
 228
 229#define CEPH_PGID_ENCODING_LEN          (1 + 8 + 4 + 4)
 230
 231static inline int ceph_decode_pgid(void **p, void *end, struct ceph_pg *pgid)
 232{
 233        __u8 version;
 234
 235        if (!ceph_has_room(p, end, CEPH_PGID_ENCODING_LEN)) {
 236                pr_warn("incomplete pg encoding\n");
 237                return -EINVAL;
 238        }
 239        version = ceph_decode_8(p);
 240        if (version > 1) {
 241                pr_warn("do not understand pg encoding %d > 1\n",
 242                        (int)version);
 243                return -EINVAL;
 244        }
 245
 246        pgid->pool = ceph_decode_64(p);
 247        pgid->seed = ceph_decode_32(p);
 248        *p += 4;        /* skip deprecated preferred value */
 249
 250        return 0;
 251}
 252
 253struct ceph_osdmap *ceph_osdmap_alloc(void);
 254extern struct ceph_osdmap *ceph_osdmap_decode(void **p, void *end);
 255struct ceph_osdmap *osdmap_apply_incremental(void **p, void *end,
 256                                             struct ceph_osdmap *map);
 257extern void ceph_osdmap_destroy(struct ceph_osdmap *map);
 258
 259struct ceph_osds {
 260        int osds[CEPH_PG_MAX_SIZE];
 261        int size;
 262        int primary; /* id, NOT index */
 263};
 264
 265static inline void ceph_osds_init(struct ceph_osds *set)
 266{
 267        set->size = 0;
 268        set->primary = -1;
 269}
 270
 271void ceph_osds_copy(struct ceph_osds *dest, const struct ceph_osds *src);
 272
 273bool ceph_pg_is_split(const struct ceph_pg *pgid, u32 old_pg_num,
 274                      u32 new_pg_num);
 275bool ceph_is_new_interval(const struct ceph_osds *old_acting,
 276                          const struct ceph_osds *new_acting,
 277                          const struct ceph_osds *old_up,
 278                          const struct ceph_osds *new_up,
 279                          int old_size,
 280                          int new_size,
 281                          int old_min_size,
 282                          int new_min_size,
 283                          u32 old_pg_num,
 284                          u32 new_pg_num,
 285                          bool old_sort_bitwise,
 286                          bool new_sort_bitwise,
 287                          bool old_recovery_deletes,
 288                          bool new_recovery_deletes,
 289                          const struct ceph_pg *pgid);
 290bool ceph_osds_changed(const struct ceph_osds *old_acting,
 291                       const struct ceph_osds *new_acting,
 292                       bool any_change);
 293
 294void __ceph_object_locator_to_pg(struct ceph_pg_pool_info *pi,
 295                                 const struct ceph_object_id *oid,
 296                                 const struct ceph_object_locator *oloc,
 297                                 struct ceph_pg *raw_pgid);
 298int ceph_object_locator_to_pg(struct ceph_osdmap *osdmap,
 299                              const struct ceph_object_id *oid,
 300                              const struct ceph_object_locator *oloc,
 301                              struct ceph_pg *raw_pgid);
 302
 303void ceph_pg_to_up_acting_osds(struct ceph_osdmap *osdmap,
 304                               struct ceph_pg_pool_info *pi,
 305                               const struct ceph_pg *raw_pgid,
 306                               struct ceph_osds *up,
 307                               struct ceph_osds *acting);
 308bool ceph_pg_to_primary_shard(struct ceph_osdmap *osdmap,
 309                              struct ceph_pg_pool_info *pi,
 310                              const struct ceph_pg *raw_pgid,
 311                              struct ceph_spg *spgid);
 312int ceph_pg_to_acting_primary(struct ceph_osdmap *osdmap,
 313                              const struct ceph_pg *raw_pgid);
 314
 315struct crush_loc {
 316        char *cl_type_name;
 317        char *cl_name;
 318};
 319
 320struct crush_loc_node {
 321        struct rb_node cl_node;
 322        struct crush_loc cl_loc;  /* pointers into cl_data */
 323        char cl_data[];
 324};
 325
 326int ceph_parse_crush_location(char *crush_location, struct rb_root *locs);
 327int ceph_compare_crush_locs(struct rb_root *locs1, struct rb_root *locs2);
 328void ceph_clear_crush_locs(struct rb_root *locs);
 329
 330int ceph_get_crush_locality(struct ceph_osdmap *osdmap, int id,
 331                            struct rb_root *locs);
 332
 333extern struct ceph_pg_pool_info *ceph_pg_pool_by_id(struct ceph_osdmap *map,
 334                                                    u64 id);
 335extern const char *ceph_pg_pool_name_by_id(struct ceph_osdmap *map, u64 id);
 336extern int ceph_pg_poolid_by_name(struct ceph_osdmap *map, const char *name);
 337u64 ceph_pg_pool_flags(struct ceph_osdmap *map, u64 id);
 338
 339#endif
 340