~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/include/linux/ceph/osdmap.h

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /* SPDX-License-Identifier: GPL-2.0 */
  2 #ifndef _FS_CEPH_OSDMAP_H
  3 #define _FS_CEPH_OSDMAP_H
  4 
  5 #include <linux/rbtree.h>
  6 #include <linux/ceph/types.h>
  7 #include <linux/ceph/decode.h>
  8 #include <linux/crush/crush.h>
  9 
 10 /*
 11  * The osd map describes the current membership of the osd cluster and
 12  * specifies the mapping of objects to placement groups and placement
 13  * groups to (sets of) osds.  That is, it completely specifies the
 14  * (desired) distribution of all data objects in the system at some
 15  * point in time.
 16  *
 17  * Each map version is identified by an epoch, which increases monotonically.
 18  *
 19  * The map can be updated either via an incremental map (diff) describing
 20  * the change between two successive epochs, or as a fully encoded map.
 21  */
 22 struct ceph_pg {
 23         uint64_t pool;
 24         uint32_t seed;
 25 };
 26 
 27 #define CEPH_SPG_NOSHARD        -1
 28 
 29 struct ceph_spg {
 30         struct ceph_pg pgid;
 31         s8 shard;
 32 };
 33 
 34 int ceph_pg_compare(const struct ceph_pg *lhs, const struct ceph_pg *rhs);
 35 int ceph_spg_compare(const struct ceph_spg *lhs, const struct ceph_spg *rhs);
 36 
 37 #define CEPH_POOL_FLAG_HASHPSPOOL       (1ULL << 0) /* hash pg seed and pool id
 38                                                        together */
 39 #define CEPH_POOL_FLAG_FULL             (1ULL << 1) /* pool is full */
 40 #define CEPH_POOL_FLAG_FULL_QUOTA       (1ULL << 10) /* pool ran out of quota,
 41                                                         will set FULL too */
 42 #define CEPH_POOL_FLAG_NEARFULL         (1ULL << 11) /* pool is nearfull */
 43 
 44 struct ceph_pg_pool_info {
 45         struct rb_node node;
 46         s64 id;
 47         u8 type; /* CEPH_POOL_TYPE_* */
 48         u8 size;
 49         u8 min_size;
 50         u8 crush_ruleset;
 51         u8 object_hash;
 52         u32 last_force_request_resend;
 53         u32 pg_num, pgp_num;
 54         int pg_num_mask, pgp_num_mask;
 55         s64 read_tier;
 56         s64 write_tier; /* wins for read+write ops */
 57         u64 flags; /* CEPH_POOL_FLAG_* */
 58         char *name;
 59 
 60         bool was_full;  /* for handle_one_map() */
 61 };
 62 
 63 static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool)
 64 {
 65         switch (pool->type) {
 66         case CEPH_POOL_TYPE_REP:
 67                 return true;
 68         case CEPH_POOL_TYPE_EC:
 69                 return false;
 70         default:
 71                 BUG();
 72         }
 73 }
 74 
 75 struct ceph_object_locator {
 76         s64 pool;
 77         struct ceph_string *pool_ns;
 78 };
 79 
 80 static inline void ceph_oloc_init(struct ceph_object_locator *oloc)
 81 {
 82         oloc->pool = -1;
 83         oloc->pool_ns = NULL;
 84 }
 85 
 86 static inline bool ceph_oloc_empty(const struct ceph_object_locator *oloc)
 87 {
 88         return oloc->pool == -1;
 89 }
 90 
 91 void ceph_oloc_copy(struct ceph_object_locator *dest,
 92                     const struct ceph_object_locator *src);
 93 void ceph_oloc_destroy(struct ceph_object_locator *oloc);
 94 
 95 /*
 96  * 51-char inline_name is long enough for all cephfs and all but one
 97  * rbd requests: <imgname> in "<imgname>.rbd"/"rbd_id.<imgname>" can be
 98  * arbitrarily long (~PAGE_SIZE).  It's done once during rbd map; all
 99  * other rbd requests fit into inline_name.
100  *
101  * Makes ceph_object_id 64 bytes on 64-bit.
102  */
103 #define CEPH_OID_INLINE_LEN 52
104 
105 /*
106  * Both inline and external buffers have space for a NUL-terminator,
107  * which is carried around.  It's not required though - RADOS object
108  * names don't have to be NUL-terminated and may contain NULs.
109  */
110 struct ceph_object_id {
111         char *name;
112         char inline_name[CEPH_OID_INLINE_LEN];
113         int name_len;
114 };
115 
116 #define __CEPH_OID_INITIALIZER(oid) { .name = (oid).inline_name }
117 
118 #define CEPH_DEFINE_OID_ONSTACK(oid)                            \
119         struct ceph_object_id oid = __CEPH_OID_INITIALIZER(oid)
120 
121 static inline void ceph_oid_init(struct ceph_object_id *oid)
122 {
123         *oid = (struct ceph_object_id) __CEPH_OID_INITIALIZER(*oid);
124 }
125 
126 static inline bool ceph_oid_empty(const struct ceph_object_id *oid)
127 {
128         return oid->name == oid->inline_name && !oid->name_len;
129 }
130 
131 void ceph_oid_copy(struct ceph_object_id *dest,
132                    const struct ceph_object_id *src);
133 __printf(2, 3)
134 void ceph_oid_printf(struct ceph_object_id *oid, const char *fmt, ...);
135 __printf(3, 4)
136 int ceph_oid_aprintf(struct ceph_object_id *oid, gfp_t gfp,
137                      const char *fmt, ...);
138 void ceph_oid_destroy(struct ceph_object_id *oid);
139 
140 struct workspace_manager {
141         struct list_head idle_ws;
142         spinlock_t ws_lock;
143         /* Number of free workspaces */
144         int free_ws;
145         /* Total number of allocated workspaces */
146         atomic_t total_ws;
147         /* Waiters for a free workspace */
148         wait_queue_head_t ws_wait;
149 };
150 
151 struct ceph_pg_mapping {
152         struct rb_node node;
153         struct ceph_pg pgid;
154 
155         union {
156                 struct {
157                         int len;
158                         int osds[];
159                 } pg_temp, pg_upmap;
160                 struct {
161                         int osd;
162                 } primary_temp;
163                 struct {
164                         int len;
165                         int from_to[][2];
166                 } pg_upmap_items;
167         };
168 };
169 
170 struct ceph_osdmap {
171         struct ceph_fsid fsid;
172         u32 epoch;
173         struct ceph_timespec created, modified;
174 
175         u32 flags;         /* CEPH_OSDMAP_* */
176 
177         u32 max_osd;       /* size of osd_state, _offload, _addr arrays */
178         u32 *osd_state;    /* CEPH_OSD_* */
179         u32 *osd_weight;   /* 0 = failed, 0x10000 = 100% normal */
180         struct ceph_entity_addr *osd_addr;
181 
182         struct rb_root pg_temp;
183         struct rb_root primary_temp;
184 
185         /* remap (post-CRUSH, pre-up) */
186         struct rb_root pg_upmap;        /* PG := raw set */
187         struct rb_root pg_upmap_items;  /* from -> to within raw set */
188 
189         u32 *osd_primary_affinity;
190 
191         struct rb_root pg_pools;
192         u32 pool_max;
193 
194         /* the CRUSH map specifies the mapping of placement groups to
195          * the list of osds that store+replicate them. */
196         struct crush_map *crush;
197 
198         struct workspace_manager crush_wsm;
199 };
200 
201 static inline bool ceph_osd_exists(struct ceph_osdmap *map, int osd)
202 {
203         return osd >= 0 && osd < map->max_osd &&
204                (map->osd_state[osd] & CEPH_OSD_EXISTS);
205 }
206 
207 static inline bool ceph_osd_is_up(struct ceph_osdmap *map, int osd)
208 {
209         return ceph_osd_exists(map, osd) &&
210                (map->osd_state[osd] & CEPH_OSD_UP);
211 }
212 
213 static inline bool ceph_osd_is_down(struct ceph_osdmap *map, int osd)
214 {
215         return !ceph_osd_is_up(map, osd);
216 }
217 
218 char *ceph_osdmap_state_str(char *str, int len, u32 state);
219 extern u32 ceph_get_primary_affinity(struct ceph_osdmap *map, int osd);
220 
221 static inline struct ceph_entity_addr *ceph_osd_addr(struct ceph_osdmap *map,
222                                                      int osd)
223 {
224         if (osd >= map->max_osd)
225                 return NULL;
226         return &map->osd_addr[osd];
227 }
228 
229 #define CEPH_PGID_ENCODING_LEN          (1 + 8 + 4 + 4)
230 
231 static inline int ceph_decode_pgid(void **p, void *end, struct ceph_pg *pgid)
232 {
233         __u8 version;
234 
235         if (!ceph_has_room(p, end, CEPH_PGID_ENCODING_LEN)) {
236                 pr_warn("incomplete pg encoding\n");
237                 return -EINVAL;
238         }
239         version = ceph_decode_8(p);
240         if (version > 1) {
241                 pr_warn("do not understand pg encoding %d > 1\n",
242                         (int)version);
243                 return -EINVAL;
244         }
245 
246         pgid->pool = ceph_decode_64(p);
247         pgid->seed = ceph_decode_32(p);
248         *p += 4;        /* skip deprecated preferred value */
249 
250         return 0;
251 }
252 
253 struct ceph_osdmap *ceph_osdmap_alloc(void);
254 struct ceph_osdmap *ceph_osdmap_decode(void **p, void *end, bool msgr2);
255 struct ceph_osdmap *osdmap_apply_incremental(void **p, void *end, bool msgr2,
256                                              struct ceph_osdmap *map);
257 extern void ceph_osdmap_destroy(struct ceph_osdmap *map);
258 
259 struct ceph_osds {
260         int osds[CEPH_PG_MAX_SIZE];
261         int size;
262         int primary; /* id, NOT index */
263 };
264 
265 static inline void ceph_osds_init(struct ceph_osds *set)
266 {
267         set->size = 0;
268         set->primary = -1;
269 }
270 
271 void ceph_osds_copy(struct ceph_osds *dest, const struct ceph_osds *src);
272 
273 bool ceph_pg_is_split(const struct ceph_pg *pgid, u32 old_pg_num,
274                       u32 new_pg_num);
275 bool ceph_is_new_interval(const struct ceph_osds *old_acting,
276                           const struct ceph_osds *new_acting,
277                           const struct ceph_osds *old_up,
278                           const struct ceph_osds *new_up,
279                           int old_size,
280                           int new_size,
281                           int old_min_size,
282                           int new_min_size,
283                           u32 old_pg_num,
284                           u32 new_pg_num,
285                           bool old_sort_bitwise,
286                           bool new_sort_bitwise,
287                           bool old_recovery_deletes,
288                           bool new_recovery_deletes,
289                           const struct ceph_pg *pgid);
290 bool ceph_osds_changed(const struct ceph_osds *old_acting,
291                        const struct ceph_osds *new_acting,
292                        bool any_change);
293 
294 void __ceph_object_locator_to_pg(struct ceph_pg_pool_info *pi,
295                                  const struct ceph_object_id *oid,
296                                  const struct ceph_object_locator *oloc,
297                                  struct ceph_pg *raw_pgid);
298 int ceph_object_locator_to_pg(struct ceph_osdmap *osdmap,
299                               const struct ceph_object_id *oid,
300                               const struct ceph_object_locator *oloc,
301                               struct ceph_pg *raw_pgid);
302 
303 void ceph_pg_to_up_acting_osds(struct ceph_osdmap *osdmap,
304                                struct ceph_pg_pool_info *pi,
305                                const struct ceph_pg *raw_pgid,
306                                struct ceph_osds *up,
307                                struct ceph_osds *acting);
308 bool ceph_pg_to_primary_shard(struct ceph_osdmap *osdmap,
309                               struct ceph_pg_pool_info *pi,
310                               const struct ceph_pg *raw_pgid,
311                               struct ceph_spg *spgid);
312 int ceph_pg_to_acting_primary(struct ceph_osdmap *osdmap,
313                               const struct ceph_pg *raw_pgid);
314 
315 struct crush_loc {
316         char *cl_type_name;
317         char *cl_name;
318 };
319 
320 struct crush_loc_node {
321         struct rb_node cl_node;
322         struct crush_loc cl_loc;  /* pointers into cl_data */
323         char cl_data[];
324 };
325 
326 int ceph_parse_crush_location(char *crush_location, struct rb_root *locs);
327 int ceph_compare_crush_locs(struct rb_root *locs1, struct rb_root *locs2);
328 void ceph_clear_crush_locs(struct rb_root *locs);
329 
330 int ceph_get_crush_locality(struct ceph_osdmap *osdmap, int id,
331                             struct rb_root *locs);
332 
333 extern struct ceph_pg_pool_info *ceph_pg_pool_by_id(struct ceph_osdmap *map,
334                                                     u64 id);
335 extern const char *ceph_pg_pool_name_by_id(struct ceph_osdmap *map, u64 id);
336 extern int ceph_pg_poolid_by_name(struct ceph_osdmap *map, const char *name);
337 u64 ceph_pg_pool_flags(struct ceph_osdmap *map, u64 id);
338 
339 #endif
340 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php