1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _BCACHEFS_BKEY_H
3 #define _BCACHEFS_BKEY_H
4
5 #include <linux/bug.h>
6 #include "bcachefs_format.h"
7 #include "bkey_types.h"
8 #include "btree_types.h"
9 #include "util.h"
10 #include "vstructs.h"
11
12 enum bch_validate_flags {
13 BCH_VALIDATE_write = (1U << 0),
14 BCH_VALIDATE_commit = (1U << 1),
15 BCH_VALIDATE_journal = (1U << 2),
16 };
17
18 #if 0
19
20 /*
21 * compiled unpack functions are disabled, pending a new interface for
22 * dynamically allocating executable memory:
23 */
24
25 #ifdef CONFIG_X86_64
26 #define HAVE_BCACHEFS_COMPILED_UNPACK 1
27 #endif
28 #endif
29
30 void bch2_bkey_packed_to_binary_text(struct printbuf *,
31 const struct bkey_format *,
32 const struct bkey_packed *);
33
34 enum bkey_lr_packed {
35 BKEY_PACKED_BOTH,
36 BKEY_PACKED_RIGHT,
37 BKEY_PACKED_LEFT,
38 BKEY_PACKED_NONE,
39 };
40
41 #define bkey_lr_packed(_l, _r) \
42 ((_l)->format + ((_r)->format << 1))
43
bkey_p_copy(struct bkey_packed * dst,const struct bkey_packed * src)44 static inline void bkey_p_copy(struct bkey_packed *dst, const struct bkey_packed *src)
45 {
46 memcpy_u64s_small(dst, src, src->u64s);
47 }
48
bkey_copy(struct bkey_i * dst,const struct bkey_i * src)49 static inline void bkey_copy(struct bkey_i *dst, const struct bkey_i *src)
50 {
51 memcpy_u64s_small(dst, src, src->k.u64s);
52 }
53
54 struct btree;
55
56 __pure
57 unsigned bch2_bkey_greatest_differing_bit(const struct btree *,
58 const struct bkey_packed *,
59 const struct bkey_packed *);
60 __pure
61 unsigned bch2_bkey_ffs(const struct btree *, const struct bkey_packed *);
62
63 __pure
64 int __bch2_bkey_cmp_packed_format_checked(const struct bkey_packed *,
65 const struct bkey_packed *,
66 const struct btree *);
67
68 __pure
69 int __bch2_bkey_cmp_left_packed_format_checked(const struct btree *,
70 const struct bkey_packed *,
71 const struct bpos *);
72
73 __pure
74 int bch2_bkey_cmp_packed(const struct btree *,
75 const struct bkey_packed *,
76 const struct bkey_packed *);
77
78 __pure
79 int __bch2_bkey_cmp_left_packed(const struct btree *,
80 const struct bkey_packed *,
81 const struct bpos *);
82
83 static inline __pure
bkey_cmp_left_packed(const struct btree * b,const struct bkey_packed * l,const struct bpos * r)84 int bkey_cmp_left_packed(const struct btree *b,
85 const struct bkey_packed *l, const struct bpos *r)
86 {
87 return __bch2_bkey_cmp_left_packed(b, l, r);
88 }
89
90 /*
91 * The compiler generates better code when we pass bpos by ref, but it's often
92 * enough terribly convenient to pass it by val... as much as I hate c++, const
93 * ref would be nice here:
94 */
95 __pure __flatten
bkey_cmp_left_packed_byval(const struct btree * b,const struct bkey_packed * l,struct bpos r)96 static inline int bkey_cmp_left_packed_byval(const struct btree *b,
97 const struct bkey_packed *l,
98 struct bpos r)
99 {
100 return bkey_cmp_left_packed(b, l, &r);
101 }
102
bpos_eq(struct bpos l,struct bpos r)103 static __always_inline bool bpos_eq(struct bpos l, struct bpos r)
104 {
105 return !((l.inode ^ r.inode) |
106 (l.offset ^ r.offset) |
107 (l.snapshot ^ r.snapshot));
108 }
109
bpos_lt(struct bpos l,struct bpos r)110 static __always_inline bool bpos_lt(struct bpos l, struct bpos r)
111 {
112 return l.inode != r.inode ? l.inode < r.inode :
113 l.offset != r.offset ? l.offset < r.offset :
114 l.snapshot != r.snapshot ? l.snapshot < r.snapshot : false;
115 }
116
bpos_le(struct bpos l,struct bpos r)117 static __always_inline bool bpos_le(struct bpos l, struct bpos r)
118 {
119 return l.inode != r.inode ? l.inode < r.inode :
120 l.offset != r.offset ? l.offset < r.offset :
121 l.snapshot != r.snapshot ? l.snapshot < r.snapshot : true;
122 }
123
bpos_gt(struct bpos l,struct bpos r)124 static __always_inline bool bpos_gt(struct bpos l, struct bpos r)
125 {
126 return bpos_lt(r, l);
127 }
128
bpos_ge(struct bpos l,struct bpos r)129 static __always_inline bool bpos_ge(struct bpos l, struct bpos r)
130 {
131 return bpos_le(r, l);
132 }
133
bpos_cmp(struct bpos l,struct bpos r)134 static __always_inline int bpos_cmp(struct bpos l, struct bpos r)
135 {
136 return cmp_int(l.inode, r.inode) ?:
137 cmp_int(l.offset, r.offset) ?:
138 cmp_int(l.snapshot, r.snapshot);
139 }
140
bpos_min(struct bpos l,struct bpos r)141 static inline struct bpos bpos_min(struct bpos l, struct bpos r)
142 {
143 return bpos_lt(l, r) ? l : r;
144 }
145
bpos_max(struct bpos l,struct bpos r)146 static inline struct bpos bpos_max(struct bpos l, struct bpos r)
147 {
148 return bpos_gt(l, r) ? l : r;
149 }
150
bkey_eq(struct bpos l,struct bpos r)151 static __always_inline bool bkey_eq(struct bpos l, struct bpos r)
152 {
153 return !((l.inode ^ r.inode) |
154 (l.offset ^ r.offset));
155 }
156
bkey_lt(struct bpos l,struct bpos r)157 static __always_inline bool bkey_lt(struct bpos l, struct bpos r)
158 {
159 return l.inode != r.inode
160 ? l.inode < r.inode
161 : l.offset < r.offset;
162 }
163
bkey_le(struct bpos l,struct bpos r)164 static __always_inline bool bkey_le(struct bpos l, struct bpos r)
165 {
166 return l.inode != r.inode
167 ? l.inode < r.inode
168 : l.offset <= r.offset;
169 }
170
bkey_gt(struct bpos l,struct bpos r)171 static __always_inline bool bkey_gt(struct bpos l, struct bpos r)
172 {
173 return bkey_lt(r, l);
174 }
175
bkey_ge(struct bpos l,struct bpos r)176 static __always_inline bool bkey_ge(struct bpos l, struct bpos r)
177 {
178 return bkey_le(r, l);
179 }
180
bkey_cmp(struct bpos l,struct bpos r)181 static __always_inline int bkey_cmp(struct bpos l, struct bpos r)
182 {
183 return cmp_int(l.inode, r.inode) ?:
184 cmp_int(l.offset, r.offset);
185 }
186
bkey_min(struct bpos l,struct bpos r)187 static inline struct bpos bkey_min(struct bpos l, struct bpos r)
188 {
189 return bkey_lt(l, r) ? l : r;
190 }
191
bkey_max(struct bpos l,struct bpos r)192 static inline struct bpos bkey_max(struct bpos l, struct bpos r)
193 {
194 return bkey_gt(l, r) ? l : r;
195 }
196
197 void bch2_bpos_swab(struct bpos *);
198 void bch2_bkey_swab_key(const struct bkey_format *, struct bkey_packed *);
199
bversion_cmp(struct bversion l,struct bversion r)200 static __always_inline int bversion_cmp(struct bversion l, struct bversion r)
201 {
202 return cmp_int(l.hi, r.hi) ?:
203 cmp_int(l.lo, r.lo);
204 }
205
206 #define ZERO_VERSION ((struct bversion) { .hi = 0, .lo = 0 })
207 #define MAX_VERSION ((struct bversion) { .hi = ~0, .lo = ~0ULL })
208
bversion_zero(struct bversion v)209 static __always_inline int bversion_zero(struct bversion v)
210 {
211 return !bversion_cmp(v, ZERO_VERSION);
212 }
213
214 #ifdef CONFIG_BCACHEFS_DEBUG
215 /* statement expressions confusing unlikely()? */
216 #define bkey_packed(_k) \
217 ({ EBUG_ON((_k)->format > KEY_FORMAT_CURRENT); \
218 (_k)->format != KEY_FORMAT_CURRENT; })
219 #else
220 #define bkey_packed(_k) ((_k)->format != KEY_FORMAT_CURRENT)
221 #endif
222
223 /*
224 * It's safe to treat an unpacked bkey as a packed one, but not the reverse
225 */
bkey_to_packed(struct bkey_i * k)226 static inline struct bkey_packed *bkey_to_packed(struct bkey_i *k)
227 {
228 return (struct bkey_packed *) k;
229 }
230
bkey_to_packed_c(const struct bkey_i * k)231 static inline const struct bkey_packed *bkey_to_packed_c(const struct bkey_i *k)
232 {
233 return (const struct bkey_packed *) k;
234 }
235
packed_to_bkey(struct bkey_packed * k)236 static inline struct bkey_i *packed_to_bkey(struct bkey_packed *k)
237 {
238 return bkey_packed(k) ? NULL : (struct bkey_i *) k;
239 }
240
packed_to_bkey_c(const struct bkey_packed * k)241 static inline const struct bkey *packed_to_bkey_c(const struct bkey_packed *k)
242 {
243 return bkey_packed(k) ? NULL : (const struct bkey *) k;
244 }
245
bkey_format_key_bits(const struct bkey_format * format)246 static inline unsigned bkey_format_key_bits(const struct bkey_format *format)
247 {
248 return format->bits_per_field[BKEY_FIELD_INODE] +
249 format->bits_per_field[BKEY_FIELD_OFFSET] +
250 format->bits_per_field[BKEY_FIELD_SNAPSHOT];
251 }
252
bpos_successor(struct bpos p)253 static inline struct bpos bpos_successor(struct bpos p)
254 {
255 if (!++p.snapshot &&
256 !++p.offset &&
257 !++p.inode)
258 BUG();
259
260 return p;
261 }
262
bpos_predecessor(struct bpos p)263 static inline struct bpos bpos_predecessor(struct bpos p)
264 {
265 if (!p.snapshot-- &&
266 !p.offset-- &&
267 !p.inode--)
268 BUG();
269
270 return p;
271 }
272
bpos_nosnap_successor(struct bpos p)273 static inline struct bpos bpos_nosnap_successor(struct bpos p)
274 {
275 p.snapshot = 0;
276
277 if (!++p.offset &&
278 !++p.inode)
279 BUG();
280
281 return p;
282 }
283
bpos_nosnap_predecessor(struct bpos p)284 static inline struct bpos bpos_nosnap_predecessor(struct bpos p)
285 {
286 p.snapshot = 0;
287
288 if (!p.offset-- &&
289 !p.inode--)
290 BUG();
291
292 return p;
293 }
294
bkey_start_offset(const struct bkey * k)295 static inline u64 bkey_start_offset(const struct bkey *k)
296 {
297 return k->p.offset - k->size;
298 }
299
bkey_start_pos(const struct bkey * k)300 static inline struct bpos bkey_start_pos(const struct bkey *k)
301 {
302 return (struct bpos) {
303 .inode = k->p.inode,
304 .offset = bkey_start_offset(k),
305 .snapshot = k->p.snapshot,
306 };
307 }
308
309 /* Packed helpers */
310
bkeyp_key_u64s(const struct bkey_format * format,const struct bkey_packed * k)311 static inline unsigned bkeyp_key_u64s(const struct bkey_format *format,
312 const struct bkey_packed *k)
313 {
314 return bkey_packed(k) ? format->key_u64s : BKEY_U64s;
315 }
316
bkeyp_u64s_valid(const struct bkey_format * f,const struct bkey_packed * k)317 static inline bool bkeyp_u64s_valid(const struct bkey_format *f,
318 const struct bkey_packed *k)
319 {
320 return ((unsigned) k->u64s - bkeyp_key_u64s(f, k) <= U8_MAX - BKEY_U64s);
321 }
322
bkeyp_key_bytes(const struct bkey_format * format,const struct bkey_packed * k)323 static inline unsigned bkeyp_key_bytes(const struct bkey_format *format,
324 const struct bkey_packed *k)
325 {
326 return bkeyp_key_u64s(format, k) * sizeof(u64);
327 }
328
bkeyp_val_u64s(const struct bkey_format * format,const struct bkey_packed * k)329 static inline unsigned bkeyp_val_u64s(const struct bkey_format *format,
330 const struct bkey_packed *k)
331 {
332 return k->u64s - bkeyp_key_u64s(format, k);
333 }
334
bkeyp_val_bytes(const struct bkey_format * format,const struct bkey_packed * k)335 static inline size_t bkeyp_val_bytes(const struct bkey_format *format,
336 const struct bkey_packed *k)
337 {
338 return bkeyp_val_u64s(format, k) * sizeof(u64);
339 }
340
set_bkeyp_val_u64s(const struct bkey_format * format,struct bkey_packed * k,unsigned val_u64s)341 static inline void set_bkeyp_val_u64s(const struct bkey_format *format,
342 struct bkey_packed *k, unsigned val_u64s)
343 {
344 k->u64s = bkeyp_key_u64s(format, k) + val_u64s;
345 }
346
347 #define bkeyp_val(_format, _k) \
348 ((struct bch_val *) ((u64 *) (_k)->_data + bkeyp_key_u64s(_format, _k)))
349
350 extern const struct bkey_format bch2_bkey_format_current;
351
352 bool bch2_bkey_transform(const struct bkey_format *,
353 struct bkey_packed *,
354 const struct bkey_format *,
355 const struct bkey_packed *);
356
357 struct bkey __bch2_bkey_unpack_key(const struct bkey_format *,
358 const struct bkey_packed *);
359
360 #ifndef HAVE_BCACHEFS_COMPILED_UNPACK
361 struct bpos __bkey_unpack_pos(const struct bkey_format *,
362 const struct bkey_packed *);
363 #endif
364
365 bool bch2_bkey_pack_key(struct bkey_packed *, const struct bkey *,
366 const struct bkey_format *);
367
368 enum bkey_pack_pos_ret {
369 BKEY_PACK_POS_EXACT,
370 BKEY_PACK_POS_SMALLER,
371 BKEY_PACK_POS_FAIL,
372 };
373
374 enum bkey_pack_pos_ret bch2_bkey_pack_pos_lossy(struct bkey_packed *, struct bpos,
375 const struct btree *);
376
bkey_pack_pos(struct bkey_packed * out,struct bpos in,const struct btree * b)377 static inline bool bkey_pack_pos(struct bkey_packed *out, struct bpos in,
378 const struct btree *b)
379 {
380 return bch2_bkey_pack_pos_lossy(out, in, b) == BKEY_PACK_POS_EXACT;
381 }
382
383 void bch2_bkey_unpack(const struct btree *, struct bkey_i *,
384 const struct bkey_packed *);
385 bool bch2_bkey_pack(struct bkey_packed *, const struct bkey_i *,
386 const struct bkey_format *);
387
388 typedef void (*compiled_unpack_fn)(struct bkey *, const struct bkey_packed *);
389
390 static inline void
__bkey_unpack_key_format_checked(const struct btree * b,struct bkey * dst,const struct bkey_packed * src)391 __bkey_unpack_key_format_checked(const struct btree *b,
392 struct bkey *dst,
393 const struct bkey_packed *src)
394 {
395 if (IS_ENABLED(HAVE_BCACHEFS_COMPILED_UNPACK)) {
396 compiled_unpack_fn unpack_fn = b->aux_data;
397 unpack_fn(dst, src);
398
399 if (IS_ENABLED(CONFIG_BCACHEFS_DEBUG) &&
400 bch2_expensive_debug_checks) {
401 struct bkey dst2 = __bch2_bkey_unpack_key(&b->format, src);
402
403 BUG_ON(memcmp(dst, &dst2, sizeof(*dst)));
404 }
405 } else {
406 *dst = __bch2_bkey_unpack_key(&b->format, src);
407 }
408 }
409
410 static inline struct bkey
bkey_unpack_key_format_checked(const struct btree * b,const struct bkey_packed * src)411 bkey_unpack_key_format_checked(const struct btree *b,
412 const struct bkey_packed *src)
413 {
414 struct bkey dst;
415
416 __bkey_unpack_key_format_checked(b, &dst, src);
417 return dst;
418 }
419
__bkey_unpack_key(const struct btree * b,struct bkey * dst,const struct bkey_packed * src)420 static inline void __bkey_unpack_key(const struct btree *b,
421 struct bkey *dst,
422 const struct bkey_packed *src)
423 {
424 if (likely(bkey_packed(src)))
425 __bkey_unpack_key_format_checked(b, dst, src);
426 else
427 *dst = *packed_to_bkey_c(src);
428 }
429
430 /**
431 * bkey_unpack_key -- unpack just the key, not the value
432 */
bkey_unpack_key(const struct btree * b,const struct bkey_packed * src)433 static inline struct bkey bkey_unpack_key(const struct btree *b,
434 const struct bkey_packed *src)
435 {
436 return likely(bkey_packed(src))
437 ? bkey_unpack_key_format_checked(b, src)
438 : *packed_to_bkey_c(src);
439 }
440
441 static inline struct bpos
bkey_unpack_pos_format_checked(const struct btree * b,const struct bkey_packed * src)442 bkey_unpack_pos_format_checked(const struct btree *b,
443 const struct bkey_packed *src)
444 {
445 #ifdef HAVE_BCACHEFS_COMPILED_UNPACK
446 return bkey_unpack_key_format_checked(b, src).p;
447 #else
448 return __bkey_unpack_pos(&b->format, src);
449 #endif
450 }
451
bkey_unpack_pos(const struct btree * b,const struct bkey_packed * src)452 static inline struct bpos bkey_unpack_pos(const struct btree *b,
453 const struct bkey_packed *src)
454 {
455 return likely(bkey_packed(src))
456 ? bkey_unpack_pos_format_checked(b, src)
457 : packed_to_bkey_c(src)->p;
458 }
459
460 /* Disassembled bkeys */
461
bkey_disassemble(const struct btree * b,const struct bkey_packed * k,struct bkey * u)462 static inline struct bkey_s_c bkey_disassemble(const struct btree *b,
463 const struct bkey_packed *k,
464 struct bkey *u)
465 {
466 __bkey_unpack_key(b, u, k);
467
468 return (struct bkey_s_c) { u, bkeyp_val(&b->format, k), };
469 }
470
471 /* non const version: */
__bkey_disassemble(const struct btree * b,struct bkey_packed * k,struct bkey * u)472 static inline struct bkey_s __bkey_disassemble(const struct btree *b,
473 struct bkey_packed *k,
474 struct bkey *u)
475 {
476 __bkey_unpack_key(b, u, k);
477
478 return (struct bkey_s) { .k = u, .v = bkeyp_val(&b->format, k), };
479 }
480
bkey_field_max(const struct bkey_format * f,enum bch_bkey_fields nr)481 static inline u64 bkey_field_max(const struct bkey_format *f,
482 enum bch_bkey_fields nr)
483 {
484 return f->bits_per_field[nr] < 64
485 ? (le64_to_cpu(f->field_offset[nr]) +
486 ~(~0ULL << f->bits_per_field[nr]))
487 : U64_MAX;
488 }
489
490 #ifdef HAVE_BCACHEFS_COMPILED_UNPACK
491
492 int bch2_compile_bkey_format(const struct bkey_format *, void *);
493
494 #else
495
bch2_compile_bkey_format(const struct bkey_format * format,void * out)496 static inline int bch2_compile_bkey_format(const struct bkey_format *format,
497 void *out) { return 0; }
498
499 #endif
500
bkey_reassemble(struct bkey_i * dst,struct bkey_s_c src)501 static inline void bkey_reassemble(struct bkey_i *dst,
502 struct bkey_s_c src)
503 {
504 dst->k = *src.k;
505 memcpy_u64s_small(&dst->v, src.v, bkey_val_u64s(src.k));
506 }
507
508 /* byte order helpers */
509
510 #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
511
high_word_offset(const struct bkey_format * f)512 static inline unsigned high_word_offset(const struct bkey_format *f)
513 {
514 return f->key_u64s - 1;
515 }
516
517 #define high_bit_offset 0
518 #define nth_word(p, n) ((p) - (n))
519
520 #elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
521
high_word_offset(const struct bkey_format * f)522 static inline unsigned high_word_offset(const struct bkey_format *f)
523 {
524 return 0;
525 }
526
527 #define high_bit_offset KEY_PACKED_BITS_START
528 #define nth_word(p, n) ((p) + (n))
529
530 #else
531 #error edit for your odd byteorder.
532 #endif
533
534 #define high_word(f, k) ((u64 *) (k)->_data + high_word_offset(f))
535 #define next_word(p) nth_word(p, 1)
536 #define prev_word(p) nth_word(p, -1)
537
538 #ifdef CONFIG_BCACHEFS_DEBUG
539 void bch2_bkey_pack_test(void);
540 #else
bch2_bkey_pack_test(void)541 static inline void bch2_bkey_pack_test(void) {}
542 #endif
543
544 #define bkey_fields() \
545 x(BKEY_FIELD_INODE, p.inode) \
546 x(BKEY_FIELD_OFFSET, p.offset) \
547 x(BKEY_FIELD_SNAPSHOT, p.snapshot) \
548 x(BKEY_FIELD_SIZE, size) \
549 x(BKEY_FIELD_VERSION_HI, version.hi) \
550 x(BKEY_FIELD_VERSION_LO, version.lo)
551
552 struct bkey_format_state {
553 u64 field_min[BKEY_NR_FIELDS];
554 u64 field_max[BKEY_NR_FIELDS];
555 };
556
557 void bch2_bkey_format_init(struct bkey_format_state *);
558
__bkey_format_add(struct bkey_format_state * s,unsigned field,u64 v)559 static inline void __bkey_format_add(struct bkey_format_state *s, unsigned field, u64 v)
560 {
561 s->field_min[field] = min(s->field_min[field], v);
562 s->field_max[field] = max(s->field_max[field], v);
563 }
564
565 /*
566 * Changes @format so that @k can be successfully packed with @format
567 */
bch2_bkey_format_add_key(struct bkey_format_state * s,const struct bkey * k)568 static inline void bch2_bkey_format_add_key(struct bkey_format_state *s, const struct bkey *k)
569 {
570 #define x(id, field) __bkey_format_add(s, id, k->field);
571 bkey_fields()
572 #undef x
573 }
574
575 void bch2_bkey_format_add_pos(struct bkey_format_state *, struct bpos);
576 struct bkey_format bch2_bkey_format_done(struct bkey_format_state *);
577
bch2_bkey_format_field_overflows(struct bkey_format * f,unsigned i)578 static inline bool bch2_bkey_format_field_overflows(struct bkey_format *f, unsigned i)
579 {
580 unsigned f_bits = f->bits_per_field[i];
581 unsigned unpacked_bits = bch2_bkey_format_current.bits_per_field[i];
582 u64 unpacked_mask = ~((~0ULL << 1) << (unpacked_bits - 1));
583 u64 field_offset = le64_to_cpu(f->field_offset[i]);
584
585 if (f_bits > unpacked_bits)
586 return true;
587
588 if ((f_bits == unpacked_bits) && field_offset)
589 return true;
590
591 u64 f_mask = f_bits
592 ? ~((~0ULL << (f_bits - 1)) << 1)
593 : 0;
594
595 if (((field_offset + f_mask) & unpacked_mask) < field_offset)
596 return true;
597 return false;
598 }
599
600 int bch2_bkey_format_invalid(struct bch_fs *, struct bkey_format *,
601 enum bch_validate_flags, struct printbuf *);
602 void bch2_bkey_format_to_text(struct printbuf *, const struct bkey_format *);
603
604 #endif /* _BCACHEFS_BKEY_H */
605