/linux/drivers/md/persistent-data/ |
H A D | Makefile | 11 dm-btree.o \ 12 dm-btree-remove.o \ 13 dm-btree-spine.o
|
/linux/fs/bcachefs/ |
H A D | btree_write_buffer_types.h | 17 enum btree_id btree:8; 19 enum btree_id btree:8; 39 enum btree_id btree:8;
|
H A D | btree_gc.c | 59 bch2_btree_id_to_text(out, p->btree); in bch2_gc_pos_to_text() 195 struct btree *prev, struct btree *cur, in btree_check_node_boundaries() 315 struct btree *prev = NULL, *cur = NULL; in bch2_btree_repair_topology_recurse() 545 struct btree *b = r->b; in bch2_check_topology() 575 unsigned level, struct btree **prev, in bch2_gc_mark_key() 583 struct btree *b = path_l(path)->b; in bch2_gc_mark_key() 663 struct btree *prev = NULL; in bch2_gc_btree() 669 gc_pos_set(c, gc_pos_btree(btree, level, k.k->p)); in bch2_gc_btree() 722 unsigned btree = i < BTREE_ID_NR ? ids[i] : i; in bch2_gc_btrees() local 727 ret = bch2_gc_btree(trans, btree, true); in bch2_gc_btrees() [all …]
|
H A D | move.c | 608 for (id = start.btree; in __bch2_move_data() 770 struct btree *b; in bch2_evacuate_bucket() 826 struct btree *b; in bch2_move_btree() 827 enum btree_id btree; in bch2_move_btree() local 838 for (btree = start.btree; in bch2_move_btree() 839 btree <= min_t(unsigned, end.btree, btree_id_nr_alive(c) - 1); in bch2_move_btree() 840 btree ++) { in bch2_move_btree() 856 if ((cmp_int(btree, end.btree) ?: in bch2_move_btree() 945 struct btree *b, in rereplicate_btree_pred() 975 struct btree *b, in rewrite_old_nodes_pred() [all …]
|
H A D | btree_io.c | 47 void bch2_btree_node_io_lock(struct btree *b) in bch2_btree_node_io_lock() 77 static void verify_no_dups(struct btree *b, in verify_no_dups() 386 struct btree *dst, in bch2_btree_sort_into() 387 struct btree *src) in bch2_btree_sort_into() 545 struct btree *b, in __btree_err() 681 struct btree *b, struct bset *i, in validate_bset() 1282 struct btree *b = rb->b; in btree_node_read_work() 1376 struct btree *b; 1429 struct btree *b = ra->b; in CLOSURE_CALLBACK() 1727 struct btree *b; in __bch2_btree_root_read() [all …]
|
H A D | bkey_methods.h | 8 struct btree; 56 int bch2_bkey_in_btree_node(struct bch_fs *, struct btree *, struct bkey_s_c, 80 enum btree_id btree, unsigned level, in bch2_key_trigger() argument 87 ? ops->trigger(trans, btree, level, old, new, flags) in bch2_key_trigger()
|
H A D | disk_accounting_format.h | 59 x(btree, 3) \ 105 x(btree, 6) \ 158 struct bch_acct_btree btree; member
|
H A D | btree_journal_iter.h | 21 struct btree *b; 68 struct btree_and_journal_iter *, struct btree *, 71 struct btree_and_journal_iter *, struct btree *);
|
H A D | btree_iter.h | 54 static inline struct btree *btree_path_node(struct btree_path *path, in btree_path_node() 61 const struct btree *b, unsigned level) in btree_node_lock_seq_matches() 66 static inline struct btree *btree_node_parent(struct btree_path *path, in btree_node_parent() 67 struct btree *b) in btree_node_parent() 180 const struct btree *b) in __path_has_node() 187 __trans_next_path_with_node(struct btree_trans *trans, struct btree *b, in __trans_next_path_with_node() 297 struct btree *, struct bkey_packed *); 299 struct btree *, struct btree_node_iter *, 383 struct btree *bch2_btree_iter_peek_node(struct btree_iter *); 385 struct btree *bch2_btree_iter_next_node(struct btree_iter *); [all …]
|
H A D | btree_gc_types.h | 11 x(btree) 21 enum btree_id btree:8;
|
H A D | btree_write_buffer.c | 208 bch2_trans_iter_init(trans, &iter, wb->btree, bkey_start_pos(&wb->k.k), in btree_write_buffered_insert() 289 wb->sorted.data[i].btree = wb->flushing.keys.data[i].btree; in bch2_btree_write_buffer_flush_locked() 342 if (path->btree_id != i->btree || in bch2_btree_write_buffer_flush_locked() 359 if (!iter.path || iter.btree_id != k->btree) { in bch2_btree_write_buffer_flush_locked() 361 bch2_trans_iter_init(trans, &iter, k->btree, k->k.k.p, in bch2_btree_write_buffer_flush_locked() 620 int bch2_accounting_key_to_wb_slowpath(struct bch_fs *c, enum btree_id btree, in bch2_accounting_key_to_wb_slowpath() argument 624 struct btree_write_buffered_key new = { .btree = btree }; in bch2_accounting_key_to_wb_slowpath() 638 enum btree_id btree, struct bkey_i *k) in bch2_journal_key_to_wb_slowpath() argument 667 wb_k->btree = btree; in bch2_journal_key_to_wb_slowpath() 720 ret = __bch2_journal_key_to_wb(c, dst, i->btree, &i->k); in bch2_journal_keys_to_write_buffer_end()
|
H A D | btree_journal_iter.c | 277 bool bch2_key_deleted_in_journal(struct btree_trans *trans, enum btree_id btree, in bch2_key_deleted_in_journal() argument 281 size_t idx = bch2_journal_key_search(keys, btree, level, pos); in bch2_key_deleted_in_journal() 287 keys->data[idx].btree_id == btree && in bch2_key_deleted_in_journal() 293 void bch2_journal_key_overwritten(struct bch_fs *c, enum btree_id btree, in bch2_journal_key_overwritten() argument 297 size_t idx = bch2_journal_key_search(keys, btree, level, pos); in bch2_journal_key_overwritten() 300 keys->data[idx].btree_id == btree && in bch2_journal_key_overwritten() 448 struct btree *b, in __bch2_btree_and_journal_iter_init_node_iter() 474 struct btree *b) in bch2_btree_and_journal_iter_init_node_iter() 601 void bch2_shoot_down_journal_keys(struct bch_fs *c, enum btree_id btree, in bch2_shoot_down_journal_keys() argument 611 if (!(i->btree_id == btree && in bch2_shoot_down_journal_keys()
|
H A D | bkey_sort.c | 9 typedef int (*sort_cmp_fn)(const struct btree *, 73 static inline int key_sort_fix_overlapping_cmp(const struct btree *b, in key_sort_fix_overlapping_cmp() 123 bch2_sort_repack(struct bset *dst, struct btree *src, in bch2_sort_repack() 157 static inline int keep_unwritten_whiteouts_cmp(const struct btree *b, in keep_unwritten_whiteouts_cmp()
|
H A D | btree_node_scan.c | 85 struct btree *b = bch2_btree_node_get_noiter(trans, &tmp.k, f->btree_id, f->level, false); in found_btree_node_is_readable() 453 bool bch2_btree_node_is_stale(struct bch_fs *c, struct btree *b) in bch2_btree_node_is_stale() 470 bool bch2_btree_has_scanned_nodes(struct bch_fs *c, enum btree_id btree) in bch2_btree_has_scanned_nodes() argument 473 .btree_id = btree, in bch2_btree_has_scanned_nodes() 484 int bch2_get_scanned_nodes(struct bch_fs *c, enum btree_id btree, in bch2_get_scanned_nodes() argument 487 if (btree_id_is_alloc(btree)) in bch2_get_scanned_nodes() 499 prt_printf(&buf, "recovering %s l=%u ", bch2_btree_id_str(btree), level); in bch2_get_scanned_nodes() 509 .btree_id = btree, in bch2_get_scanned_nodes() 535 ret = bch2_journal_key_insert(c, btree, level + 1, &tmp.k); in bch2_get_scanned_nodes()
|
H A D | btree_update.c | 550 enum btree_id btree, in bch2_btree_insert_clone_trans() argument 559 return bch2_btree_insert_trans(trans, btree, n, 0); in bch2_btree_insert_clone_trans() 589 enum btree_id btree, struct bpos end) in bch2_bkey_get_empty_slot() argument 594 bch2_trans_iter_init(trans, iter, btree, POS_MAX, BTREE_ITER_intent); in bch2_bkey_get_empty_slot() 627 enum btree_id btree, struct bkey_i *k, in bch2_btree_insert_nonextent() argument 633 bch2_trans_iter_init(trans, &iter, btree, k->k.p, in bch2_btree_insert_nonextent() 697 enum btree_id btree, struct bpos pos, in bch2_btree_delete() argument 703 bch2_trans_iter_init(trans, &iter, btree, pos, in bch2_btree_delete() 797 int bch2_btree_bit_mod(struct btree_trans *trans, enum btree_id btree, in bch2_btree_bit_mod() argument 810 bch2_trans_iter_init(trans, &iter, btree, pos, BTREE_ITER_intent); in bch2_btree_bit_mod() [all …]
|
H A D | snapshot.c | 300 enum btree_id btree, unsigned level, in __bch2_mark_snapshot() argument 930 enum btree_id btree; member 946 return r->btree == BTREE_ID_inodes in same_snapshot() 1003 for (unsigned btree = 0; btree < BTREE_ID_NR; btree++) { in bch2_reconstruct_snapshots() local 1004 if (btree_type_has_snapshots(btree)) { in bch2_reconstruct_snapshots() 1005 r.btree = btree; in bch2_reconstruct_snapshots() 1605 for (unsigned btree = 0; btree < BTREE_ID_NR; btree++) { in bch2_delete_dead_snapshots() local 1610 if (!btree_type_has_snapshots(btree)) in bch2_delete_dead_snapshots() 1614 btree, POS_MIN, in bch2_delete_dead_snapshots() 1750 enum btree_id btree, in bch2_propagate_key_to_snapshot_leaf() argument [all …]
|
H A D | bkey_cmp.h | 91 const struct btree *b) in __bch2_bkey_cmp_packed_format_checked_inlined() 109 int bch2_bkey_cmp_packed_inlined(const struct btree *b, in bch2_bkey_cmp_packed_inlined()
|
H A D | trace.h | 96 TP_PROTO(struct bch_fs *c, struct btree *b), 121 TP_PROTO(struct btree_trans *trans, struct btree *b), 434 TP_PROTO(struct bch_fs *c, struct btree *b), 461 TP_PROTO(struct btree_trans *trans, struct btree *b), 572 struct btree *b = btree_path_node(path, level); 968 struct btree *b), 1331 __field(unsigned, btree ) 1340 __entry->btree = path->btree_id; 1349 bch2_btree_id_str(__entry->btree), 1702 struct btree *b = path->l[0].b; [all …]
|
H A D | bkey.c | 348 void bch2_bkey_unpack(const struct btree *b, struct bkey_i *dst, in bch2_bkey_unpack() 403 const struct btree *b, in bkey_packed_successor() 470 const struct btree *b) in bch2_bkey_pack_pos_lossy() 703 unsigned bch2_bkey_greatest_differing_bit(const struct btree *b, in bch2_bkey_greatest_differing_bit() 747 unsigned bch2_bkey_ffs(const struct btree *b, const struct bkey_packed *k) in bch2_bkey_ffs() 1019 const struct btree *b) in __bch2_bkey_cmp_packed_format_checked() 1025 int __bch2_bkey_cmp_left_packed_format_checked(const struct btree *b, in __bch2_bkey_cmp_left_packed_format_checked() 1033 int bch2_bkey_cmp_packed(const struct btree *b, in bch2_bkey_cmp_packed() 1041 int __bch2_bkey_cmp_left_packed(const struct btree *b, in __bch2_bkey_cmp_left_packed()
|
/linux/include/trace/events/ |
H A D | bcache.h | 64 TP_PROTO(struct btree *b), 258 TP_PROTO(struct btree *b), 263 TP_PROTO(struct btree *b), 283 TP_PROTO(struct btree *b), 293 TP_PROTO(struct btree *b), 333 TP_PROTO(struct btree *b, struct bkey *k, unsigned op, unsigned status), 366 TP_PROTO(struct btree *b, unsigned keys), 383 TP_PROTO(struct btree *b, unsigned keys), 388 TP_PROTO(struct btree *b, unsigned keys), 393 TP_PROTO(struct btree *b),
|
/linux/drivers/md/bcache/ |
H A D | debug.h | 11 void bch_btree_verify(struct btree *b); 20 static inline void bch_btree_verify(struct btree *b) {} in bch_btree_verify()
|
/linux/fs/xfs/ |
H A D | xfs_attr_list.c | 220 struct xfs_da_node_entry *btree; in xfs_attr_node_list_lookup() local 268 btree = nodehdr.btree; in xfs_attr_node_list_lookup() 269 for (i = 0; i < nodehdr.count; btree++, i++) { in xfs_attr_node_list_lookup() 270 if (cursor->hashval <= be32_to_cpu(btree->hashval)) { in xfs_attr_node_list_lookup() 271 cursor->blkno = be32_to_cpu(btree->before); in xfs_attr_node_list_lookup() 273 btree); in xfs_attr_node_list_lookup()
|
/linux/fs/befs/ |
H A D | ChangeLog | 27 * Did the string comparison really right this time (btree.c) [WD] 30 a pointer value. (btree.c) [WD] 39 before. (btree.c) [Sergey S. Kostyliov <rathamahata@php4.ru>] 56 (btree.c) [WD] 105 * Removed notion of btree handle from btree.c. It was unnecessary, as the 128 (btree.c) [WD] 133 seekleaf() in btree.c [WD] 148 (datastream.c, btree.c super.c inode.c) [WD] 259 * Added btree searching function. 263 * Additional comments in btree.c [all …]
|
/linux/fs/xfs/scrub/ |
H A D | listxattr.c | 139 struct xfs_da_node_entry *btree; in xchk_xattr_find_leftmost_leaf() local 192 btree = nodehdr.btree; in xchk_xattr_find_leftmost_leaf() 193 blkno = be32_to_cpu(btree->before); in xchk_xattr_find_leftmost_leaf()
|
/linux/Documentation/filesystems/xfs/ |
H A D | xfs-online-fsck-design.rst | 1229 of btree? 1314 `refcount btree 1316 `inode btree 1318 `rmap btree 2218 `in-memory btree 2281 btree type. 2903 btree repair: 3795 a. Create a btree cursor for the in-memory btree. 3806 a. Create a btree cursor for the in-memory btree. 3819 shadow btree, like all other btree rebuilding functions. [all …]
|