/linux/fs/bcachefs/ |
H A D | btree_cache.c | 64 struct btree_cache *bc = &c->btree_cache; in btree_node_data_free() 143 struct btree_cache *bc = &c->btree_cache; in __bch2_btree_node_mem_alloc() 250 struct btree_cache *bc = &c->btree_cache; in __btree_node_reclaim() 371 struct btree_cache *bc = &c->btree_cache; in bch2_btree_cache_scan() 472 struct btree_cache *bc = &c->btree_cache; in bch2_btree_cache_count() 482 struct btree_cache *bc = &c->btree_cache; in bch2_fs_btree_cache_exit() 536 struct btree_cache *bc = &c->btree_cache; in bch2_fs_btree_cache_init() 590 struct btree_cache *bc = &c->btree_cache; in bch2_btree_cache_cannibalize_unlock() 602 struct btree_cache *bc = &c->btree_cache; in bch2_btree_cache_cannibalize_lock() 634 struct btree_cache *bc = &c->btree_cache; in btree_node_cannibalize() [all …]
|
H A D | btree_cache.h | 15 void bch2_btree_node_hash_remove(struct btree_cache *, struct btree *); 16 int __bch2_btree_node_hash_insert(struct btree_cache *, struct btree *); 17 int bch2_btree_node_hash_insert(struct btree_cache *, struct btree *, 43 void bch2_fs_btree_cache_init_early(struct btree_cache *); 75 for ((_tbl) = rht_dereference_rcu((_c)->btree_cache.table.tbl, \ 76 &(_c)->btree_cache.table), \ 137 void bch2_btree_cache_to_text(struct printbuf *, const struct btree_cache *);
|
H A D | btree_update_interior.c | 240 mutex_lock(&c->btree_cache.lock); in __btree_node_free() 242 mutex_unlock(&c->btree_cache.lock); in __btree_node_free() 285 mutex_lock(&c->btree_cache.lock); in bch2_btree_node_free_never_used() 288 mutex_unlock(&c->btree_cache.lock); in bch2_btree_node_free_never_used() 1297 mutex_lock(&c->btree_cache.lock); in bch2_btree_set_root_inmem() 1299 mutex_unlock(&c->btree_cache.lock); in bch2_btree_set_root_inmem() 1907 mutex_lock(&c->btree_cache.lock); in __btree_increase_depth() 1909 mutex_unlock(&c->btree_cache.lock); in __btree_increase_depth() 2396 mutex_lock(&c->btree_cache.lock); in __bch2_btree_node_update_key() 2414 mutex_lock(&c->btree_cache.lock); in __bch2_btree_node_update_key() [all …]
|
H A D | sysfs.c | 169 read_attribute(btree_cache); 244 mutex_lock(&c->btree_cache.lock); in bch2_btree_cache_size() 245 list_for_each_entry(b, &c->btree_cache.live, list) in bch2_btree_cache_size() 248 mutex_unlock(&c->btree_cache.lock); in bch2_btree_cache_size() 386 bch2_btree_cache_to_text(out, &c->btree_cache); in SHOW() 475 c->btree_cache.shrink->scan_objects(c->btree_cache.shrink, &sc); in STORE() 483 c->btree_key_cache.shrink->scan_objects(c->btree_cache.shrink, &sc); in STORE()
|
H A D | backpointers.c | 781 c->btree_cache.pinned_nodes_leaf_mask = btree_leaf_mask; in bch2_get_btree_in_memory_pos() 782 c->btree_cache.pinned_nodes_interior_mask = btree_interior_mask; in bch2_get_btree_in_memory_pos() 783 c->btree_cache.pinned_nodes_start = start; in bch2_get_btree_in_memory_pos() 784 c->btree_cache.pinned_nodes_end = *end = BBPOS_MAX; in bch2_get_btree_in_memory_pos() 802 c->btree_cache.pinned_nodes_end = *end = in bch2_get_btree_in_memory_pos() 897 c->btree_cache.pinned_nodes_leaf_mask = 0; in bch2_check_extents_to_backpointers() 898 c->btree_cache.pinned_nodes_interior_mask = 0; in bch2_check_extents_to_backpointers() 1005 c->btree_cache.pinned_nodes_leaf_mask = 0; in bch2_check_backpointers_to_extents() 1006 c->btree_cache.pinned_nodes_interior_mask = 0; in bch2_check_backpointers_to_extents()
|
H A D | btree_io.h | 21 atomic_inc(&c->btree_cache.dirty); in set_btree_node_dirty_acct() 27 atomic_dec(&c->btree_cache.dirty); in clear_btree_node_dirty_acct()
|
H A D | journal_reclaim.c | 672 if (atomic_read(&c->btree_cache.dirty) * 2 > c->btree_cache.used) in __bch2_journal_reclaim() 680 atomic_read(&c->btree_cache.dirty), in __bch2_journal_reclaim() 681 c->btree_cache.used, in __bch2_journal_reclaim()
|
H A D | btree_gc.c | 167 mutex_lock(&c->btree_cache.lock); in set_node_max() 168 bch2_btree_node_hash_remove(&c->btree_cache, b); in set_node_max() 171 ret = __bch2_btree_node_hash_insert(&c->btree_cache, b); in set_node_max() 173 mutex_unlock(&c->btree_cache.lock); in set_node_max() 533 bch2_btree_node_hash_remove(&c->btree_cache, b); in bch2_check_topology() 534 mutex_lock(&c->btree_cache.lock); in bch2_check_topology() 535 list_move(&b->list, &c->btree_cache.freeable); in bch2_check_topology() 536 mutex_unlock(&c->btree_cache.lock); in bch2_check_topology()
|
H A D | Makefile | 13 btree_cache.o \
|
H A D | btree_io.c | 1765 BUG_ON(bch2_btree_node_hash_insert(&c->btree_cache, b, level, id)); in __bch2_btree_root_read() 1772 bch2_btree_node_hash_remove(&c->btree_cache, b); in __bch2_btree_root_read() 1774 mutex_lock(&c->btree_cache.lock); in __bch2_btree_root_read() 1775 list_move(&b->list, &c->btree_cache.freeable); in __bch2_btree_root_read() 1776 mutex_unlock(&c->btree_cache.lock); in __bch2_btree_root_read() 2056 atomic_dec(&c->btree_cache.dirty); in __bch2_btree_node_write()
|
H A D | bcachefs.h | 800 struct btree_cache btree_cache; member
|
H A D | debug.c | 542 tbl = rht_dereference_rcu(c->btree_cache.table.tbl, in bch2_cached_btree_nodes_read() 543 &c->btree_cache.table); in bch2_cached_btree_nodes_read()
|
H A D | btree_types.h | 141 struct btree_cache { struct
|
H A D | super.c | 375 BUG_ON(atomic_read(&c->btree_cache.dirty)); in bch2_fs_read_only() 818 bch2_fs_btree_cache_init_early(&c->btree_cache); in bch2_fs_alloc()
|
/linux/drivers/md/bcache/ |
H A D | btree.c | 556 list_move(&b->list, &b->c->btree_cache); in mca_data_alloc() 715 list_for_each_entry_safe_reverse(b, t, &c->btree_cache, list) { in bch_mca_scan() 762 list_move(&c->verify_data->list, &c->btree_cache); in bch_btree_cache_free() 768 &c->btree_cache); in bch_btree_cache_free() 770 while (!list_empty(&c->btree_cache)) { in bch_btree_cache_free() 771 b = list_first_entry(&c->btree_cache, struct btree, list); in bch_btree_cache_free() 804 list_splice_init(&c->btree_cache, in bch_btree_cache_alloc() 896 list_for_each_entry_reverse(b, &c->btree_cache, list) in mca_cannibalize() 900 list_for_each_entry_reverse(b, &c->btree_cache, list) in mca_cannibalize() 966 list_move(&b->list, &c->btree_cache); in mca_alloc()
|
H A D | bcache.h | 581 struct list_head btree_cache; member
|
H A D | super.c | 1729 list_add(&c->root->list, &c->btree_cache); in CLOSURE_CALLBACK() 1736 list_for_each_entry(b, &c->btree_cache, list) { in CLOSURE_CALLBACK() 1912 INIT_LIST_HEAD(&c->btree_cache); in bch_cache_set_alloc()
|
H A D | journal.c | 455 list_for_each_entry_safe_reverse(b, t, &c->btree_cache, list) { in btree_flush_write()
|
H A D | sysfs.c | 688 list_for_each_entry(b, &c->btree_cache, list) in bch_cache_size()
|