/linux/drivers/md/bcache/ |
H A D | btree.h | 126 struct cache_set *c; 195 static inline void set_gc_sectors(struct cache_set *c) in set_gc_sectors() 200 void bkey_put(struct cache_set *c, struct bkey *k); 231 struct cache_set *c; 280 int bch_gc_thread_start(struct cache_set *c); 281 void bch_initial_gc_finish(struct cache_set *c); 282 void bch_moving_gc(struct cache_set *c); 283 int bch_btree_check(struct cache_set *c); 285 void bch_cannibalize_unlock(struct cache_set *c); 287 static inline void wake_up_gc(struct cache_set *c) in wake_up_gc() [all …]
|
H A D | bcache.h | 255 struct cache_set *c; 419 struct cache_set *set; 517 struct cache_set { struct 974 struct bio *bch_bbio_alloc(struct cache_set *c); 1022 int bch_uuid_write(struct cache_set *c); 1023 void bcache_write_super(struct cache_set *c); 1034 void bch_cache_set_stop(struct cache_set *c); 1037 void bch_btree_cache_free(struct cache_set *c); 1038 int bch_btree_cache_alloc(struct cache_set *c); 1040 int bch_open_buckets_alloc(struct cache_set *c); [all …]
|
H A D | journal.h | 97 struct cache_set *c; 171 struct cache_set; 175 atomic_t *bch_journal(struct cache_set *c, 179 void bch_journal_mark(struct cache_set *c, struct list_head *list); 180 void bch_journal_meta(struct cache_set *c, struct closure *cl); 181 int bch_journal_read(struct cache_set *c, struct list_head *list); 182 int bch_journal_replay(struct cache_set *c, struct list_head *list); 184 void bch_journal_free(struct cache_set *c); 185 int bch_journal_alloc(struct cache_set *c);
|
H A D | stats.h | 40 struct cache_set; 54 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d, 56 void bch_mark_cache_miss_collision(struct cache_set *c, 58 void bch_mark_sectors_bypassed(struct cache_set *c,
|
H A D | extents.h | 9 struct cache_set; 12 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k); 13 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k);
|
H A D | io.c | 17 void bch_bbio_free(struct bio *bio, struct cache_set *c) in bch_bbio_free() 24 struct bio *bch_bbio_alloc(struct cache_set *c) in bch_bbio_alloc() 35 void __bch_submit_bbio(struct bio *bio, struct cache_set *c) in __bch_submit_bbio() 46 void bch_submit_bbio(struct bio *bio, struct cache_set *c, in bch_submit_bbio() 137 void bch_bbio_count_io_errors(struct cache_set *c, struct bio *bio, in bch_bbio_count_io_errors() 167 void bch_bbio_endio(struct cache_set *c, struct bio *bio, in bch_bbio_endio()
|
H A D | debug.h | 7 struct cache_set; 30 void bch_debug_init_cache_set(struct cache_set *c); 32 static inline void bch_debug_init_cache_set(struct cache_set *c) {} in bch_debug_init_cache_set()
|
H A D | journal.c | 339 static bool is_discard_enabled(struct cache_set *s) in is_discard_enabled() 416 static void btree_flush_write(struct cache_set *c) in btree_flush_write() 649 static void journal_reclaim(struct cache_set *c) in journal_reclaim() 742 closure_type(c, struct cache_set, journal.io); 751 closure_type(c, struct cache_set, journal.io); 828 closure_type(c, struct cache_set, journal.io); in CLOSURE_CALLBACK() 834 static void journal_try_write(struct cache_set *c) in journal_try_write() 909 struct cache_set, in journal_write_work() 924 atomic_t *bch_journal(struct cache_set *c, in bch_journal() 974 void bch_journal_free(struct cache_set *c) in bch_journal_free() [all …]
|
H A D | features.c | 56 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_compat() 63 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_ro_compat() 70 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size) in bch_print_cache_set_feature_incompat()
|
H A D | super.c | 403 struct cache_set *c = container_of(cl, struct cache_set, uuid_write); in uuid_endio() 525 int bch_uuid_write(struct cache_set *c) in bch_uuid_write() 1136 struct cache_set *c = dc->disk.c; in cached_dev_detach_finish() 1446 struct cache_set *c; in register_bdev() 1660 struct cache_set *c = container_of(kobj, struct cache_set, kobj); in bch_cache_set_release() 1854 struct cache_set *c = kzalloc(sizeof(struct cache_set), GFP_KERNEL); in bch_cache_set_alloc() 2144 struct cache_set *c; in register_cache_set() 2406 struct cache_set *c, *tc; in bch_is_open_backing() 2421 struct cache_set *c, *tc; in bch_is_open_cache() 2665 struct cache_set *c, *tc; in bch_pending_bdevs_cleanup() [all …]
|
H A D | features.h | 109 int bch_print_cache_set_feature_compat(struct cache_set *c, char *buf, int size); 110 int bch_print_cache_set_feature_ro_compat(struct cache_set *c, char *buf, int size); 111 int bch_print_cache_set_feature_incompat(struct cache_set *c, char *buf, int size);
|
H A D | request.h | 7 struct cache_set *c; 36 unsigned int bch_get_congested(const struct cache_set *c);
|
H A D | btree.c | 128 void bkey_put(struct cache_set *c, struct bkey *k) in bkey_put() 670 struct cache_set *c = shrink->private_data; in bch_mca_scan() 737 struct cache_set *c = shrink->private_data; in bch_mca_count() 748 void bch_btree_cache_free(struct cache_set *c) in bch_btree_cache_free() 796 int bch_btree_cache_alloc(struct cache_set *c) in bch_btree_cache_alloc() 1207 struct cache_set *c = b->c; in btree_check_reserve() 1833 static void bch_btree_gc(struct cache_set *c) in bch_btree_gc() 1892 struct cache_set *c = arg; in bch_gc_thread() 1912 int bch_gc_thread_start(struct cache_set *c) in bch_gc_thread_start() 1964 struct cache_set *c = check_state->c; in bch_btree_check_thread() [all …]
|
H A D | sysfs.c | 296 struct cache_set *c; in STORE() 632 static int bch_bset_print_stats(struct cache_set *c, char *buf) in bch_bset_print_stats() 658 static unsigned int bch_root_usage(struct cache_set *c) in bch_root_usage() 684 static size_t bch_cache_size(struct cache_set *c) in bch_cache_size() 697 static unsigned int bch_cache_max_chain(struct cache_set *c) in bch_cache_max_chain() 714 static unsigned int bch_btree_used(struct cache_set *c) in bch_btree_used() 720 static unsigned int bch_average_key_size(struct cache_set *c) in bch_average_key_size() 729 struct cache_set *c = container_of(kobj, struct cache_set, kobj); in SHOW() 816 struct cache_set *c = container_of(kobj, struct cache_set, kobj); in STORE() 937 struct cache_set *c = container_of(kobj, struct cache_set, internal); in SHOW() [all …]
|
H A D | movinggc.c | 24 struct cache_set *c = container_of(buf, struct cache_set, in moving_pred() 126 static void read_moving(struct cache_set *c) in read_moving() 208 void bch_moving_gc(struct cache_set *c) in bch_moving_gc() 269 void bch_moving_init_cache_set(struct cache_set *c) in bch_moving_init_cache_set()
|
H A D | alloc.c | 86 void bch_rescale_priorities(struct cache_set *c, int sectors) in bch_rescale_priorities() 507 void bch_bucket_free(struct cache_set *c, struct bkey *k) in bch_bucket_free() 515 int __bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, in __bch_bucket_alloc_set() 543 int bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, in bch_bucket_alloc_set() 588 static struct open_bucket *pick_data_bucket(struct cache_set *c, in pick_data_bucket() 629 bool bch_alloc_sectors(struct cache_set *c, in bch_alloc_sectors() 720 void bch_open_buckets_free(struct cache_set *c) in bch_open_buckets_free() 732 int bch_open_buckets_alloc(struct cache_set *c) in bch_open_buckets_alloc()
|
H A D | writeback.c | 20 static void update_gc_after_writeback(struct cache_set *c) in update_gc_after_writeback() 32 struct cache_set *c = dc->disk.c; in __calc_target_rate() 101 struct cache_set *c = dc->disk.c; in __update_writeback_rate() 160 static bool idle_counter_exceeded(struct cache_set *c) in idle_counter_exceeded() 207 static bool set_at_max_writeback_rate(struct cache_set *c, in set_at_max_writeback_rate() 247 struct cache_set *c = dc->disk.c; in update_writeback_rate() 596 void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned int inode, in bcache_dev_sectors_dirty_add() 742 struct cache_set *c = dc->disk.c; in bch_writeback_thread() 872 static int bch_root_node_dirty_init(struct cache_set *c, in bch_root_node_dirty_init() 910 struct cache_set *c = state->c; in bch_dirty_init_thread() [all …]
|
H A D | stats.c | 198 void bch_mark_cache_accounting(struct cache_set *c, struct bcache_device *d, in bch_mark_cache_accounting() 207 void bch_mark_cache_miss_collision(struct cache_set *c, struct bcache_device *d) in bch_mark_cache_miss_collision() 215 void bch_mark_sectors_bypassed(struct cache_set *c, struct cached_dev *dc, in bch_mark_sectors_bypassed()
|
H A D | writeback.h | 38 struct cache_set *c; 148 void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned int inode,
|
H A D | extents.c | 48 static bool __ptr_invalid(struct cache_set *c, const struct bkey *k) in __ptr_invalid() 69 static const char *bch_ptr_status(struct cache_set *c, const struct bkey *k) in bch_ptr_status() 150 bool __bch_btree_ptr_invalid(struct cache_set *c, const struct bkey *k) in __bch_btree_ptr_invalid() 329 struct cache_set *c, in bch_subtract_dirty() 343 struct cache_set *c = container_of(b, struct btree, keys)->c; in bch_extent_insert_fixup() 495 bool __bch_extent_invalid(struct cache_set *c, const struct bkey *k) in __bch_extent_invalid()
|
H A D | debug.c | 163 struct cache_set *c; 209 struct cache_set *c = inode->i_private; in bch_dump_open() 237 void bch_debug_init_cache_set(struct cache_set *c) in bch_debug_init_cache_set()
|
H A D | request.c | 91 struct cache_set *c) in bch_keylist_realloc() 324 unsigned int bch_get_congested(const struct cache_set *c) in bch_get_congested() 365 struct cache_set *c = dc->disk.c; in check_should_bypass() 1144 static void quit_max_writeback_rate(struct cache_set *c, in quit_max_writeback_rate()
|
/linux/include/trace/events/ |
H A D | bcache.h | 194 DECLARE_EVENT_CLASS(cache_set, 195 TP_PROTO(struct cache_set *c), 214 DEFINE_EVENT(cache_set, bcache_journal_full, 215 TP_PROTO(struct cache_set *c), 220 TP_PROTO(struct cache_set *c), 253 TP_PROTO(struct cache_set *c), 288 TP_PROTO(struct cache_set *c), 312 DEFINE_EVENT(cache_set, bcache_gc_start, 313 TP_PROTO(struct cache_set *c), 317 DEFINE_EVENT(cache_set, bcache_gc_end, [all …]
|