Home
last modified time | relevance | path

Searched refs:bucket (Results 1 – 25 of 216) sorted by relevance

123456789

/linux/drivers/md/dm-vdo/
H A Dint-map.c276 struct bucket *bucket = dereference_hop(neighborhood, next_hop); in insert_in_hop_list() local
325 struct bucket *bucket, in search_hop_list() argument
337 struct bucket *entry = dereference_hop(bucket, next_hop); in search_hop_list()
466 struct bucket *bucket; in move_empty_bucket() local
468 for (bucket = &hole[1 - NEIGHBORHOOD]; bucket < hole; bucket++) { in move_empty_bucket()
473 struct bucket *new_hole = dereference_hop(bucket, bucket->first_hop); in move_empty_bucket()
531 struct bucket *bucket = search_hop_list(map, neighborhood, key, NULL); in update_mapping() local
613 struct bucket *neighborhood, *bucket; in vdo_int_map_put() local
657 bucket->key = key; in vdo_int_map_put()
680 struct bucket *bucket = select_bucket(map, key); in vdo_int_map_remove() local
[all …]
H A Dpriority-table.c23 struct bucket { struct
44 struct bucket buckets[];
69 struct bucket *bucket = &table->buckets[priority]; in vdo_make_priority_table() local
71 bucket->priority = priority; in vdo_make_priority_table()
72 INIT_LIST_HEAD(&bucket->queue); in vdo_make_priority_table()
140 static inline void mark_bucket_empty(struct priority_table *table, struct bucket *bucket) in mark_bucket_empty() argument
157 struct bucket *bucket; in vdo_priority_table_dequeue() local
173 bucket = &table->buckets[top_priority]; in vdo_priority_table_dequeue()
174 entry = bucket->queue.next; in vdo_priority_table_dequeue()
178 if (list_empty(&bucket->queue)) in vdo_priority_table_dequeue()
[all …]
/linux/net/mptcp/
H A Dtoken.c111 struct token_bucket *bucket; in mptcp_token_new_request() local
122 bucket = token_bucket(token); in mptcp_token_new_request()
123 spin_lock_bh(&bucket->lock); in mptcp_token_new_request()
130 bucket->chain_len++; in mptcp_token_new_request()
157 struct token_bucket *bucket; in mptcp_token_new_connect() local
164 spin_lock_bh(&bucket->lock); in mptcp_token_new_connect()
177 bucket->chain_len++; in mptcp_token_new_connect()
196 struct token_bucket *bucket; in mptcp_token_accept() local
200 spin_lock_bh(&bucket->lock); in mptcp_token_accept()
364 bucket->chain_len--; in mptcp_token_destroy_request()
[all …]
/linux/net/ceph/crush/
H A Dmapper.c88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose()
89 bucket->size; in bucket_perm_choose()
113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose()
114 (bucket->size - p); in bucket_perm_choose()
131 return bucket->items[s]; in bucket_perm_choose()
148 __u64 w = crush_hash32_4(bucket->h.hash, x, bucket->h.items[i], in bucket_list_choose()
149 r, bucket->h.id); in bucket_list_choose()
153 i, x, r, bucket->h.items[i], bucket->item_weights[i], in bucket_list_choose()
164 return bucket->h.items[0]; in bucket_list_choose()
237 draw = crush_hash32_3(bucket->h.hash, x, bucket->h.items[i], r); in bucket_straw_choose()
[all …]
/linux/drivers/interconnect/qcom/
H A Dbcm-voter.c65 int bucket, i; in bcm_aggregate_mask() local
67 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate_mask()
68 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask()
69 bcm->vote_y[bucket] = 0; in bcm_aggregate_mask()
75 if (node->sum_avg[bucket] || node->max_peak[bucket]) { in bcm_aggregate_mask()
76 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask()
77 bcm->vote_y[bucket] = bcm->enable_mask; in bcm_aggregate_mask()
94 size_t i, bucket; in bcm_aggregate() local
99 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate()
104 agg_avg[bucket] = max(agg_avg[bucket], temp); in bcm_aggregate()
[all …]
/linux/block/
H A Dblk-stat.c55 int bucket, cpu; in blk_stat_add() local
66 bucket = cb->bucket_fn(rq); in blk_stat_add()
67 if (bucket < 0) in blk_stat_add()
80 unsigned int bucket; in blk_stat_timer_fn() local
83 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
84 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn()
90 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
91 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn()
92 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn()
136 unsigned int bucket; in blk_stat_add_callback() local
[all …]
/linux/net/sched/
H A Dsch_hhf.c333 bucket->head = skb->next; in dequeue_head()
341 if (bucket->head == NULL) in bucket_add()
342 bucket->head = skb; in bucket_add()
344 bucket->tail->next = skb; in bucket_add()
345 bucket->tail = skb; in bucket_add()
352 struct wdrr_bucket *bucket; in hhf_drop() local
356 if (!bucket->head) in hhf_drop()
359 if (bucket->head) { in hhf_drop()
381 bucket = &q->buckets[idx]; in hhf_enqueue()
382 bucket_add(bucket, skb); in hhf_enqueue()
[all …]
/linux/fs/bcachefs/
H A Dbackpointers.h52 *bucket = bp_pos_to_bucket(ca, bp_pos); in bp_pos_to_bucket_nodev_noerror()
59 return !bch2_fs_inconsistent_on(!bp_pos_to_bucket_nodev_noerror(c, bp_pos, bucket), in bp_pos_to_bucket_nodev()
64 struct bpos bucket, in bucket_pos_to_bp_noerror()
67 return POS(bucket.inode, in bucket_pos_to_bp_noerror()
68 (bucket_to_sector(ca, bucket.offset) << in bucket_pos_to_bp_noerror()
76 struct bpos bucket, in bucket_pos_to_bp()
79 struct bpos ret = bucket_pos_to_bp_noerror(ca, bucket, bucket_offset); in bucket_pos_to_bp()
80 EBUG_ON(!bkey_eq(bucket, bp_pos_to_bucket(ca, ret))); in bucket_pos_to_bp()
85 struct bpos bucket, struct bch_backpointer, struct bkey_s_c, bool);
89 struct bpos bucket, in bch2_bucket_backpointer_mod() argument
[all …]
H A Dalloc_foreground.h121 unsigned dev, u64 bucket) in open_bucket_hashslot() argument
124 (jhash_3words(dev, bucket, bucket >> 32, 0) & in open_bucket_hashslot()
128 static inline bool bch2_bucket_is_open(struct bch_fs *c, unsigned dev, u64 bucket) in bch2_bucket_is_open() argument
130 open_bucket_idx_t slot = *open_bucket_hashslot(c, dev, bucket); in bch2_bucket_is_open()
135 if (ob->dev == dev && ob->bucket == bucket) in bch2_bucket_is_open()
144 static inline bool bch2_bucket_is_open_safe(struct bch_fs *c, unsigned dev, u64 bucket) in bch2_bucket_is_open_safe() argument
148 if (bch2_bucket_is_open(c, dev, bucket)) in bch2_bucket_is_open_safe()
152 ret = bch2_bucket_is_open(c, dev, bucket); in bch2_bucket_is_open_safe()
H A Dbackpointers.c19 struct bpos bucket, in extent_matches_bp() argument
39 if (bpos_eq(bucket, bucket2) && in extent_matches_bp()
97 bch2_bpos_to_text(out, bucket); in bch2_backpointer_k_to_text()
167 struct bpos bucket, in bch2_bucket_backpointer_mod_nowritebuffer() argument
280 struct bpos bucket; in backpointer_not_found() local
287 bch2_bpos_to_text(&buf, bucket); in backpointer_not_found()
314 struct bpos bucket; in bch2_backpointer_get_key() local
355 struct bpos bucket; in bch2_backpointer_get_node() local
373 bucket, bp)) in bch2_backpointer_get_node()
396 struct bpos bucket; in bch2_check_btree_backpointer() local
[all …]
H A Dalloc_background.h25 static inline u64 bucket_to_u64(struct bpos bucket) in bucket_to_u64() argument
27 return (bucket.inode << 48) | bucket.offset; in bucket_to_u64()
30 static inline struct bpos u64_to_bucket(u64 bucket) in u64_to_bucket() argument
32 return POS(bucket >> 48, bucket & ~(~0ULL << 48)); in u64_to_bucket()
40 static inline void alloc_to_bucket(struct bucket *dst, struct bch_alloc_v4 src) in alloc_to_bucket()
49 static inline void __bucket_m_to_alloc(struct bch_alloc_v4 *dst, struct bucket src) in __bucket_m_to_alloc()
58 static inline struct bch_alloc_v4 bucket_m_to_alloc(struct bucket b) in bucket_m_to_alloc()
76 static inline bool bucket_data_type_mismatch(enum bch_data_type bucket, in bucket_data_type_mismatch() argument
79 return !data_type_is_empty(bucket) && in bucket_data_type_mismatch()
80 bucket_data_type(bucket) != bucket_data_type(ptr); in bucket_data_type_mismatch()
H A Dnocow_locking.h28 struct bpos bucket, int flags) in bch2_bucket_nocow_lock() argument
30 u64 dev_bucket = bucket_to_u64(bucket); in bch2_bucket_nocow_lock()
37 struct bpos bucket, int flags) in bch2_bucket_nocow_trylock() argument
39 u64 dev_bucket = bucket_to_u64(bucket); in bch2_bucket_nocow_trylock()
/linux/drivers/infiniband/sw/rdmavt/
H A Dtrace_qp.h18 TP_PROTO(struct rvt_qp *qp, u32 bucket),
19 TP_ARGS(qp, bucket),
23 __field(u32, bucket)
28 __entry->bucket = bucket;
34 __entry->bucket
39 TP_PROTO(struct rvt_qp *qp, u32 bucket),
40 TP_ARGS(qp, bucket));
43 TP_PROTO(struct rvt_qp *qp, u32 bucket),
44 TP_ARGS(qp, bucket));
/linux/fs/nfs/
H A Dnfs42xattr.c238 entry->bucket = NULL; in nfs4_xattr_alloc_entry()
396 spin_lock(&bucket->lock); in nfs4_xattr_discard_cache()
397 bucket->draining = true; in nfs4_xattr_discard_cache()
534 entry->bucket = bucket; in nfs4_xattr_hash_add()
536 spin_lock(&bucket->lock); in nfs4_xattr_hash_add()
538 if (bucket->draining) { in nfs4_xattr_hash_add()
555 spin_unlock(&bucket->lock); in nfs4_xattr_hash_add()
571 spin_lock(&bucket->lock); in nfs4_xattr_hash_remove()
594 spin_lock(&bucket->lock); in nfs4_xattr_hash_find()
878 bucket = entry->bucket; in entry_lru_isolate()
[all …]
H A Dpnfs_nfs.c65 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { in pnfs_free_bucket_lseg()
67 bucket->lseg = NULL; in pnfs_free_bucket_lseg()
91 if (bucket) in pnfs_generic_clear_request_commit()
414 lseg = pnfs_get_lseg(bucket->lseg); in pnfs_bucket_get_committing()
435 struct pnfs_commit_bucket *bucket; in pnfs_generic_retry_commit() local
439 for (bucket = buckets; idx < nbuckets; bucket++, idx++) { in pnfs_generic_retry_commit()
456 struct pnfs_commit_bucket *bucket; in pnfs_bucket_alloc_ds_commits() local
461 for (i = 0, bucket = buckets; i < nbuckets; i++, bucket++) { in pnfs_bucket_alloc_ds_commits()
1207 struct pnfs_commit_bucket *bucket; in pnfs_layout_mark_request_commit() local
1214 list = &bucket->written; in pnfs_layout_mark_request_commit()
[all …]
/linux/net/9p/
H A Derror.c179 int bucket; in p9_error_init() local
182 for (bucket = 0; bucket < ERRHASHSZ; bucket++) in p9_error_init()
183 INIT_HLIST_HEAD(&hash_errmap[bucket]); in p9_error_init()
188 bucket = jhash(c->name, c->namelen, 0) % ERRHASHSZ; in p9_error_init()
190 hlist_add_head(&c->list, &hash_errmap[bucket]); in p9_error_init()
208 int bucket; in p9_errstr2errno() local
212 bucket = jhash(errstr, len, 0) % ERRHASHSZ; in p9_errstr2errno()
213 hlist_for_each_entry(c, &hash_errmap[bucket], list) { in p9_errstr2errno()
/linux/net/vmw_vsock/
H A Ddiag.c52 unsigned int bucket; in vsock_diag_dump() local
63 bucket = cb->args[1]; in vsock_diag_dump()
72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump()
73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump()
94 bucket++; in vsock_diag_dump()
98 bucket = 0; in vsock_diag_dump()
102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump()
103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump()
128 bucket++; in vsock_diag_dump()
135 cb->args[1] = bucket; in vsock_diag_dump()
/linux/drivers/cpuidle/governors/
H A Dmenu.c116 unsigned int bucket; member
124 int bucket = 0; in which_bucket() local
133 bucket = BUCKETS/2; in which_bucket()
136 return bucket; in which_bucket()
138 return bucket + 1; in which_bucket()
140 return bucket + 2; in which_bucket()
142 return bucket + 3; in which_bucket()
144 return bucket + 4; in which_bucket()
145 return bucket + 5; in which_bucket()
298 data->correction_factor[data->bucket], in menu_select()
[all …]
/linux/kernel/dma/
H A Ddebug.c267 __releases(&bucket->lock) in put_hash_bucket()
590 struct hash_bucket *bucket; in add_dma_entry() local
595 hash_bucket_add(bucket, entry); in add_dma_entry()
596 put_hash_bucket(bucket, flags); in add_dma_entry()
964 struct hash_bucket *bucket; in check_unmap() local
1054 put_hash_bucket(bucket, flags); in check_unmap()
1097 struct hash_bucket *bucket; in check_sync() local
1164 put_hash_bucket(bucket, flags); in check_sync()
1249 struct hash_bucket *bucket; in debug_dma_mapping_error() local
1279 put_hash_bucket(bucket, flags); in debug_dma_mapping_error()
[all …]
/linux/Documentation/userspace-api/media/v4l/
H A Dmetafmt-vsp1-hgt.rst28 The Saturation position **n** (0 - 31) of the bucket in the matrix is
33 The Hue position **m** (0 - 5) of the bucket in the matrix depends on
101 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0]
103 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0]
107 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0]
109 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0]
113 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0]
117 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0]
121 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0]
125 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0]
[all …]
/linux/fs/ocfs2/
H A Dxattr.c324 if (bucket) { in ocfs2_xattr_bucket_new()
329 return bucket; in ocfs2_xattr_bucket_new()
344 if (bucket) { in ocfs2_xattr_bucket_free()
347 kfree(bucket); in ocfs2_xattr_bucket_free()
363 bucket->bu_bhs[i] = sb_getblk(bucket->bu_inode->i_sb, in ocfs2_init_xattr_bucket()
396 bucket->bu_blocks, bucket->bu_bhs, 0, in ocfs2_read_xattr_bucket()
440 bucket->bu_bhs, bucket->bu_blocks, in ocfs2_xattr_bucket_journal_dirty()
3973 if (!bucket) { in ocfs2_iterate_xattr_buckets()
5562 xs->bucket); in ocfs2_xattr_set_entry_bucket()
5806 bucket = xbs->bucket; in ocfs2_prepare_refcount_xattr()
[all …]
/linux/drivers/md/
H A Ddm-clone-target.c588 bucket = clone->ht + i; in hash_table_init()
883 bucket_lock_irq(bucket); in hydrate_bio_region()
889 bucket_unlock_irq(bucket); in hydrate_bio_region()
895 bucket_unlock_irq(bucket); in hydrate_bio_region()
904 bucket_unlock_irq(bucket); in hydrate_bio_region()
909 bucket_lock_irq(bucket); in hydrate_bio_region()
913 bucket_unlock_irq(bucket); in hydrate_bio_region()
923 bucket_unlock_irq(bucket); in hydrate_bio_region()
935 bucket_unlock_irq(bucket); in hydrate_bio_region()
949 bucket_unlock_irq(bucket); in hydrate_bio_region()
[all …]
/linux/drivers/md/bcache/
H A Dalloc.c76 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen()
89 struct bucket *b; in bch_rescale_priorities()
179 struct bucket *b; in invalidate_buckets_lru()
216 struct bucket *b; in invalidate_buckets_fifo()
239 struct bucket *b; in invalidate_buckets_random()
305 if (fifo_push(&ca->free[RESERVE_PRIO], bucket)) in bch_allocator_push()
309 if (fifo_push(&ca->free[i], bucket)) in bch_allocator_push()
328 long bucket; in bch_allocator_thread() local
330 if (!fifo_pop(&ca->free_inc, bucket)) in bch_allocator_thread()
336 bucket_to_sector(ca->set, bucket), in bch_allocator_thread()
[all …]
/linux/tools/testing/selftests/drivers/net/hw/
H A Dethtool_rmon.sh35 local bucket=$1; shift
50 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val")
58 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val")
78 while read -r -a bucket; do
82 if ! ensure_mtu $if ${bucket[0]}; then
88 if ! bucket_test $iface $neigh $set $nbuckets ${bucket[0]}; then
/linux/Documentation/networking/
H A Dnexthop-group-resilient.rst49 to choose a hash table bucket, then reads the next hop that this bucket
92 through a bucket, this timer is updated to current jiffies value. One
104 upkeep changes the next hop that the bucket references to one of the
135 - Single-bucket notifications of the type
145 hop associated with the bucket was removed, and the bucket really must be
150 bucket should be migrated, but the HW discovers that the bucket has in fact
153 A second way for the HW to report that a bucket is busy is through the
212 ``NHA_RES_BUCKET_INDEX`` Index of bucket in the resilient table.
255 Changing next-hop weights leads to change in bucket allocation::
261 # ip nexthop bucket show id 10
[all …]

123456789