/linux/drivers/md/dm-vdo/ |
H A D | int-map.c | 276 struct bucket *bucket = dereference_hop(neighborhood, next_hop); in insert_in_hop_list() local 325 struct bucket *bucket, in search_hop_list() argument 337 struct bucket *entry = dereference_hop(bucket, next_hop); in search_hop_list() 466 struct bucket *bucket; in move_empty_bucket() local 468 for (bucket = &hole[1 - NEIGHBORHOOD]; bucket < hole; bucket++) { in move_empty_bucket() 473 struct bucket *new_hole = dereference_hop(bucket, bucket->first_hop); in move_empty_bucket() 531 struct bucket *bucket = search_hop_list(map, neighborhood, key, NULL); in update_mapping() local 613 struct bucket *neighborhood, *bucket; in vdo_int_map_put() local 657 bucket->key = key; in vdo_int_map_put() 680 struct bucket *bucket = select_bucket(map, key); in vdo_int_map_remove() local [all …]
|
H A D | priority-table.c | 23 struct bucket { struct 44 struct bucket buckets[]; 69 struct bucket *bucket = &table->buckets[priority]; in vdo_make_priority_table() local 71 bucket->priority = priority; in vdo_make_priority_table() 72 INIT_LIST_HEAD(&bucket->queue); in vdo_make_priority_table() 140 static inline void mark_bucket_empty(struct priority_table *table, struct bucket *bucket) in mark_bucket_empty() argument 157 struct bucket *bucket; in vdo_priority_table_dequeue() local 173 bucket = &table->buckets[top_priority]; in vdo_priority_table_dequeue() 174 entry = bucket->queue.next; in vdo_priority_table_dequeue() 178 if (list_empty(&bucket->queue)) in vdo_priority_table_dequeue() [all …]
|
/linux/net/mptcp/ |
H A D | token.c | 111 struct token_bucket *bucket; in mptcp_token_new_request() local 122 bucket = token_bucket(token); in mptcp_token_new_request() 123 spin_lock_bh(&bucket->lock); in mptcp_token_new_request() 130 bucket->chain_len++; in mptcp_token_new_request() 157 struct token_bucket *bucket; in mptcp_token_new_connect() local 164 spin_lock_bh(&bucket->lock); in mptcp_token_new_connect() 177 bucket->chain_len++; in mptcp_token_new_connect() 196 struct token_bucket *bucket; in mptcp_token_accept() local 200 spin_lock_bh(&bucket->lock); in mptcp_token_accept() 364 bucket->chain_len--; in mptcp_token_destroy_request() [all …]
|
/linux/net/ceph/crush/ |
H A D | mapper.c | 88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose() 89 bucket->size; in bucket_perm_choose() 113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose() 114 (bucket->size - p); in bucket_perm_choose() 131 return bucket->items[s]; in bucket_perm_choose() 148 __u64 w = crush_hash32_4(bucket->h.hash, x, bucket->h.items[i], in bucket_list_choose() 149 r, bucket->h.id); in bucket_list_choose() 153 i, x, r, bucket->h.items[i], bucket->item_weights[i], in bucket_list_choose() 164 return bucket->h.items[0]; in bucket_list_choose() 237 draw = crush_hash32_3(bucket->h.hash, x, bucket->h.items[i], r); in bucket_straw_choose() [all …]
|
/linux/drivers/interconnect/qcom/ |
H A D | bcm-voter.c | 65 int bucket, i; in bcm_aggregate_mask() local 67 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate_mask() 68 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask() 69 bcm->vote_y[bucket] = 0; in bcm_aggregate_mask() 75 if (node->sum_avg[bucket] || node->max_peak[bucket]) { in bcm_aggregate_mask() 76 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask() 77 bcm->vote_y[bucket] = bcm->enable_mask; in bcm_aggregate_mask() 94 size_t i, bucket; in bcm_aggregate() local 99 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate() 104 agg_avg[bucket] = max(agg_avg[bucket], temp); in bcm_aggregate() [all …]
|
/linux/block/ |
H A D | blk-stat.c | 55 int bucket, cpu; in blk_stat_add() local 66 bucket = cb->bucket_fn(rq); in blk_stat_add() 67 if (bucket < 0) in blk_stat_add() 80 unsigned int bucket; in blk_stat_timer_fn() local 83 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn() 84 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn() 90 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn() 91 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn() 92 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn() 136 unsigned int bucket; in blk_stat_add_callback() local [all …]
|
/linux/net/sched/ |
H A D | sch_hhf.c | 333 bucket->head = skb->next; in dequeue_head() 341 if (bucket->head == NULL) in bucket_add() 342 bucket->head = skb; in bucket_add() 344 bucket->tail->next = skb; in bucket_add() 345 bucket->tail = skb; in bucket_add() 352 struct wdrr_bucket *bucket; in hhf_drop() local 356 if (!bucket->head) in hhf_drop() 359 if (bucket->head) { in hhf_drop() 381 bucket = &q->buckets[idx]; in hhf_enqueue() 382 bucket_add(bucket, skb); in hhf_enqueue() [all …]
|
/linux/fs/bcachefs/ |
H A D | backpointers.h | 52 *bucket = bp_pos_to_bucket(ca, bp_pos); in bp_pos_to_bucket_nodev_noerror() 59 return !bch2_fs_inconsistent_on(!bp_pos_to_bucket_nodev_noerror(c, bp_pos, bucket), in bp_pos_to_bucket_nodev() 64 struct bpos bucket, in bucket_pos_to_bp_noerror() 67 return POS(bucket.inode, in bucket_pos_to_bp_noerror() 68 (bucket_to_sector(ca, bucket.offset) << in bucket_pos_to_bp_noerror() 76 struct bpos bucket, in bucket_pos_to_bp() 79 struct bpos ret = bucket_pos_to_bp_noerror(ca, bucket, bucket_offset); in bucket_pos_to_bp() 80 EBUG_ON(!bkey_eq(bucket, bp_pos_to_bucket(ca, ret))); in bucket_pos_to_bp() 85 struct bpos bucket, struct bch_backpointer, struct bkey_s_c, bool); 89 struct bpos bucket, in bch2_bucket_backpointer_mod() argument [all …]
|
H A D | alloc_foreground.h | 121 unsigned dev, u64 bucket) in open_bucket_hashslot() argument 124 (jhash_3words(dev, bucket, bucket >> 32, 0) & in open_bucket_hashslot() 128 static inline bool bch2_bucket_is_open(struct bch_fs *c, unsigned dev, u64 bucket) in bch2_bucket_is_open() argument 130 open_bucket_idx_t slot = *open_bucket_hashslot(c, dev, bucket); in bch2_bucket_is_open() 135 if (ob->dev == dev && ob->bucket == bucket) in bch2_bucket_is_open() 144 static inline bool bch2_bucket_is_open_safe(struct bch_fs *c, unsigned dev, u64 bucket) in bch2_bucket_is_open_safe() argument 148 if (bch2_bucket_is_open(c, dev, bucket)) in bch2_bucket_is_open_safe() 152 ret = bch2_bucket_is_open(c, dev, bucket); in bch2_bucket_is_open_safe()
|
H A D | backpointers.c | 19 struct bpos bucket, in extent_matches_bp() argument 39 if (bpos_eq(bucket, bucket2) && in extent_matches_bp() 97 bch2_bpos_to_text(out, bucket); in bch2_backpointer_k_to_text() 167 struct bpos bucket, in bch2_bucket_backpointer_mod_nowritebuffer() argument 280 struct bpos bucket; in backpointer_not_found() local 287 bch2_bpos_to_text(&buf, bucket); in backpointer_not_found() 314 struct bpos bucket; in bch2_backpointer_get_key() local 355 struct bpos bucket; in bch2_backpointer_get_node() local 373 bucket, bp)) in bch2_backpointer_get_node() 396 struct bpos bucket; in bch2_check_btree_backpointer() local [all …]
|
H A D | alloc_background.h | 25 static inline u64 bucket_to_u64(struct bpos bucket) in bucket_to_u64() argument 27 return (bucket.inode << 48) | bucket.offset; in bucket_to_u64() 30 static inline struct bpos u64_to_bucket(u64 bucket) in u64_to_bucket() argument 32 return POS(bucket >> 48, bucket & ~(~0ULL << 48)); in u64_to_bucket() 40 static inline void alloc_to_bucket(struct bucket *dst, struct bch_alloc_v4 src) in alloc_to_bucket() 49 static inline void __bucket_m_to_alloc(struct bch_alloc_v4 *dst, struct bucket src) in __bucket_m_to_alloc() 58 static inline struct bch_alloc_v4 bucket_m_to_alloc(struct bucket b) in bucket_m_to_alloc() 76 static inline bool bucket_data_type_mismatch(enum bch_data_type bucket, in bucket_data_type_mismatch() argument 79 return !data_type_is_empty(bucket) && in bucket_data_type_mismatch() 80 bucket_data_type(bucket) != bucket_data_type(ptr); in bucket_data_type_mismatch()
|
H A D | nocow_locking.h | 28 struct bpos bucket, int flags) in bch2_bucket_nocow_lock() argument 30 u64 dev_bucket = bucket_to_u64(bucket); in bch2_bucket_nocow_lock() 37 struct bpos bucket, int flags) in bch2_bucket_nocow_trylock() argument 39 u64 dev_bucket = bucket_to_u64(bucket); in bch2_bucket_nocow_trylock()
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | trace_qp.h | 18 TP_PROTO(struct rvt_qp *qp, u32 bucket), 19 TP_ARGS(qp, bucket), 23 __field(u32, bucket) 28 __entry->bucket = bucket; 34 __entry->bucket 39 TP_PROTO(struct rvt_qp *qp, u32 bucket), 40 TP_ARGS(qp, bucket)); 43 TP_PROTO(struct rvt_qp *qp, u32 bucket), 44 TP_ARGS(qp, bucket));
|
/linux/fs/nfs/ |
H A D | nfs42xattr.c | 238 entry->bucket = NULL; in nfs4_xattr_alloc_entry() 396 spin_lock(&bucket->lock); in nfs4_xattr_discard_cache() 397 bucket->draining = true; in nfs4_xattr_discard_cache() 534 entry->bucket = bucket; in nfs4_xattr_hash_add() 536 spin_lock(&bucket->lock); in nfs4_xattr_hash_add() 538 if (bucket->draining) { in nfs4_xattr_hash_add() 555 spin_unlock(&bucket->lock); in nfs4_xattr_hash_add() 571 spin_lock(&bucket->lock); in nfs4_xattr_hash_remove() 594 spin_lock(&bucket->lock); in nfs4_xattr_hash_find() 878 bucket = entry->bucket; in entry_lru_isolate() [all …]
|
H A D | pnfs_nfs.c | 65 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { in pnfs_free_bucket_lseg() 67 bucket->lseg = NULL; in pnfs_free_bucket_lseg() 91 if (bucket) in pnfs_generic_clear_request_commit() 414 lseg = pnfs_get_lseg(bucket->lseg); in pnfs_bucket_get_committing() 435 struct pnfs_commit_bucket *bucket; in pnfs_generic_retry_commit() local 439 for (bucket = buckets; idx < nbuckets; bucket++, idx++) { in pnfs_generic_retry_commit() 456 struct pnfs_commit_bucket *bucket; in pnfs_bucket_alloc_ds_commits() local 461 for (i = 0, bucket = buckets; i < nbuckets; i++, bucket++) { in pnfs_bucket_alloc_ds_commits() 1207 struct pnfs_commit_bucket *bucket; in pnfs_layout_mark_request_commit() local 1214 list = &bucket->written; in pnfs_layout_mark_request_commit() [all …]
|
/linux/net/9p/ |
H A D | error.c | 179 int bucket; in p9_error_init() local 182 for (bucket = 0; bucket < ERRHASHSZ; bucket++) in p9_error_init() 183 INIT_HLIST_HEAD(&hash_errmap[bucket]); in p9_error_init() 188 bucket = jhash(c->name, c->namelen, 0) % ERRHASHSZ; in p9_error_init() 190 hlist_add_head(&c->list, &hash_errmap[bucket]); in p9_error_init() 208 int bucket; in p9_errstr2errno() local 212 bucket = jhash(errstr, len, 0) % ERRHASHSZ; in p9_errstr2errno() 213 hlist_for_each_entry(c, &hash_errmap[bucket], list) { in p9_errstr2errno()
|
/linux/net/vmw_vsock/ |
H A D | diag.c | 52 unsigned int bucket; in vsock_diag_dump() local 63 bucket = cb->args[1]; in vsock_diag_dump() 72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump() 73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump() 94 bucket++; in vsock_diag_dump() 98 bucket = 0; in vsock_diag_dump() 102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump() 103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump() 128 bucket++; in vsock_diag_dump() 135 cb->args[1] = bucket; in vsock_diag_dump()
|
/linux/drivers/cpuidle/governors/ |
H A D | menu.c | 116 unsigned int bucket; member 124 int bucket = 0; in which_bucket() local 133 bucket = BUCKETS/2; in which_bucket() 136 return bucket; in which_bucket() 138 return bucket + 1; in which_bucket() 140 return bucket + 2; in which_bucket() 142 return bucket + 3; in which_bucket() 144 return bucket + 4; in which_bucket() 145 return bucket + 5; in which_bucket() 298 data->correction_factor[data->bucket], in menu_select() [all …]
|
/linux/kernel/dma/ |
H A D | debug.c | 267 __releases(&bucket->lock) in put_hash_bucket() 590 struct hash_bucket *bucket; in add_dma_entry() local 595 hash_bucket_add(bucket, entry); in add_dma_entry() 596 put_hash_bucket(bucket, flags); in add_dma_entry() 964 struct hash_bucket *bucket; in check_unmap() local 1054 put_hash_bucket(bucket, flags); in check_unmap() 1097 struct hash_bucket *bucket; in check_sync() local 1164 put_hash_bucket(bucket, flags); in check_sync() 1249 struct hash_bucket *bucket; in debug_dma_mapping_error() local 1279 put_hash_bucket(bucket, flags); in debug_dma_mapping_error() [all …]
|
/linux/Documentation/userspace-api/media/v4l/ |
H A D | metafmt-vsp1-hgt.rst | 28 The Saturation position **n** (0 - 31) of the bucket in the matrix is 33 The Hue position **m** (0 - 5) of the bucket in the matrix depends on 101 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0] 103 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0] 107 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0] 109 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0] 113 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0] 117 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0] 121 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0] 125 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0] [all …]
|
/linux/fs/ocfs2/ |
H A D | xattr.c | 324 if (bucket) { in ocfs2_xattr_bucket_new() 329 return bucket; in ocfs2_xattr_bucket_new() 344 if (bucket) { in ocfs2_xattr_bucket_free() 347 kfree(bucket); in ocfs2_xattr_bucket_free() 363 bucket->bu_bhs[i] = sb_getblk(bucket->bu_inode->i_sb, in ocfs2_init_xattr_bucket() 396 bucket->bu_blocks, bucket->bu_bhs, 0, in ocfs2_read_xattr_bucket() 440 bucket->bu_bhs, bucket->bu_blocks, in ocfs2_xattr_bucket_journal_dirty() 3973 if (!bucket) { in ocfs2_iterate_xattr_buckets() 5562 xs->bucket); in ocfs2_xattr_set_entry_bucket() 5806 bucket = xbs->bucket; in ocfs2_prepare_refcount_xattr() [all …]
|
/linux/drivers/md/ |
H A D | dm-clone-target.c | 588 bucket = clone->ht + i; in hash_table_init() 883 bucket_lock_irq(bucket); in hydrate_bio_region() 889 bucket_unlock_irq(bucket); in hydrate_bio_region() 895 bucket_unlock_irq(bucket); in hydrate_bio_region() 904 bucket_unlock_irq(bucket); in hydrate_bio_region() 909 bucket_lock_irq(bucket); in hydrate_bio_region() 913 bucket_unlock_irq(bucket); in hydrate_bio_region() 923 bucket_unlock_irq(bucket); in hydrate_bio_region() 935 bucket_unlock_irq(bucket); in hydrate_bio_region() 949 bucket_unlock_irq(bucket); in hydrate_bio_region() [all …]
|
/linux/drivers/md/bcache/ |
H A D | alloc.c | 76 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen() 89 struct bucket *b; in bch_rescale_priorities() 179 struct bucket *b; in invalidate_buckets_lru() 216 struct bucket *b; in invalidate_buckets_fifo() 239 struct bucket *b; in invalidate_buckets_random() 305 if (fifo_push(&ca->free[RESERVE_PRIO], bucket)) in bch_allocator_push() 309 if (fifo_push(&ca->free[i], bucket)) in bch_allocator_push() 328 long bucket; in bch_allocator_thread() local 330 if (!fifo_pop(&ca->free_inc, bucket)) in bch_allocator_thread() 336 bucket_to_sector(ca->set, bucket), in bch_allocator_thread() [all …]
|
/linux/tools/testing/selftests/drivers/net/hw/ |
H A D | ethtool_rmon.sh | 35 local bucket=$1; shift 50 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val") 58 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val") 78 while read -r -a bucket; do 82 if ! ensure_mtu $if ${bucket[0]}; then 88 if ! bucket_test $iface $neigh $set $nbuckets ${bucket[0]}; then
|
/linux/Documentation/networking/ |
H A D | nexthop-group-resilient.rst | 49 to choose a hash table bucket, then reads the next hop that this bucket 92 through a bucket, this timer is updated to current jiffies value. One 104 upkeep changes the next hop that the bucket references to one of the 135 - Single-bucket notifications of the type 145 hop associated with the bucket was removed, and the bucket really must be 150 bucket should be migrated, but the HW discovers that the bucket has in fact 153 A second way for the HW to report that a bucket is busy is through the 212 ``NHA_RES_BUCKET_INDEX`` Index of bucket in the resilient table. 255 Changing next-hop weights leads to change in bucket allocation:: 261 # ip nexthop bucket show id 10 [all …]
|