/freebsd/sys/dev/mlx5/mlx5_en/ |
H A D | mlx5_en_tx.c | 39 if (sq->cev_counter >= sq->cev_factor) { in mlx5e_do_send_cqe_inline() 56 u16 pi = sq->pc & sq->wq.sz_m1; in mlx5e_send_nop() 74 sq->pc += sq->mbuf[pi].num_wqebbs; in mlx5e_send_nop() 129 if (sq != NULL && READ_ONCE(sq->running) != 0) in mlx5e_select_queue_by_send_tag() 171 sq = &priv->channel[ch].sq[tc]; in mlx5e_select_queue() 567 pi = sq->pc & sq->wq.sz_m1; in mlx5e_sq_dump_xmit() 653 wqe_last = mlx5_wq_cyc_get_wqe(&sq->wq, (sq->pc - 1) & sq->wq.sz_m1); in mlx5e_sq_dump_xmit() 709 pi = ((~sq->pc) & sq->wq.sz_m1); in mlx5e_sq_xmit() 713 pi = ((~sq->pc) & sq->wq.sz_m1); in mlx5e_sq_xmit() 738 pi = sq->pc & sq->wq.sz_m1; in mlx5e_sq_xmit() [all …]
|
H A D | mlx5_en_rl.c | 137 sq->wq.db = &sq->wq.db[MLX5_SND_DBR]; in mlx5e_rl_create_sq() 233 callout_init_mtx(&sq->cev_callout, &sq->lock, 0); in mlx5e_rl_chan_mtx_init() 251 sq = malloc(sizeof(*sq), M_MLX5EN, M_WAITOK | M_ZERO); in mlx5e_rl_open_channel() 262 err = mlx5e_rl_open_sq(priv, sq, &cparam->sq, eq_ix); in mlx5e_rl_open_channel() 267 *ppsq = sq; in mlx5e_rl_open_channel() 270 sq->cq.mcq.comp(&sq->cq.mcq, NULL); in mlx5e_rl_open_channel() 477 iq_channel = &sq->priv->channel[sq->sqn % sq->priv->params.num_channels]; in mlx5e_rl_remap_sq() 573 sq = channel->sq; in mlx5e_rlw_channel_set_rate_locked() 574 if (sq != NULL && READ_ONCE(sq->running) != 0) { in mlx5e_rlw_channel_set_rate_locked() 1329 sq = channel->sq; in mlx5e_rl_refresh_channel_params() [all …]
|
H A D | mlx5_en_hw_tls.c | 504 pi = sq->pc & sq->wq.sz_m1; in mlx5e_tls_send_static_parameters() 528 memcpy(sq->doorbell.d32, &wqe->ctrl, sizeof(sq->doorbell.d32)); in mlx5e_tls_send_static_parameters() 530 sq->mbuf[pi].mbuf = NULL; in mlx5e_tls_send_static_parameters() 535 sq->pc += sq->mbuf[pi].num_wqebbs; in mlx5e_tls_send_static_parameters() 549 pi = sq->pc & sq->wq.sz_m1; in mlx5e_tls_send_progress_parameters() 566 memcpy(sq->doorbell.d32, &wqe->ctrl, sizeof(sq->doorbell.d32)); in mlx5e_tls_send_progress_parameters() 568 sq->mbuf[pi].mbuf = NULL; in mlx5e_tls_send_progress_parameters() 573 sq->pc += sq->mbuf[pi].num_wqebbs; in mlx5e_tls_send_progress_parameters() 583 pi = sq->pc & sq->wq.sz_m1; in mlx5e_tls_send_nop() 596 memcpy(sq->doorbell.d32, &wqe->ctrl, sizeof(sq->doorbell.d32)); in mlx5e_tls_send_nop() [all …]
|
H A D | mlx5_en_main.c | 879 struct mlx5e_sq *sq = channel->sq; in mlx5e_update_stats_locked() local 1610 sq->max_inline = sq->priv->params.tx_max_inline; in mlx5e_update_sq_inline() 1692 sq->wq.db = &sq->wq.db[MLX5_SND_DBR]; in mlx5e_create_sq() 1806 mlx5_core_destroy_sq(sq->priv->mdev, sq->sqn); in mlx5e_disable_sq() 1932 while (sq->cc != sq->pc && in mlx5e_drain_sq() 1938 sq->cq.mcq.comp(&sq->cq.mcq, NULL); in mlx5e_drain_sq() 1952 while (sq->cc != sq->pc && in mlx5e_drain_sq() 1957 sq->cq.mcq.comp(&sq->cq.mcq, NULL); in mlx5e_drain_sq() 2187 struct mlx5e_sq *sq = c->sq + tc; in mlx5e_chan_static_init() local 4012 sq->doorbell.d32[1] = cpu_to_be32(sq->sqn << 8); in mlx5e_reset_sq_doorbell_record() [all …]
|
/freebsd/lib/libthr/thread/ |
H A D | thr_sleepq.c | 63 struct sleepqueue *sq; in _sleepq_alloc() local 68 return (sq); in _sleepq_alloc() 74 free(sq); in _sleepq_free() 100 struct sleepqueue *sq; in lookup() local 104 return (sq); in lookup() 118 struct sleepqueue *sq; in _sleepq_add() local 122 if (sq != NULL) { in _sleepq_add() 125 sq = td->sleepqueue; in _sleepq_add() 127 sq->sq_wchan = wchan; in _sleepq_add() 146 td->sleepqueue = sq; in _sleepq_remove() [all …]
|
/freebsd/contrib/unbound/services/ |
H A D | outside_network.c | 2254 pend->sq = sq; 2530 w->sq = sq; 2617 sq->node.key = sq; 2897 if(!infra_host(sq->outnet->infra, &sq->addr, sq->addrlen, sq->zone, 3082 sq->addrlen, sq->zone, sq->zonelen); 3118 sq->addrlen, sq->zone, sq->zonelen, -1, 3127 sq->addrlen, sq->zone, sq->zonelen, sq->qtype, 3142 sq->addrlen, sq->zone, sq->zonelen, sq->qtype, 3186 if(!infra_host(sq->outnet->infra, &sq->addr, sq->addrlen, sq->zone, 3279 sq->zone, sq->zonelen, sq->qtype, -1, sq->last_rtt, [all …]
|
/freebsd/contrib/unbound/edns-subnet/ |
H A D | subnetmod.c | 95 if(!sq) in subnet_new_qstate() 98 memset(sq, 0, sizeof(*sq)); in subnet_new_qstate() 163 if(sq->ecs_server_out.subnet_validdata && ((sq->subnet_downstream && in ecs_whitelist_check() 463 memset(&sq->ecs_client_out, 0, sizeof(sq->ecs_client_out)); in lookup_and_reply() 465 if (sq) { in lookup_and_reply() 545 if (!sq->subnet_sent && !sq->subnet_sent_no_subnet) { in eval_response() 734 memset(&sq->ecs_server_out, 0, sizeof(sq->ecs_server_out)); in ecs_query_response() 783 sq->subnet_sent && sq->ecs_server_in.subnet_validdata) { in ecs_edns_back_parsed() 791 if(!sq->track_max_scope || (sq->track_max_scope && in ecs_edns_back_parsed() 794 sq->max_scope = sq->ecs_server_in.subnet_scope_mask; in ecs_edns_back_parsed() [all …]
|
/freebsd/sys/kern/ |
H A D | subr_sleepqueue.c | 284 return (sq); in sleepq_lookup() 339 if (sq == NULL) { in sleepq_add() 430 if (sq == NULL) in sleepq_sleepcnt() 726 if (sq == NULL) in sleepq_type() 878 MPASS(sq != NULL); in sleepq_remove_nested() 895 sq = mem; in sleepq_dtor() 913 sq = mem; in sleepq_init() 938 if (sq == NULL) { in sleepq_signal() 999 if (sq == NULL) in sleepq_broadcast() 1115 MPASS(sq != NULL); in sleepq_remove() [all …]
|
/freebsd/sys/dev/vnic/ |
H A D | nicvf_queues.c | 703 sq = &nic->qs->sq[cqe_tx->sq_idx]; in nicvf_snd_pkt_handler() 1074 sq->desc = sq->dmem.base; in nicvf_init_snd_queue() 1075 sq->head = sq->tail = 0; in nicvf_init_snd_queue() 1414 sq = &qs->sq[qidx]; in nicvf_snd_queue_config() 1433 mbx.sq.cfg = (sq->cq_qs << 3) | sq->cq_idx; in nicvf_snd_queue_config() 1668 sq->tail &= (sq->dmem.q_len - 1); in nicvf_get_sq_desc() 1680 sq->head &= (sq->dmem.q_len - 1); in nicvf_put_sq_desc() 1728 snd_buff = &sq->snd_buff[sq->head]; in nicvf_sq_free_used_descs() 1732 sq->snd_buff[sq->head].mbuf = NULL; in nicvf_sq_free_used_descs() 1914 snd_buff = &sq->snd_buff[sq->tail]; in nicvf_tx_mbuf_locked() [all …]
|
/freebsd/contrib/ofed/libcxgb4/ |
H A D | cq.c | 105 wq->sq.flush_cidx = wq->sq.cidx; in c4iw_flush_sq() 106 idx = wq->sq.flush_cidx; in c4iw_flush_sq() 122 if (wq->sq.flush_cidx >= wq->sq.size) in c4iw_flush_sq() 123 wq->sq.flush_cidx -= wq->sq.size; in c4iw_flush_sq() 132 wq->sq.flush_cidx = wq->sq.cidx; in flush_completed_wrs() 178 u32 rptr = wq->sq.oldest_read - wq->sq.sw_sq + 1; in advance_oldest_read() 183 wq->sq.oldest_read = &wq->sq.sw_sq[rptr]; in advance_oldest_read() 514 wq->sq.in_use -= wq->sq.size + idx - wq->sq.cidx; in poll_cq() 516 wq->sq.in_use -= idx - wq->sq.cidx; in poll_cq() 517 BUG_ON(wq->sq.in_use <= 0 || wq->sq.in_use >= wq->sq.size); in poll_cq() [all …]
|
H A D | verbs.c | 356 qhp->wq.sq.queue = mmap(NULL, qhp->wq.sq.memsize, in create_qp_v0() 384 qhp->wq.sq.udb, qhp->wq.sq.queue, in create_qp_v0() 385 qhp->wq.sq.size, qhp->wq.sq.memsize, in create_qp_v0() 403 munmap((void *)qhp->wq.sq.queue, qhp->wq.sq.memsize); in create_qp_v0() 482 qhp->wq.sq.queue = mmap(NULL, qhp->wq.sq.memsize, in create_qp() 537 qhp->wq.sq.udb, qhp->wq.sq.queue, in create_qp() 538 qhp->wq.sq.size, qhp->wq.sq.memsize, in create_qp() 558 munmap((void *)qhp->wq.sq.queue, qhp->wq.sq.memsize); in create_qp() 585 qhp->wq.sq.wq_pidx = qhp->wq.sq.pidx = qhp->wq.sq.in_use = 0; in reset_qp() 588 memset(qhp->wq.sq.queue, 0, qhp->wq.sq.memsize); in reset_qp() [all …]
|
H A D | t4.h | 413 return wq->sq.in_use == (wq->sq.size - 1); in t4_sq_full() 418 return wq->sq.size - 1 - wq->sq.in_use; in t4_sq_avail() 429 if (++wq->sq.pidx == wq->sq.size) in t4_sq_produce() 432 if (wq->sq.wq_pidx >= wq->sq.size * T4_SQ_NUM_SLOTS) in t4_sq_produce() 433 wq->sq.wq_pidx %= wq->sq.size * T4_SQ_NUM_SLOTS; in t4_sq_produce() 437 wq->sq.queue[wq->sq.size].status.host_pidx = (wq->sq.pidx); in t4_sq_produce() 444 if (wq->sq.cidx == wq->sq.flush_cidx) in t4_sq_consume() 447 if (++wq->sq.cidx == wq->sq.size) in t4_sq_consume() 449 assert((wq->sq.cidx != wq->sq.pidx) || wq->sq.in_use == 0); in t4_sq_consume() 453 wq->sq.queue[wq->sq.size].status.host_cidx = wq->sq.cidx; in t4_sq_consume() [all …]
|
H A D | qp.c | 52 dst = &wq->sq.queue->flits[wq->sq.wq_pidx * in copy_wr_to_sq() 67 end = (uintptr_t)&wq->sq.queue[wq->sq.size]; in copy_wr_to_sq() 74 memcpy(wq->sq.queue, src + len, total - len); in copy_wr_to_sq() 298 if (qid == qhp->wq.sq.qid) { in ring_kernel_db() 348 swsqe = &qhp->wq.sq.sw_sq[qhp->wq.sq.pidx]; in c4iw_post_send() 374 if (!qhp->wq.sq.oldest_read) in c4iw_post_send() 375 qhp->wq.sq.oldest_read = swsqe; in c4iw_post_send() 386 swsqe->idx = qhp->wq.sq.pidx; in c4iw_post_send() 407 ring_kernel_db(qhp, qhp->wq.sq.qid, idx); in c4iw_post_send() 410 qhp->wq.sq.queue[qhp->wq.sq.size].status.host_wq_pidx = \ in c4iw_post_send() [all …]
|
/freebsd/sys/contrib/ck/include/ |
H A D | ck_sequence.h | 42 ck_sequence_init(struct ck_sequence *sq) in ck_sequence_init() argument 45 ck_pr_store_uint(&sq->sequence, 0); in ck_sequence_init() 50 ck_sequence_read_begin(const struct ck_sequence *sq) in ck_sequence_read_begin() argument 55 version = ck_pr_load_uint(&sq->sequence); in ck_sequence_read_begin() 77 ck_sequence_read_retry(const struct ck_sequence *sq, unsigned int version) in ck_sequence_read_retry() argument 85 return ck_pr_load_uint(&sq->sequence) != version; in ck_sequence_read_retry() 97 ck_sequence_write_begin(struct ck_sequence *sq) in ck_sequence_write_begin() argument 104 ck_pr_store_uint(&sq->sequence, sq->sequence + 1); in ck_sequence_write_begin() 113 ck_sequence_write_end(struct ck_sequence *sq) in ck_sequence_write_end() argument 121 ck_pr_store_uint(&sq->sequence, sq->sequence + 1); in ck_sequence_write_end()
|
/freebsd/sys/dev/ice/ |
H A D | ice_controlq.c | 94 if (cq->sq.len && cq->sq.len_mask && cq->sq.len_ena_mask) in ice_check_sq_alive() 95 return (rd32(hw, cq->sq.len) & (cq->sq.len_mask | in ice_check_sq_alive() 112 cq->sq.desc_buf.va = ice_alloc_dma_mem(hw, &cq->sq.desc_buf, size); in ice_alloc_ctrlq_sq_ring() 227 cq->sq.r.sq_bi = (struct ice_dma_mem *)cq->sq.dma_head; in ice_alloc_sq_bufs() 826 struct ice_ctl_q_ring *sq = &cq->sq; in ice_clean_sq() local 931 return rd32(hw, cq->sq.head) == cq->sq.next_to_use; in ice_sq_done() 1010 desc_on_ring = ICE_CTL_Q_DESC(cq->sq, cq->sq.next_to_use); in ice_sq_send_cmd_nolock() 1018 dma_buf = &cq->sq.r.sq_bi[cq->sq.next_to_use]; in ice_sq_send_cmd_nolock() 1037 if (cq->sq.next_to_use == cq->sq.count) in ice_sq_send_cmd_nolock() 1039 wr32(hw, cq->sq.tail, cq->sq.next_to_use); in ice_sq_send_cmd_nolock() [all …]
|
/freebsd/contrib/bearssl/src/symcipher/ |
H A D | aes_ct_cbcdec.c | 53 uint32_t q[8], sq[8]; in br_aes_ct_cbcdec_run() local 70 memcpy(sq, q, sizeof q); in br_aes_ct_cbcdec_run() 79 iv0 = sq[0]; in br_aes_ct_cbcdec_run() 80 iv1 = sq[2]; in br_aes_ct_cbcdec_run() 81 iv2 = sq[4]; in br_aes_ct_cbcdec_run() 82 iv3 = sq[6]; in br_aes_ct_cbcdec_run() 85 br_enc32le(buf + 16, q[1] ^ sq[0]); in br_aes_ct_cbcdec_run() 89 iv0 = sq[1]; in br_aes_ct_cbcdec_run() 90 iv1 = sq[3]; in br_aes_ct_cbcdec_run() 91 iv2 = sq[5]; in br_aes_ct_cbcdec_run() [all …]
|
/freebsd/usr.bin/fortune/tools/ |
H A D | Troff.mac | 24 \(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq\|\(sq
|
/freebsd/sys/dev/cxgbe/iw_cxgbe/ |
H A D | cq.c | 265 wq->sq.flush_cidx = wq->sq.cidx; in c4iw_flush_sq() 266 idx = wq->sq.flush_cidx; in c4iw_flush_sq() 282 if (wq->sq.flush_cidx >= wq->sq.size) in c4iw_flush_sq() 283 wq->sq.flush_cidx -= wq->sq.size; in c4iw_flush_sq() 293 wq->sq.flush_cidx = wq->sq.cidx; in flush_completed_wrs() 339 u32 rptr = wq->sq.oldest_read - wq->sq.sw_sq + 1; in advance_oldest_read() 344 wq->sq.oldest_read = &wq->sq.sw_sq[rptr]; in advance_oldest_read() 666 wq->sq.in_use -= wq->sq.size + idx - wq->sq.cidx; in poll_cq() 668 wq->sq.in_use -= idx - wq->sq.cidx; in poll_cq() 669 BUG_ON(wq->sq.in_use <= 0 && wq->sq.in_use >= wq->sq.size); in poll_cq() [all …]
|
H A D | t4.h | 361 struct t4_sq sq; member 427 return wq->sq.in_use == (wq->sq.size - 1); in t4_sq_full() 432 return wq->sq.size - 1 - wq->sq.in_use; in t4_sq_avail() 437 wq->sq.in_use++; in t4_sq_produce() 438 if (++wq->sq.pidx == wq->sq.size) in t4_sq_produce() 441 if (wq->sq.wq_pidx >= wq->sq.size * T4_SQ_NUM_SLOTS) in t4_sq_produce() 442 wq->sq.wq_pidx %= wq->sq.size * T4_SQ_NUM_SLOTS; in t4_sq_produce() 448 if (wq->sq.cidx == wq->sq.flush_cidx) in t4_sq_consume() 450 wq->sq.in_use--; in t4_sq_consume() 451 if (++wq->sq.cidx == wq->sq.size) in t4_sq_consume() [all …]
|
H A D | qp.c | 115 wq->sq.memsize, wq->sq.queue, in destroy_qp() 152 wq->sq.sw_sq = kzalloc(wq->sq.size * sizeof *wq->sq.sw_sq, in create_qp() 184 wq->sq.phys_addr = vtophys(wq->sq.queue); in create_qp() 186 memset(wq->sq.queue, 0, wq->sq.memsize); in create_qp() 320 wq->sq.memsize, wq->sq.queue, in create_qp() 351 if (dstp == (u8 *)&sq->queue[sq->size]) in build_immd() 356 len = (u8 *)&sq->queue[sq->size] - dstp; in build_immd() 451 (__be64 *)&sq->queue[sq->size], in build_rdma_send() 494 (__be64 *)&sq->queue[sq->size], in build_rdma_write() 754 if (++p == (__be64 *)&sq->queue[sq->size]) in build_memreg() [all …]
|
/freebsd/contrib/ofed/libmlx4/ |
H A D | qp.c | 66 return qp->buf.buf + qp->sq.offset + (n << qp->sq.wqe_shift); in get_send_wqe() 86 qp->sq.head = 0; in mlx4_init_qp_indices() 87 qp->sq.tail = 0; in mlx4_init_qp_indices() 231 ind = qp->sq.head; in mlx4_post_send() 253 qp->sq.wrid[ind & (qp->sq.wqe_cnt - 1)] = wr->wr_id; in mlx4_post_send() 475 ++qp->sq.head; in mlx4_post_send() 648 for (qp->sq.wqe_shift = 6; 1 << qp->sq.wqe_shift < size; in mlx4_calc_sq_wqe_size() 659 qp->sq.wrid = malloc(qp->sq.wqe_cnt * sizeof (uint64_t)); in mlx4_alloc_qp_buf() 678 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in mlx4_alloc_qp_buf() 683 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in mlx4_alloc_qp_buf() [all …]
|
/freebsd/bin/sh/tests/parser/ |
H A D | heredoc2.0 | 11 s='ast*que?non' sq=\' dq=\" 19 ${s+'$sq'x'$sq'} 21 )" = ${sq}x${sq}' 34 ${s##'$sq'ast'$sq'}
|
/freebsd/sys/dev/bnxt/bnxt_re/ |
H A D | qplib_fp.c | 911 struct bnxt_qplib_q *sq = &qp->sq; in bnxt_qplib_create_qp1() local 1001 sq->dbinfo.hwq = &sq->hwq; in bnxt_qplib_create_qp1() 1046 sq = &qp->sq; in bnxt_qplib_init_psn_ptr() 1064 struct bnxt_qplib_q *sq = &qp->sq; in bnxt_qplib_create_qp() local 1247 sq->dbinfo.hwq = &sq->hwq; in bnxt_qplib_create_qp() 1681 struct bnxt_qplib_q *sq = &qp->sq; in bnxt_qplib_get_qp1_sq_buf() local 1923 struct bnxt_qplib_q *sq = &qp->sq; in bnxt_qplib_post_send_db() local 1933 struct bnxt_qplib_q *sq = &qp->sq; in bnxt_qplib_post_send() local 2881 sq = &qp->sq; in bnxt_qplib_cq_process_req() 2900 swq = &sq->swq[sq->swq_last]; in bnxt_qplib_cq_process_req() [all …]
|
/freebsd/sys/dev/mthca/ |
H A D | mthca_qp.c | 829 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in __mthca_modify_qp() 1032 for (qp->sq.wqe_shift = 6; 1 << qp->sq.wqe_shift < size; in mthca_alloc_wqe_buf() 1048 (qp->sq.max << qp->sq.wqe_shift)); in mthca_alloc_wqe_buf() 1071 (qp->sq.max << qp->sq.wqe_shift)), in mthca_free_wqe_buf() 1228 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in mthca_alloc_qp_common() 1642 qp->sq.head, qp->sq.tail, in mthca_tavor_post_send() 1792 mthca_write64(((qp->sq.next_ind << qp->sq.wqe_shift) + in mthca_tavor_post_send() 1951 ind = qp->sq.head & (qp->sq.max - 1); in mthca_arbel_post_send() 1967 *qp->sq.db = cpu_to_be32(qp->sq.head & 0xffff); in mthca_arbel_post_send() 1983 qp->sq.head, qp->sq.tail, in mthca_arbel_post_send() [all …]
|
/freebsd/sys/net80211/ |
H A D | ieee80211_superg.c | 643 if (sq->depth == 0) { in ieee80211_ff_age() 650 head = sq->head; in ieee80211_ff_age() 660 sq->depth--; in ieee80211_ff_age() 663 sq->tail = NULL; in ieee80211_ff_age() 680 if (sq->tail != NULL) { in stageq_add() 684 sq->head = m; in stageq_add() 692 sq->tail = m; in stageq_add() 693 sq->depth++; in stageq_add() 710 if (sq->tail == m) in stageq_remove() 711 sq->tail = mprev; in stageq_remove() [all …]
|