/freebsd/sys/dev/sfxge/ |
H A D | sfxge_tx.c | 155 &txq->stmp[txq->ptr_mask])) in sfxge_next_stmp() 326 txq->reaped = txq->completed; in sfxge_tx_qreap() 348 txq->reaped, &txq->added); in sfxge_tx_qlist_post() 923 txq = sc->txq[index]; in sfxge_if_transmit() 1477 txq = sc->txq[index]; in sfxge_tx_qstop() 1523 txq->pending = txq->added; in sfxge_tx_qstop() 1623 txq = sc->txq[index]; in sfxge_tx_qstart() 1678 txq->added = txq->pending = txq->completed = txq->reaped = desc_index; in sfxge_tx_qstart() 1778 txq = sc->txq[index]; in sfxge_tx_qfini() 1823 sc->txq[txq_index] = txq; in sfxge_tx_qinit() [all …]
|
H A D | sfxge_ev.c | 54 struct sfxge_txq *txq; in sfxge_ev_qcomplete() local 62 if ((txq = evq->txq) != NULL) { in sfxge_ev_qcomplete() 63 evq->txq = NULL; in sfxge_ev_qcomplete() 69 next = txq->next; in sfxge_ev_qcomplete() 70 txq->next = NULL; in sfxge_ev_qcomplete() 75 if (txq->pending != txq->completed) in sfxge_ev_qcomplete() 78 txq = next; in sfxge_ev_qcomplete() 300 id = txq->pending & txq->ptr_mask; in sfxge_ev_tx() 309 *(evq->txqs) = txq; in sfxge_ev_tx() 313 if (txq->pending - txq->completed >= SFXGE_TX_BATCH) in sfxge_ev_tx() [all …]
|
/freebsd/sys/contrib/dev/iwlwifi/queue/ |
H A D | tx.c | 796 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_txq_gen2_tx() 856 txq->tfds, txq->dma_addr); in iwl_txq_gen2_free_memory() 886 txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_free() 997 txq->read_ptr, txq->write_ptr); in iwl_txq_log_scd_error() 1006 txq->need_update, txq->frozen, txq->ampdu, in iwl_txq_log_scd_error() 1023 txq->read_ptr, txq->write_ptr, in iwl_txq_log_scd_error() 1201 trans->txqs.txq[qid] = txq; in iwl_txq_alloc_response() 1576 if (txq->read_ptr == txq->write_ptr) in iwl_txq_progress() 1622 txq->write_ptr, txq->read_ptr); in iwl_txq_reclaim() 1715 txq->read_ptr = txq->write_ptr; in iwl_txq_set_q_ptrs() [all …]
|
H A D | tx.h | 16 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) in iwl_txq_get_first_tb_dma() argument 18 return txq->first_tb_dma + in iwl_txq_get_first_tb_dma() 30 struct iwl_txq *txq) in iwl_wake_queue() argument 39 struct iwl_txq *txq, int idx) in iwl_txq_get_tfd() argument 42 idx = iwl_txq_get_cmd_index(txq, idx); in iwl_txq_get_tfd() 64 iwl_op_mode_queue_full(trans->op_mode, txq->id); in iwl_txq_stop() 68 txq->id); in iwl_txq_stop() 169 struct iwl_txq *txq, int index); 171 struct iwl_txq *txq); 173 struct iwl_txq *txq, u16 byte_cnt, [all …]
|
/freebsd/sys/dev/ath/ |
H A D | if_ath_tx_edma.c | 148 aq.txq = htobe32(txq); in ath_tx_alq_edma_push() 184 txq->axq_qnum, in ath_tx_edma_push_staging_list() 280 __func__, sqdepth, txq->fifo.axq_depth, txq->axq_fifo_depth); in ath_tx_edma_push_staging_list() 309 txq->axq_qnum, in ath_edma_tx_fifo_fill() 312 txq->axq_depth, in ath_edma_tx_fifo_fill() 378 else if (txq->axq_aggr_depth > 0 && txq->axq_fifo_depth < 2) in ath_edma_tx_fifo_fill() 411 txq->axq_qnum); in ath_edma_dma_restart() 500 txq->axq_qnum, in ath_edma_dma_restart() 532 ATH_TXQ_LOCK(txq); in ath_edma_xmit_handoff_hw() 573 ATH_TXQ_LOCK(txq); in ath_edma_xmit_handoff_mcast() [all …]
|
H A D | if_ath.c | 3990 txq->axq_ac = 0; in ath_txq_init() 4078 if (txq != NULL) { in ath_tx_setup() 4079 txq->axq_ac = ac; in ath_tx_setup() 4447 txq->axq_link); in ath_tx_processq() 4451 txq->axq_qnum, in ath_tx_processq() 4453 txq->axq_link, in ath_tx_processq() 4985 if (txq->axq_fifo_depth != 0 || txq->fifo.axq_depth != 0) { in ath_tx_draintxq_get_one() 5069 ATH_TXQ_LOCK(txq); in ath_tx_draintxq() 5091 txq->axq_qnum, in ath_tx_stopdma() 5096 txq->axq_link, in ath_tx_stopdma() [all …]
|
H A D | if_ath_tx.c | 713 txq->axq_qnum); in ath_tx_handoff_mcast() 716 ATH_TXQ_LOCK(txq); in ath_tx_handoff_mcast() 777 ATH_TXQ_LOCK(txq); in ath_tx_handoff_hw() 806 txq->axq_qnum, in ath_tx_handoff_hw() 820 txq->axq_qnum, txq->axq_link, in ath_tx_handoff_hw() 826 txq->axq_qnum, txq->axq_link, in ath_tx_handoff_hw() 854 txq->axq_qnum, in ath_tx_handoff_hw() 925 txq->axq_qnum, in ath_legacy_tx_dma_restart() 1742 txq, in ath_tx_normal_setup() 3220 if (txq->axq_depth + txq->fifo.axq_depth == 0) { in ath_tx_swq() [all …]
|
/freebsd/sys/contrib/dev/iwlwifi/pcie/ |
H A D | tx.c | 126 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local 218 while (txq->write_ptr != txq->read_ptr) { in iwl_pcie_txq_unmap() 233 if (txq->read_ptr == txq->write_ptr && in iwl_pcie_txq_unmap() 281 txq->tfds, txq->dma_addr); in iwl_pcie_txq_free() 296 memset(txq, 0, sizeof(*txq)); in iwl_pcie_txq_free() 689 txq->write_ptr, txq->read_ptr); in iwl_pcie_cmdq_reclaim() 704 if (txq->read_ptr == txq->write_ptr) in iwl_pcie_cmdq_reclaim() 1152 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_pcie_enqueue_hcmd() 1461 txq = trans->txqs.txq[txq_id]; in iwl_trans_pcie_tx() 1511 txq->entries[txq->write_ptr].skb = skb; in iwl_trans_pcie_tx() [all …]
|
H A D | tx-gen2.c | 33 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() local 114 spin_lock_irqsave(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd() 116 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 117 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 129 out_cmd = txq->entries[idx].cmd; in iwl_pcie_gen2_enqueue_hcmd() 130 out_meta = &txq->entries[idx].meta; in iwl_pcie_gen2_enqueue_hcmd() 146 INDEX_TO_SEQ(txq->write_ptr)); in iwl_pcie_gen2_enqueue_hcmd() 246 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_pcie_gen2_enqueue_hcmd() 247 mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); in iwl_pcie_gen2_enqueue_hcmd() 251 txq->write_ptr = iwl_txq_inc_wrap(trans, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() [all …]
|
/freebsd/sys/dev/xen/netfront/ |
H A D | netfront.c | 884 txq = &info->txq[q]; in setup_txqs() 902 txq->xennet_tag[i].txq = txq; in setup_txqs() 916 SLIST_INSERT_HEAD(&txq->tags, &txq->xennet_tag[i], in setup_txqs() 944 TASK_INIT(&txq->defrtask, 0, xn_txq_tq_deferred, txq); in setup_txqs() 946 txq->tq = taskqueue_create(txq->name, M_WAITOK, in setup_txqs() 1398 &txq->gref_head, txq->grant_ref[id]); in xn_txeof() 2034 txq = &np->txq[i]; in xn_connect() 2063 txq = &np->txq[i]; in xn_kick_rings() 2249 txq = &np->txq[i]; in xn_txq_mq_start() 2256 taskqueue_enqueue(txq->tq, &txq->defrtask); in xn_txq_mq_start() [all …]
|
/freebsd/sys/net/ |
H A D | iflib.c | 438 #define TXQ_AVAIL(txq) (txq->ift_size - get_inuse(txq->ift_size, txq->ift_cidx, txq->ift_pidx, txq-… argument 1924 txq->ift_pidx = txq->ift_cidx = txq->ift_npending = 0; in iflib_txq_setup() 2440 txq, txq->ift_timer.c_cpu); in iflib_timer() 2446 txq->ift_id, TXQ_AVAIL(txq), txq->ift_pidx); in iflib_timer() 2632 txq->ift_processed = txq->ift_cleaned = txq->ift_cidx_processed = 0; in iflib_stop() 2633 txq->ift_in_use = txq->ift_gen = txq->ift_no_desc_avail = 0; in iflib_stop() 2635 txq->ift_cidx = txq->ift_pidx; in iflib_stop() 2637 txq->ift_cidx = txq->ift_pidx = 0; in iflib_stop() 2639 txq->ift_closed = txq->ift_mbuf_defrag = txq->ift_mbuf_defrag_failed = 0; in iflib_stop() 3105 #define TXD_NOTIFY_COUNT(txq) (((txq)->ift_size / (txq)->ift_update_freq)-1) argument [all …]
|
/freebsd/sys/dev/cxgb/ |
H A D | cxgb_sge.c | 262 txq = &qs->txq[TXQ_ETH]; in check_pkt_coalesce() 978 txq = &qs->txq[0]; in sge_timer_cb() 1147 (txq->pidx >= txq->cidx)) || in txq_prod() 1149 (txq->pidx >= txq-> cidx)) || in txq_prod() 1153 txqs->pidx, txq->pidx, txq->cidx); in txq_prod() 1155 if (txq->pidx >= txq->size) { in txq_prod() 1156 txq->pidx -= txq->size; in txq_prod() 1375 txq = &qs->txq[TXQ_ETH]; in t3_encap() 1376 txd = &txq->desc[txq->pidx]; in t3_encap() 1377 txsd = &txq->sdesc[txq->pidx]; in t3_encap() [all …]
|
/freebsd/sys/dev/ice/ |
H A D | ice_iflib_txrx.c | 159 mask = txq->desc_count - 1; in _ice_ift_txd_encap() 163 txd = &txq->tx_base[i]; in _ice_ift_txd_encap() 185 txq->tx_rsq[txq->tx_rs_pidx] = pidx_last; in _ice_ift_txd_encap() 186 txq->tx_rs_pidx = (txq->tx_rs_pidx+1) & mask; in _ice_ift_txd_encap() 187 MPASS(txq->tx_rs_pidx != txq->tx_rs_cidx); in _ice_ift_txd_encap() 191 ++txq->stats.tx_packets; in _ice_ift_txd_encap() 232 wr32(hw, txq->tail, pidx); in ice_ift_txd_flush() 257 rs_cidx = txq->tx_rs_cidx; in _ice_ift_txd_credits_update() 260 cur = txq->tx_rsq[rs_cidx]; in _ice_ift_txd_credits_update() 270 ntxd = txq->desc_count; in _ice_ift_txd_credits_update() [all …]
|
H A D | ice_common_txrx.h | 153 txd = (struct ice_tx_ctx_desc *)&txq->tx_base[idx]; in ice_tso_setup() 160 txq->stats.mss_too_small++; in ice_tso_setup() 172 txq->tso++; in ice_tso_setup() 174 return ((idx + 1) & (txq->desc_count-1)); in ice_tso_setup() 201 txq->stats.cso[ICE_CSO_STAT_TX_IP4]++; in ice_tx_setup_offload() 214 txq->stats.cso[ICE_CSO_STAT_TX_IP6]++; in ice_tx_setup_offload() 218 txq->stats.cso[ICE_CSO_STAT_TX_L3_ERR]++; in ice_tx_setup_offload() 234 txq->stats.cso[ICE_CSO_STAT_TX_TCP]++; in ice_tx_setup_offload() 242 txq->stats.cso[ICE_CSO_STAT_TX_UDP]++; in ice_tx_setup_offload() 250 txq->stats.cso[ICE_CSO_STAT_TX_SCTP]++; in ice_tx_setup_offload() [all …]
|
H A D | if_ice_iflib.c | 1054 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_if_tx_queues_alloc() 1078 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_if_tx_queues_alloc() 1080 txq->me = txq->q_handle = i; in ice_if_tx_queues_alloc() 1081 txq->vsi = vsi; in ice_if_tx_queues_alloc() 1099 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_if_tx_queues_alloc() 1975 txq->tx_rs_cidx = txq->tx_rs_pidx = 0; in ice_init_tx_tracking() 1982 txq->tx_cidx_processed = txq->desc_count - 1; in ice_init_tx_tracking() 3747 IFLIB_INTR_TX, txq, txq->me, irq_name); in ice_wire_mirror_intrs() 3910 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_subif_if_tx_queues_alloc() 3926 for (i = 0, txq = vsi->tx_queues; i < ntxqsets; i++, txq++) { in ice_subif_if_tx_queues_alloc() [all …]
|
/freebsd/sys/dev/virtio/network/ |
H A D | if_vtnet.c | 865 mtx_init(&txq->vtntx_mtx, txq->vtntx_name, NULL, MTX_DEF); in vtnet_init_txq() 1046 vtnet_tx_vq_intr, txq, &txq->vtntx_vq, in vtnet_alloc_virtqueues() 2675 taskqueue_enqueue(txq->vtntx_tq, &txq->vtntx_intrtask); in vtnet_start_locked() 2755 taskqueue_enqueue(txq->vtntx_tq, &txq->vtntx_intrtask); in vtnet_txq_mq_start_locked() 2783 taskqueue_enqueue(txq->vtntx_tq, &txq->vtntx_defrtask); in vtnet_txq_mq_start() 2795 txq = xtxq; in vtnet_txq_tq_deferred() 2831 txq = xtxq; in vtnet_txq_tq_intr() 2886 txq = xtxq; in vtnet_tx_vq_intr() 2981 if (txq->vtntx_watchdog == 0 || --txq->vtntx_watchdog) { in vtnet_watchdog() 3147 taskqueue_drain(txq->vtntx_tq, &txq->vtntx_intrtask); in vtnet_drain_taskqueues() [all …]
|
/freebsd/sys/dev/mana/ |
H A D | mana_en.c | 843 txq = &apc->tx_qp[txq_id].txq; in mana_start_xmit() 1435 struct mana_txq *txq = cq->txq; in mana_poll_tx_cq() local 1575 if (txq->alt_txq_idx != txq->idx) { in mana_poll_tx_cq() 1580 txq->alt_txq_idx = txq->idx; in mana_poll_tx_cq() 1594 txq->idx, txq->pending_sends); in mana_poll_tx_cq() 1962 if (txq->next_to_use != txq->next_to_complete) { in mana_deinit_txq() 1966 txq->next_to_use, txq->next_to_complete); in mana_deinit_txq() 2067 txq = &apc->tx_qp[i].txq; in mana_create_txq() 2086 cq->txq = txq; in mana_create_txq() 2124 txq->gdma_txq_id = txq->gdma_sq->id; in mana_create_txq() [all …]
|
/freebsd/sys/dev/cxgbe/crypto/ |
H A D | t6_kern_tls.c | 510 txq = &sc->sge.txq[vi->first_txq]; in t6_tls_tag_alloc() 514 tlsp->txq = txq; in t6_tls_tag_alloc() 570 TXQ_LOCK(txq); in t6_tls_tag_alloc() 575 TXQ_UNLOCK(txq); in t6_tls_tag_alloc() 1178 txq->txcsum++; in ktls_write_tcp_options() 1189 txq->imm_wrs++; in ktls_write_tcp_options() 1278 txq->txcsum++; in ktls_write_tunnel_packet() 1293 txq->imm_wrs++; in ktls_write_tunnel_packet() 1334 MPASS(tlsp->txq == txq); in ktls_write_tls_wr() 1920 txq->txcsum++; in ktls_write_tcp_fin() [all …]
|
/freebsd/sys/dev/malo/ |
H A D | if_malo.c | 573 txq->nfree = 0; in malo_txdma_setup() 587 txq->nfree++; in malo_txdma_setup() 652 txq->nfree = 0; in malo_txdma_cleanup() 745 txq->qnum = qnum; in malo_txq_init() 749 bf->bf_txq = txq; in malo_txq_init() 778 MALO_TXQ_LOCK(txq); in malo_getbuf() 782 txq->nfree--; in malo_getbuf() 992 txq->nfree++; in malo_tx_processq() 1147 txq = bf->bf_txq; in malo_tx_start() 1607 txq->nfree++; in malo_tx_draintxq() [all …]
|
/freebsd/sys/contrib/dev/mediatek/mt76/ |
H A D | tx.c | 11 if (!txq->sta) in mt76_txq_get_qid() 14 return txq->ac; in mt76_txq_get_qid() 21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local 30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn() 416 struct ieee80211_txq *txq = sta->txq[i]; in mt76_release_buffered_frames() local 479 ieee80211_get_tx_rates(txq->vif, txq->sta, skb, in mt76_txq_send_burst() 501 ieee80211_get_tx_rates(txq->vif, txq->sta, skb, in mt76_txq_send_burst() 542 if (!txq) in mt76_txq_schedule_list() 555 u8 tid = txq->tid; in mt76_txq_schedule_list() 643 struct ieee80211_txq *txq = sta->txq[i]; in mt76_stop_tx_queues() local [all …]
|
/freebsd/sys/contrib/dev/athk/ath10k/ |
H A D | txrx.c | 50 struct ieee80211_txq *txq; in ath10k_txrx_tx_unref() local 76 txq = skb_cb->txq; in ath10k_txrx_tx_unref() 78 if (txq) { in ath10k_txrx_tx_unref() 79 artxq = (void *)txq->drv_priv; in ath10k_txrx_tx_unref() 89 if (txq && txq->sta && skb_cb->airtime_est) in ath10k_txrx_tx_unref() 90 ieee80211_sta_register_airtime(txq->sta, txq->tid, in ath10k_txrx_tx_unref() 138 if (txq) in ath10k_txrx_tx_unref() 139 status.sta = txq->sta; in ath10k_txrx_tx_unref()
|
/freebsd/sys/dev/iavf/ |
H A D | iavf_vc_iflib.c | 88 vqpi->txq.vsi_id = vqci->vsi_id; in iavf_configure_queues() 89 vqpi->txq.queue_id = i; in iavf_configure_queues() 90 vqpi->txq.ring_len = scctx->isc_ntxd[0]; in iavf_configure_queues() 91 vqpi->txq.dma_ring_addr = txr->tx_paddr; in iavf_configure_queues() 94 vqpi->txq.headwb_enabled = 0; in iavf_configure_queues() 95 vqpi->txq.dma_headwb_addr = 0; in iavf_configure_queues() 97 vqpi->txq.headwb_enabled = 1; in iavf_configure_queues() 98 vqpi->txq.dma_headwb_addr = txr->tx_paddr + in iavf_configure_queues()
|
/freebsd/sys/dev/ral/ |
H A D | rt2661.c | 866 txq = (qid <= 3) ? &sc->txq[qid] : &sc->mgtq; in rt2661_tx_intr() 869 data = &txq->data[txq->stat]; in rt2661_tx_intr() 932 desc = &txq->desc[txq->next]; in rt2661_tx_dma_intr() 933 data = &txq->data[txq->next]; in rt2661_tx_dma_intr() 1358 struct rt2661_tx_ring *txq = &sc->txq[ac]; in rt2661_sendprot() local 1374 data = &txq->data[txq->cur]; in rt2661_sendprot() 1375 desc = &txq->desc[txq->cur]; in rt2661_sendprot() 1403 txq->cur = (txq->cur + 1) % RT2661_TX_RING_COUNT; in rt2661_sendprot() 1414 struct rt2661_tx_ring *txq = &sc->txq[ac]; in rt2661_tx_data() local 1472 data = &txq->data[txq->cur]; in rt2661_tx_data() [all …]
|
/freebsd/sys/dev/mwl/ |
H A D | if_mwl.c | 1272 MWL_TXQ_LOCK(txq); in mwl_gettxbuf() 1276 txq->nfree--; in mwl_gettxbuf() 1296 MWL_TXQ_LOCK(txq); in mwl_puttxbuf_head() 1298 txq->nfree++; in mwl_puttxbuf_head() 1307 MWL_TXQ_LOCK(txq); in mwl_puttxbuf_tail() 1309 txq->nfree++; in mwl_puttxbuf_tail() 2016 txq->nfree = i; in mwl_txq_reset() 2075 txq->nfree = 0; in mwl_txdma_cleanup() 2831 txq->qnum = qnum; in mwl_txq_init() 2838 bf->bf_txq = txq; in mwl_txq_init() [all …]
|
/freebsd/sys/dev/vmware/vmxnet3/ |
H A D | if_vmx.c | 723 txq = &sc->vmx_txq[i]; in vmxnet3_set_interrupt_idx() 724 txs = txq->vxtxq_ts; in vmxnet3_set_interrupt_idx() 763 txq = &sc->vmx_txq[q]; in vmxnet3_init_txq() 768 snprintf(txq->vxtxq_name, sizeof(txq->vxtxq_name), "%s-tx%d", in vmxnet3_init_txq() 771 txq->vxtxq_sc = sc; in vmxnet3_init_txq() 772 txq->vxtxq_id = q; in vmxnet3_init_txq() 822 txq = &sc->vmx_txq[q]; in vmxnet3_tx_queues_alloc() 1114 txq = &sc->vmx_txq[i]; in vmxnet3_init_shared_data() 1115 txs = txq->vxtxq_ts; in vmxnet3_init_shared_data() 1385 txq = &sc->vmx_txq[txqid]; in vmxnet3_isc_txd_flush() [all …]
|