Home
last modified time | relevance | path

Searched refs:txqs (Results 1 – 25 of 63) sorted by relevance

123

/linux/drivers/net/wireless/intel/iwlwifi/
H A Diwl-trans.c46 trans->txqs.tfd.addr_size = 64; in iwl_trans_alloc()
50 trans->txqs.tfd.addr_size = 36; in iwl_trans_alloc()
51 trans->txqs.tfd.max_tbs = IWL_NUM_OF_TBS; in iwl_trans_alloc()
82 trans->txqs.bc_tbl_size = in iwl_trans_init()
85 trans->txqs.bc_tbl_size = in iwl_trans_init()
96 trans->txqs.bc_tbl_size, in iwl_trans_init()
98 if (!trans->txqs.bc_pool) in iwl_trans_init()
103 WARN_ON(trans->txqs.tfd.addr_size != in iwl_trans_init()
116 if (!trans->txqs.tso_hdr_page) { in iwl_trans_init()
131 if (trans->txqs.tso_hdr_page) { in iwl_trans_free()
[all …]
H A Diwl-trans.h174 #define IWL_TRANS_MAX_FRAGS(trans) ((trans)->txqs.tfd.max_tbs - 3)
615 void (*freeze_txq_timer)(struct iwl_trans *trans, unsigned long txqs,
1172 struct iwl_trans_txqs txqs; member
1434 unsigned long txqs, in iwl_trans_freeze_txq_timer() argument
1443 trans->ops->freeze_txq_timer(trans, txqs, freeze); in iwl_trans_freeze_txq_timer()
1447 u32 txqs) in iwl_trans_wait_tx_queues_empty() argument
1458 return trans->ops->wait_tx_queues_empty(trans, txqs); in iwl_trans_wait_tx_queues_empty()
/linux/drivers/net/wireless/intel/iwlwifi/pcie/
H A Dtx.c151 trans->txqs.tfd.max_tbs); in iwl_pcie_txq_build_tfd()
258 trans->txqs.tfd.size * in iwl_pcie_txq_free()
290 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_pcie_tx_start()
313 trans->txqs.cmd.fifo, in iwl_pcie_tx_start()
314 trans->txqs.cmd.wdg_timeout); in iwl_pcie_tx_start()
427 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_pcie_tx_stop()
451 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_pcie_tx_free()
915 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_enqueue_hcmd()
1199 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_hcmd_complete()
1331 trans->txqs.tfd.size, in iwl_fill_data_tbs_amsdu()
[all …]
H A Dtx-gen2.c31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd()
146 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | in iwl_pcie_gen2_enqueue_hcmd()
194 cmd_size, txq->write_ptr, idx, trans->txqs.cmd.q_id); in iwl_pcie_gen2_enqueue_hcmd()
H A Dtrans-gen2.c250 if (iwl_txq_gen2_init(trans, trans->txqs.cmd.q_id, queue_size)) in iwl_pcie_gen2_nic_init()
349 memset(trans->txqs.queue_stopped, 0, in iwl_trans_pcie_gen2_fw_alive()
350 sizeof(trans->txqs.queue_stopped)); in iwl_trans_pcie_gen2_fw_alive()
351 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_trans_pcie_gen2_fw_alive()
H A Dtrans.c1959 trans->txqs.cmd.q_id = trans_cfg->cmd_queue; in iwl_trans_pcie_configure()
1960 trans->txqs.cmd.fifo = trans_cfg->cmd_fifo; in iwl_trans_pcie_configure()
2404 if (!test_bit(txq_idx, trans->txqs.queue_used)) in iwl_trans_pcie_wait_txq_empty()
2408 txq = trans->txqs.txq[txq_idx]; in iwl_trans_pcie_wait_txq_empty()
2464 if (cnt == trans->txqs.cmd.q_id) in iwl_trans_pcie_wait_txqs_empty()
2466 if (!test_bit(cnt, trans->txqs.queue_used)) in iwl_trans_pcie_wait_txqs_empty()
2654 if (state->pos == trans->txqs.cmd.q_id) in iwl_dbgfs_tx_queue_seq_show()
3074 for (i = 0; i < trans->txqs.tfd.max_tbs; i++) in iwl_trans_pcie_get_cmdlen()
3343 struct iwl_txq *cmdq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_trans_pcie_dump_data()
3410 u16 tfd_size = trans->txqs.tfd.size; in iwl_trans_pcie_dump_data()
[all …]
H A Dctxt-info.c221 cpu_to_le64(trans->txqs.txq[trans->txqs.cmd.q_id]->dma_addr); in iwl_pcie_ctxt_info_init()
/linux/drivers/net/wireless/intel/iwlwifi/queue/
H A Dtx.c50 WARN_ON(trans->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl()
111 trans->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb()
891 txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_free()
908 trans->txqs.txq[txq_id] = NULL; in iwl_txq_gen2_free()
1209 trans->txqs.txq[qid] = txq; in iwl_txq_alloc_response()
1321 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_txq_gen2_tx_free()
1325 if (!trans->txqs.txq[i]) in iwl_txq_gen2_tx_free()
1338 if (!trans->txqs.txq[txq_id]) { in iwl_txq_gen2_init()
1467 if (trans->txqs.bc_table_dword) in iwl_txq_gen1_update_byte_cnt_tbl()
1683 trans->txqs.dev_cmd_offs); in iwl_txq_reclaim()
[all …]
H A Dtx.h32 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_wake_queue()
44 return (u8 *)txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd()
63 if (!test_and_set_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_txq_stop()
186 void iwl_trans_txq_freeze_timer(struct iwl_trans *trans, unsigned long txqs,
/linux/drivers/net/wireless/ath/ath5k/
H A Dmac80211-ops.c69 ath5k_tx_queue(hw, skb, &ah->txqs[qnum], control); in ath5k_tx()
745 *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; in ath5k_get_ringparam()
765 for (qnum = 0; qnum < ARRAY_SIZE(ah->txqs); qnum++) { in ath5k_set_ringparam()
766 if (!ah->txqs[qnum].setup) in ath5k_set_ringparam()
768 if (ah->txqs[qnum].qnum < AR5K_TX_QUEUE_ID_DATA_MIN || in ath5k_set_ringparam()
769 ah->txqs[qnum].qnum > AR5K_TX_QUEUE_ID_DATA_MAX) in ath5k_set_ringparam()
772 ah->txqs[qnum].txq_max = tx; in ath5k_set_ringparam()
773 if (ah->txqs[qnum].txq_len >= ah->txqs[qnum].txq_max) in ath5k_set_ringparam()
774 ieee80211_stop_queue(hw, ah->txqs[qnum].qnum); in ath5k_set_ringparam()
H A Dbase.c1055 txq = &ah->txqs[qnum]; in ath5k_txq_setup()
1067 return &ah->txqs[qnum]; in ath5k_txq_setup()
1161 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_drain_tx_buffs()
1162 if (ah->txqs[i].setup) { in ath5k_drain_tx_buffs()
1163 txq = &ah->txqs[i]; in ath5k_drain_tx_buffs()
1186 struct ath5k_txq *txq = ah->txqs; in ath5k_txq_release()
1189 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) in ath5k_txq_release()
1839 ath5k_tx_processq(ah, &ah->txqs[i]); in ath5k_tasklet_tx()
2519 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_tx_complete_poll_work()
2520 if (ah->txqs[i].setup) { in ath5k_tx_complete_poll_work()
[all …]
/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_main.c355 for (i = start; i < nqs && txqs[i]; i++) in free_txqs()
356 txqs[i] = funeth_txq_free(txqs[i], state); in free_txqs()
369 state, &txqs[i]); in alloc_txqs()
453 qset->txqs = fp->txqs; in fun_free_rings()
465 fp->txqs = NULL; in fun_free_rings()
481 struct funeth_txq **xdpqs = NULL, **txqs; in fun_alloc_rings() local
514 qset->txqs = txqs; in fun_alloc_rings()
796 fp->txqs = qset->txqs; in fun_up()
1630 .txqs = fp->txqs, in fun_change_num_queues()
1659 memcpy(newqs.txqs, fp->txqs, keep_tx * sizeof(*fp->txqs)); in fun_change_num_queues()
[all …]
H A Dfuneth.h67 struct funeth_txq **txqs; member
88 struct funeth_txq **txqs; member
/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_main.c134 if (nic_dev->txqs) in create_txqs()
139 if (!nic_dev->txqs) in create_txqs()
165 hinic_clean_txq(&nic_dev->txqs[i]); in create_txqs()
169 hinic_clean_txq(&nic_dev->txqs[j]); in create_txqs()
184 napi_enable(&nic_dev->txqs[i].napi); in enable_txqs_napi()
193 napi_disable(&nic_dev->txqs[i].napi); in disable_txqs_napi()
205 if (!nic_dev->txqs) in free_txqs()
210 hinic_clean_txq(&nic_dev->txqs[i]); in free_txqs()
216 nic_dev->txqs = NULL; in free_txqs()
834 nic_dev->txqs[q_id].napi.state); in hinic_tx_timeout()
[all …]
H A Dhinic_dev.h97 struct hinic_txq *txqs; member
/linux/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c697 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init()
701 if (!priv->txqs) in hfi1_ipoib_txreq_init()
705 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init()
756 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init()
765 kfree(priv->txqs); in hfi1_ipoib_txreq_init()
766 priv->txqs = NULL; in hfi1_ipoib_txreq_init()
799 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_deinit()
812 kfree(priv->txqs); in hfi1_ipoib_txreq_deinit()
813 priv->txqs = NULL; in hfi1_ipoib_txreq_deinit()
822 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_napi_tx_enable()
[all …]
H A Dipoib.h125 struct hfi1_ipoib_txq *txqs; member
/linux/net/
H A Ddevres.c22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument
30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c153 idpf_tx_desc_rel(txq_grp->txqs[j], true); in idpf_tx_desc_rel_all()
814 kfree(txq_grp->txqs[j]); in idpf_txq_group_rel()
815 txq_grp->txqs[j] = NULL; in idpf_txq_group_rel()
904 kfree(vport->txqs); in idpf_vport_queues_rel()
905 vport->txqs = NULL; in idpf_vport_queues_rel()
925 if (!vport->txqs) in idpf_vport_init_fast_path_txqs()
932 vport->txqs[k] = tx_grp->txqs[j]; in idpf_vport_init_fast_path_txqs()
933 vport->txqs[k]->idx = k; in idpf_vport_init_fast_path_txqs()
1181 tx_qgrp->txqs[j] = kzalloc(sizeof(*tx_qgrp->txqs[j]), in idpf_txq_group_alloc()
1183 if (!tx_qgrp->txqs[j]) { in idpf_txq_group_alloc()
[all …]
H A Didpf_virtchnl.c753 set_bit(__IDPF_Q_SW_MARKER, vport->txqs[i]->flags); in idpf_wait_for_marker_event()
1104 tx_qgrp->txqs[j]->tail = in __idpf_queue_reg_init()
1448 cpu_to_le32(tx_qgrp->txqs[j]->q_id); in idpf_send_config_tx_queues_msg()
1452 cpu_to_le32(tx_qgrp->txqs[j]->q_type); in idpf_send_config_tx_queues_msg()
1454 cpu_to_le16(tx_qgrp->txqs[j]->desc_count); in idpf_send_config_tx_queues_msg()
1456 cpu_to_le64(tx_qgrp->txqs[j]->dma); in idpf_send_config_tx_queues_msg()
1458 struct idpf_queue *q = tx_qgrp->txqs[j]; in idpf_send_config_tx_queues_msg()
1862 cpu_to_le16(tx_qgrp->txqs[j]->q_vector->v_idx); in idpf_send_map_unmap_queue_vector_msg()
1978 set_bit(__IDPF_Q_POLL_MODE, vport->txqs[i]->flags); in idpf_send_disable_queues_msg()
3254 tx_qgrp->txqs[j]->q_id = qids[k]; in __idpf_vport_queue_ids_init()
[all …]
/linux/drivers/net/ethernet/netronome/nfp/abm/
H A Dmain.c85 unsigned int txqs; in nfp_abm_spawn_repr() local
90 txqs = 1; in nfp_abm_spawn_repr()
93 txqs = alink->vnic->max_rx_rings; in nfp_abm_spawn_repr()
96 netdev = nfp_repr_alloc_mqs(app, txqs, 1); in nfp_abm_spawn_repr()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_base.c874 u8 buf_len = struct_size(qg_buf, txqs, 1); in ice_vsi_cfg_txq()
890 qg_buf->txqs[0].txq_id = cpu_to_le16(pf_q); in ice_vsi_cfg_txq()
891 ice_set_ctx(hw, (u8 *)&tlan_ctx, qg_buf->txqs[0].txq_ctx, in ice_vsi_cfg_txq()
927 txq = &qg_buf->txqs[0]; in ice_vsi_cfg_txq()
937 DEFINE_RAW_FLEX(struct ice_aqc_add_tx_qgrp, qg_buf, txqs, 1); in ice_vsi_cfg_single_txq()
959 DEFINE_RAW_FLEX(struct ice_aqc_add_tx_qgrp, qg_buf, txqs, 1); in ice_vsi_cfg_txqs()
/linux/drivers/net/ethernet/netronome/nfp/
H A Dnfp_net_repr.h102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
/linux/net/ethernet/
H A Deth.c379 struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs, in alloc_etherdev_mqs() argument
383 ether_setup, txqs, rxqs); in alloc_etherdev_mqs()
/linux/include/linux/
H A Detherdevice.h56 struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs,
62 unsigned int txqs,

123