Lines Matching refs:adapter

71 	struct ena_adapter *adapter = que->adapter;  in ena_cleanup()  local
72 if_t ifp = adapter->ifp; in ena_cleanup()
83 ena_log_io(adapter->pdev, DBG, "MSI-X TX/RX routine\n"); in ena_cleanup()
89 io_cq = &adapter->ena_dev->io_cq_queues[ena_qid]; in ena_cleanup()
116 if_t ifp = tx_ring->adapter->ifp; in ena_deferred_mq_start()
129 struct ena_adapter *adapter = if_getsoftc(ifp); in ena_mq_start() local
137 if (unlikely((if_getdrvflags(adapter->ifp) & IFF_DRV_RUNNING) == 0)) in ena_mq_start()
150 i = bucket_id % adapter->num_io_queues; in ena_mq_start()
153 i = m->m_pkthdr.flowid % adapter->num_io_queues; in ena_mq_start()
155 i = curcpu % adapter->num_io_queues; in ena_mq_start()
157 tx_ring = &adapter->tx_ring[i]; in ena_mq_start()
180 struct ena_adapter *adapter = if_getsoftc(ifp); in ena_qflush() local
181 struct ena_ring *tx_ring = adapter->tx_ring; in ena_qflush()
184 for (i = 0; i < adapter->num_io_queues; ++i, ++tx_ring) in ena_qflush()
202 struct ena_adapter *adapter = tx_ring->adapter; in ena_get_tx_req_id() local
210 ena_log(adapter->pdev, ERR, "Invalid req_id %hu in qid %hu\n", in ena_get_tx_req_id()
219 ena_log(adapter->pdev, ERR, in ena_get_tx_req_id()
223 ena_trigger_reset(adapter, ENA_REGS_RESET_INV_TX_REQ_ID); in ena_get_tx_req_id()
242 struct ena_adapter *adapter; in ena_tx_cleanup() local
254 adapter = tx_ring->que->adapter; in ena_tx_cleanup()
256 io_cq = &adapter->ena_dev->io_cq_queues[ena_qid]; in ena_tx_cleanup()
260 if (netmap_tx_irq(adapter->ifp, tx_ring->qid) != NM_IRQ_PASS) in ena_tx_cleanup()
279 bus_dmamap_sync(adapter->tx_buf_tag, tx_info->dmamap, in ena_tx_cleanup()
281 bus_dmamap_unload(adapter->tx_buf_tag, tx_info->dmamap); in ena_tx_cleanup()
283 ena_log_io(adapter->pdev, DBG, "tx: q %d mbuf %p completed\n", in ena_tx_cleanup()
299 &adapter->ena_dev->io_sq_queues[ena_qid], in ena_tx_cleanup()
307 ena_log_io(adapter->pdev, DBG, "tx: q %d done. total pkts: %d\n", in ena_tx_cleanup()
313 ena_com_comp_ack(&adapter->ena_dev->io_sq_queues[ena_qid], in ena_tx_cleanup()
347 struct ena_adapter *adapter = rx_ring->adapter; in ena_rx_hash_mbuf() local
349 if (likely(ENA_FLAG_ISSET(ENA_FLAG_RSS_ACTIVE, adapter))) { in ena_rx_hash_mbuf()
358 if (adapter->ena_dev->rss.hash_func != ENA_ADMIN_TOEPLITZ && in ena_rx_hash_mbuf()
422 struct ena_adapter *adapter; in ena_rx_mbuf() local
428 adapter = rx_ring->adapter; in ena_rx_mbuf()
429 pdev = adapter->pdev; in ena_rx_mbuf()
442 bus_dmamap_sync(adapter->rx_buf_tag, rx_info->map, in ena_rx_mbuf()
451 mbuf->m_pkthdr.rcvif = rx_ring->que->adapter->ifp; in ena_rx_mbuf()
460 bus_dmamap_unload(rx_ring->adapter->rx_buf_tag, rx_info->map); in ena_rx_mbuf()
492 bus_dmamap_sync(adapter->rx_buf_tag, rx_info->map, in ena_rx_mbuf()
504 bus_dmamap_unload(rx_ring->adapter->rx_buf_tag, rx_info->map); in ena_rx_mbuf()
524 device_t pdev = rx_ring->adapter->pdev; in ena_rx_checksum()
559 struct ena_adapter *adapter; in ena_rx_cleanup() local
579 adapter = rx_ring->que->adapter; in ena_rx_cleanup()
580 pdev = adapter->pdev; in ena_rx_cleanup()
581 ifp = adapter->ifp; in ena_rx_cleanup()
584 io_cq = &adapter->ena_dev->io_cq_queues[ena_qid]; in ena_rx_cleanup()
585 io_sq = &adapter->ena_dev->io_sq_queues[ena_qid]; in ena_rx_cleanup()
589 if (netmap_rx_irq(adapter->ifp, rx_ring->qid, &done) != NM_IRQ_PASS) in ena_rx_cleanup()
597 ena_rx_ctx.max_bufs = adapter->max_rx_sgl_size; in ena_rx_cleanup()
614 ena_trigger_reset(adapter, reset_reason); in ena_rx_cleanup()
650 counter_u64_add_protected(adapter->hw_stats.rx_bytes, in ena_rx_cleanup()
679 counter_u64_add_protected(adapter->hw_stats.rx_packets, 1); in ena_rx_cleanup()
814 struct ena_adapter *adapter; in ena_check_and_collapse_mbuf() local
818 adapter = tx_ring->adapter; in ena_check_and_collapse_mbuf()
822 if (num_frags < adapter->max_tx_sgl_size) in ena_check_and_collapse_mbuf()
825 if ((num_frags == adapter->max_tx_sgl_size) && in ena_check_and_collapse_mbuf()
832 adapter->max_tx_sgl_size - 1); in ena_check_and_collapse_mbuf()
848 struct ena_adapter *adapter = tx_ring->adapter; in ena_tx_map_mbuf() local
864 rc = bus_dmamap_load_mbuf_sg(adapter->tx_buf_tag, in ena_tx_map_mbuf()
867 ena_log_io(adapter->pdev, WARN, in ena_tx_map_mbuf()
901 ena_log_io(adapter->pdev, DBG, in ena_tx_map_mbuf()
962 struct ena_adapter *adapter; in ena_xmit_mbuf() local
977 adapter = tx_ring->que->adapter; in ena_xmit_mbuf()
978 pdev = adapter->pdev; in ena_xmit_mbuf()
979 ena_dev = adapter->ena_dev; in ena_xmit_mbuf()
996 ENA_WARN(tx_info->mbuf != NULL, adapter->ena_dev, in ena_xmit_mbuf()
1012 ena_tx_csum(&ena_tx_ctx, *mbuf, adapter->disable_meta_caching); in ena_xmit_mbuf()
1030 ena_trigger_reset(adapter, in ena_xmit_mbuf()
1042 counter_u64_add_protected(adapter->hw_stats.tx_packets, 1); in ena_xmit_mbuf()
1043 counter_u64_add_protected(adapter->hw_stats.tx_bytes, in ena_xmit_mbuf()
1059 adapter->max_tx_sgl_size + 2))) { in ena_xmit_mbuf()
1081 bus_dmamap_sync(adapter->tx_buf_tag, tx_info->dmamap, in ena_xmit_mbuf()
1088 bus_dmamap_unload(adapter->tx_buf_tag, tx_info->dmamap); in ena_xmit_mbuf()
1097 struct ena_adapter *adapter = tx_ring->adapter; in ena_start_xmit() local
1102 if (unlikely((if_getdrvflags(adapter->ifp) & IFF_DRV_RUNNING) == 0)) in ena_start_xmit()
1105 if (unlikely(!ENA_FLAG_ISSET(ENA_FLAG_LINK_UP, adapter))) in ena_start_xmit()
1108 while ((mbuf = drbr_peek(adapter->ifp, tx_ring->br)) != NULL) { in ena_start_xmit()
1109 ena_log_io(adapter->pdev, DBG, in ena_start_xmit()
1114 drbr_putback(adapter->ifp, tx_ring->br, mbuf); in ena_start_xmit()
1120 drbr_putback(adapter->ifp, tx_ring->br, mbuf); in ena_start_xmit()
1122 drbr_putback(adapter->ifp, tx_ring->br, mbuf); in ena_start_xmit()
1125 drbr_advance(adapter->ifp, tx_ring->br); in ena_start_xmit()
1131 drbr_advance(adapter->ifp, tx_ring->br); in ena_start_xmit()
1133 if (unlikely((if_getdrvflags(adapter->ifp) & IFF_DRV_RUNNING) == 0)) in ena_start_xmit()
1138 BPF_MTAP(adapter->ifp, mbuf); in ena_start_xmit()