/freebsd/sys/dev/smartpqi/ |
H A D | smartpqi_queue.c | 155 dma_addr_t dma_addr = 0; in pqisrc_allocate_and_init_inbound_q() local 179 dma_addr = dma_mem->dma_addr; in pqisrc_allocate_and_init_inbound_q() 181 ASSERT(!(dma_addr & PQI_ADDR_ALIGN_MASK)); in pqisrc_allocate_and_init_inbound_q() 186 ib_q->array_dma_addr = dma_addr; in pqisrc_allocate_and_init_inbound_q() 191 dma_addr += ib_array_size; in pqisrc_allocate_and_init_inbound_q() 196 ib_q->ci_dma_addr = dma_addr; in pqisrc_allocate_and_init_inbound_q() 227 dma_addr_t dma_addr = 0; in pqisrc_allocate_and_init_outbound_q() local 251 dma_addr = dma_mem->dma_addr; in pqisrc_allocate_and_init_outbound_q() 256 ob_q->array_dma_addr = dma_addr; in pqisrc_allocate_and_init_outbound_q() 261 dma_addr += ob_array_size; in pqisrc_allocate_and_init_outbound_q() [all …]
|
H A D | smartpqi_mem.c | 109 os_dma_map, &dma_mem->dma_addr, 0)) != 0) { in os_dma_mem_alloc() 145 if(dma_mem->dma_addr) { in os_dma_mem_free() 147 dma_mem->dma_addr = 0; in os_dma_mem_free()
|
H A D | smartpqi_event.c | 357 request->sg_desc.addr = buff->dma_addr; in pqi_event_configure() 405 DBG_INFO("buf_report_event.dma_addr = %p \n",(void*)buf_report_event.dma_addr); in pqisrc_report_event_config() 467 DBG_INFO("buf_set_event.dma_addr = %p\n",(void*)buf_set_event.dma_addr); in pqisrc_set_event_config()
|
/freebsd/sys/compat/linuxkpi/common/include/linux/ |
H A D | dma-mapping.h | 86 int (*mapping_error)(struct device *dev, dma_addr_t dma_addr); 100 void linux_dma_unmap(struct device *dev, dma_addr_t dma_addr, size_t size); 173 dma_addr_t dma_addr) in dma_free_coherent() argument 176 linux_dma_unmap(dev, dma_addr, size); in dma_free_coherent() 293 dma_mapping_error(struct device *dev, dma_addr_t dma_addr) in dma_mapping_error() argument 296 if (dma_addr == 0 || dma_addr == DMA_MAPPING_ERROR) in dma_mapping_error() 339 #define dma_unmap_single_attrs(dev, dma_addr, size, dir, attrs) \ argument 340 _dma_unmap_single_attrs(dev, dma_addr, size, dir, 0)
|
H A D | dmapool.h | 46 dma_addr_t dma_addr); 97 dma_pool_free(struct dma_pool *pool, void *vaddr, dma_addr_t dma_addr) in dma_pool_free() argument 100 linux_dma_pool_free(pool, vaddr, dma_addr); in dma_pool_free()
|
H A D | usb.h | 299 uint16_t mem_flags, uint8_t *dma_addr); 302 void usb_buffer_free(struct usb_device *dev, usb_size_t size, void *addr, uint8_t dma_addr);
|
/freebsd/sys/dev/iser/ |
H A D | iser_initiator.c | 108 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc() 201 u64 dma_addr; in iser_alloc_rx_descriptors() local 224 dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc, in iser_alloc_rx_descriptors() 226 if (ib_dma_mapping_error(device->ib_device, dma_addr)) in iser_alloc_rx_descriptors() 229 rx_desc->dma_addr = dma_addr; in iser_alloc_rx_descriptors() 232 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors() 244 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_alloc_rx_descriptors() 268 ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr, in iser_free_rx_descriptors() 483 rx_dma = rx_desc->dma_addr; in iser_rcv_completion()
|
H A D | icl_iser.c | 93 u64 dma_addr; in iser_initialize_headers() local 96 dma_addr = ib_dma_map_single(device->ib_device, (void *)tx_desc, in iser_initialize_headers() 98 if (ib_dma_mapping_error(device->ib_device, dma_addr)) { in iser_initialize_headers() 104 tx_desc->dma_addr = dma_addr; in iser_initialize_headers() 105 tx_desc->tx_sg[0].addr = tx_desc->dma_addr; in iser_initialize_headers() 504 ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr, in iser_conn_task_done()
|
H A D | icl_iser.h | 276 u64 dma_addr; member 298 u64 dma_addr; member
|
/freebsd/sys/dev/mlx5/mlx5_core/ |
H A D | mlx5_pagealloc.c | 63 fwp->dma_addr = segs->ds_addr; in mlx5_fwp_load_mem_cb() 194 return ((fwp + index)->dma_addr + (offset % MLX5_ADAPTER_PAGE_SIZE)); in mlx5_fwp_get_dma() 217 if (tfp->dma_addr < nfp->dma_addr) in mlx5_insert_fw_page_locked() 219 else if (tfp->dma_addr > nfp->dma_addr) in mlx5_insert_fw_page_locked() 240 if (tfp->dma_addr < addr) { in mlx5_remove_fw_page_locked() 242 } else if (tfp->dma_addr > addr) { in mlx5_remove_fw_page_locked() 276 *addr = fwp->dma_addr; in alloc_4k() 408 MLX5_ARRAY_SET64(manage_pages_out, out, pas, i, fwp->dma_addr); in reclaim_pages_cmd()
|
/freebsd/sys/dev/liquidio/ |
H A D | lio_network.h | 220 vm_paddr_t dma_addr; in lio_map_ring() local 222 dma_addr = vtophys(((struct mbuf *)buf)->m_data); in lio_map_ring() 223 return ((uint64_t)dma_addr); in lio_map_ring()
|
/freebsd/sys/dev/liquidio/base/ |
H A D | lio_request_manager.c | 725 uint64_t dma_addr; in lio_setup_sc_buffer_pool() local 734 lio_dma_alloc(LIO_SOFT_COMMAND_BUFFER_SIZE, (vm_paddr_t *)&dma_addr); in lio_setup_sc_buffer_pool() 740 sc->dma_addr = dma_addr; in lio_setup_sc_buffer_pool() 779 uint64_t dma_addr; in lio_alloc_soft_command() local 804 dma_addr = sc->dma_addr; in lio_alloc_soft_command() 809 sc->dma_addr = dma_addr; in lio_alloc_soft_command() 822 sc->dmadptr = dma_addr + offset; in lio_alloc_soft_command() 832 sc->dmarptr = dma_addr + offset; in lio_alloc_soft_command()
|
/freebsd/sys/ofed/drivers/infiniband/core/ |
H A D | ib_umem_odp.c | 442 dma_addr_t dma_addr; in ib_umem_odp_map_dma_single_page() local 457 dma_addr = ib_dma_map_page(dev, in ib_umem_odp_map_dma_single_page() 461 if (ib_dma_mapping_error(dev, dma_addr)) { in ib_umem_odp_map_dma_single_page() 465 umem->odp_data->dma_list[page_index] = dma_addr | access_mask; in ib_umem_odp_map_dma_single_page() 642 dma_addr_t dma_addr = dma & ODP_DMA_ADDR_MASK; in ib_umem_odp_unmap_dma_pages() local 644 WARN_ON(!dma_addr); in ib_umem_odp_unmap_dma_pages() 646 ib_dma_unmap_page(dev, dma_addr, PAGE_SIZE, in ib_umem_odp_unmap_dma_pages()
|
/freebsd/sys/dev/mpi3mr/ |
H A D | mpi3mr_app.c | 324 dma_addr = dma_buff->dma_desc[count].dma_addr; in mpi3mr_app_build_nvme_prp() 325 if (dma_addr & page_mask) { in mpi3mr_app_build_nvme_prp() 332 dma_addr = dma_buff->dma_desc[0].dma_addr; in mpi3mr_app_build_nvme_prp() 410 offset = dma_addr & page_mask; in mpi3mr_app_build_nvme_prp() 418 *prp1_entry = dma_addr; in mpi3mr_app_build_nvme_prp() 463 *prp2_entry = dma_addr; in mpi3mr_app_build_nvme_prp() 479 *prp_entry = dma_addr; in mpi3mr_app_build_nvme_prp() 495 dma_addr += entry_len; in mpi3mr_app_build_nvme_prp() 505 dma_addr = in mpi3mr_app_build_nvme_prp() 566 dma_buffers->dma_desc[i].dma_addr = sc->ioctl_sge[desc_count].dma_addr; in mpi3mr_map_data_buffer_dma() [all …]
|
/freebsd/sys/contrib/dev/mediatek/mt76/ |
H A D | dma.c | 61 t->dma_addr = addr; in mt76_alloc_txwi() 165 dma_unmap_single(dev->dma_dev, t->dma_addr, dev->drv->txwi_size, in mt76_free_pending_txwi() 246 entry->dma_addr[0] = buf->addr; in mt76_dma_add_rx_buf() 286 entry->dma_addr[0] = buf[0].addr; in mt76_dma_add_buf() 291 entry->dma_addr[1] = buf[1].addr; in mt76_dma_add_buf() 327 dma_unmap_single(dev->dma_dev, e->dma_addr[0], e->dma_len[0], in mt76_dma_tx_cleanup_idx() 331 dma_unmap_single(dev->dma_dev, e->dma_addr[1], e->dma_len[1], in mt76_dma_tx_cleanup_idx() 416 dma_sync_single_for_cpu(dev->dma_dev, t->dma_addr, in mt76_dma_get_buf() 421 t->dma_addr = 0; in mt76_dma_get_buf() 437 dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0], in mt76_dma_get_buf() [all …]
|
/freebsd/sys/compat/linuxkpi/common/src/ |
H A D | linux_pci.c | 1356 uint64_t dma_addr; member 1452 obj->dma_addr = seg.ds_addr; in linux_dma_map_phys_common() 1463 return (obj->dma_addr); in linux_dma_map_phys_common() 1485 linux_dma_unmap(struct device *dev, dma_addr_t dma_addr, size_t len) in linux_dma_unmap() argument 1496 obj = LINUX_DMA_PCTRIE_LOOKUP(&priv->ptree, dma_addr); in linux_dma_unmap() 1501 LINUX_DMA_PCTRIE_REMOVE(&priv->ptree, dma_addr); in linux_dma_unmap() 1604 obj = LINUX_DMA_PCTRIE_LOOKUP(&priv->ptree, dma_addr); in linuxkpi_dma_sync() 1730 obj->dma_addr = seg.ds_addr; in dma_pool_obj_ctor() 1862 *handle = obj->dma_addr; in linux_dma_pool_alloc() 1872 obj = LINUX_DMA_PCTRIE_LOOKUP(&pool->pool_ptree, dma_addr); in linux_dma_pool_free() [all …]
|
/freebsd/sys/dev/sfxge/common/ |
H A D | ef10_ev.c | 152 efx_qword_t *dma_addr; in efx_mcdi_init_evq() local 231 dma_addr = MCDI_IN2(req, efx_qword_t, INIT_EVQ_IN_DMA_ADDR); in efx_mcdi_init_evq() 235 EFX_POPULATE_QWORD_2(*dma_addr, in efx_mcdi_init_evq() 239 dma_addr++; in efx_mcdi_init_evq() 289 efx_qword_t *dma_addr; in efx_mcdi_init_evq_v2() local 356 dma_addr = MCDI_IN2(req, efx_qword_t, INIT_EVQ_V2_IN_DMA_ADDR); in efx_mcdi_init_evq_v2() 360 EFX_POPULATE_QWORD_2(*dma_addr, in efx_mcdi_init_evq_v2() 364 dma_addr++; in efx_mcdi_init_evq_v2()
|
H A D | ef10_tx.c | 60 efx_qword_t *dma_addr; in efx_mcdi_init_txq() local 109 dma_addr = MCDI_IN2(req, efx_qword_t, INIT_TXQ_IN_DMA_ADDR); in efx_mcdi_init_txq() 113 EFX_POPULATE_QWORD_2(*dma_addr, in efx_mcdi_init_txq() 117 dma_addr++; in efx_mcdi_init_txq()
|
/freebsd/sys/dev/sfxge/ |
H A D | sfxge_tx.c | 951 uint64_t dma_addr; /* DMA address of current position */ member 1155 uint64_t dma_addr = tso->dma_addr; in tso_fill_packet_with_fragment() local 1183 tso->dma_addr += n; in tso_fill_packet_with_fragment() 1195 efx_tx_qdesc_dma_create(txq->common, dma_addr, n, eop, desc); in tso_fill_packet_with_fragment() 1222 uint64_t dma_addr; in tso_start_new_packet() local 1278 dma_addr = (txq->tsoh_buffer[page_index].esm_addr + in tso_start_new_packet() 1293 tso_map_long_header, &dma_addr, in tso_start_new_packet() 1295 if (__predict_false(dma_addr == 0)) { in tso_start_new_packet() 1346 dma_addr, in tso_start_new_packet() 1379 tso.dma_addr = dma_seg->ds_addr + (tso.header_len - skipped); in sfxge_tx_queue_tso() [all …]
|
/freebsd/sys/dev/mlx5/mlx5_fpga/ |
H A D | mlx5fpga_conn.c | 56 buf->sg[0].dma_addr = dma_map_single(dma_device, buf->sg[0].data, in mlx5_fpga_conn_map_buf() 58 err = dma_mapping_error(dma_device, buf->sg[0].dma_addr); in mlx5_fpga_conn_map_buf() 68 buf->sg[1].dma_addr = dma_map_single(dma_device, buf->sg[1].data, in mlx5_fpga_conn_map_buf() 70 err = dma_mapping_error(dma_device, buf->sg[1].dma_addr); in mlx5_fpga_conn_map_buf() 73 dma_unmap_single(dma_device, buf->sg[0].dma_addr, in mlx5_fpga_conn_map_buf() 89 dma_unmap_single(dma_device, buf->sg[1].dma_addr, in mlx5_fpga_conn_unmap_buf() 93 dma_unmap_single(dma_device, buf->sg[0].dma_addr, in mlx5_fpga_conn_unmap_buf() 117 data->addr = cpu_to_be64(buf->sg[0].dma_addr); in mlx5_fpga_conn_post_recv() 157 data->addr = cpu_to_be64(buf->sg[sgi].dma_addr); in mlx5_fpga_conn_post_send()
|
/freebsd/sys/dev/qlnx/qlnxe/ |
H A D | qlnx_os.h | 143 bus_addr_t dma_addr; member
|
/freebsd/sys/dev/qat/include/common/ |
H A D | adf_transport_internal.h | 19 bus_addr_t dma_addr; member
|
/freebsd/sys/dev/bnxt/bnxt_re/ |
H A D | qplib_sp.c | 102 &sbuf.dma_addr, GFP_KERNEL); in bnxt_qplib_get_dev_attr() 204 sbuf.sb, sbuf.dma_addr); in bnxt_qplib_get_dev_attr() 948 &sbuf.dma_addr, GFP_KERNEL); in bnxt_qplib_query_cc_param() 991 sbuf.sb, sbuf.dma_addr); in bnxt_qplib_query_cc_param() 1014 &sbuf.dma_addr, GFP_KERNEL); in bnxt_qplib_get_roce_error_stats() 1113 sbuf.sb, sbuf.dma_addr); in bnxt_qplib_get_roce_error_stats() 1169 &sbuf.dma_addr, GFP_KERNEL); in bnxt_qplib_qext_stat() 1181 req.resp_addr = cpu_to_le64(sbuf.dma_addr); in bnxt_qplib_qext_stat() 1232 sbuf.sb, sbuf.dma_addr); in bnxt_qplib_qext_stat()
|
/freebsd/sys/contrib/ena-com/ |
H A D | ena_com.h | 201 dma_addr_t dma_addr; member 210 dma_addr_t dma_addr; member 272 dma_addr_t dma_addr; member
|
/freebsd/sys/dev/qlxgbe/ |
H A D | ql_os.h | 167 bus_addr_t dma_addr; member
|