Home
last modified time | relevance | path

Searched refs:dma_mem (Results 1 – 18 of 18) sorted by relevance

/linux/kernel/dma/
H A Dcoherent.c25 if (dev && dev->dma_mem) in dev_get_coherent_memory()
26 return dev->dma_mem; in dev_get_coherent_memory()
53 if (!dma_mem) in dma_init_coherent_memory()
56 if (!dma_mem->bitmap) in dma_init_coherent_memory()
62 dma_mem->size = pages; in dma_init_coherent_memory()
66 return dma_mem; in dma_init_coherent_memory()
69 kfree(dma_mem); in dma_init_coherent_memory()
93 if (dev->dma_mem) in dma_assign_coherent_memory()
96 dev->dma_mem = mem; in dma_assign_coherent_memory()
137 dev->dma_mem = NULL; in dma_release_coherent_memory()
[all …]
/linux/drivers/infiniband/hw/irdma/
H A Dhmc.c478 struct irdma_dma_mem dma_mem; in irdma_add_sd_table_entry() local
489 dma_mem.size = ALIGN(alloc_len, IRDMA_HMC_PD_BP_BUF_ALIGNMENT); in irdma_add_sd_table_entry()
490 dma_mem.va = dma_alloc_coherent(hw->device, dma_mem.size, in irdma_add_sd_table_entry()
491 &dma_mem.pa, GFP_KERNEL); in irdma_add_sd_table_entry()
492 if (!dma_mem.va) in irdma_add_sd_table_entry()
501 dma_free_coherent(hw->device, dma_mem.size, in irdma_add_sd_table_entry()
502 dma_mem.va, dma_mem.pa); in irdma_add_sd_table_entry()
503 dma_mem.va = NULL; in irdma_add_sd_table_entry()
508 memcpy(&sd_entry->u.pd_table.pd_page_addr, &dma_mem, in irdma_add_sd_table_entry()
511 memcpy(&sd_entry->u.bp.addr, &dma_mem, in irdma_add_sd_table_entry()
H A Dverbs.h151 struct irdma_dma_mem dma_mem; member
H A Dutils.c1162 dma_free_coherent(rf->sc_dev.hw->device, iwqp->kqp.dma_mem.size, in irdma_free_qp_rsrc()
1163 iwqp->kqp.dma_mem.va, iwqp->kqp.dma_mem.pa); in irdma_free_qp_rsrc()
1164 iwqp->kqp.dma_mem.va = NULL; in irdma_free_qp_rsrc()
H A Dverbs.c667 struct irdma_dma_mem *mem = &iwqp->kqp.dma_mem; in irdma_setup_kmode_qp()
/linux/drivers/scsi/be2iscsi/
H A Dbe.h31 struct be_dma_mem dma_mem; member
53 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node()
58 return q->dma_mem.va + wrb_num * q->entry_size; in queue_get_wrb()
63 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node()
H A Dbe_cmds.c741 struct be_dma_mem *q_mem = &eq->dma_mem; in beiscsi_cmd_eq_create()
783 struct be_dma_mem *q_mem = &cq->dma_mem; in beiscsi_cmd_cq_create()
856 struct be_dma_mem *q_mem = &mccq->dma_mem; in beiscsi_cmd_mccq_create()
982 struct be_dma_mem *q_mem = &dq->dma_mem; in be_cmd_create_default_pdu_queue()
H A Dbe_main.c2998 struct be_dma_mem *mem = &q->dma_mem; in be_fill_queue()
3030 mem = &eq->dma_mem; in beiscsi_create_eqs()
3067 mem = &eq->dma_mem; in beiscsi_create_eqs()
3096 mem = &cq->dma_mem; in beiscsi_create_cqs()
3130 mem = &cq->dma_mem; in beiscsi_create_cqs()
3155 mem = &dq->dma_mem; in beiscsi_create_def_hdr()
3207 mem = &dataq->dma_mem; in beiscsi_create_def_data()
3322 struct be_dma_mem *mem = &q->dma_mem; in be_queue_free()
3333 struct be_dma_mem *mem = &q->dma_mem; in be_queue_alloc()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_virtchnl.c223 if (dma_mem) in idpf_mb_clean()
225 dma_mem->va, dma_mem->pa); in idpf_mb_clean()
227 kfree(dma_mem); in idpf_mb_clean()
271 dma_mem = kzalloc(sizeof(*dma_mem), GFP_ATOMIC); in idpf_send_mb_msg()
272 if (!dma_mem) { in idpf_send_mb_msg()
303 dma_free_coherent(&adapter->pdev->dev, dma_mem->size, dma_mem->va, in idpf_send_mb_msg()
304 dma_mem->pa); in idpf_send_mb_msg()
306 kfree(dma_mem); in idpf_send_mb_msg()
711 dma_mem = NULL; in idpf_recv_mb_msg()
726 if (dma_mem) in idpf_recv_mb_msg()
[all …]
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_cxt.c776 if (!p_t2 || !p_t2->dma_mem) in qed_cxt_src_t2_free()
780 if (p_t2->dma_mem[i].virt_addr) in qed_cxt_src_t2_free()
782 p_t2->dma_mem[i].size, in qed_cxt_src_t2_free()
783 p_t2->dma_mem[i].virt_addr, in qed_cxt_src_t2_free()
784 p_t2->dma_mem[i].phys_addr); in qed_cxt_src_t2_free()
786 kfree(p_t2->dma_mem); in qed_cxt_src_t2_free()
787 p_t2->dma_mem = NULL; in qed_cxt_src_t2_free()
797 if (!p_t2 || !p_t2->dma_mem) in qed_cxt_t2_alloc_pages()
808 if (!p_t2->dma_mem[i].virt_addr) in qed_cxt_t2_alloc_pages()
812 p_t2->dma_mem[i].size = size; in qed_cxt_t2_alloc_pages()
[all …]
H A Dqed_cxt.h303 struct phys_mem_desc *dma_mem; member
/linux/drivers/scsi/
H A Dstex.c315 void *dma_mem; member
422 return (struct req_msg *)(hba->dma_mem + in stex_ss_alloc_req()
1148 msg_h = (struct st_msg_header *)hba->dma_mem; in stex_ss_handshake()
1738 hba->dma_mem = dma_alloc_coherent(&pdev->dev, in stex_probe()
1740 if (!hba->dma_mem) { in stex_probe()
1749 hba->dma_mem = dma_alloc_coherent(&pdev->dev, in stex_probe()
1753 if (!hba->dma_mem) { in stex_probe()
1770 hba->scratch = (__le32 *)(hba->dma_mem + scratch_offset); in stex_probe()
1772 hba->copy_buffer = hba->dma_mem + cp_offset; in stex_probe()
1841 hba->dma_mem, hba->dma_handle); in stex_probe()
[all …]
/linux/drivers/net/ethernet/emulex/benet/
H A Dbe.h130 struct be_dma_mem dma_mem; member
152 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node()
157 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node()
162 return q->dma_mem.va + index * q->entry_size; in queue_index_node()
H A Dbe_cmds.c999 struct be_dma_mem *q_mem = &eqo->q.dma_mem; in be_cmd_eq_create()
1164 struct be_dma_mem *q_mem = &cq->dma_mem; in be_cmd_cq_create()
1242 struct be_dma_mem *q_mem = &mccq->dma_mem; in be_cmd_mccq_ext_create()
1307 struct be_dma_mem *q_mem = &mccq->dma_mem; in be_cmd_mccq_org_create()
1366 struct be_dma_mem *q_mem = &txq->dma_mem; in be_cmd_txq_create()
1414 struct be_dma_mem *q_mem = &rxq->dma_mem; in be_cmd_rxq_create()
H A Dbe_main.c146 struct be_dma_mem *mem = &q->dma_mem; in be_queue_free()
158 struct be_dma_mem *mem = &q->dma_mem; in be_queue_alloc()
1436 entry = txo->q.dma_mem.va; in be_tx_timeout()
1446 entry = txo->cq.dma_mem.va; in be_tx_timeout()
/linux/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-eth.c3686 void *dma_mem, *key, *mask; in dpaa2_eth_set_vlan_qos() local
3700 dma_mem = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL); in dpaa2_eth_set_vlan_qos()
3701 if (!dma_mem) in dpaa2_eth_set_vlan_qos()
3710 err = dpni_prepare_key_cfg(&kg_cfg, dma_mem); in dpaa2_eth_set_vlan_qos()
3719 qos_cfg.key_cfg_iova = dma_map_single(dev, dma_mem, in dpaa2_eth_set_vlan_qos()
3786 kfree(dma_mem); in dpaa2_eth_set_vlan_qos()
4232 u8 *dma_mem; in dpaa2_eth_set_dist_key() local
4262 dma_mem = kzalloc(DPAA2_CLASSIFIER_DMA_SIZE, GFP_KERNEL); in dpaa2_eth_set_dist_key()
4263 if (!dma_mem) in dpaa2_eth_set_dist_key()
4266 err = dpni_prepare_key_cfg(&cls_cfg, dma_mem); in dpaa2_eth_set_dist_key()
[all …]
H A Ddpaa2-switch.c846 u8 *dma_mem; in dpaa2_switch_fdb_iterate() local
850 dma_mem = kzalloc(fdb_dump_size, GFP_KERNEL); in dpaa2_switch_fdb_iterate()
851 if (!dma_mem) in dpaa2_switch_fdb_iterate()
854 fdb_dump_iova = dma_map_single(dev, dma_mem, fdb_dump_size, in dpaa2_switch_fdb_iterate()
872 fdb_entries = (struct fdb_dump_entry *)dma_mem; in dpaa2_switch_fdb_iterate()
882 kfree(dma_mem); in dpaa2_switch_fdb_iterate()
889 kfree(dma_mem); in dpaa2_switch_fdb_iterate()
/linux/include/linux/
H A Ddevice.h768 struct dma_coherent_mem *dma_mem; /* internal for coherent mem member