/linux/drivers/media/common/videobuf2/ |
H A D | videobuf2-dma-contig.c | 331 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_attach() 364 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_detach() 387 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_map() 411 return sgt; in vb2_dc_dmabuf_ops_map() 481 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in vb2_dc_get_base_sgt() 482 if (!sgt) { in vb2_dc_get_base_sgt() 495 return sgt; in vb2_dc_get_base_sgt() 538 if (sgt) { in vb2_dc_put_userptr() 626 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in vb2_dc_get_userptr() 627 if (!sgt) { in vb2_dc_get_userptr() [all …]
|
H A D | videobuf2-dma-sg.c | 105 struct sg_table *sgt; in vb2_dma_sg_alloc() local 145 sgt = &buf->sg_table; in vb2_dma_sg_alloc() 382 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_attach() 393 wr = sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach() 415 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_detach() 420 sg_free_table(sgt); in vb2_dma_sg_dmabuf_ops_detach() 431 sgt = &attach->sgt; in vb2_dma_sg_dmabuf_ops_map() 434 return sgt; in vb2_dma_sg_dmabuf_ops_map() 450 return sgt; in vb2_dma_sg_dmabuf_ops_map() 472 dma_sync_sg_for_cpu(buf->dev, sgt->sgl, sgt->nents, buf->dma_dir); in vb2_dma_sg_dmabuf_ops_begin_cpu_access() [all …]
|
H A D | videobuf2-vmalloc.c | 208 struct sg_table sgt; member 218 struct sg_table *sgt; in vb2_vmalloc_dmabuf_ops_attach() local 228 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_attach() 238 sg_free_table(sgt); in vb2_vmalloc_dmabuf_ops_attach() 255 struct sg_table *sgt; in vb2_vmalloc_dmabuf_ops_detach() local 260 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_detach() 265 sg_free_table(sgt); in vb2_vmalloc_dmabuf_ops_detach() 274 struct sg_table *sgt; in vb2_vmalloc_dmabuf_ops_map() local 276 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_map() 279 return sgt; in vb2_vmalloc_dmabuf_ops_map() [all …]
|
/linux/drivers/gpu/drm/tests/ |
H A D | drm_gem_shmem_test.c | 38 struct sg_table *sgt = ptr; in sg_free_table_wrapper() local 40 sg_free_table(sgt); in sg_free_table_wrapper() 83 struct sg_table *sgt; in drm_gem_shmem_test_obj_create_private() local 91 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in drm_gem_shmem_test_obj_create_private() 92 KUNIT_ASSERT_NOT_NULL(test, sgt); in drm_gem_shmem_test_obj_create_private() 120 KUNIT_EXPECT_PTR_EQ(test, shmem->sgt, sgt); in drm_gem_shmem_test_obj_create_private() 203 struct sg_table *sgt; in drm_gem_shmem_test_get_pages_sgt() local 224 for_each_sgtable_sg(sgt, sg, si) { in drm_gem_shmem_test_get_pages_sgt() 242 struct sg_table *sgt; in drm_gem_shmem_test_get_sg_table() local 257 KUNIT_EXPECT_PTR_EQ(test, sgt, shmem->sgt); in drm_gem_shmem_test_get_sg_table() [all …]
|
/linux/drivers/gpu/drm/tegra/ |
H A D | gem.c | 49 return sg_dma_count_chunks(sgt->sgl, sgt->nents); in sgt_dma_count_chunks() 92 map->sgt = NULL; in tegra_bo_pin() 106 map->sgt = kzalloc(sizeof(*map->sgt), GFP_KERNEL); in tegra_bo_pin() 107 if (!map->sgt) { in tegra_bo_pin() 156 kfree(map->sgt); in tegra_bo_pin() 367 kfree(bo->sgt); in tegra_bo_get_pages() 632 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in tegra_gem_prime_map_dma_buf() 633 if (!sgt) in tegra_gem_prime_map_dma_buf() 649 return sgt; in tegra_gem_prime_map_dma_buf() 653 kfree(sgt); in tegra_gem_prime_map_dma_buf() [all …]
|
/linux/drivers/gpu/drm/armada/ |
H A D | armada_gem.c | 68 if (dobj->sgt) in armada_gem_free_object() 392 struct sg_table *sgt; in armada_gem_prime_map_dma_buf() local 395 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in armada_gem_prime_map_dma_buf() 396 if (!sgt) in armada_gem_prime_map_dma_buf() 439 return sgt; in armada_gem_prime_map_dma_buf() 446 sg_free_table(sgt); in armada_gem_prime_map_dma_buf() 448 kfree(sgt); in armada_gem_prime_map_dma_buf() 469 sg_free_table(sgt); in armada_gem_prime_unmap_dma_buf() 470 kfree(sgt); in armada_gem_prime_unmap_dma_buf() 544 if (IS_ERR(dobj->sgt)) { in armada_gem_map_import() [all …]
|
/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_vram.c | 75 struct sg_table *sgt; in virtio_gpu_vram_map_dma_buf() local 79 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in virtio_gpu_vram_map_dma_buf() 80 if (!sgt) in virtio_gpu_vram_map_dma_buf() 90 return sgt; in virtio_gpu_vram_map_dma_buf() 108 return sgt; in virtio_gpu_vram_map_dma_buf() 110 sg_free_table(sgt); in virtio_gpu_vram_map_dma_buf() 111 kfree(sgt); in virtio_gpu_vram_map_dma_buf() 116 struct sg_table *sgt, in virtio_gpu_vram_unmap_dma_buf() argument 119 if (sgt->nents) { in virtio_gpu_vram_unmap_dma_buf() 124 sg_free_table(sgt); in virtio_gpu_vram_unmap_dma_buf() [all …]
|
/linux/drivers/gpu/drm/i915/ |
H A D | i915_mm.c | 37 struct sgt_iter sgt; member 46 return (r->sgt.dma + r->sgt.curr + r->iobase) >> PAGE_SHIFT; in sgt_pfn() 48 return r->sgt.pfn + (r->sgt.curr >> PAGE_SHIFT); in sgt_pfn() 55 if (GEM_WARN_ON(!r->sgt.sgp)) in remap_sg() 63 r->sgt.curr += PAGE_SIZE; in remap_sg() 64 if (r->sgt.curr >= r->sgt.max) in remap_sg() 65 r->sgt = __sgt_iter(__sg_next(r->sgt.sgp), use_dma(r->iobase)); in remap_sg() 136 .sgt = __sgt_iter(sgl, use_dma(iobase)), in remap_io_sg()
|
/linux/drivers/gpu/drm/mediatek/ |
H A D | mtk_gem.c | 200 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in mtk_gem_prime_get_sg_table() 201 if (!sgt) in mtk_gem_prime_get_sg_table() 209 kfree(sgt); in mtk_gem_prime_get_sg_table() 213 return sgt; in mtk_gem_prime_get_sg_table() 247 if (IS_ERR(sgt)) in mtk_gem_prime_vmap() 253 sg_free_table(sgt); in mtk_gem_prime_vmap() 254 kfree(sgt); in mtk_gem_prime_vmap() 263 sg_free_table(sgt); in mtk_gem_prime_vmap() 264 kfree(sgt); in mtk_gem_prime_vmap() 268 sg_free_table(sgt); in mtk_gem_prime_vmap() [all …]
|
/linux/drivers/hwtracing/intel_th/ |
H A D | msu-sink.c | 51 static int msu_sink_alloc_window(void *data, struct sg_table **sgt, size_t size) in msu_sink_alloc_window() argument 64 ret = sg_alloc_table(*sgt, nents, GFP_KERNEL); in msu_sink_alloc_window() 68 priv->sgts[priv->nr_sgts++] = *sgt; in msu_sink_alloc_window() 70 for_each_sg((*sgt)->sgl, sg_ptr, nents, i) { in msu_sink_alloc_window() 84 static void msu_sink_free_window(void *data, struct sg_table *sgt) in msu_sink_free_window() argument 90 for_each_sg(sgt->sgl, sg_ptr, sgt->nents, i) { in msu_sink_free_window() 95 sg_free_table(sgt); in msu_sink_free_window() 99 static int msu_sink_ready(void *data, struct sg_table *sgt, size_t bytes) in msu_sink_ready() argument 103 intel_th_msc_window_unlock(priv->dev, sgt); in msu_sink_ready()
|
/linux/drivers/xen/ |
H A D | gntdev-dmabuf.c | 201 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in dmabuf_pages_to_sgt() 202 if (!sgt) { in dmabuf_pages_to_sgt() 213 return sgt; in dmabuf_pages_to_sgt() 216 kfree(sgt); in dmabuf_pages_to_sgt() 241 struct sg_table *sgt = gntdev_dmabuf_attach->sgt; in dmabuf_exp_ops_detach() local 243 if (sgt) { in dmabuf_exp_ops_detach() 251 kfree(sgt); in dmabuf_exp_ops_detach() 288 kfree(sgt); in dmabuf_exp_ops_map_dma_buf() 291 gntdev_dmabuf_attach->sgt = sgt; in dmabuf_exp_ops_map_dma_buf() 297 return sgt; in dmabuf_exp_ops_map_dma_buf() [all …]
|
/linux/kernel/dma/ |
H A D | mapping.c | 669 sgt = kmalloc(sizeof(*sgt), gfp); in alloc_single_sgt() 670 if (!sgt) in alloc_single_sgt() 678 sg_dma_len(sgt->sgl) = sgt->sgl->length; in alloc_single_sgt() 679 return sgt; in alloc_single_sgt() 683 kfree(sgt); in alloc_single_sgt() 703 if (sgt) { in dma_alloc_noncontiguous() 704 sgt->nents = 1; in dma_alloc_noncontiguous() 705 debug_dma_map_sg(dev, sgt->sgl, sgt->orig_nents, 1, dir, attrs); in dma_alloc_noncontiguous() 707 return sgt; in dma_alloc_noncontiguous() 717 kfree(sgt); in free_single_sgt() [all …]
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_dmabuf.c | 32 struct sg_table *sgt; in i915_gem_map_dma_buf() local 40 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in i915_gem_map_dma_buf() 41 if (!sgt) { in i915_gem_map_dma_buf() 50 dst = sgt->sgl; in i915_gem_map_dma_buf() 60 return sgt; in i915_gem_map_dma_buf() 63 sg_free_table(sgt); in i915_gem_map_dma_buf() 65 kfree(sgt); in i915_gem_map_dma_buf() 240 struct sg_table *sgt; in i915_gem_object_get_pages_dmabuf() local 246 if (IS_ERR(sgt)) in i915_gem_object_get_pages_dmabuf() 247 return PTR_ERR(sgt); in i915_gem_object_get_pages_dmabuf() [all …]
|
/linux/drivers/media/platform/nvidia/tegra-vde/ |
H A D | dmabuf-cache.c | 27 struct sg_table *sgt; member 72 struct sg_table *sgt; in tegra_vde_dmabuf_cache_map() local 93 *addrp = sg_dma_address(entry->sgt->sgl); in tegra_vde_dmabuf_cache_map() 105 sgt = dma_buf_map_attachment_unlocked(attachment, dma_dir); in tegra_vde_dmabuf_cache_map() 106 if (IS_ERR(sgt)) { in tegra_vde_dmabuf_cache_map() 108 err = PTR_ERR(sgt); in tegra_vde_dmabuf_cache_map() 112 if (!vde->domain && sgt->nents > 1) { in tegra_vde_dmabuf_cache_map() 125 err = tegra_vde_iommu_map(vde, sgt, &iova, dmabuf->size); in tegra_vde_dmabuf_cache_map() 131 *addrp = sg_dma_address(sgt->sgl); in tegra_vde_dmabuf_cache_map() 141 entry->sgt = sgt; in tegra_vde_dmabuf_cache_map() [all …]
|
/linux/drivers/gpu/drm/ |
H A D | drm_gem_shmem_helper.c | 148 if (shmem->sgt) { in drm_gem_shmem_free() 152 kfree(shmem->sgt); in drm_gem_shmem_free() 454 kfree(shmem->sgt); in drm_gem_shmem_purge() 455 shmem->sgt = NULL; in drm_gem_shmem_purge() 683 if (shmem->sgt) in drm_gem_shmem_get_pages_sgt_locked() 693 if (IS_ERR(sgt)) { in drm_gem_shmem_get_pages_sgt_locked() 702 shmem->sgt = sgt; in drm_gem_shmem_get_pages_sgt_locked() 704 return sgt; in drm_gem_shmem_get_pages_sgt_locked() 708 kfree(sgt); in drm_gem_shmem_get_pages_sgt_locked() 741 return sgt; in drm_gem_shmem_get_pages_sgt() [all …]
|
H A D | drm_prime.c | 641 if (IS_ERR(sgt)) in drm_gem_map_dma_buf() 642 return sgt; in drm_gem_map_dma_buf() 647 sg_free_table(sgt); in drm_gem_map_dma_buf() 648 kfree(sgt); in drm_gem_map_dma_buf() 649 sgt = ERR_PTR(ret); in drm_gem_map_dma_buf() 652 return sgt; in drm_gem_map_dma_buf() 668 if (!sgt) in drm_gem_unmap_dma_buf() 672 sg_free_table(sgt); in drm_gem_unmap_dma_buf() 673 kfree(sgt); in drm_gem_unmap_dma_buf() 941 if (IS_ERR(sgt)) { in drm_gem_prime_import_dev() [all …]
|
H A D | drm_gem_dma_helper.c | 234 drm_prime_gem_destroy(gem_obj, dma_obj->sgt); in drm_gem_dma_free() 424 struct sg_table *sgt; in drm_gem_dma_get_sg_table() local 427 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in drm_gem_dma_get_sg_table() 428 if (!sgt) in drm_gem_dma_get_sg_table() 431 ret = dma_get_sgtable(obj->dev->dev, sgt, dma_obj->vaddr, in drm_gem_dma_get_sg_table() 436 return sgt; in drm_gem_dma_get_sg_table() 439 kfree(sgt); in drm_gem_dma_get_sg_table() 464 struct sg_table *sgt) in drm_gem_dma_prime_import_sg_table() argument 477 dma_obj->dma_addr = sg_dma_address(sgt->sgl); in drm_gem_dma_prime_import_sg_table() 478 dma_obj->sgt = sgt; in drm_gem_dma_prime_import_sg_table() [all …]
|
/linux/net/ceph/ |
H A D | crypto.c | 173 memset(sgt, 0, sizeof(*sgt)); in setup_sgtable() 183 ret = sg_alloc_table(sgt, chunk_cnt, GFP_NOFS); in setup_sgtable() 189 sgt->sgl = prealloc_sg; in setup_sgtable() 190 sgt->nents = sgt->orig_nents = 1; in setup_sgtable() 193 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) { in setup_sgtable() 213 static void teardown_sgtable(struct sg_table *sgt) in teardown_sgtable() argument 215 if (sgt->orig_nents > 1) in teardown_sgtable() 216 sg_free_table(sgt); in teardown_sgtable() 223 struct sg_table sgt; in ceph_aes_crypt() local 240 skcipher_request_set_crypt(req, sgt.sgl, sgt.sgl, crypt_len, iv); in ceph_aes_crypt() [all …]
|
/linux/include/linux/ |
H A D | scatterlist.h | 46 struct sg_table sgt; /* The scatter list table */ member 201 #define for_each_sgtable_sg(sgt, sg, i) \ argument 202 for_each_sg((sgt)->sgl, sg, (sgt)->orig_nents, i) 209 #define for_each_sgtable_dma_sg(sgt, sg, i) \ argument 210 for_each_sg((sgt)->sgl, sg, (sgt)->nents, i) 437 void sg_free_append_table(struct sg_append_table *sgt); 441 int sg_alloc_append_table_from_pages(struct sg_append_table *sgt, 637 #define for_each_sgtable_page(sgt, piter, pgoffset) \ argument 638 for_each_sg_page((sgt)->sgl, piter, (sgt)->orig_nents, pgoffset) 651 #define for_each_sgtable_dma_page(sgt, dma_iter, pgoffset) \ argument [all …]
|
/linux/drivers/infiniband/core/ |
H A D | umem_dmabuf.c | 17 struct sg_table *sgt; in ib_umem_dmabuf_map_pages() local 26 if (umem_dmabuf->sgt) in ib_umem_dmabuf_map_pages() 29 sgt = dma_buf_map_attachment(umem_dmabuf->attach, in ib_umem_dmabuf_map_pages() 31 if (IS_ERR(sgt)) in ib_umem_dmabuf_map_pages() 32 return PTR_ERR(sgt); in ib_umem_dmabuf_map_pages() 39 for_each_sgtable_dma_sg(sgt, sg, i) { in ib_umem_dmabuf_map_pages() 62 umem_dmabuf->umem.sgt_append.sgt.sgl = umem_dmabuf->first_sg; in ib_umem_dmabuf_map_pages() 63 umem_dmabuf->umem.sgt_append.sgt.nents = nmap; in ib_umem_dmabuf_map_pages() 64 umem_dmabuf->sgt = sgt; in ib_umem_dmabuf_map_pages() 87 if (!umem_dmabuf->sgt) in ib_umem_dmabuf_unmap_pages() [all …]
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_dma_buf.c | 95 struct sg_table *sgt; in xe_dma_buf_map() local 112 sgt = drm_prime_pages_to_sg(obj->dev, in xe_dma_buf_map() 115 if (IS_ERR(sgt)) in xe_dma_buf_map() 116 return sgt; in xe_dma_buf_map() 128 dir, &sgt); in xe_dma_buf_map() 136 return sgt; in xe_dma_buf_map() 139 sg_free_table(sgt); in xe_dma_buf_map() 140 kfree(sgt); in xe_dma_buf_map() 145 struct sg_table *sgt, in xe_dma_buf_unmap() argument 153 sg_free_table(sgt); in xe_dma_buf_unmap() [all …]
|
/linux/drivers/gpu/drm/imagination/ |
H A D | pvr_gem.c | 216 if (shmem_obj->sgt) in pvr_gem_object_vmap() 217 dma_sync_sgtable_for_cpu(dev, shmem_obj->sgt, DMA_BIDIRECTIONAL); in pvr_gem_object_vmap() 256 if (shmem_obj->sgt) in pvr_gem_object_vunmap() 257 dma_sync_sgtable_for_device(dev, shmem_obj->sgt, DMA_BIDIRECTIONAL); in pvr_gem_object_vunmap() 341 struct sg_table *sgt; in pvr_gem_object_create() local 357 sgt = drm_gem_shmem_get_pages_sgt(shmem_obj); in pvr_gem_object_create() 358 if (IS_ERR(sgt)) { in pvr_gem_object_create() 359 err = PTR_ERR(sgt); in pvr_gem_object_create() 363 dma_sync_sgtable_for_device(shmem_obj->base.dev->dev, sgt, in pvr_gem_object_create() 400 WARN_ON(!shmem_obj->sgt); in pvr_gem_get_dma_addr() [all …]
|
/linux/drivers/gpu/drm/lima/ |
H A D | lima_gem.c | 29 struct sg_table sgt; in lima_heap_alloc() local 72 if (bo->base.sgt) { in lima_heap_alloc() 74 sg_free_table(bo->base.sgt); in lima_heap_alloc() 76 bo->base.sgt = kmalloc(sizeof(*bo->base.sgt), GFP_KERNEL); in lima_heap_alloc() 77 if (!bo->base.sgt) { in lima_heap_alloc() 87 *bo->base.sgt = sgt; in lima_heap_alloc() 101 kfree(bo->base.sgt); in lima_heap_alloc() 102 bo->base.sgt = NULL; in lima_heap_alloc() 104 sg_free_table(&sgt); in lima_heap_alloc() 138 if (IS_ERR(sgt)) { in lima_gem_create_handle() [all …]
|
/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem.c | 23 struct sg_table *sgt = etnaviv_obj->sgt; in etnaviv_gem_scatter_map() local 36 struct sg_table *sgt = etnaviv_obj->sgt; in etnaviv_gem_scatterlist_unmap() local 75 if (etnaviv_obj->sgt) { in put_pages() 78 kfree(etnaviv_obj->sgt); in put_pages() 79 etnaviv_obj->sgt = NULL; in put_pages() 101 if (!etnaviv_obj->sgt) { in etnaviv_gem_get_pages() 104 struct sg_table *sgt; in etnaviv_gem_get_pages() local 107 if (IS_ERR(sgt)) { in etnaviv_gem_get_pages() 109 PTR_ERR(sgt)); in etnaviv_gem_get_pages() 110 return ERR_CAST(sgt); in etnaviv_gem_get_pages() [all …]
|
/linux/drivers/gpu/drm/rockchip/ |
H A D | rockchip_drm_gem.c | 92 if (IS_ERR(rk_obj->sgt)) { in rockchip_gem_get_pages() 93 ret = PTR_ERR(rk_obj->sgt); in rockchip_gem_get_pages() 118 sg_free_table(rk_obj->sgt); in rockchip_gem_put_pages() 119 kfree(rk_obj->sgt); in rockchip_gem_put_pages() 430 struct sg_table *sgt; in rockchip_gem_prime_get_sg_table() local 436 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in rockchip_gem_prime_get_sg_table() 437 if (!sgt) in rockchip_gem_prime_get_sg_table() 445 kfree(sgt); in rockchip_gem_prime_get_sg_table() 449 return sgt; in rockchip_gem_prime_get_sg_table() 458 rk_obj->sgt = sg; in rockchip_gem_iommu_map_sg() [all …]
|