/dragonfly/sys/dev/drm/radeon/ |
H A D | radeon_ttm.c | 549 unsigned pinned = 0, nents; 590 nents = dma_map_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction); 591 if (nents != ttm->sg->nents) 622 dma_unmap_sg(rdev->dev, ttm->sg->sgl, ttm->sg->nents, direction); 624 for_each_sg_page(ttm->sg->sgl, &sg_iter, ttm->sg->nents, 0) {
|
/dragonfly/sys/dev/drm/ |
H A D | linux_scatterlist.c | 56 __sg_alloc_table(struct sg_table *table, unsigned int nents, in __sg_alloc_table() argument 64 if (nents == 0) in __sg_alloc_table() 66 left = nents; in __sg_alloc_table() 88 table->nents = ++table->orig_nents; in __sg_alloc_table() 94 table->nents = table->orig_nents += sg_size; in __sg_alloc_table() 154 sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents, in sg_pcopy_from_buffer() argument 165 for_each_sg_page(sgl, &iter, nents, 0) { in sg_pcopy_from_buffer() 190 sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents, in sg_pcopy_to_buffer() argument 201 for_each_sg_page(sgl, &iter, nents, 0) { in sg_pcopy_to_buffer()
|
H A D | drm_prime.c | 221 dma_unmap_sg(attach->dev, sgt->sgl, sgt->nents, in drm_gem_map_detach() 280 if (!dma_map_sg(attach->dev, sgt->sgl, sgt->nents, dir)) { in drm_gem_map_dma_buf() 846 for_each_sg(sgt->sgl, sg, sgt->nents, count) { in drm_prime_sg_to_page_addr_arrays()
|
H A D | drm_cache.c | 113 for_each_sg_page(st->sgl, &sg_iter, st->nents, 0) in drm_clflush_sg()
|
/dragonfly/sys/dev/drm/include/linux/ |
H A D | scatterlist.h | 111 bzero(sg, sizeof(*sg) * nents); in sg_init_table() 112 sg[nents - 1].flags = SG_END; in sg_init_table() 230 int nents; in sg_nents() local 231 for (nents = 0; sg; sg = sg_next(sg)) in sg_nents() 232 nents++; in sg_nents() 233 return nents; in sg_nents() 238 struct scatterlist *sglist, unsigned int nents, in __sg_page_iter_start() argument 242 piter->__nents = nents; in __sg_page_iter_start() 305 unsigned int nents, unsigned long pgoffset) in _sg_iter_init() argument 307 if (nents) { in _sg_iter_init() [all …]
|
H A D | dma-mapping.h | 61 int nents, enum dma_data_direction dir) in dma_map_sg() argument 66 for_each_sg(sg, s, nents, i) in dma_map_sg() 69 return nents; in dma_map_sg() 74 int nents, enum dma_data_direction dir) in dma_unmap_sg() argument
|
/dragonfly/sys/dev/drm/i915/ |
H A D | intel_guc_fw.c | 133 sg_pcopy_to_buffer(sg->sgl, sg->nents, rsa, sizeof(rsa), offset); in guc_ucode_xfer_dma()
|
H A D | i915_gem.c | 2618 if (orig_st->nents == orig_st->orig_nents) in i915_sg_trim() 2621 if (sg_alloc_table(&new_st, orig_st->nents, GFP_KERNEL | __GFP_NOWARN)) in i915_sg_trim() 2625 for_each_sg(orig_st->sgl, sg, orig_st->nents, i) { in i915_sg_trim() 2685 st->nents = 0; in i915_gem_object_get_pages_gtt() 2743 st->nents++; in i915_gem_object_get_pages_gtt()
|
H A D | i915_gem_dmabuf.c | 58 ret = sg_alloc_table(st, obj->mm.pages->nents, GFP_KERNEL); in i915_gem_map_dma_buf() 64 for (i = 0; i < obj->mm.pages->nents; i++) { in i915_gem_map_dma_buf() 70 if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { in i915_gem_map_dma_buf() 93 dma_unmap_sg(attachment->dev, sg->sgl, sg->nents, dir); in i915_gem_unmap_dma_buf()
|
H A D | i915_gem_gtt.c | 2304 pages->sgl, pages->nents, in i915_gem_gtt_prepare_pages() 2692 dma_unmap_sg(kdev, pages->sgl, pages->nents, PCI_DMA_BIDIRECTIONAL); in i915_gem_gtt_finish_pages() 3668 st->nents++; in rotate_pages() 3720 st->nents = 0; in intel_rotate_pages() 3769 st->nents = 0; in intel_partial_pages() 3780 st->nents++; in intel_partial_pages()
|
H A D | i915_gem_internal.c | 90 st->nents = 0; in i915_gem_object_get_pages_internal() 110 st->nents++; in i915_gem_object_get_pages_internal()
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_ttm.c | 939 unsigned nents; in amdgpu_ttm_tt_pin_userptr() local 955 nents = dma_map_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction); in amdgpu_ttm_tt_pin_userptr() 956 if (nents != ttm->sg->nents) in amdgpu_ttm_tt_pin_userptr() 988 dma_unmap_sg(adev->dev, ttm->sg->sgl, ttm->sg->nents, direction); in amdgpu_ttm_tt_unpin_userptr()
|
/dragonfly/sys/dev/agp/ |
H A D | intel-gtt.c | 1425 for_each_sg(st->sgl, sg, st->nents, j) { in intel_gtt_insert_sg_entries()
|
/dragonfly/contrib/gmp/mpn/generic/ |
H A D | powm_sec.c | 310 mp_size_t nents, mp_size_t which) in mpn_tabselect() argument 316 for (k = 0; k < nents; k++) in mpn_tabselect()
|