Home
last modified time | relevance | path

Searched refs:ndw (Results 1 – 13 of 13) sorted by relevance

/dragonfly/sys/dev/drm/radeon/
H A Dni_dma.c318 unsigned ndw; in cayman_dma_vm_copy_pages() local
321 ndw = count * 2; in cayman_dma_vm_copy_pages()
323 ndw = 0xFFFFE; in cayman_dma_vm_copy_pages()
332 pe += ndw * 4; in cayman_dma_vm_copy_pages()
333 src += ndw * 4; in cayman_dma_vm_copy_pages()
358 unsigned ndw; in cayman_dma_vm_write_pages() local
363 ndw = 0xFFFFE; in cayman_dma_vm_write_pages()
370 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in cayman_dma_vm_write_pages()
406 unsigned ndw; in cayman_dma_vm_set_pages() local
411 ndw = 0xFFFFE; in cayman_dma_vm_set_pages()
[all …]
H A Dsi_dma.c112 unsigned ndw; in si_dma_vm_write_pages() local
115 ndw = count * 2; in si_dma_vm_write_pages()
116 if (ndw > 0xFFFFE) in si_dma_vm_write_pages()
117 ndw = 0xFFFFE; in si_dma_vm_write_pages()
123 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in si_dma_vm_write_pages()
159 unsigned ndw; in si_dma_vm_set_pages() local
162 ndw = count * 2; in si_dma_vm_set_pages()
163 if (ndw > 0xFFFFE) in si_dma_vm_set_pages()
164 ndw = 0xFFFFE; in si_dma_vm_set_pages()
181 pe += ndw * 4; in si_dma_vm_set_pages()
[all …]
H A Dcik_sdma.c846 unsigned ndw; in cik_sdma_vm_write_pages() local
849 ndw = count * 2; in cik_sdma_vm_write_pages()
850 if (ndw > 0xFFFFE) in cik_sdma_vm_write_pages()
851 ndw = 0xFFFFE; in cik_sdma_vm_write_pages()
859 for (; ndw > 0; ndw -= 2, --count, pe += 8) { in cik_sdma_vm_write_pages()
895 unsigned ndw; in cik_sdma_vm_set_pages() local
898 ndw = count; in cik_sdma_vm_set_pages()
899 if (ndw > 0x7FFFF) in cik_sdma_vm_set_pages()
900 ndw = 0x7FFFF; in cik_sdma_vm_set_pages()
919 pe += ndw * 8; in cik_sdma_vm_set_pages()
[all …]
H A Dradeon_ring.c104 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) in radeon_ring_alloc() argument
109 if (ndw > (ring->ring_size / 4)) in radeon_ring_alloc()
114 ndw = (ndw + ring->align_mask) & ~ring->align_mask; in radeon_ring_alloc()
115 while (ndw > (ring->ring_free_dw - 1)) { in radeon_ring_alloc()
117 if (ndw < ring->ring_free_dw) { in radeon_ring_alloc()
124 ring->count_dw = ndw; in radeon_ring_alloc()
140 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *ring, unsigned ndw) in radeon_ring_lock() argument
145 r = radeon_ring_alloc(rdev, ring, ndw); in radeon_ring_lock()
H A Dradeon_vm.c653 ndw = 64; in radeon_vm_update_page_directory()
656 ndw += vm->max_pde_used * 6; in radeon_vm_update_page_directory()
659 if (ndw > 0xfffff) in radeon_vm_update_page_directory()
918 unsigned nptes, ncmds, ndw; in radeon_vm_bo_update() local
974 ndw = 64; in radeon_vm_bo_update()
979 ndw += ncmds * 7; in radeon_vm_bo_update()
983 ndw += ncmds * 4; in radeon_vm_bo_update()
986 ndw += nptes * 2; in radeon_vm_bo_update()
990 ndw += ncmds * 10; in radeon_vm_bo_update()
993 ndw += 2 * 10; in radeon_vm_bo_update()
[all …]
H A Dr100.c896 unsigned ndw; in r100_copy_blit() local
908 ndw = 64 + (10 * num_loops); in r100_copy_blit()
909 r = radeon_ring_lock(rdev, ring, ndw); in r100_copy_blit()
911 DRM_ERROR("radeon: moving bo (%d) asking for %u dw.\n", r, ndw); in r100_copy_blit()
H A Dradeon.h1032 int radeon_ring_alloc(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
1033 int radeon_ring_lock(struct radeon_device *rdev, struct radeon_ring *cp, unsigned ndw);
/dragonfly/sys/dev/drm/amd/amdgpu/
H A Damdgpu_ring.c64 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw) in amdgpu_ring_alloc() argument
68 ndw = (ndw + ring->funcs->align_mask) & ~ring->funcs->align_mask; in amdgpu_ring_alloc()
73 if (WARN_ON_ONCE(ndw > ring->max_dw)) in amdgpu_ring_alloc()
76 ring->count_dw = ndw; in amdgpu_ring_alloc()
H A Damdgpu_vm.c1118 unsigned ndw = 0; in amdgpu_vm_update_directories() local
1143 ndw = 512 * 8; in amdgpu_vm_update_directories()
1175 (ndw - params.ib->length_dw) < 32) in amdgpu_vm_update_directories()
1196 WARN_ON(params.ib->length_dw > ndw); in amdgpu_vm_update_directories()
1451 unsigned nptes, ncmds, ndw; in amdgpu_vm_bo_update_mapping() local
1499 ndw = 64; in amdgpu_vm_bo_update_mapping()
1506 ndw += nptes * 2; in amdgpu_vm_bo_update_mapping()
1512 ndw += ncmds * 10; in amdgpu_vm_bo_update_mapping()
1523 r = amdgpu_job_alloc_with_ib(adev, ndw * 4, &job); in amdgpu_vm_bo_update_mapping()
1534 i = ndw - nptes * 2; in amdgpu_vm_bo_update_mapping()
[all …]
H A Dsdma_v2_4.c754 unsigned ndw = count * 2; in sdma_v2_4_vm_write_pte() local
760 ib->ptr[ib->length_dw++] = ndw; in sdma_v2_4_vm_write_pte()
761 for (; ndw > 0; ndw -= 2) { in sdma_v2_4_vm_write_pte()
H A Dsdma_v3_0.c1026 unsigned ndw = count * 2; in sdma_v3_0_vm_write_pte() local
1032 ib->ptr[ib->length_dw++] = ndw; in sdma_v3_0_vm_write_pte()
1033 for (; ndw > 0; ndw -= 2) { in sdma_v3_0_vm_write_pte()
H A Dsdma_v4_0.c1092 unsigned ndw = count * 2; in sdma_v4_0_vm_write_pte() local
1098 ib->ptr[ib->length_dw++] = ndw - 1; in sdma_v4_0_vm_write_pte()
1099 for (; ndw > 0; ndw -= 2) { in sdma_v4_0_vm_write_pte()
H A Damdgpu_ring.h224 int amdgpu_ring_alloc(struct amdgpu_ring *ring, unsigned ndw);