/dragonfly/sys/dev/drm/radeon/ |
H A D | ni_dma.c | 59 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cayman_dma_get_rptr() 83 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cayman_dma_get_wptr() 104 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cayman_dma_set_wptr() 159 if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) || in cayman_dma_stop() 173 rdev->ring[R600_RING_TYPE_DMA_INDEX].ready = false; in cayman_dma_stop() 195 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in cayman_dma_resume() 255 if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) || in cayman_dma_resume() 272 radeon_ring_fini(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX]); in cayman_dma_fini() 290 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cayman_dma_is_lockup()
|
H A D | r600_dma.c | 101 if (rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) in r600_dma_stop() 107 rdev->ring[R600_RING_TYPE_DMA_INDEX].ready = false; in r600_dma_stop() 120 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in r600_dma_resume() 172 r = radeon_ring_test(rdev, R600_RING_TYPE_DMA_INDEX, ring); in r600_dma_resume() 178 if (rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) in r600_dma_resume() 194 radeon_ring_fini(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX]); in r600_dma_fini() 237 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in r600_dma_ring_test() 344 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in r600_dma_ib_test()
|
H A D | radeon_asic.c | 969 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1055 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1148 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1254 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1374 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1468 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1561 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1709 .dma_ring_index = R600_RING_TYPE_DMA_INDEX, 1946 [R600_RING_TYPE_DMA_INDEX] = &si_dma_ring, 2116 [R600_RING_TYPE_DMA_INDEX] = &ci_dma_ring, [all …]
|
H A D | cik_sdma.c | 69 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_get_rptr() 93 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_get_wptr() 114 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_set_wptr() 175 if (ridx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_hdp_flush_ring_emit() 253 if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) || in cik_sdma_gfx_stop() 267 rdev->ring[R600_RING_TYPE_DMA_INDEX].ready = false; in cik_sdma_gfx_stop() 373 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in cik_sdma_gfx_resume() 559 radeon_ring_fini(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX]); in cik_sdma_fini() 652 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_ring_test() 709 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in cik_sdma_ib_test() [all …]
|
H A D | radeon_test.c | 119 if (ring == R600_RING_TYPE_DMA_INDEX) in radeon_do_test_moves() 170 if (ring == R600_RING_TYPE_DMA_INDEX) in radeon_do_test_moves()
|
H A D | radeon_vm.c | 407 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, 256); in radeon_vm_clear_bo() 662 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); in radeon_vm_update_page_directory() 1000 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); in radeon_vm_bo_update()
|
H A D | si_dma.c | 46 if (ring->idx == R600_RING_TYPE_DMA_INDEX) in si_dma_is_lockup()
|
H A D | rv770.c | 1786 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX); in rv770_startup() 1815 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in rv770_startup() 1967 rdev->ring[R600_RING_TYPE_DMA_INDEX].ring_obj = NULL; in rv770_init() 1968 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX], 64 * 1024); in rv770_init()
|
H A D | radeon_cs.c | 227 p->ring = R600_RING_TYPE_DMA_INDEX; in radeon_cs_get_ring() 231 p->ring = R600_RING_TYPE_DMA_INDEX; in radeon_cs_get_ring()
|
H A D | radeon_ring.c | 520 static int radeon_dma1_index = R600_RING_TYPE_DMA_INDEX;
|
H A D | ni.c | 2236 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX); in cayman_startup() 2268 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in cayman_startup() 2436 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in cayman_init()
|
H A D | evergreen.c | 4524 if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { in evergreen_irq_set() 4863 radeon_fence_process(rdev, R600_RING_TYPE_DMA_INDEX); in evergreen_irq_process() 5038 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX); in evergreen_startup() 5067 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in evergreen_startup() 5239 rdev->ring[R600_RING_TYPE_DMA_INDEX].ring_obj = NULL; in evergreen_init() 5240 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_DMA_INDEX], 64 * 1024); in evergreen_init()
|
H A D | radeon_fence.c | 1044 case R600_RING_TYPE_DMA_INDEX: return "radeon.dma"; in radeon_fence_get_timeline_name()
|
H A D | radeon_kms.c | 512 *value = rdev->ring[R600_RING_TYPE_DMA_INDEX].ready; in radeon_info_ioctl()
|
H A D | si.c | 6079 if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { in si_irq_set() 6396 radeon_fence_process(rdev, R600_RING_TYPE_DMA_INDEX); in si_irq_process() 6664 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX); in si_startup() 6712 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in si_startup() 6888 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in si_init()
|
H A D | cik.c | 7138 if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { in cik_irq_set() 7981 radeon_fence_process(rdev, R600_RING_TYPE_DMA_INDEX); in cik_irq_process() 8341 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_DMA_INDEX); in cik_startup() 8410 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in cik_startup() 8625 ring = &rdev->ring[R600_RING_TYPE_DMA_INDEX]; in cik_init()
|
H A D | r600.c | 3839 if (atomic_read(&rdev->irq.ring_int[R600_RING_TYPE_DMA_INDEX])) { in r600_irq_set() 4313 radeon_fence_process(rdev, R600_RING_TYPE_DMA_INDEX); in r600_irq_process()
|
H A D | radeon.h | 150 #define R600_RING_TYPE_DMA_INDEX 3 macro
|