/dragonfly/sys/dev/drm/radeon/ |
H A D | uvd_v1_0.c | 159 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in uvd_v1_0_init() 174 r = radeon_ring_test(rdev, R600_RING_TYPE_UVD_INDEX, ring); in uvd_v1_0_init() 250 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in uvd_v1_0_fini() 265 struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in uvd_v1_0_start()
|
H A D | radeon_asic.c | 1038 [R600_RING_TYPE_UVD_INDEX] = &rv6xx_uvd_ring, 1131 [R600_RING_TYPE_UVD_INDEX] = &rv6xx_uvd_ring, 1237 [R600_RING_TYPE_UVD_INDEX] = &rv770_uvd_ring, 1357 [R600_RING_TYPE_UVD_INDEX] = &rv770_uvd_ring, 1451 [R600_RING_TYPE_UVD_INDEX] = &rv770_uvd_ring, 1544 [R600_RING_TYPE_UVD_INDEX] = &rv770_uvd_ring, 1692 [R600_RING_TYPE_UVD_INDEX] = &cayman_uvd_ring, 1810 [R600_RING_TYPE_UVD_INDEX] = &cayman_uvd_ring, 1948 [R600_RING_TYPE_UVD_INDEX] = &cayman_uvd_ring, 2118 [R600_RING_TYPE_UVD_INDEX] = &cayman_uvd_ring, [all …]
|
H A D | radeon_cs.c | 126 if (p->ring == R600_RING_TYPE_UVD_INDEX && in radeon_cs_parser_relocs() 237 p->ring = R600_RING_TYPE_UVD_INDEX; in radeon_cs_get_ring() 479 if (parser->ring == R600_RING_TYPE_UVD_INDEX) in radeon_cs_ib_chunk() 561 if (parser->ring == R600_RING_TYPE_UVD_INDEX) in radeon_cs_ib_vm_chunk()
|
H A D | radeon_uvd.c | 244 radeon_ring_fini(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX]); in radeon_uvd_fini() 264 R600_RING_TYPE_UVD_INDEX, handle, &fence); in radeon_uvd_suspend() 339 R600_RING_TYPE_UVD_INDEX, handle, &fence); in radeon_uvd_free_handles() 875 if (radeon_fence_count_emitted(rdev, R600_RING_TYPE_UVD_INDEX) == 0) { in radeon_uvd_idle_work_handler()
|
H A D | rv770.c | 1702 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL; in rv770_uvd_init() 1703 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096); in rv770_uvd_init() 1718 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX); in rv770_uvd_start() 1726 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0; in rv770_uvd_start() 1734 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size) in rv770_uvd_resume() 1737 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in rv770_uvd_resume()
|
H A D | ni.c | 2032 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL; in cayman_uvd_init() 2033 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096); in cayman_uvd_init() 2048 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX); in cayman_uvd_start() 2056 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0; in cayman_uvd_start() 2064 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size) in cayman_uvd_resume() 2067 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in cayman_uvd_resume()
|
H A D | radeon_fence.c | 834 if (ring != R600_RING_TYPE_UVD_INDEX) { in radeon_fence_driver_start_ring() 1046 case R600_RING_TYPE_UVD_INDEX: return "radeon.uvd"; in radeon_fence_get_timeline_name()
|
H A D | radeon_ring.c | 522 static int r600_uvd_index = R600_RING_TYPE_UVD_INDEX;
|
H A D | r600.c | 3069 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL; in r600_uvd_init() 3070 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096); in r600_uvd_init() 3085 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX); in r600_uvd_start() 3093 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0; in r600_uvd_start() 3101 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size) in r600_uvd_resume() 3104 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in r600_uvd_resume() 4299 radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX); in r600_irq_process()
|
H A D | radeon_test.c | 268 if (ring->idx == R600_RING_TYPE_UVD_INDEX) { in radeon_test_create_and_emit_fence()
|
H A D | evergreen.c | 4821 radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX); in evergreen_irq_process() 4932 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL; in evergreen_uvd_init() 4933 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096); in evergreen_uvd_init() 4948 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX); in evergreen_uvd_start() 4956 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0; in evergreen_uvd_start() 4964 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size) in evergreen_uvd_resume() 4967 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in evergreen_uvd_resume()
|
H A D | si.c | 6354 radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX); in si_irq_process() 6464 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL; in si_uvd_init() 6465 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096); in si_uvd_init() 6480 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX); in si_uvd_start() 6488 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0; in si_uvd_start() 6496 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size) in si_uvd_resume() 6499 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in si_uvd_resume()
|
H A D | radeon_kms.c | 516 *value = rdev->ring[R600_RING_TYPE_UVD_INDEX].ready; in radeon_info_ioctl()
|
H A D | radeon_object.c | 573 if (ring == R600_RING_TYPE_UVD_INDEX) in radeon_bo_list_validate()
|
H A D | cik.c | 7871 radeon_fence_process(rdev, R600_RING_TYPE_UVD_INDEX); in cik_irq_process() 8111 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_obj = NULL; in cik_uvd_init() 8112 r600_ring_init(rdev, &rdev->ring[R600_RING_TYPE_UVD_INDEX], 4096); in cik_uvd_init() 8132 r = radeon_fence_driver_start_ring(rdev, R600_RING_TYPE_UVD_INDEX); in cik_uvd_start() 8140 rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size = 0; in cik_uvd_start() 8148 if (!rdev->has_uvd || !rdev->ring[R600_RING_TYPE_UVD_INDEX].ring_size) in cik_uvd_resume() 8151 ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX]; in cik_uvd_resume()
|
H A D | radeon.h | 155 #define R600_RING_TYPE_UVD_INDEX 5 macro
|