/dragonfly/sys/dev/drm/radeon/ |
H A D | sumo_smc.c | 41 for (i = 0; i < rdev->usec_timeout; i++) { in sumo_send_msg_to_smu() 50 for (i = 0; i < rdev->usec_timeout; i++) { in sumo_send_msg_to_smu() 56 for (i = 0; i < rdev->usec_timeout; i++) { in sumo_send_msg_to_smu() 62 for (i = 0; i < rdev->usec_timeout; i++) { in sumo_send_msg_to_smu()
|
H A D | trinity_smc.c | 36 for (i = 0; i < rdev->usec_timeout; i++) { in trinity_notify_message_to_smu() 117 for (i = 0; i < rdev->usec_timeout; i++) { in trinity_acquire_mutex()
|
H A D | r600_dma.c | 258 for (i = 0; i < rdev->usec_timeout; i++) { in r600_dma_ring_test() 265 if (i < rdev->usec_timeout) { in r600_dma_ring_test() 379 for (i = 0; i < rdev->usec_timeout; i++) { in r600_dma_ib_test() 385 if (i < rdev->usec_timeout) { in r600_dma_ib_test()
|
H A D | si_smc.c | 182 for (i = 0; i < rdev->usec_timeout; i++) { in si_send_msg_to_smc() 201 for (i = 0; i < rdev->usec_timeout; i++) { in si_wait_for_smc_inactive()
|
H A D | cik_sdma.c | 674 for (i = 0; i < rdev->usec_timeout; i++) { in cik_sdma_ring_test() 681 if (i < rdev->usec_timeout) { in cik_sdma_ring_test() 748 for (i = 0; i < rdev->usec_timeout; i++) { in cik_sdma_ib_test() 754 if (i < rdev->usec_timeout) { in cik_sdma_ib_test()
|
H A D | rv770_smc.c | 435 for (i = 0; i < rdev->usec_timeout; i++) { in rv770_send_msg_to_smc() 458 for (i = 0; i < rdev->usec_timeout; i++) { in rv770_wait_for_smc_inactive()
|
H A D | uvd_v1_0.c | 437 for (i = 0; i < rdev->usec_timeout; i++) { in uvd_v1_0_ring_test() 444 if (i < rdev->usec_timeout) { in uvd_v1_0_ring_test()
|
H A D | r600_dpm.c | 254 for (i = 0; i < rdev->usec_timeout; i++) { in r600_gfx_clockgating_enable() 332 for (i = 0; i < rdev->usec_timeout; i++) { in r600_wait_for_spll_change() 655 for (i = 0; i < rdev->usec_timeout; i++) { in r600_wait_for_power_level_unequal() 661 for (i = 0; i < rdev->usec_timeout; i++) { in r600_wait_for_power_level_unequal() 673 for (i = 0; i < rdev->usec_timeout; i++) { in r600_wait_for_power_level() 679 for (i = 0; i < rdev->usec_timeout; i++) { in r600_wait_for_power_level()
|
H A D | kv_smc.c | 37 for (i = 0; i < rdev->usec_timeout; i++) { in kv_notify_message_to_smu()
|
H A D | rs400.c | 61 unsigned int timeout = rdev->usec_timeout; in rs400_gart_tlb_flush() 240 for (i = 0; i < rdev->usec_timeout; i++) { in rs400_mc_wait_for_idle()
|
H A D | ci_smc.c | 175 for (i = 0; i < rdev->usec_timeout; i++) {
|
H A D | r520.c | 41 for (i = 0; i < rdev->usec_timeout; i++) { in r520_mc_wait_for_idle()
|
H A D | rv515.c | 134 for (i = 0; i < rdev->usec_timeout; i++) { in rv515_mc_wait_for_idle() 317 for (j = 0; j < rdev->usec_timeout; j++) { in rv515_mc_stop() 420 for (j = 0; j < rdev->usec_timeout; j++) { in rv515_mc_resume() 451 for (j = 0; j < rdev->usec_timeout; j++) { in rv515_mc_resume()
|
H A D | radeon_vce.c | 771 for (i = 0; i < rdev->usec_timeout; i++) { in radeon_vce_ring_test() 777 if (i < rdev->usec_timeout) { in radeon_vce_ring_test()
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_vcn.c | 267 for (i = 0; i < adev->usec_timeout; i++) { in amdgpu_vcn_dec_ring_test_ring() 274 if (i < adev->usec_timeout) { in amdgpu_vcn_dec_ring_test_ring() 450 for (i = 0; i < adev->usec_timeout; i++) { in amdgpu_vcn_enc_ring_test_ring() 456 if (i < adev->usec_timeout) { in amdgpu_vcn_enc_ring_test_ring() 625 for (i = 0; i < adev->usec_timeout; i++) { in amdgpu_vcn_jpeg_ring_test_ring() 632 if (i < adev->usec_timeout) { in amdgpu_vcn_jpeg_ring_test_ring() 708 for (i = 0; i < adev->usec_timeout; i++) { in amdgpu_vcn_jpeg_ring_test_ib() 715 if (i < adev->usec_timeout) in amdgpu_vcn_jpeg_ring_test_ib()
|
H A D | uvd_v5_0.c | 511 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v5_0_ring_test_ring() 518 if (i < adev->usec_timeout) { in uvd_v5_0_ring_test_ring() 573 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v5_0_wait_for_idle()
|
H A D | soc15_common.h | 53 uint32_t loop = adev->usec_timeout; \
|
H A D | uvd_v6_0.c | 189 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v6_0_enc_ring_test_ring() 195 if (i < adev->usec_timeout) { in uvd_v6_0_enc_ring_test_ring() 987 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v6_0_ring_test_ring() 994 if (i < adev->usec_timeout) { in uvd_v6_0_ring_test_ring() 1140 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v6_0_wait_for_idle()
|
H A D | uvd_v7_0.c | 197 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v7_0_enc_ring_test_ring() 203 if (i < adev->usec_timeout) { in uvd_v7_0_enc_ring_test_ring() 1244 for (i = 0; i < adev->usec_timeout; i++) { in uvd_v7_0_ring_test_ring() 1251 if (i < adev->usec_timeout) { in uvd_v7_0_ring_test_ring() 1448 for (i = 0; i < adev->usec_timeout; i++) {
|
H A D | gmc_v8_0.c | 346 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v8_0_tonga_mc_load_microcode() 352 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v8_0_tonga_mc_load_microcode() 416 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v8_0_polaris_mc_load_microcode() 1312 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v8_0_wait_for_idle()
|
H A D | gfx_v9_0.c | 365 for (i = 0; i < adev->usec_timeout; i++) { in gfx_v9_0_ring_test_ring() 371 if (i < adev->usec_timeout) { in gfx_v9_0_ring_test_ring() 1821 if (k == adev->usec_timeout) { in gfx_v9_0_wait_for_rlc_serdes() 1838 for (k = 0; k < adev->usec_timeout; k++) { in gfx_v9_0_wait_for_rlc_serdes() 2670 for (i = 0; i < adev->usec_timeout; i++) { in gfx_v9_0_kiq_kcq_enable() 2676 if (i >= adev->usec_timeout) { in gfx_v9_0_kiq_kcq_enable() 2849 for (j = 0; j < adev->usec_timeout; j++) { in gfx_v9_0_kiq_init_register() 2939 for (j = 0; j < adev->usec_timeout; j++) { in gfx_v9_0_kiq_fini_register() 3211 for (i = 0; i < adev->usec_timeout; i++) { in gfx_v9_0_kcq_disable() 3217 if (i >= adev->usec_timeout) { in gfx_v9_0_kcq_disable() [all …]
|
H A D | sdma_v2_4.c | 619 for (i = 0; i < adev->usec_timeout; i++) { in sdma_v2_4_ring_test_ring() 626 if (i < adev->usec_timeout) { in sdma_v2_4_ring_test_ring() 1008 for (i = 0; i < adev->usec_timeout; i++) { in sdma_v2_4_wait_for_idle()
|
H A D | gmc_v7_0.c | 221 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v7_0_mc_load_microcode() 227 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v7_0_mc_load_microcode() 1178 for (i = 0; i < adev->usec_timeout; i++) { in gmc_v7_0_wait_for_idle()
|
/dragonfly/sys/dev/drm/amd/powerplay/hwmgr/ |
H A D | smu_helper.c | 73 for (i = 0; i < hwmgr->usec_timeout; i++) { in phm_wait_on_register() 81 if (i == hwmgr->usec_timeout) in phm_wait_on_register() 117 for (i = 0; i < hwmgr->usec_timeout; i++) { in phm_wait_for_register_unequal() 126 if (i == hwmgr->usec_timeout) in phm_wait_for_register_unequal()
|
/dragonfly/sys/dev/drm/amd/powerplay/smumgr/ |
H A D | smu8_smumgr.c | 158 for (i = 0; i < hwmgr->usec_timeout; i++) { in smu8_check_fw_load_finish() 165 if (i >= hwmgr->usec_timeout) { in smu8_check_fw_load_finish()
|