/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_gfx.h | 64 int mec, int pipe, int queue) in amdgpu_gfx_queue_to_bit() argument 68 bit += mec * adev->gfx.mec.num_pipe_per_mec in amdgpu_gfx_queue_to_bit() 69 * adev->gfx.mec.num_queue_per_pipe; in amdgpu_gfx_queue_to_bit() 70 bit += pipe * adev->gfx.mec.num_queue_per_pipe; in amdgpu_gfx_queue_to_bit() 77 int *mec, int *pipe, int *queue) in amdgpu_gfx_bit_to_queue() argument 79 *queue = bit % adev->gfx.mec.num_queue_per_pipe; in amdgpu_gfx_bit_to_queue() 81 % adev->gfx.mec.num_pipe_per_mec; in amdgpu_gfx_bit_to_queue() 82 *mec = (bit / adev->gfx.mec.num_queue_per_pipe) in amdgpu_gfx_bit_to_queue() 83 / adev->gfx.mec.num_pipe_per_mec; in amdgpu_gfx_bit_to_queue() 87 int mec, int pipe, int queue) in amdgpu_gfx_is_mec_queue_enabled() argument [all …]
|
H A D | amdgpu_gfx.c | 125 return adev->gfx.mec.num_mec > 1; in amdgpu_gfx_is_multipipe_capable() 130 int i, queue, pipe, mec; in amdgpu_gfx_compute_queue_acquire() local 137 % adev->gfx.mec.num_pipe_per_mec; in amdgpu_gfx_compute_queue_acquire() 138 mec = (i / adev->gfx.mec.num_queue_per_pipe) in amdgpu_gfx_compute_queue_acquire() 139 / adev->gfx.mec.num_pipe_per_mec; in amdgpu_gfx_compute_queue_acquire() 142 if (mec >= adev->gfx.mec.num_mec) in amdgpu_gfx_compute_queue_acquire() 147 if (mec == 0 && queue < 2) in amdgpu_gfx_compute_queue_acquire() 151 if (mec == 0 && pipe == 0) in amdgpu_gfx_compute_queue_acquire() 170 int mec, pipe, queue; in amdgpu_gfx_kiq_acquire() local 172 queue_bit = adev->gfx.mec.num_mec in amdgpu_gfx_kiq_acquire() [all …]
|
H A D | amdgpu_amdkfd_gfx_v8.c | 217 uint32_t mec = (pipe_id / adev->gfx.mec.num_pipe_per_mec) + 1; in acquire_queue() local 220 lock_srbm(kgd, mec, pipe, queue_id, 0); in acquire_queue() 276 uint32_t mec; in kgd_init_interrupts() local 279 mec = (pipe_id / adev->gfx.mec.num_pipe_per_mec) + 1; in kgd_init_interrupts() 280 pipe = (pipe_id % adev->gfx.mec.num_pipe_per_mec); in kgd_init_interrupts() 282 lock_srbm(kgd, mec, pipe, 0, 0); in kgd_init_interrupts() 329 uint32_t value, mec, pipe; in kgd_hqd_load() local 331 mec = (pipe_id / adev->gfx.mec.num_pipe_per_mec) + 1; in kgd_hqd_load() 332 pipe = (pipe_id % adev->gfx.mec.num_pipe_per_mec); in kgd_hqd_load() 335 mec, pipe, queue_id); in kgd_hqd_load() [all …]
|
H A D | amdgpu_amdkfd_gfx_v9.c | 236 soc15_grbm_select(adev, mec, pipe, queue, vmid); in lock_srbm() 252 uint32_t mec = (pipe_id / adev->gfx.mec.num_pipe_per_mec) + 1; in acquire_queue() local 255 lock_srbm(kgd, mec, pipe, queue_id, 0); in acquire_queue() 353 uint32_t mec; in kgd_init_interrupts() local 356 mec = (pipe_id / adev->gfx.mec.num_pipe_per_mec) + 1; in kgd_init_interrupts() 357 pipe = (pipe_id % adev->gfx.mec.num_pipe_per_mec); in kgd_init_interrupts() 359 lock_srbm(kgd, mec, pipe, 0, 0); in kgd_init_interrupts() 416 uint32_t value, mec, pipe; in kgd_hqd_load() local 418 mec = (pipe_id / adev->gfx.mec.num_pipe_per_mec) + 1; in kgd_hqd_load() 422 mec, pipe, queue_id); in kgd_hqd_load() [all …]
|
H A D | amdgpu_amdkfd.c | 158 .num_pipe_per_mec = adev->gfx.mec.num_pipe_per_mec, in amdgpu_amdkfd_device_init() 159 .num_queue_per_pipe = adev->gfx.mec.num_queue_per_pipe, in amdgpu_amdkfd_device_init() 169 adev->gfx.mec.queue_bitmap, in amdgpu_amdkfd_device_init() 183 * adev->gfx.mec.num_pipe_per_mec in amdgpu_amdkfd_device_init() 184 * adev->gfx.mec.num_queue_per_pipe; in amdgpu_amdkfd_device_init()
|
H A D | gfx_v9_0.c | 1050 &adev->gfx.mec.hpd_eop_obj, in gfx_v9_0_mec_init() 1073 &adev->gfx.mec.mec_fw_obj, in gfx_v9_0_mec_init() 1074 &adev->gfx.mec.mec_fw_gpu_addr, in gfx_v9_0_mec_init() 1448 int mec, int pipe, int queue) in gfx_v9_0_compute_ring_init() argument 1457 ring->me = mec + 1; in gfx_v9_0_compute_ring_init() 1494 adev->gfx.mec.num_mec = 2; in gfx_v9_0_sw_init() 1497 adev->gfx.mec.num_mec = 1; in gfx_v9_0_sw_init() 1501 adev->gfx.mec.num_pipe_per_mec = 4; in gfx_v9_0_sw_init() 1502 adev->gfx.mec.num_queue_per_pipe = 8; in gfx_v9_0_sw_init() 2978 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v9_0_kiq_init_queue() [all …]
|
H A D | gfx_v8_0.c | 1450 &adev->gfx.mec.hpd_eop_obj, in gfx_v8_0_mec_init() 1988 int mec, int pipe, int queue) in gfx_v8_0_compute_ring_init() argument 1997 ring->me = mec + 1; in gfx_v8_0_compute_ring_init() 2039 adev->gfx.mec.num_mec = 2; in gfx_v8_0_sw_init() 2044 adev->gfx.mec.num_mec = 1; in gfx_v8_0_sw_init() 2048 adev->gfx.mec.num_pipe_per_mec = 4; in gfx_v8_0_sw_init() 2049 adev->gfx.mec.num_queue_per_pipe = 8; in gfx_v8_0_sw_init() 4903 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v8_0_kiq_init_queue() 4925 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v8_0_kiq_init_queue() 4948 if (adev->gfx.mec.mqd_backup[mqd_idx]) in gfx_v8_0_kcq_init_queue() [all …]
|
H A D | amdgpu.h | 905 struct amdgpu_mec mec; member
|
/dragonfly/usr.bin/sort/ |
H A D | sort.c | 542 int fo_index, mec; in check_mutually_exclusive_flags() local 549 mec = mutually_exclusive_flags[i]; in check_mutually_exclusive_flags() 551 if (mec != c) { in check_mutually_exclusive_flags() 554 errx(1, "%c:%c: %s", c, mec, getstr(1)); in check_mutually_exclusive_flags()
|
/dragonfly/sys/dev/drm/radeon/ |
H A D | cik.c | 4402 if (rdev->mec.hpd_eop_obj) { in cik_mec_fini() 4406 radeon_bo_unpin(rdev->mec.hpd_eop_obj); in cik_mec_fini() 4410 rdev->mec.hpd_eop_obj = NULL; in cik_mec_fini() 4427 rdev->mec.num_mec = 1; in cik_mec_init() 4428 rdev->mec.num_pipe = 1; in cik_mec_init() 4429 rdev->mec.num_queue = rdev->mec.num_mec * rdev->mec.num_pipe * 8; in cik_mec_init() 4431 if (rdev->mec.hpd_eop_obj == NULL) { in cik_mec_init() 4433 rdev->mec.num_mec *rdev->mec.num_pipe * MEC_HPD_SIZE * 2, in cik_mec_init() 4436 &rdev->mec.hpd_eop_obj); in cik_mec_init() 4449 &rdev->mec.hpd_eop_gpu_addr); in cik_mec_init() [all …]
|
H A D | radeon.h | 2422 struct radeon_mec mec; member
|