/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager.c | 342 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_nocpsch() 885 retval = dqm->asic_ops.update_qpd(dqm, qpd); in register_process() 1002 return pm_init(&dqm->packets, dqm); in start_nocpsch() 1166 retval = pm_init(&dqm->packets, dqm); in start_cpsch() 1184 dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr; in start_cpsch() 1217 kfd_gtt_sa_free(dqm->dev, dqm->fence_mem); in stop_cpsch() 1302 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_cpsch() 1610 dqm, in set_cache_memory_policy() 1834 dqm = kzalloc(sizeof(*dqm), GFP_KERNEL); in device_queue_manager_init() 1955 if (!dqm->ops.initialize(dqm)) in device_queue_manager_init() [all …]
|
H A D | kfd_device_queue_manager.h | 87 int (*create_queue)(struct device_queue_manager *dqm, 95 int (*update_queue)(struct device_queue_manager *dqm, 104 int (*initialize)(struct device_queue_manager *dqm); 105 int (*start)(struct device_queue_manager *dqm); 106 int (*stop)(struct device_queue_manager *dqm); 107 void (*pre_reset)(struct device_queue_manager *dqm); 140 int (*update_qpd)(struct device_queue_manager *dqm, 240 mutex_lock(&dqm->lock_hidden); in dqm_lock() 241 dqm->saved_flags = memalloc_noreclaim_save(); in dqm_lock() 245 memalloc_noreclaim_restore(dqm->saved_flags); in dqm_unlock() [all …]
|
H A D | kfd_process_queue_manager.c | 74 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device() 233 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue() 255 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd); in pqm_create_queue() 287 retval = dev->dqm->ops.create_kernel_queue(dev->dqm, in pqm_create_queue() 335 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); in pqm_create_queue() 347 dqm = NULL; in pqm_destroy_queue() 373 dqm = pqn->kq->dev->dqm; in pqm_destroy_queue() 374 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); in pqm_destroy_queue() 380 dqm = pqn->q->device->dqm; in pqm_destroy_queue() 381 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); in pqm_destroy_queue() [all …]
|
H A D | kfd_device_queue_manager_vi.c | 29 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, 35 static bool set_cache_memory_policy_vi_tonga(struct device_queue_manager *dqm, 41 static int update_qpd_vi(struct device_queue_manager *dqm, 43 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, 45 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, 47 static void init_sdma_vm_tonga(struct device_queue_manager *dqm, 97 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, in set_cache_memory_policy_vi() argument 153 static int update_qpd_vi(struct device_queue_manager *dqm, in update_qpd_vi() argument 194 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, in update_qpd_vi_tonga() argument 228 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument [all …]
|
H A D | kfd_device_queue_manager_cik.c | 29 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, 35 static int update_qpd_cik(struct device_queue_manager *dqm, 37 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, 39 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, 41 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, 90 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, in set_cache_memory_policy_cik() argument 116 static int update_qpd_cik(struct device_queue_manager *dqm, in update_qpd_cik() argument 150 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, in update_qpd_cik_hawaii() argument 180 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument 196 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, in init_sdma_vm_hawaii() argument
|
H A D | kfd_packet_manager.c | 47 struct kfd_dev *dev = pm->dqm->dev; in pm_calc_rlib_size() 49 process_count = pm->dqm->processes_count; in pm_calc_rlib_size() 50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size() 51 compute_queue_count = pm->dqm->active_cp_queue_count; in pm_calc_rlib_size() 52 gws_queue_count = pm->dqm->gws_queue_count; in pm_calc_rlib_size() 145 pm->dqm->processes_count, pm->dqm->active_queue_count); in pm_create_runlist_ib() 151 if (proccesses_mapped >= pm->dqm->processes_count) { in pm_create_runlist_ib() 226 switch (dqm->dev->device_info->asic_family) { in pm_init() 257 dqm->dev->device_info->asic_family); in pm_init() 261 pm->dqm = dqm; in pm_init() [all …]
|
H A D | kfd_mqd_manager.c | 56 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd() 57 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd() 58 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd() 76 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd() 78 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size; in allocate_sdma_mqd() 80 mqd_mem_obj->gtt_mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.gtt_mem in allocate_sdma_mqd() 82 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr + offset; in allocate_sdma_mqd() 84 dev->dqm->hiq_sdma_mqd.cpu_ptr + offset); in allocate_sdma_mqd()
|
H A D | kfd_device_queue_manager_v9.c | 30 static int update_qpd_v9(struct device_queue_manager *dqm, 32 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, 52 static int update_qpd_v9(struct device_queue_manager *dqm, in update_qpd_v9() argument 64 if (dqm->dev->noretry && in update_qpd_v9() 65 !dqm->dev->use_iommu_v2) in update_qpd_v9() 80 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm_v9() argument
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager.c | 342 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_nocpsch() 885 retval = dqm->asic_ops.update_qpd(dqm, qpd); in register_process() 1002 return pm_init(&dqm->packets, dqm); in start_nocpsch() 1166 retval = pm_init(&dqm->packets, dqm); in start_cpsch() 1184 dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr; in start_cpsch() 1217 kfd_gtt_sa_free(dqm->dev, dqm->fence_mem); in stop_cpsch() 1302 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_cpsch() 1610 dqm, in set_cache_memory_policy() 1834 dqm = kzalloc(sizeof(*dqm), GFP_KERNEL); in device_queue_manager_init() 1955 if (!dqm->ops.initialize(dqm)) in device_queue_manager_init() [all …]
|
H A D | kfd_device_queue_manager.h | 87 int (*create_queue)(struct device_queue_manager *dqm, 95 int (*update_queue)(struct device_queue_manager *dqm, 104 int (*initialize)(struct device_queue_manager *dqm); 105 int (*start)(struct device_queue_manager *dqm); 106 int (*stop)(struct device_queue_manager *dqm); 107 void (*pre_reset)(struct device_queue_manager *dqm); 140 int (*update_qpd)(struct device_queue_manager *dqm, 240 mutex_lock(&dqm->lock_hidden); in dqm_lock() 241 dqm->saved_flags = memalloc_noreclaim_save(); in dqm_lock() 245 memalloc_noreclaim_restore(dqm->saved_flags); in dqm_unlock() [all …]
|
H A D | kfd_process_queue_manager.c | 74 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device() 233 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue() 255 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd); in pqm_create_queue() 287 retval = dev->dqm->ops.create_kernel_queue(dev->dqm, in pqm_create_queue() 335 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); in pqm_create_queue() 347 dqm = NULL; in pqm_destroy_queue() 373 dqm = pqn->kq->dev->dqm; in pqm_destroy_queue() 374 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); in pqm_destroy_queue() 380 dqm = pqn->q->device->dqm; in pqm_destroy_queue() 381 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); in pqm_destroy_queue() [all …]
|
H A D | kfd_device_queue_manager_vi.c | 29 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, 35 static bool set_cache_memory_policy_vi_tonga(struct device_queue_manager *dqm, 41 static int update_qpd_vi(struct device_queue_manager *dqm, 43 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, 45 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, 47 static void init_sdma_vm_tonga(struct device_queue_manager *dqm, 97 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, in set_cache_memory_policy_vi() argument 153 static int update_qpd_vi(struct device_queue_manager *dqm, in update_qpd_vi() argument 194 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, in update_qpd_vi_tonga() argument 228 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument [all …]
|
H A D | kfd_device_queue_manager_cik.c | 29 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, 35 static int update_qpd_cik(struct device_queue_manager *dqm, 37 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, 39 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, 41 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, 90 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, in set_cache_memory_policy_cik() argument 116 static int update_qpd_cik(struct device_queue_manager *dqm, in update_qpd_cik() argument 150 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, in update_qpd_cik_hawaii() argument 180 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument 196 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, in init_sdma_vm_hawaii() argument
|
H A D | kfd_packet_manager.c | 47 struct kfd_dev *dev = pm->dqm->dev; in pm_calc_rlib_size() 49 process_count = pm->dqm->processes_count; in pm_calc_rlib_size() 50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size() 51 compute_queue_count = pm->dqm->active_cp_queue_count; in pm_calc_rlib_size() 52 gws_queue_count = pm->dqm->gws_queue_count; in pm_calc_rlib_size() 145 pm->dqm->processes_count, pm->dqm->active_queue_count); in pm_create_runlist_ib() 151 if (proccesses_mapped >= pm->dqm->processes_count) { in pm_create_runlist_ib() 226 switch (dqm->dev->device_info->asic_family) { in pm_init() 257 dqm->dev->device_info->asic_family); in pm_init() 261 pm->dqm = dqm; in pm_init() [all …]
|
H A D | kfd_mqd_manager.c | 56 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd() 57 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd() 58 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd() 76 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd() 78 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size; in allocate_sdma_mqd() 80 mqd_mem_obj->gtt_mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.gtt_mem in allocate_sdma_mqd() 82 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr + offset; in allocate_sdma_mqd() 84 dev->dqm->hiq_sdma_mqd.cpu_ptr + offset); in allocate_sdma_mqd()
|
H A D | kfd_device_queue_manager_v9.c | 30 static int update_qpd_v9(struct device_queue_manager *dqm, 32 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, 52 static int update_qpd_v9(struct device_queue_manager *dqm, in update_qpd_v9() argument 64 if (dqm->dev->noretry && in update_qpd_v9() 65 !dqm->dev->use_iommu_v2) in update_qpd_v9() 80 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm_v9() argument
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager.c | 342 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_nocpsch() 885 retval = dqm->asic_ops.update_qpd(dqm, qpd); in register_process() 1002 return pm_init(&dqm->packets, dqm); in start_nocpsch() 1166 retval = pm_init(&dqm->packets, dqm); in start_cpsch() 1184 dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr; in start_cpsch() 1217 kfd_gtt_sa_free(dqm->dev, dqm->fence_mem); in stop_cpsch() 1302 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_cpsch() 1610 dqm, in set_cache_memory_policy() 1834 dqm = kzalloc(sizeof(*dqm), GFP_KERNEL); in device_queue_manager_init() 1955 if (!dqm->ops.initialize(dqm)) in device_queue_manager_init() [all …]
|
H A D | kfd_device_queue_manager.h | 87 int (*create_queue)(struct device_queue_manager *dqm, 95 int (*update_queue)(struct device_queue_manager *dqm, 104 int (*initialize)(struct device_queue_manager *dqm); 105 int (*start)(struct device_queue_manager *dqm); 106 int (*stop)(struct device_queue_manager *dqm); 107 void (*pre_reset)(struct device_queue_manager *dqm); 140 int (*update_qpd)(struct device_queue_manager *dqm, 240 mutex_lock(&dqm->lock_hidden); in dqm_lock() 241 dqm->saved_flags = memalloc_noreclaim_save(); in dqm_lock() 245 memalloc_noreclaim_restore(dqm->saved_flags); in dqm_unlock() [all …]
|
H A D | kfd_process_queue_manager.c | 74 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device() 233 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue() 255 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd); in pqm_create_queue() 287 retval = dev->dqm->ops.create_kernel_queue(dev->dqm, in pqm_create_queue() 335 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); in pqm_create_queue() 347 dqm = NULL; in pqm_destroy_queue() 373 dqm = pqn->kq->dev->dqm; in pqm_destroy_queue() 374 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); in pqm_destroy_queue() 380 dqm = pqn->q->device->dqm; in pqm_destroy_queue() 381 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); in pqm_destroy_queue() [all …]
|
H A D | kfd_device_queue_manager_vi.c | 29 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, 35 static bool set_cache_memory_policy_vi_tonga(struct device_queue_manager *dqm, 41 static int update_qpd_vi(struct device_queue_manager *dqm, 43 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, 45 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, 47 static void init_sdma_vm_tonga(struct device_queue_manager *dqm, 97 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, in set_cache_memory_policy_vi() argument 153 static int update_qpd_vi(struct device_queue_manager *dqm, in update_qpd_vi() argument 194 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, in update_qpd_vi_tonga() argument 228 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument [all …]
|
H A D | kfd_device_queue_manager_cik.c | 29 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, 35 static int update_qpd_cik(struct device_queue_manager *dqm, 37 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, 39 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, 41 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, 90 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, in set_cache_memory_policy_cik() argument 116 static int update_qpd_cik(struct device_queue_manager *dqm, in update_qpd_cik() argument 150 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, in update_qpd_cik_hawaii() argument 180 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument 196 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, in init_sdma_vm_hawaii() argument
|
H A D | kfd_packet_manager.c | 47 struct kfd_dev *dev = pm->dqm->dev; in pm_calc_rlib_size() 49 process_count = pm->dqm->processes_count; in pm_calc_rlib_size() 50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size() 51 compute_queue_count = pm->dqm->active_cp_queue_count; in pm_calc_rlib_size() 52 gws_queue_count = pm->dqm->gws_queue_count; in pm_calc_rlib_size() 145 pm->dqm->processes_count, pm->dqm->active_queue_count); in pm_create_runlist_ib() 151 if (proccesses_mapped >= pm->dqm->processes_count) { in pm_create_runlist_ib() 226 switch (dqm->dev->device_info->asic_family) { in pm_init() 257 dqm->dev->device_info->asic_family); in pm_init() 261 pm->dqm = dqm; in pm_init() [all …]
|
H A D | kfd_mqd_manager.c | 56 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd() 57 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd() 58 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd() 76 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd() 78 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size; in allocate_sdma_mqd() 80 mqd_mem_obj->gtt_mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.gtt_mem in allocate_sdma_mqd() 82 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr + offset; in allocate_sdma_mqd() 84 dev->dqm->hiq_sdma_mqd.cpu_ptr + offset); in allocate_sdma_mqd()
|
H A D | kfd_device_queue_manager_v9.c | 30 static int update_qpd_v9(struct device_queue_manager *dqm, 32 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, 52 static int update_qpd_v9(struct device_queue_manager *dqm, in update_qpd_v9() argument 64 if (dqm->dev->noretry && in update_qpd_v9() 65 !dqm->dev->use_iommu_v2) in update_qpd_v9() 80 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm_v9() argument
|
/dports/games/libretro-paralleln64/parallel-n64-6e26fbb/mupen64plus-rsp-paraLLEl/arch/simd/rsp/ |
H A D | rsp_core.cpp | 337 dqm = sse2_pshufb(dqm, sll_b2l_keys[element]); in rsp_vload_group1() 340 dqm = _mm_shuffle_epi8(dqm, ekey); in rsp_vload_group1() 437 dqm = _mm_cmpeq_epi8(_mm_setzero_si128(), dqm); in rsp_vload_group4() 441 dqm = sse2_pshufb(dqm, ror_b2l_keys[ror & 0xF]); in rsp_vload_group4() 445 dqm = _mm_shuffle_epi8(dqm, dkey); in rsp_vload_group4() 478 dqm = sse2_pshufb(dqm, sll_l2b_keys[offset]); in rsp_vstore_group1() 481 dqm = _mm_shuffle_epi8(dqm, ekey); in rsp_vstore_group1() 508 reg = _mm_and_si128(dqm, reg); in rsp_vstore_group1() 527 reg = _mm_and_si128(dqm, reg); in rsp_vstore_group1() 584 else dqm = _mm_cmpeq_epi8(_mm_setzero_si128(), dqm); in rsp_vstore_group4() [all …]
|