Home
last modified time | relevance | path

Searched refs:dqm (Results 1 – 25 of 239) sorted by relevance

12345678910

/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager.c342 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_nocpsch()
885 retval = dqm->asic_ops.update_qpd(dqm, qpd); in register_process()
1002 return pm_init(&dqm->packets, dqm); in start_nocpsch()
1166 retval = pm_init(&dqm->packets, dqm); in start_cpsch()
1184 dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr; in start_cpsch()
1217 kfd_gtt_sa_free(dqm->dev, dqm->fence_mem); in stop_cpsch()
1302 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_cpsch()
1610 dqm, in set_cache_memory_policy()
1834 dqm = kzalloc(sizeof(*dqm), GFP_KERNEL); in device_queue_manager_init()
1955 if (!dqm->ops.initialize(dqm)) in device_queue_manager_init()
[all …]
H A Dkfd_device_queue_manager.h87 int (*create_queue)(struct device_queue_manager *dqm,
95 int (*update_queue)(struct device_queue_manager *dqm,
104 int (*initialize)(struct device_queue_manager *dqm);
105 int (*start)(struct device_queue_manager *dqm);
106 int (*stop)(struct device_queue_manager *dqm);
107 void (*pre_reset)(struct device_queue_manager *dqm);
140 int (*update_qpd)(struct device_queue_manager *dqm,
240 mutex_lock(&dqm->lock_hidden); in dqm_lock()
241 dqm->saved_flags = memalloc_noreclaim_save(); in dqm_lock()
245 memalloc_noreclaim_restore(dqm->saved_flags); in dqm_unlock()
[all …]
H A Dkfd_process_queue_manager.c74 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device()
233 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue()
255 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd); in pqm_create_queue()
287 retval = dev->dqm->ops.create_kernel_queue(dev->dqm, in pqm_create_queue()
335 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); in pqm_create_queue()
347 dqm = NULL; in pqm_destroy_queue()
373 dqm = pqn->kq->dev->dqm; in pqm_destroy_queue()
374 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); in pqm_destroy_queue()
380 dqm = pqn->q->device->dqm; in pqm_destroy_queue()
381 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); in pqm_destroy_queue()
[all …]
H A Dkfd_device_queue_manager_vi.c29 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm,
35 static bool set_cache_memory_policy_vi_tonga(struct device_queue_manager *dqm,
41 static int update_qpd_vi(struct device_queue_manager *dqm,
43 static int update_qpd_vi_tonga(struct device_queue_manager *dqm,
45 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q,
47 static void init_sdma_vm_tonga(struct device_queue_manager *dqm,
97 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, in set_cache_memory_policy_vi() argument
153 static int update_qpd_vi(struct device_queue_manager *dqm, in update_qpd_vi() argument
194 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, in update_qpd_vi_tonga() argument
228 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument
[all …]
H A Dkfd_device_queue_manager_cik.c29 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm,
35 static int update_qpd_cik(struct device_queue_manager *dqm,
37 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm,
39 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q,
41 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm,
90 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, in set_cache_memory_policy_cik() argument
116 static int update_qpd_cik(struct device_queue_manager *dqm, in update_qpd_cik() argument
150 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, in update_qpd_cik_hawaii() argument
180 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument
196 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, in init_sdma_vm_hawaii() argument
H A Dkfd_packet_manager.c47 struct kfd_dev *dev = pm->dqm->dev; in pm_calc_rlib_size()
49 process_count = pm->dqm->processes_count; in pm_calc_rlib_size()
50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size()
51 compute_queue_count = pm->dqm->active_cp_queue_count; in pm_calc_rlib_size()
52 gws_queue_count = pm->dqm->gws_queue_count; in pm_calc_rlib_size()
145 pm->dqm->processes_count, pm->dqm->active_queue_count); in pm_create_runlist_ib()
151 if (proccesses_mapped >= pm->dqm->processes_count) { in pm_create_runlist_ib()
226 switch (dqm->dev->device_info->asic_family) { in pm_init()
257 dqm->dev->device_info->asic_family); in pm_init()
261 pm->dqm = dqm; in pm_init()
[all …]
H A Dkfd_mqd_manager.c56 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd()
57 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd()
58 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd()
76 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd()
78 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size; in allocate_sdma_mqd()
80 mqd_mem_obj->gtt_mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.gtt_mem in allocate_sdma_mqd()
82 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr + offset; in allocate_sdma_mqd()
84 dev->dqm->hiq_sdma_mqd.cpu_ptr + offset); in allocate_sdma_mqd()
H A Dkfd_device_queue_manager_v9.c30 static int update_qpd_v9(struct device_queue_manager *dqm,
32 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
52 static int update_qpd_v9(struct device_queue_manager *dqm, in update_qpd_v9() argument
64 if (dqm->dev->noretry && in update_qpd_v9()
65 !dqm->dev->use_iommu_v2) in update_qpd_v9()
80 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm_v9() argument
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager.c342 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_nocpsch()
885 retval = dqm->asic_ops.update_qpd(dqm, qpd); in register_process()
1002 return pm_init(&dqm->packets, dqm); in start_nocpsch()
1166 retval = pm_init(&dqm->packets, dqm); in start_cpsch()
1184 dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr; in start_cpsch()
1217 kfd_gtt_sa_free(dqm->dev, dqm->fence_mem); in stop_cpsch()
1302 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_cpsch()
1610 dqm, in set_cache_memory_policy()
1834 dqm = kzalloc(sizeof(*dqm), GFP_KERNEL); in device_queue_manager_init()
1955 if (!dqm->ops.initialize(dqm)) in device_queue_manager_init()
[all …]
H A Dkfd_device_queue_manager.h87 int (*create_queue)(struct device_queue_manager *dqm,
95 int (*update_queue)(struct device_queue_manager *dqm,
104 int (*initialize)(struct device_queue_manager *dqm);
105 int (*start)(struct device_queue_manager *dqm);
106 int (*stop)(struct device_queue_manager *dqm);
107 void (*pre_reset)(struct device_queue_manager *dqm);
140 int (*update_qpd)(struct device_queue_manager *dqm,
240 mutex_lock(&dqm->lock_hidden); in dqm_lock()
241 dqm->saved_flags = memalloc_noreclaim_save(); in dqm_lock()
245 memalloc_noreclaim_restore(dqm->saved_flags); in dqm_unlock()
[all …]
H A Dkfd_process_queue_manager.c74 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device()
233 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue()
255 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd); in pqm_create_queue()
287 retval = dev->dqm->ops.create_kernel_queue(dev->dqm, in pqm_create_queue()
335 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); in pqm_create_queue()
347 dqm = NULL; in pqm_destroy_queue()
373 dqm = pqn->kq->dev->dqm; in pqm_destroy_queue()
374 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); in pqm_destroy_queue()
380 dqm = pqn->q->device->dqm; in pqm_destroy_queue()
381 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); in pqm_destroy_queue()
[all …]
H A Dkfd_device_queue_manager_vi.c29 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm,
35 static bool set_cache_memory_policy_vi_tonga(struct device_queue_manager *dqm,
41 static int update_qpd_vi(struct device_queue_manager *dqm,
43 static int update_qpd_vi_tonga(struct device_queue_manager *dqm,
45 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q,
47 static void init_sdma_vm_tonga(struct device_queue_manager *dqm,
97 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, in set_cache_memory_policy_vi() argument
153 static int update_qpd_vi(struct device_queue_manager *dqm, in update_qpd_vi() argument
194 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, in update_qpd_vi_tonga() argument
228 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument
[all …]
H A Dkfd_device_queue_manager_cik.c29 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm,
35 static int update_qpd_cik(struct device_queue_manager *dqm,
37 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm,
39 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q,
41 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm,
90 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, in set_cache_memory_policy_cik() argument
116 static int update_qpd_cik(struct device_queue_manager *dqm, in update_qpd_cik() argument
150 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, in update_qpd_cik_hawaii() argument
180 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument
196 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, in init_sdma_vm_hawaii() argument
H A Dkfd_packet_manager.c47 struct kfd_dev *dev = pm->dqm->dev; in pm_calc_rlib_size()
49 process_count = pm->dqm->processes_count; in pm_calc_rlib_size()
50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size()
51 compute_queue_count = pm->dqm->active_cp_queue_count; in pm_calc_rlib_size()
52 gws_queue_count = pm->dqm->gws_queue_count; in pm_calc_rlib_size()
145 pm->dqm->processes_count, pm->dqm->active_queue_count); in pm_create_runlist_ib()
151 if (proccesses_mapped >= pm->dqm->processes_count) { in pm_create_runlist_ib()
226 switch (dqm->dev->device_info->asic_family) { in pm_init()
257 dqm->dev->device_info->asic_family); in pm_init()
261 pm->dqm = dqm; in pm_init()
[all …]
H A Dkfd_mqd_manager.c56 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd()
57 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd()
58 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd()
76 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd()
78 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size; in allocate_sdma_mqd()
80 mqd_mem_obj->gtt_mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.gtt_mem in allocate_sdma_mqd()
82 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr + offset; in allocate_sdma_mqd()
84 dev->dqm->hiq_sdma_mqd.cpu_ptr + offset); in allocate_sdma_mqd()
H A Dkfd_device_queue_manager_v9.c30 static int update_qpd_v9(struct device_queue_manager *dqm,
32 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
52 static int update_qpd_v9(struct device_queue_manager *dqm, in update_qpd_v9() argument
64 if (dqm->dev->noretry && in update_qpd_v9()
65 !dqm->dev->use_iommu_v2) in update_qpd_v9()
80 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm_v9() argument
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager.c342 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_nocpsch()
885 retval = dqm->asic_ops.update_qpd(dqm, qpd); in register_process()
1002 return pm_init(&dqm->packets, dqm); in start_nocpsch()
1166 retval = pm_init(&dqm->packets, dqm); in start_cpsch()
1184 dqm->fence_gpu_addr = dqm->fence_mem->gpu_addr; in start_cpsch()
1217 kfd_gtt_sa_free(dqm->dev, dqm->fence_mem); in stop_cpsch()
1302 dqm->asic_ops.init_sdma_vm(dqm, q, qpd); in create_queue_cpsch()
1610 dqm, in set_cache_memory_policy()
1834 dqm = kzalloc(sizeof(*dqm), GFP_KERNEL); in device_queue_manager_init()
1955 if (!dqm->ops.initialize(dqm)) in device_queue_manager_init()
[all …]
H A Dkfd_device_queue_manager.h87 int (*create_queue)(struct device_queue_manager *dqm,
95 int (*update_queue)(struct device_queue_manager *dqm,
104 int (*initialize)(struct device_queue_manager *dqm);
105 int (*start)(struct device_queue_manager *dqm);
106 int (*stop)(struct device_queue_manager *dqm);
107 void (*pre_reset)(struct device_queue_manager *dqm);
140 int (*update_qpd)(struct device_queue_manager *dqm,
240 mutex_lock(&dqm->lock_hidden); in dqm_lock()
241 dqm->saved_flags = memalloc_noreclaim_save(); in dqm_lock()
245 memalloc_noreclaim_restore(dqm->saved_flags); in dqm_unlock()
[all …]
H A Dkfd_process_queue_manager.c74 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device()
233 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue()
255 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd); in pqm_create_queue()
287 retval = dev->dqm->ops.create_kernel_queue(dev->dqm, in pqm_create_queue()
335 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); in pqm_create_queue()
347 dqm = NULL; in pqm_destroy_queue()
373 dqm = pqn->kq->dev->dqm; in pqm_destroy_queue()
374 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); in pqm_destroy_queue()
380 dqm = pqn->q->device->dqm; in pqm_destroy_queue()
381 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); in pqm_destroy_queue()
[all …]
H A Dkfd_device_queue_manager_vi.c29 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm,
35 static bool set_cache_memory_policy_vi_tonga(struct device_queue_manager *dqm,
41 static int update_qpd_vi(struct device_queue_manager *dqm,
43 static int update_qpd_vi_tonga(struct device_queue_manager *dqm,
45 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q,
47 static void init_sdma_vm_tonga(struct device_queue_manager *dqm,
97 static bool set_cache_memory_policy_vi(struct device_queue_manager *dqm, in set_cache_memory_policy_vi() argument
153 static int update_qpd_vi(struct device_queue_manager *dqm, in update_qpd_vi() argument
194 static int update_qpd_vi_tonga(struct device_queue_manager *dqm, in update_qpd_vi_tonga() argument
228 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument
[all …]
H A Dkfd_device_queue_manager_cik.c29 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm,
35 static int update_qpd_cik(struct device_queue_manager *dqm,
37 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm,
39 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q,
41 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm,
90 static bool set_cache_memory_policy_cik(struct device_queue_manager *dqm, in set_cache_memory_policy_cik() argument
116 static int update_qpd_cik(struct device_queue_manager *dqm, in update_qpd_cik() argument
150 static int update_qpd_cik_hawaii(struct device_queue_manager *dqm, in update_qpd_cik_hawaii() argument
180 static void init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm() argument
196 static void init_sdma_vm_hawaii(struct device_queue_manager *dqm, in init_sdma_vm_hawaii() argument
H A Dkfd_packet_manager.c47 struct kfd_dev *dev = pm->dqm->dev; in pm_calc_rlib_size()
49 process_count = pm->dqm->processes_count; in pm_calc_rlib_size()
50 queue_count = pm->dqm->active_queue_count; in pm_calc_rlib_size()
51 compute_queue_count = pm->dqm->active_cp_queue_count; in pm_calc_rlib_size()
52 gws_queue_count = pm->dqm->gws_queue_count; in pm_calc_rlib_size()
145 pm->dqm->processes_count, pm->dqm->active_queue_count); in pm_create_runlist_ib()
151 if (proccesses_mapped >= pm->dqm->processes_count) { in pm_create_runlist_ib()
226 switch (dqm->dev->device_info->asic_family) { in pm_init()
257 dqm->dev->device_info->asic_family); in pm_init()
261 pm->dqm = dqm; in pm_init()
[all …]
H A Dkfd_mqd_manager.c56 mqd_mem_obj->gtt_mem = dev->dqm->hiq_sdma_mqd.gtt_mem; in allocate_hiq_mqd()
57 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr; in allocate_hiq_mqd()
58 mqd_mem_obj->cpu_ptr = dev->dqm->hiq_sdma_mqd.cpu_ptr; in allocate_hiq_mqd()
76 dev->dqm->mqd_mgrs[KFD_MQD_TYPE_SDMA]->mqd_size; in allocate_sdma_mqd()
78 offset += dev->dqm->mqd_mgrs[KFD_MQD_TYPE_HIQ]->mqd_size; in allocate_sdma_mqd()
80 mqd_mem_obj->gtt_mem = (void *)((uint64_t)dev->dqm->hiq_sdma_mqd.gtt_mem in allocate_sdma_mqd()
82 mqd_mem_obj->gpu_addr = dev->dqm->hiq_sdma_mqd.gpu_addr + offset; in allocate_sdma_mqd()
84 dev->dqm->hiq_sdma_mqd.cpu_ptr + offset); in allocate_sdma_mqd()
H A Dkfd_device_queue_manager_v9.c30 static int update_qpd_v9(struct device_queue_manager *dqm,
32 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
52 static int update_qpd_v9(struct device_queue_manager *dqm, in update_qpd_v9() argument
64 if (dqm->dev->noretry && in update_qpd_v9()
65 !dqm->dev->use_iommu_v2) in update_qpd_v9()
80 static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, in init_sdma_vm_v9() argument
/dports/games/libretro-paralleln64/parallel-n64-6e26fbb/mupen64plus-rsp-paraLLEl/arch/simd/rsp/
H A Drsp_core.cpp337 dqm = sse2_pshufb(dqm, sll_b2l_keys[element]); in rsp_vload_group1()
340 dqm = _mm_shuffle_epi8(dqm, ekey); in rsp_vload_group1()
437 dqm = _mm_cmpeq_epi8(_mm_setzero_si128(), dqm); in rsp_vload_group4()
441 dqm = sse2_pshufb(dqm, ror_b2l_keys[ror & 0xF]); in rsp_vload_group4()
445 dqm = _mm_shuffle_epi8(dqm, dkey); in rsp_vload_group4()
478 dqm = sse2_pshufb(dqm, sll_l2b_keys[offset]); in rsp_vstore_group1()
481 dqm = _mm_shuffle_epi8(dqm, ekey); in rsp_vstore_group1()
508 reg = _mm_and_si128(dqm, reg); in rsp_vstore_group1()
527 reg = _mm_and_si128(dqm, reg); in rsp_vstore_group1()
584 else dqm = _mm_cmpeq_epi8(_mm_setzero_si128(), dqm); in rsp_vstore_group4()
[all …]

12345678910