/linux/drivers/gpu/drm/ |
H A D | drm_gpuvm.c | 1018 gpuvm->ops = ops; in drm_gpuvm_init() 1035 __drm_gpuva_insert(gpuvm, &gpuvm->kernel_alloc_node); in drm_gpuvm_init() 1048 drm_WARN(gpuvm->drm, !RB_EMPTY_ROOT(&gpuvm->rb.tree.rb_root), in drm_gpuvm_fini() 1051 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->extobj.list), in drm_gpuvm_fini() 1053 drm_WARN(gpuvm->drm, !list_empty(&gpuvm->evict.list), in drm_gpuvm_fini() 1066 if (drm_WARN_ON(gpuvm->drm, !gpuvm->ops->vm_free)) in drm_gpuvm_free() 1069 gpuvm->ops->vm_free(gpuvm); in drm_gpuvm_free() 1083 if (gpuvm) in drm_gpuvm_put() 1728 va->vm = gpuvm; in __drm_gpuva_insert() 2534 args.vm = gpuvm; in drm_gpuvm_sm_map_ops_create() [all …]
|
H A D | drm_debugfs.c | 200 struct drm_gpuvm *gpuvm) in drm_debugfs_gpuva_info() argument 202 struct drm_gpuva *va, *kva = &gpuvm->kernel_alloc_node; in drm_debugfs_gpuva_info() 204 if (!gpuvm->name) in drm_debugfs_gpuva_info() 208 gpuvm->name, gpuvm->mm_start, gpuvm->mm_start + gpuvm->mm_range); in drm_debugfs_gpuva_info() 214 drm_gpuvm_for_each_va(va, gpuvm) { in drm_debugfs_gpuva_info()
|
/linux/include/drm/ |
H A D | drm_gpuvm.h | 353 drm_gpuvm_get(struct drm_gpuvm *gpuvm) in drm_gpuvm_get() argument 355 kref_get(&gpuvm->kref); in drm_gpuvm_get() 357 return gpuvm; in drm_gpuvm_get() 360 void drm_gpuvm_put(struct drm_gpuvm *gpuvm); 378 return gpuvm->flags & DRM_GPUVM_RESV_PROTECTED; in drm_gpuvm_resv_protected() 421 drm_gpuvm_is_extobj(struct drm_gpuvm *gpuvm, in drm_gpuvm_is_extobj() argument 708 drm_gpuvm_bo_create(struct drm_gpuvm *gpuvm, 712 drm_gpuvm_bo_obtain(struct drm_gpuvm *gpuvm, 736 drm_gpuvm_bo_find(struct drm_gpuvm *gpuvm, 1095 void (*vm_free)(struct drm_gpuvm *gpuvm); [all …]
|
H A D | drm_debugfs.h | 155 struct drm_gpuvm *gpuvm); 180 struct drm_gpuvm *gpuvm) in drm_debugfs_gpuva_info() argument
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 33 drm_gpuvm_get(&vm->gpuvm); in xe_vm_get() 39 drm_gpuvm_put(&vm->gpuvm); in xe_vm_put() 83 static inline struct xe_vm *gpuvm_to_vm(struct drm_gpuvm *gpuvm) in gpuvm_to_vm() argument 85 return container_of(gpuvm, struct xe_vm, gpuvm); in gpuvm_to_vm() 135 return container_of(vma->gpuva.vm, struct xe_vm, gpuvm); in xe_vma_vm() 258 return drm_gpuvm_resv(&vm->gpuvm); in xe_vm_resv()
|
H A D | xe_exec.c | 103 struct xe_vm *vm = container_of(vm_exec->vm, struct xe_vm, gpuvm); in xe_exec_fn() 228 vm_exec.vm = &vm->gpuvm; in xe_exec_ioctl() 294 drm_gpuvm_resv_add_fence(&vm->gpuvm, exec, &job->drm.s_fence->finished, in xe_exec_ioctl()
|
H A D | xe_vm.c | 45 return vm->gpuvm.r_obj; in xe_vm_obj() 225 .vm = &vm->gpuvm, in xe_vm_add_compute_exec_queue() 410 ret = drm_gpuvm_validate(&vm->gpuvm, exec); in xe_vm_validate_rebind() 417 } while (!list_empty(&vm->gpuvm.evict.list)); in xe_vm_validate_rebind() 797 vma->gpuva.vm = &vm->gpuvm; in xe_vma_create() 1494 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); in xe_vm_free() 2609 drm_gpuva_ops_free(&vm->gpuvm, op->ops); in xe_vma_op_cleanup() 2691 drm_gpuva_ops_free(&vm->gpuvm, __ops); in vm_bind_ioctl_ops_unwind() 3221 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) { in xe_analyze_vm() 3274 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) { in xe_vm_snapshot_capture() [all …]
|
H A D | xe_vm_types.h | 128 struct drm_gpuvm gpuvm; member
|
/linux/drivers/gpu/drm/imagination/ |
H A D | pvr_vm.c | 68 struct pvr_vm_context *to_pvr_vm_context(struct drm_gpuvm *gpuvm) in to_pvr_vm_context() argument 70 return container_of(gpuvm, struct pvr_vm_context, gpuvm_mgr); in to_pvr_vm_context() 494 static void pvr_gpuvm_free(struct drm_gpuvm *gpuvm) in pvr_gpuvm_free() argument 496 kfree(to_pvr_vm_context(gpuvm)); in pvr_gpuvm_free()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/ |
H A D | display_mode_lib.c | 188 dml_print("DML PARAMS: gpuvm = %d\n", pipe_src->gpuvm); in dml_log_pipe_params()
|
H A D | display_mode_structs.h | 392 bool gpuvm; // gpuvm enabled member
|
H A D | display_mode_vba.c | 882 …mode_lib->vba.GPUVMEnable = mode_lib->vba.GPUVMEnable || !!pipes[k].pipe.src.gpuvm || !!pipes[k].p… in fetch_pipe_params()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn32/ |
H A D | display_rq_dlg_calc_32.c | 401 if (src->dynamic_metadata_enable && src->gpuvm) in dml32_rq_dlg_get_dlg_reg()
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_uvmm.c | 1794 nouveau_uvmm_free(struct drm_gpuvm *gpuvm) in nouveau_uvmm_free() argument 1796 struct nouveau_uvmm *uvmm = uvmm_from_gpuvm(gpuvm); in nouveau_uvmm_free()
|
/linux/drivers/gpu/drm/amd/display/dc/resource/dcn31/ |
H A D | dcn31_resource.c | 1630 pipes[i].pipe.src.gpuvm = 1; in dcn31x_populate_dml_pipes_from_context() 1675 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn31_populate_dml_pipes_from_context()
|
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.c | 1802 static void panthor_vm_free(struct drm_gpuvm *gpuvm) in panthor_vm_free() argument 1804 struct panthor_vm *vm = container_of(gpuvm, struct panthor_vm, base); in panthor_vm_free()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn30/ |
H A D | display_rq_dlg_calc_30.c | 1441 if (src->dynamic_metadata_enable && src->gpuvm) { in dml_rq_dlg_get_dlg_params()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn31/ |
H A D | display_rq_dlg_calc_31.c | 1289 if (src->dynamic_metadata_enable && src->gpuvm) in dml_rq_dlg_get_dlg_params()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn314/ |
H A D | display_rq_dlg_calc_314.c | 1377 if (src->dynamic_metadata_enable && src->gpuvm) in dml_rq_dlg_get_dlg_params()
|
/linux/drivers/gpu/drm/amd/display/dc/resource/dcn32/ |
H A D | dcn32_resource.c | 1874 pipes[pipe_cnt].pipe.src.gpuvm = true; in dcn32_populate_dml_pipes_from_context()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn20/ |
H A D | dcn20_fpu.c | 2168 pipes[i].pipe.src.gpuvm = 1; in dcn21_populate_dml_pipes_from_context()
|