/dragonfly/sys/dev/drm/radeon/ |
H A D | radeon_vm.c | 300 return bo_va; in radeon_vm_bo_find() 329 bo_va->vm = vm; in radeon_vm_bo_add() 330 bo_va->bo = bo; in radeon_vm_bo_add() 342 return bo_va; in radeon_vm_bo_add() 496 if (bo_va->it.start || bo_va->it.last) { in radeon_vm_bo_set_addr() 967 nptes = bo_va->it.last - bo_va->it.start + 1; in radeon_vm_bo_update() 1127 if (bo_va->it.start || bo_va->it.last) in radeon_vm_bo_rmv() 1132 if (bo_va->it.start || bo_va->it.last) { in radeon_vm_bo_rmv() 1133 bo_va->bo = radeon_bo_ref(bo_va->bo); in radeon_vm_bo_rmv() 1161 (bo_va->it.start || bo_va->it.last)) in radeon_vm_bo_invalidate() [all …]
|
H A D | radeon_gem.c | 154 struct radeon_bo_va *bo_va; in radeon_gem_object_open() local 168 if (!bo_va) { in radeon_gem_object_open() 171 ++bo_va->ref_count; in radeon_gem_object_open() 185 struct radeon_bo_va *bo_va; in radeon_gem_object_close() local 200 if (bo_va) { in radeon_gem_object_close() 566 tv.bo = &bo_va->bo->tbo; in radeon_gem_va_update_vm() 591 if (bo_va->it.start) in radeon_gem_va_update_vm() 592 r = radeon_vm_bo_update(rdev, bo_va, &bo_va->bo->tbo.mem); in radeon_gem_va_update_vm() 615 struct radeon_bo_va *bo_va; in radeon_gem_va_ioctl() local 679 if (!bo_va) { in radeon_gem_va_ioctl() [all …]
|
H A D | radeon_trace.h | 42 #define trace_radeon_vm_bo_update(bo_va) argument
|
H A D | radeon_cs.c | 496 struct radeon_bo_va *bo_va; in radeon_bo_vm_update_pte() local 521 bo_va = radeon_vm_bo_find(vm, bo); in radeon_bo_vm_update_pte() 522 if (bo_va == NULL) { in radeon_bo_vm_update_pte() 527 r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem); in radeon_bo_vm_update_pte() 531 radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); in radeon_bo_vm_update_pte()
|
H A D | radeon.h | 2875 struct radeon_bo_va *bo_va, 2885 struct radeon_bo_va *bo_va, 2889 struct radeon_bo_va *bo_va);
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 878 return bo_va; in amdgpu_vm_bo_find() 1754 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() 1757 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() 1791 list_splice_init(&bo_va->invalids, &bo_va->valids); in amdgpu_vm_bo_update() 2085 return bo_va; in amdgpu_vm_bo_add() 2105 mapping->bo_va = bo_va; in amdgpu_vm_bo_insert_map() 2352 before->bo_va = tmp->bo_va; in amdgpu_vm_bo_clear_mappings() 2363 after->bo_va = tmp->bo_va; in amdgpu_vm_bo_clear_mappings() 2383 tmp->bo_va = NULL; in amdgpu_vm_bo_clear_mappings() 2445 if (mapping->bo_va && mapping->bo_va->base.bo) { in amdgpu_vm_bo_trace_cs() [all …]
|
H A D | amdgpu_gem.c | 127 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_open() local 144 if (!bo_va) { in amdgpu_gem_object_open() 147 ++bo_va->ref_count; in amdgpu_gem_object_open() 165 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_close() local 183 bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_gem_object_close() 184 if (bo_va && --bo_va->ref_count == 0) { in amdgpu_gem_object_close() 185 amdgpu_vm_bo_rmv(adev, bo_va); in amdgpu_gem_object_close() 564 struct amdgpu_bo_va *bo_va; in amdgpu_gem_va_ioctl() local 642 if (!bo_va) { in amdgpu_gem_va_ioctl() 647 bo_va = fpriv->prt_va; in amdgpu_gem_va_ioctl() [all …]
|
H A D | amdgpu_virt.c | 77 struct amdgpu_bo_va **bo_va) in amdgpu_map_static_csa() argument 100 *bo_va = amdgpu_vm_bo_add(adev, vm, adev->virt.csa_obj); in amdgpu_map_static_csa() 101 if (!*bo_va) { in amdgpu_map_static_csa() 107 r = amdgpu_vm_alloc_pts(adev, (*bo_va)->base.vm, csa_addr, in amdgpu_map_static_csa() 111 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa() 116 r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, AMDGPU_CSA_SIZE, in amdgpu_map_static_csa() 122 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa()
|
H A D | amdgpu_vm.h | 296 struct amdgpu_bo_va *bo_va, 306 struct amdgpu_bo_va *bo_va, 310 struct amdgpu_bo_va *bo_va, 314 struct amdgpu_bo_va *bo_va, 323 struct amdgpu_bo_va *bo_va);
|
H A D | amdgpu_cs.c | 721 e->bo_va = amdgpu_vm_bo_find(vm, e->robj); in amdgpu_cs_parser_bos() 821 struct amdgpu_bo_va *bo_va; in amdgpu_bo_vm_update_pte() local 841 bo_va = fpriv->csa_va; in amdgpu_bo_vm_update_pte() 842 BUG_ON(!bo_va); in amdgpu_bo_vm_update_pte() 843 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_bo_vm_update_pte() 847 f = bo_va->last_pt_update; in amdgpu_bo_vm_update_pte() 861 bo_va = e->bo_va; in amdgpu_bo_vm_update_pte() 862 if (bo_va == NULL) in amdgpu_bo_vm_update_pte() 869 f = bo_va->last_pt_update; in amdgpu_bo_vm_update_pte() 1657 if (!mapping || !mapping->bo_va || !mapping->bo_va->base.bo) in amdgpu_cs_find_mapping() [all …]
|
H A D | amdgpu_trace.h | 41 #define trace_amdgpu_vm_bo_unmap(bo_va, mapping) argument
|
H A D | amdgpu_bo_list.h | 37 struct amdgpu_bo_va *bo_va; member
|
H A D | amdgpu_amdkfd.h | 43 struct amdgpu_bo_va *bo_va; member
|
H A D | amdgpu_virt.h | 286 struct amdgpu_bo_va **bo_va);
|
H A D | amdgpu_object.h | 49 struct amdgpu_bo_va *bo_va; member
|