/dports/emulators/citra/citra-ac98458e0/src/core/hle/kernel/ |
H A D | process.cpp | 43 ar& vm_manager; in serialize() local 184 kernel.MapSharedPages(vm_manager); in Run() 186 kernel.HandleSpecialMapping(vm_manager, mapping); in Run() 191 vm_manager.LogLayout(Log::Level::Debug); in Run() 220 auto vma = vm_manager.FindVMA(target); in HeapAllocate() 244 vm_manager.Reprotect(vma.Unwrap(), perms); in HeapAllocate() 272 ResultCode result = vm_manager.UnmapRange(target, size); in HeapFree() 318 vm_manager.Reprotect(vma.Unwrap(), perms); in LinearAllocate() 366 auto vma = vm_manager.FindVMA(target); in Map() 403 vm_manager.Reprotect(target_vma.Unwrap(), perms); in Map() [all …]
|
H A D | shared_memory.cpp | 24 owner_process->vm_manager.ChangeMemoryState(base_address, size, MemoryState::Locked, in ~SharedMemory() 59 auto& vm_manager = shared_memory->owner_process->vm_manager; in CreateSharedMemory() local 62 CASCADE_CODE(vm_manager.ChangeMemoryState(address, size, MemoryState::Private, in CreateSharedMemory() 66 auto backing_blocks = vm_manager.GetBackingBlocksForRange(address, size); in CreateSharedMemory() 161 auto vma = target_process.vm_manager.FindVMA(target_address); in Map() 173 auto vma = target_process.vm_manager.MapBackingMemory(interval_target, interval.first, in Map() 176 target_process.vm_manager.Reprotect(vma.Unwrap(), ConvertPermissions(permissions)); in Map() 186 return target_process.vm_manager.UnmapRange(address, size); in Unmap()
|
H A D | ipc.cpp | 177 auto& prev_vma = src_process->vm_manager.FindVMA(prev_reserve)->second; in TranslateCommandBuffer() 178 auto& next_vma = src_process->vm_manager.FindVMA(next_reserve)->second; in TranslateCommandBuffer() 183 ResultCode result = src_process->vm_manager.UnmapRange( in TranslateCommandBuffer() 200 dst_process->vm_manager.MapBackingMemoryToBase( in TranslateCommandBuffer() 210 dst_process->vm_manager in TranslateCommandBuffer() 218 dst_process->vm_manager.MapBackingMemoryToBase( in TranslateCommandBuffer()
|
H A D | kernel.cpp | 68 SetCurrentMemoryPageTable(process->vm_manager.page_table); in SetCurrentProcess() 74 SetCurrentMemoryPageTable(process->vm_manager.page_table); in SetCurrentProcessForCPU() 77 thread_managers[core_id]->cpu->SetPageTable(process->vm_manager.page_table); in SetCurrentProcessForCPU() 186 process->vm_manager.Unlock(); in serialize()
|
/dports/emulators/citra-qt5/citra-ac98458e0/src/core/hle/kernel/ |
H A D | process.cpp | 43 ar& vm_manager; in serialize() local 184 kernel.MapSharedPages(vm_manager); in Run() 186 kernel.HandleSpecialMapping(vm_manager, mapping); in Run() 191 vm_manager.LogLayout(Log::Level::Debug); in Run() 220 auto vma = vm_manager.FindVMA(target); in HeapAllocate() 244 vm_manager.Reprotect(vma.Unwrap(), perms); in HeapAllocate() 272 ResultCode result = vm_manager.UnmapRange(target, size); in HeapFree() 318 vm_manager.Reprotect(vma.Unwrap(), perms); in LinearAllocate() 366 auto vma = vm_manager.FindVMA(target); in Map() 403 vm_manager.Reprotect(target_vma.Unwrap(), perms); in Map() [all …]
|
H A D | shared_memory.cpp | 24 owner_process->vm_manager.ChangeMemoryState(base_address, size, MemoryState::Locked, in ~SharedMemory() 59 auto& vm_manager = shared_memory->owner_process->vm_manager; in CreateSharedMemory() local 62 CASCADE_CODE(vm_manager.ChangeMemoryState(address, size, MemoryState::Private, in CreateSharedMemory() 66 auto backing_blocks = vm_manager.GetBackingBlocksForRange(address, size); in CreateSharedMemory() 161 auto vma = target_process.vm_manager.FindVMA(target_address); in Map() 173 auto vma = target_process.vm_manager.MapBackingMemory(interval_target, interval.first, in Map() 176 target_process.vm_manager.Reprotect(vma.Unwrap(), ConvertPermissions(permissions)); in Map() 186 return target_process.vm_manager.UnmapRange(address, size); in Unmap()
|
H A D | ipc.cpp | 177 auto& prev_vma = src_process->vm_manager.FindVMA(prev_reserve)->second; in TranslateCommandBuffer() 178 auto& next_vma = src_process->vm_manager.FindVMA(next_reserve)->second; in TranslateCommandBuffer() 183 ResultCode result = src_process->vm_manager.UnmapRange( in TranslateCommandBuffer() 200 dst_process->vm_manager.MapBackingMemoryToBase( in TranslateCommandBuffer() 210 dst_process->vm_manager in TranslateCommandBuffer() 218 dst_process->vm_manager.MapBackingMemoryToBase( in TranslateCommandBuffer()
|
H A D | kernel.cpp | 68 SetCurrentMemoryPageTable(process->vm_manager.page_table); in SetCurrentProcess() 74 SetCurrentMemoryPageTable(process->vm_manager.page_table); in SetCurrentProcessForCPU() 77 thread_managers[core_id]->cpu->SetPageTable(process->vm_manager.page_table); in SetCurrentProcessForCPU() 186 process->vm_manager.Unlock(); in serialize()
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 130 adev->vm_manager.block_size; in amdgpu_vm_level_shift() 151 adev->vm_manager.root_level); in amdgpu_vm_num_entries() 2735 adev->vm_manager.block_size = in amdgpu_vm_adjust_size() 2740 adev->vm_manager.block_size = 9; in amdgpu_vm_adjust_size() 2751 adev->vm_manager.block_size, in amdgpu_vm_adjust_size() 2752 adev->vm_manager.fragment_size); in amdgpu_vm_adjust_size() 2805 adev->vm_manager.vm_pte_scheds, in amdgpu_vm_init() 2811 adev->vm_manager.vm_pte_scheds, in amdgpu_vm_init() 3159 adev->vm_manager.fence_context = in amdgpu_vm_manager_init() 3162 adev->vm_manager.seqno[i] = 0; in amdgpu_vm_manager_init() [all …]
|
H A D | amdgpu_ids.c | 203 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_grab_idle() 219 struct amdgpu_ring *r = adev->vm_manager.concurrent_flush ? in amdgpu_vmid_grab_idle() 230 u64 fence_context = adev->vm_manager.fence_context + ring->idx; in amdgpu_vmid_grab_idle() 231 unsigned seqno = ++adev->vm_manager.seqno[ring->idx]; in amdgpu_vmid_grab_idle() 298 if (adev->vm_manager.concurrent_flush) in amdgpu_vmid_grab_reserved() 376 if (needs_flush && !adev->vm_manager.concurrent_flush) in amdgpu_vmid_grab_used() 478 id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_alloc_reserved() 556 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_reset_all() 576 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_mgr_init() 583 id_mgr->num_ids = adev->vm_manager.first_kfd_vmid; in amdgpu_vmid_mgr_init() [all …]
|
H A D | amdgpu_vm.h | 51 #define AMDGPU_VM_PTE_COUNT(adev) (1 << (adev)->vm_manager.block_size) 370 #define amdgpu_vm_copy_pte(adev, ib, pe, src, count) ((adev)->vm_manager.vm_pte_funcs->copy_pte((ib… 371 #define amdgpu_vm_write_pte(adev, ib, pe, value, count, incr) ((adev)->vm_manager.vm_pte_funcs->wri… 372 #define amdgpu_vm_set_pte_pde(adev, ib, pe, addr, count, incr, flags) ((adev)->vm_manager.vm_pte_fu…
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 130 adev->vm_manager.block_size; in amdgpu_vm_level_shift() 151 adev->vm_manager.root_level); in amdgpu_vm_num_entries() 2735 adev->vm_manager.block_size = in amdgpu_vm_adjust_size() 2740 adev->vm_manager.block_size = 9; in amdgpu_vm_adjust_size() 2751 adev->vm_manager.block_size, in amdgpu_vm_adjust_size() 2752 adev->vm_manager.fragment_size); in amdgpu_vm_adjust_size() 2805 adev->vm_manager.vm_pte_scheds, in amdgpu_vm_init() 2811 adev->vm_manager.vm_pte_scheds, in amdgpu_vm_init() 3159 adev->vm_manager.fence_context = in amdgpu_vm_manager_init() 3162 adev->vm_manager.seqno[i] = 0; in amdgpu_vm_manager_init() [all …]
|
H A D | amdgpu_ids.c | 203 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_grab_idle() 219 struct amdgpu_ring *r = adev->vm_manager.concurrent_flush ? in amdgpu_vmid_grab_idle() 230 u64 fence_context = adev->vm_manager.fence_context + ring->idx; in amdgpu_vmid_grab_idle() 231 unsigned seqno = ++adev->vm_manager.seqno[ring->idx]; in amdgpu_vmid_grab_idle() 298 if (adev->vm_manager.concurrent_flush) in amdgpu_vmid_grab_reserved() 376 if (needs_flush && !adev->vm_manager.concurrent_flush) in amdgpu_vmid_grab_used() 478 id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_alloc_reserved() 556 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_reset_all() 576 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_mgr_init() 583 id_mgr->num_ids = adev->vm_manager.first_kfd_vmid; in amdgpu_vmid_mgr_init() [all …]
|
H A D | amdgpu_vm.h | 51 #define AMDGPU_VM_PTE_COUNT(adev) (1 << (adev)->vm_manager.block_size) 370 #define amdgpu_vm_copy_pte(adev, ib, pe, src, count) ((adev)->vm_manager.vm_pte_funcs->copy_pte((ib… 371 #define amdgpu_vm_write_pte(adev, ib, pe, value, count, incr) ((adev)->vm_manager.vm_pte_funcs->wri… 372 #define amdgpu_vm_set_pte_pde(adev, ib, pe, addr, count, incr, flags) ((adev)->vm_manager.vm_pte_fu…
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 130 adev->vm_manager.block_size; in amdgpu_vm_level_shift() 151 adev->vm_manager.root_level); in amdgpu_vm_num_entries() 2735 adev->vm_manager.block_size = in amdgpu_vm_adjust_size() 2740 adev->vm_manager.block_size = 9; in amdgpu_vm_adjust_size() 2751 adev->vm_manager.block_size, in amdgpu_vm_adjust_size() 2752 adev->vm_manager.fragment_size); in amdgpu_vm_adjust_size() 2805 adev->vm_manager.vm_pte_scheds, in amdgpu_vm_init() 2811 adev->vm_manager.vm_pte_scheds, in amdgpu_vm_init() 3159 adev->vm_manager.fence_context = in amdgpu_vm_manager_init() 3162 adev->vm_manager.seqno[i] = 0; in amdgpu_vm_manager_init() [all …]
|
H A D | amdgpu_ids.c | 203 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_grab_idle() 219 struct amdgpu_ring *r = adev->vm_manager.concurrent_flush ? in amdgpu_vmid_grab_idle() 230 u64 fence_context = adev->vm_manager.fence_context + ring->idx; in amdgpu_vmid_grab_idle() 231 unsigned seqno = ++adev->vm_manager.seqno[ring->idx]; in amdgpu_vmid_grab_idle() 298 if (adev->vm_manager.concurrent_flush) in amdgpu_vmid_grab_reserved() 376 if (needs_flush && !adev->vm_manager.concurrent_flush) in amdgpu_vmid_grab_used() 478 id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_alloc_reserved() 556 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_reset_all() 576 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_mgr_init() 583 id_mgr->num_ids = adev->vm_manager.first_kfd_vmid; in amdgpu_vmid_mgr_init() [all …]
|
/dports/emulators/citra/citra-ac98458e0/src/tests/core/memory/ |
H A D | memory.cpp | 31 kernel.MapSharedPages(process->vm_manager); 39 kernel.HandleSpecialMapping(process->vm_manager, 46 process->vm_manager, {Memory::IO_AREA_VADDR, Memory::IO_AREA_SIZE, false, false}); 53 kernel.MapSharedPages(process->vm_manager); 54 process->vm_manager.UnmapRange(Memory::CONFIG_MEMORY_VADDR, Memory::CONFIG_MEMORY_SIZE);
|
/dports/emulators/citra-qt5/citra-ac98458e0/src/tests/core/memory/ |
H A D | memory.cpp | 31 kernel.MapSharedPages(process->vm_manager); 39 kernel.HandleSpecialMapping(process->vm_manager, 46 process->vm_manager, {Memory::IO_AREA_VADDR, Memory::IO_AREA_SIZE, false, false}); 53 kernel.MapSharedPages(process->vm_manager); 54 process->vm_manager.UnmapRange(Memory::CONFIG_MEMORY_VADDR, Memory::CONFIG_MEMORY_SIZE);
|
/dports/emulators/citra/citra-ac98458e0/src/tests/core/hle/kernel/ |
H A D | hle_ipc.cpp | 145 auto result = process->vm_manager.MapBackingMemory(target_address, buffer, buffer.GetSize(), 159 REQUIRE(process->vm_manager.UnmapRange(target_address, buffer.GetSize()) == RESULT_SUCCESS); 168 auto result = process->vm_manager.MapBackingMemory(target_address, buffer, buffer.GetSize(), 197 auto result = process->vm_manager.MapBackingMemory( 202 result = process->vm_manager.MapBackingMemory( 233 REQUIRE(process->vm_manager.UnmapRange(target_address_static, buffer_static.GetSize()) == 235 REQUIRE(process->vm_manager.UnmapRange(target_address_mapped, buffer_mapped.GetSize()) == 327 auto result = process->vm_manager.MapBackingMemory( 346 REQUIRE(process->vm_manager.UnmapRange(target_address, output_buffer.GetSize()) == 358 auto result = process->vm_manager.MapBackingMemory( [all …]
|
/dports/emulators/citra-qt5/citra-ac98458e0/src/tests/core/hle/kernel/ |
H A D | hle_ipc.cpp | 145 auto result = process->vm_manager.MapBackingMemory(target_address, buffer, buffer.GetSize(), 159 REQUIRE(process->vm_manager.UnmapRange(target_address, buffer.GetSize()) == RESULT_SUCCESS); 168 auto result = process->vm_manager.MapBackingMemory(target_address, buffer, buffer.GetSize(), 197 auto result = process->vm_manager.MapBackingMemory( 202 result = process->vm_manager.MapBackingMemory( 233 REQUIRE(process->vm_manager.UnmapRange(target_address_static, buffer_static.GetSize()) == 235 REQUIRE(process->vm_manager.UnmapRange(target_address_mapped, buffer_mapped.GetSize()) == 327 auto result = process->vm_manager.MapBackingMemory( 346 REQUIRE(process->vm_manager.UnmapRange(target_address, output_buffer.GetSize()) == 358 auto result = process->vm_manager.MapBackingMemory( [all …]
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | radeon_vm.c | 89 if (!rdev->vm_manager.enabled) { in radeon_vm_manager_init() 94 rdev->vm_manager.enabled = true; in radeon_vm_manager_init() 110 if (!rdev->vm_manager.enabled) in radeon_vm_manager_fini() 114 radeon_fence_unref(&rdev->vm_manager.active[i]); in radeon_vm_manager_fini() 116 rdev->vm_manager.enabled = false; in radeon_vm_manager_fini() 194 for (i = 1; i < rdev->vm_manager.nvm; ++i) { in radeon_vm_grab_id() 214 return rdev->vm_manager.active[choices[i]]; in radeon_vm_grab_id() 267 radeon_fence_unref(&rdev->vm_manager.active[vm->id]); in radeon_vm_fence() 436 if (last_pfn > rdev->vm_manager.max_pfn) { in radeon_vm_bo_set_addr() 438 last_pfn, rdev->vm_manager.max_pfn); in radeon_vm_bo_set_addr() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_vm.c | 89 if (!rdev->vm_manager.enabled) { in radeon_vm_manager_init() 94 rdev->vm_manager.enabled = true; in radeon_vm_manager_init() 110 if (!rdev->vm_manager.enabled) in radeon_vm_manager_fini() 114 radeon_fence_unref(&rdev->vm_manager.active[i]); in radeon_vm_manager_fini() 116 rdev->vm_manager.enabled = false; in radeon_vm_manager_fini() 196 for (i = 1; i < rdev->vm_manager.nvm; ++i) { in radeon_vm_grab_id() 216 return rdev->vm_manager.active[choices[i]]; in radeon_vm_grab_id() 275 radeon_fence_unref(&rdev->vm_manager.active[vm_id]); in radeon_vm_fence() 467 if (last_pfn >= rdev->vm_manager.max_pfn) { in radeon_vm_bo_set_addr() 469 last_pfn, rdev->vm_manager.max_pfn); in radeon_vm_bo_set_addr() [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_vm.c | 89 if (!rdev->vm_manager.enabled) { in radeon_vm_manager_init() 94 rdev->vm_manager.enabled = true; in radeon_vm_manager_init() 110 if (!rdev->vm_manager.enabled) in radeon_vm_manager_fini() 114 radeon_fence_unref(&rdev->vm_manager.active[i]); in radeon_vm_manager_fini() 116 rdev->vm_manager.enabled = false; in radeon_vm_manager_fini() 196 for (i = 1; i < rdev->vm_manager.nvm; ++i) { in radeon_vm_grab_id() 216 return rdev->vm_manager.active[choices[i]]; in radeon_vm_grab_id() 275 radeon_fence_unref(&rdev->vm_manager.active[vm_id]); in radeon_vm_fence() 467 if (last_pfn >= rdev->vm_manager.max_pfn) { in radeon_vm_bo_set_addr() 469 last_pfn, rdev->vm_manager.max_pfn); in radeon_vm_bo_set_addr() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/radeon/ |
H A D | radeon_vm.c | 89 if (!rdev->vm_manager.enabled) { in radeon_vm_manager_init() 94 rdev->vm_manager.enabled = true; in radeon_vm_manager_init() 110 if (!rdev->vm_manager.enabled) in radeon_vm_manager_fini() 114 radeon_fence_unref(&rdev->vm_manager.active[i]); in radeon_vm_manager_fini() 116 rdev->vm_manager.enabled = false; in radeon_vm_manager_fini() 196 for (i = 1; i < rdev->vm_manager.nvm; ++i) { in radeon_vm_grab_id() 216 return rdev->vm_manager.active[choices[i]]; in radeon_vm_grab_id() 275 radeon_fence_unref(&rdev->vm_manager.active[vm_id]); in radeon_vm_fence() 467 if (last_pfn >= rdev->vm_manager.max_pfn) { in radeon_vm_bo_set_addr() 469 last_pfn, rdev->vm_manager.max_pfn); in radeon_vm_bo_set_addr() [all …]
|
/dports/net/py-softlayer/SoftLayer-5.9.7/SoftLayer/CLI/virt/placementgroup/ |
H A D | delete.py | 38 vm_manager = VSManager(env.client) 41 vm_manager.cancel_instance(guest['id'])
|