Home
last modified time | relevance | path

Searched refs:cpu_addr (Results 1 – 23 of 23) sorted by relevance

/dragonfly/sys/dev/drm/amd/amdgpu/
H A Damdgpu_fence.c100 if (drv->cpu_addr) in amdgpu_fence_write()
101 *drv->cpu_addr = cpu_to_le32(seq); in amdgpu_fence_write()
117 if (drv->cpu_addr) in amdgpu_fence_read()
118 seq = le32_to_cpu(*drv->cpu_addr); in amdgpu_fence_read()
383 ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs]; in amdgpu_fence_driver_start_ring()
388 ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index; in amdgpu_fence_driver_start_ring()
402 ring->fence_drv.gpu_addr, ring->fence_drv.cpu_addr); in amdgpu_fence_driver_start_ring()
426 ring->fence_drv.cpu_addr = NULL; in amdgpu_fence_driver_init_ring()
689 le32_to_cpu(*(ring->fence_drv.cpu_addr + 2))); in amdgpu_debugfs_fence_info()
692 le32_to_cpu(*(ring->fence_drv.cpu_addr + 4))); in amdgpu_debugfs_fence_info()
[all …]
H A Damdgpu_object.c255 u64 *gpu_addr, void **cpu_addr) in amdgpu_bo_create_reserved() argument
301 if (cpu_addr) { in amdgpu_bo_create_reserved()
302 r = amdgpu_bo_kmap(*bo_ptr, cpu_addr); in amdgpu_bo_create_reserved()
344 u64 *gpu_addr, void **cpu_addr) in amdgpu_bo_create_kernel() argument
349 gpu_addr, cpu_addr); in amdgpu_bo_create_kernel()
369 void **cpu_addr) in amdgpu_bo_free_kernel() argument
375 if (cpu_addr) in amdgpu_bo_free_kernel()
386 if (cpu_addr) in amdgpu_bo_free_kernel()
387 *cpu_addr = NULL; in amdgpu_bo_free_kernel()
H A Damdgpu_virt.c347 (void *)&adev->virt.mm_table.cpu_addr); in amdgpu_virt_alloc_mm_table()
353 memset((void *)adev->virt.mm_table.cpu_addr, 0, PAGE_SIZE); in amdgpu_virt_alloc_mm_table()
356 adev->virt.mm_table.cpu_addr); in amdgpu_virt_alloc_mm_table()
372 (void *)&adev->virt.mm_table.cpu_addr); in amdgpu_virt_free_mm_table()
H A Damdgpu_object.h247 u64 *gpu_addr, void **cpu_addr);
251 u64 *gpu_addr, void **cpu_addr);
253 void **cpu_addr);
H A Damdgpu_vcn.c120 (u64 *)&adev->vcn.gpu_addr, &adev->vcn.cpu_addr); in amdgpu_vcn_sw_init()
137 (void **)&adev->vcn.cpu_addr); in amdgpu_vcn_sw_fini()
162 ptr = adev->vcn.cpu_addr; in amdgpu_vcn_suspend()
182 ptr = adev->vcn.cpu_addr; in amdgpu_vcn_resume()
195 memcpy_toio(adev->vcn.cpu_addr, adev->vcn.fw->data + offset, in amdgpu_vcn_resume()
H A Damdgpu_vce.c183 (u64 *)&adev->vce.gpu_addr, &adev->vce.cpu_addr); in amdgpu_vce_sw_init()
217 (void **)&adev->vce.cpu_addr); in amdgpu_vce_sw_fini()
285 void *cpu_addr; in amdgpu_vce_resume() local
299 r = amdgpu_bo_kmap(adev->vce.vcpu_bo, &cpu_addr); in amdgpu_vce_resume()
308 memcpy_toio(cpu_addr, (adev->vce.fw->data) + offset, in amdgpu_vce_resume()
H A Damdgpu_vcn.h61 void *cpu_addr; member
H A Damdgpu_uvd.h42 void *cpu_addr; member
H A Damdgpu_vce.h38 void *cpu_addr; member
H A Damdgpu_uvd.c260 (u64 *)&adev->uvd.inst[j].gpu_addr, &adev->uvd.inst[j].cpu_addr); in amdgpu_uvd_sw_init()
309 (void **)&adev->uvd.inst[j].cpu_addr); in amdgpu_uvd_sw_fini()
369 ptr = adev->uvd.inst[j].cpu_addr; in amdgpu_uvd_suspend()
393 ptr = adev->uvd.inst[i].cpu_addr; in amdgpu_uvd_resume()
406 memcpy_toio(adev->uvd.inst[i].cpu_addr, adev->uvd.fw->data + offset, in amdgpu_uvd_resume()
H A Dvce_v4_0.c157 struct mmsch_v1_0_init_header *header = (struct mmsch_v1_0_init_header *)table->cpu_addr; in vce_v4_0_mmsch_start()
213 uint32_t *init_table = adev->virt.mm_table.cpu_addr; in vce_v4_0_sriov_start()
567 void *ptr = adev->vce.cpu_addr; in vce_v4_0_suspend()
589 void *ptr = adev->vce.cpu_addr; in vce_v4_0_resume()
H A Damdgpu_virt.h35 uint32_t *cpu_addr; member
H A Damdgpu_ring.h73 volatile uint32_t *cpu_addr; member
H A Duvd_v7_0.c719 struct mmsch_v1_0_init_header *header = (struct mmsch_v1_0_init_header *)table->cpu_addr; in uvd_v7_0_mmsch_start()
779 uint32_t *init_table = adev->virt.mm_table.cpu_addr; in uvd_v7_0_sriov_start()
/dragonfly/sys/dev/drm/radeon/
H A Dradeon_fence.c65 if (drv->cpu_addr) { in radeon_fence_write()
66 *drv->cpu_addr = cpu_to_le32(seq); in radeon_fence_write()
88 if (drv->cpu_addr) { in radeon_fence_read()
89 seq = le32_to_cpu(*drv->cpu_addr); in radeon_fence_read()
836 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring()
843 rdev->fence_drv[ring].cpu_addr = (void*)((uint8_t*)rdev->uvd.cpu_addr + index); in radeon_fence_driver_start_ring()
856 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring()
862 ring, rdev->fence_drv[ring].gpu_addr, rdev->fence_drv[ring].cpu_addr); in radeon_fence_driver_start_ring()
881 rdev->fence_drv[ring].cpu_addr = NULL; in radeon_fence_driver_init_ring()
H A Dradeon_vce.c223 void *cpu_addr; in radeon_vce_resume() local
235 r = radeon_bo_kmap(rdev->vce.vcpu_bo, &cpu_addr); in radeon_vce_resume()
242 memset(cpu_addr, 0, radeon_bo_size(rdev->vce.vcpu_bo)); in radeon_vce_resume()
244 r = vce_v1_0_load_fw(rdev, cpu_addr); in radeon_vce_resume()
246 memcpy(cpu_addr, rdev->vce_fw->data, rdev->vce_fw->datasize); in radeon_vce_resume()
H A Dradeon_uvd.c211 r = radeon_bo_kmap(rdev->uvd.vcpu_bo, &rdev->uvd.cpu_addr); in radeon_uvd_init()
289 memcpy(rdev->uvd.cpu_addr, rdev->uvd_fw->data, rdev->uvd_fw->datasize); in radeon_uvd_resume()
294 ptr = rdev->uvd.cpu_addr; in radeon_uvd_resume()
784 uint32_t *msg = rdev->uvd.cpu_addr + offs; in radeon_uvd_get_create_msg()
820 uint32_t *msg = rdev->uvd.cpu_addr + offs; in radeon_uvd_get_destroy_msg()
H A Duvd_v1_0.c145 WREG32(UVD_FW_START, *((uint32_t*)rdev->uvd.cpu_addr)); in uvd_v1_0_resume()
H A Dradeon.h368 volatile uint32_t *cpu_addr; member
1682 void *cpu_addr; member
/dragonfly/sys/dev/drm/
H A Dlinux_dma.c55 dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, in dma_free_coherent() argument
59 kmem_free(kernel_map, (vm_offset_t)cpu_addr, size); in dma_free_coherent()
/dragonfly/sys/dev/drm/include/linux/
H A Ddma-mapping.h83 dma_free_coherent(struct device *dev, size_t size, void *cpu_addr,
/dragonfly/sys/dev/drm/amd/display/amdgpu_dm/
H A Damdgpu_dm.h76 void *cpu_addr; member
H A Damdgpu_dm.c381 (u64 *)&compressor->gpu_addr, &compressor->cpu_addr); in amdgpu_dm_fbc_init()