/dragonfly/sys/dev/drm/radeon/ |
H A D | radeon_gem.c | 37 if (robj) { in radeon_gem_object_free() 40 drm_prime_gem_destroy(&robj->gem_base, robj->tbo.sg); in radeon_gem_object_free() 43 radeon_bo_unref(&robj); in radeon_gem_object_free() 52 struct radeon_bo *robj; in radeon_gem_object_create() local 86 *obj = &robj->gem_base; in radeon_gem_object_create() 99 struct radeon_bo *robj; in radeon_gem_set_domain() local 388 struct radeon_bo *robj; in radeon_gem_set_domain_ioctl() local 416 struct radeon_bo *robj; in radeon_mode_dumb_mmap() local 447 struct radeon_bo *robj; in radeon_gem_busy_ioctl() local 475 struct radeon_bo *robj; in radeon_gem_wait_idle_ioctl() local [all …]
|
H A D | radeon_gart.c | 131 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_alloc() 160 r = radeon_bo_pin(rdev->gart.robj, in radeon_gart_table_vram_pin() 163 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin() 168 radeon_bo_unpin(rdev->gart.robj); in radeon_gart_table_vram_pin() 169 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin() 199 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_unpin() 204 radeon_bo_kunmap(rdev->gart.robj); in radeon_gart_table_vram_unpin() 205 radeon_bo_unpin(rdev->gart.robj); in radeon_gart_table_vram_unpin() 206 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_unpin() 222 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_free() [all …]
|
H A D | evergreen_cs.c | 1220 track->db_z_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1232 track->db_z_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1244 track->db_s_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1256 track->db_s_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1280 track->vgt_strmout_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1504 track->cb_color_fmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1562 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1578 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1590 track->htile_bo = reloc->robj; in evergreen_cs_handle_reg() 2375 texture = reloc->robj; in evergreen_packet3_check() [all …]
|
H A D | radeon_cursor.c | 285 struct radeon_bo *robj; in radeon_crtc_cursor_set2() local 307 robj = gem_to_radeon_bo(obj); in radeon_crtc_cursor_set2() 308 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 314 ret = radeon_bo_pin_restricted(robj, RADEON_GEM_DOMAIN_VRAM, in radeon_crtc_cursor_set2() 317 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2() 349 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_crtc_cursor_set2() local 350 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 352 radeon_bo_unpin(robj); in radeon_crtc_cursor_set2() 353 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2()
|
H A D | r100_track.h | 13 struct radeon_bo *robj; member 20 struct radeon_bo *robj; member 25 struct radeon_bo *robj; member 36 struct radeon_bo *robj; member
|
H A D | r600_cs.c | 1086 track->vgt_strmout_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1213 track->cb_color_frag_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1244 track->cb_color_tile_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1284 track->cb_color_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1297 track->db_bo = reloc->robj; in r600_cs_check_reg() 1310 track->htile_bo = reloc->robj; in r600_cs_check_reg() 1810 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() 1840 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() 1972 texture = reloc->robj; in r600_packet3_check() 1980 mipmap = reloc->robj; in r600_packet3_check() [all …]
|
H A D | r200.c | 188 track->zb.robj = reloc->robj; in r200_packet0_check() 201 track->cb[0].robj = reloc->robj; in r200_packet0_check() 231 track->textures[i].robj = reloc->robj; in r200_packet0_check() 275 track->textures[i].cube_info[face - 1].robj = reloc->robj; in r200_packet0_check()
|
H A D | r100.c | 1344 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1354 track->arrays[i + 1].robj = reloc->robj; in r100_packet3_load_vbpntr() 1368 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1605 track->zb.robj = reloc->robj; in r100_packet0_check() 1618 track->cb[0].robj = reloc->robj; in r100_packet0_check() 1645 track->textures[i].robj = reloc->robj; in r100_packet0_check() 1663 track->textures[0].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1681 track->textures[1].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1699 track->textures[2].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1963 track->arrays[0].robj = reloc->robj; in r100_packet3_check() [all …]
|
H A D | radeon_device.c | 1658 struct radeon_bo *robj; in radeon_suspend_kms() local 1662 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1664 radeon_bo_unpin(robj); in radeon_suspend_kms() 1665 radeon_bo_unreserve(robj); in radeon_suspend_kms() 1672 robj = gem_to_radeon_bo(rfb->obj); in radeon_suspend_kms() 1675 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1677 radeon_bo_unpin(robj); in radeon_suspend_kms() 1678 radeon_bo_unreserve(robj); in radeon_suspend_kms() 1797 r = radeon_bo_reserve(robj, false); in radeon_resume_kms() 1800 r = radeon_bo_pin_restricted(robj, in radeon_resume_kms() [all …]
|
H A D | radeon_cs.c | 110 p->relocs[i].robj = gem_to_radeon_bo(gobj); in radeon_cs_parser_relocs() 157 if (radeon_ttm_tt_has_userptr(p->relocs[i].robj->tbo.ttm)) { in radeon_cs_parser_relocs() 172 if (p->relocs[i].robj->prime_shared_count) { in radeon_cs_parser_relocs() 181 p->relocs[i].tv.bo = &p->relocs[i].robj->tbo; in radeon_cs_parser_relocs() 255 resv = reloc->robj->tbo.resv; in radeon_cs_sync_rings() 399 return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages; in cmp_size_smaller_first() 437 struct radeon_bo *bo = parser->relocs[i].robj; in radeon_cs_parser_fini() 520 bo = p->relocs[i].robj; in radeon_bo_vm_update_pte()
|
H A D | r300.c | 125 if (rdev->gart.robj) { in rv370_pcie_gart_init() 149 if (rdev->gart.robj == NULL) { in rv370_pcie_gart_enable() 666 track->cb[i].robj = reloc->robj; in r300_packet0_check() 679 track->zb.robj = reloc->robj; in r300_packet0_check() 724 track->textures[i].robj = reloc->robj; in r300_packet0_check() 1123 track->aa.robj = reloc->robj; in r300_packet0_check() 1192 r = r100_cs_track_check_pkt3_indx_buffer(p, pkt, reloc->robj); in r300_packet3_check()
|
H A D | radeon_fb.c | 433 bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj) in radeon_fbdev_robj_is_fb() argument 438 if (robj == gem_to_radeon_bo(rdev->mode_info.rfbdev->rfb.obj)) in radeon_fbdev_robj_is_fb()
|
H A D | r600.c | 1107 if (rdev->gart.robj) { in r600_pcie_gart_init() 1124 if (rdev->gart.robj == NULL) { in r600_pcie_gart_enable() 1502 if (rdev->vram_scratch.robj == NULL) { in r600_vram_scratch_init() 1514 r = radeon_bo_pin(rdev->vram_scratch.robj, in r600_vram_scratch_init() 1520 r = radeon_bo_kmap(rdev->vram_scratch.robj, in r600_vram_scratch_init() 1523 radeon_bo_unpin(rdev->vram_scratch.robj); in r600_vram_scratch_init() 1524 radeon_bo_unreserve(rdev->vram_scratch.robj); in r600_vram_scratch_init() 1533 if (rdev->vram_scratch.robj == NULL) { in r600_vram_scratch_fini() 1538 radeon_bo_kunmap(rdev->vram_scratch.robj); in r600_vram_scratch_fini() 1539 radeon_bo_unpin(rdev->vram_scratch.robj); in r600_vram_scratch_fini() [all …]
|
H A D | radeon_object.c | 549 struct radeon_bo *bo = lobj->robj; in radeon_bo_list_validate() 596 lobj->gpu_offset = radeon_bo_gpu_offset(lobj->robj); in radeon_bo_list_validate() 597 lobj->tiling_flags = lobj->robj->tiling_flags; in radeon_bo_list_validate()
|
H A D | radeon_vm.c | 141 list[0].robj = vm->page_directory; in radeon_vm_get_bos() 153 list[idx].robj = vm->page_tables[i].bo; in radeon_vm_get_bos() 156 list[idx].tv.bo = &list[idx].robj->tbo; in radeon_vm_get_bos()
|
H A D | radeon_uvd.c | 593 end = start + radeon_bo_size(reloc->robj); in radeon_uvd_cs_reloc() 637 r = radeon_uvd_cs_msg(p, reloc->robj, offset, buf_sizes); in radeon_uvd_cs_reloc()
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_gem.c | 39 if (robj) { in amdgpu_gem_object_free() 40 amdgpu_mn_unregister(robj); in amdgpu_gem_object_free() 41 amdgpu_bo_unref(&robj); in amdgpu_gem_object_free() 381 struct amdgpu_bo *robj; in amdgpu_mode_dumb_mmap() local 440 struct amdgpu_bo *robj; in amdgpu_gem_wait_idle_ioctl() local 473 struct amdgpu_bo *robj; in amdgpu_gem_metadata_ioctl() local 505 amdgpu_bo_unreserve(robj); in amdgpu_gem_metadata_ioctl() 705 struct amdgpu_bo *robj; in amdgpu_gem_op_ioctl() local 727 amdgpu_bo_unreserve(robj); in amdgpu_gem_op_ioctl() 746 robj->allowed_domains = robj->preferred_domains; in amdgpu_gem_op_ioctl() [all …]
|
H A D | amdgpu_gart.c | 115 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_alloc() 153 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_pin() 158 amdgpu_bo_unpin(adev->gart.robj); in amdgpu_gart_table_vram_pin() 159 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_pin() 176 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_unpin() 179 r = amdgpu_bo_reserve(adev->gart.robj, true); in amdgpu_gart_table_vram_unpin() 181 amdgpu_bo_kunmap(adev->gart.robj); in amdgpu_gart_table_vram_unpin() 182 amdgpu_bo_unpin(adev->gart.robj); in amdgpu_gart_table_vram_unpin() 183 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_unpin() 199 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_free() [all …]
|
H A D | amdgpu_bo_list.c | 53 amdgpu_bo_unref(&e->robj); in amdgpu_bo_list_free() 115 entry->robj = bo; in amdgpu_bo_list_create() 118 entry->tv.bo = &entry->robj->tbo; in amdgpu_bo_list_create() 119 entry->tv.shared = !entry->robj->prime_shared_count; in amdgpu_bo_list_create() 122 list->gds_obj = entry->robj; in amdgpu_bo_list_create() 124 list->gws_obj = entry->robj; in amdgpu_bo_list_create() 126 list->oa_obj = entry->robj; in amdgpu_bo_list_create() 128 total_size += amdgpu_bo_size(entry->robj); in amdgpu_bo_list_create() 129 trace_amdgpu_bo_list_set(list, entry->robj); in amdgpu_bo_list_create() 142 amdgpu_bo_unref(&array[i].robj); in amdgpu_bo_list_create() [all …]
|
H A D | amdgpu_cs.c | 72 amdgpu_bo_unref(&p->uf_entry.robj); in amdgpu_cs_user_fence_chunk() 231 if (p->uf_entry.robj) in amdgpu_cs_parser_init() 466 if (candidate->robj == validated) in amdgpu_cs_try_evict() 531 struct amdgpu_bo *bo = lobj->robj; in amdgpu_cs_list_validate() 608 if (p->uf_entry.robj && !p->uf_entry.robj->parent) in amdgpu_cs_parser_bos() 624 struct amdgpu_bo *bo = e->robj; in amdgpu_cs_parser_bos() 643 amdgpu_bo_unreserve(e->robj); in amdgpu_cs_parser_bos() 736 if (!r && p->uf_entry.robj) { in amdgpu_cs_parser_bos() 857 bo = e->robj; in amdgpu_bo_vm_update_pte() 891 if (!e->robj) in amdgpu_bo_vm_update_pte() [all …]
|
H A D | amdgpu_fb.c | 410 struct amdgpu_bo *robj; in amdgpu_fbdev_total_size() local 416 robj = gem_to_amdgpu_bo(adev->mode_info.rfbdev->rfb.base.obj[0]); in amdgpu_fbdev_total_size() 417 size += amdgpu_bo_size(robj); in amdgpu_fbdev_total_size() 421 bool amdgpu_fbdev_robj_is_fb(struct amdgpu_device *adev, struct amdgpu_bo *robj) in amdgpu_fbdev_robj_is_fb() argument 425 if (robj == gem_to_amdgpu_bo(adev->mode_info.rfbdev->rfb.base.obj[0])) in amdgpu_fbdev_robj_is_fb()
|
H A D | amdgpu_gart.h | 44 struct amdgpu_bo *robj; member
|
H A D | amdgpu_bo_list.h | 35 struct amdgpu_bo *robj; member
|
H A D | amdgpu_device.c | 434 &adev->vram_scratch.robj, in amdgpu_device_vram_scratch_init() 448 amdgpu_bo_free_kernel(&adev->vram_scratch.robj, NULL, NULL); in amdgpu_device_vram_scratch_fini() 2782 struct amdgpu_bo *robj; in amdgpu_device_suspend() local 2796 robj = gem_to_amdgpu_bo(fb->obj[0]); in amdgpu_device_suspend() 2798 if (!amdgpu_fbdev_robj_is_fb(adev, robj)) { in amdgpu_device_suspend() 2799 r = amdgpu_bo_reserve(robj, true); in amdgpu_device_suspend() 2801 amdgpu_bo_unpin(robj); in amdgpu_device_suspend() 2802 amdgpu_bo_unreserve(robj); in amdgpu_device_suspend()
|
H A D | gmc_v9_0.c | 784 if (adev->gart.robj) { in gmc_v9_0_gart_init() 1031 if (adev->gart.robj == NULL) { in gmc_v9_0_gart_enable()
|