/dragonfly/sys/dev/drm/i915/ |
H A D | i915_vma.c | 49 list_move_tail(&vma->vm_link, &vma->vm->inactive_list); in i915_vma_retire() 227 if (!vma) in i915_vma_instance() 254 GEM_BUG_ON(vma->size > vma->node.size); in i915_vma_bind() 349 if (!vma) in i915_vma_unpin_and_release() 404 mappable = vma->node.start + vma->fence_size <= i915_vm_to_ggtt(vma->vm)->mappable_end; in __i915_vma_set_map_and_fenceable() 512 ret = vma->vm->set_pages(vma); in i915_vma_insert() 563 ret = i915_gem_gtt_insert(vma->vm, &vma->node, in i915_vma_insert() 587 vma->vm->clear_pages(vma); in i915_vma_insert() 602 vma->vm->clear_pages(vma); in i915_vma_remove() 691 rb_erase(&vma->obj_node, &vma->obj->vma_tree); in i915_vma_close() [all …]
|
H A D | i915_vma.h | 169 return vma->active; in i915_vma_get_active() 180 vma->active |= BIT(engine); in i915_vma_set_active() 200 GEM_BUG_ON(upper_32_bits(vma->node.start + vma->node.size - 1)); in i915_ggtt_offset() 207 return vma; in i915_vma_get() 234 cmp = vma->ggtt_view.type; in i915_vma_compare() 303 vma->flags++; in __i915_vma_pin() 309 vma->flags--; in __i915_vma_unpin() 316 __i915_vma_unpin(vma); in i915_vma_unpin() 348 GEM_BUG_ON(!vma->pages); in i915_vma_first_page() 373 vma->fence->pin_count--; in __i915_vma_unpin_fence() [all …]
|
H A D | i915_gem_fence_reg.c | 79 if (vma) { in i965_write_fence_reg() 122 if (vma) { in i915_write_fence_reg() 162 if (vma) { in i830_write_fence_reg() 215 if (vma) { in fence_update() 239 if (fence->vma && fence->vma != vma) { in fence_update() 262 if (fence->vma != vma) { in fence_update() 264 fence->vma = vma; in fence_update() 350 GEM_BUG_ON(fence->vma != vma); in i915_vma_pin_fence() 473 struct i915_vma *vma = reg->vma; in i915_gem_restore_fences() local 475 GEM_BUG_ON(vma && vma->fence != reg); in i915_gem_restore_fences() [all …]
|
H A D | i915_gem_evict.c | 81 struct i915_vma *vma, in mark_free() argument 85 if (i915_vma_is_pinned(vma)) in mark_free() 243 __i915_vma_pin(vma); in i915_gem_evict_something() 251 __i915_vma_unpin(vma); in i915_gem_evict_something() 320 vma = container_of(node, typeof(*vma), node); in i915_gem_evict_for_node() 340 (i915_vma_is_pinned(vma) || i915_vma_is_active(vma))) { in i915_gem_evict_for_node() 353 if (vma->exec_flags && in i915_gem_evict_for_node() 367 __i915_vma_pin(vma); in i915_gem_evict_for_node() 372 __i915_vma_unpin(vma); in i915_gem_evict_for_node() 424 __i915_vma_pin(vma); in i915_gem_evict_vm() [all …]
|
H A D | i915_gem_execbuffer.c | 343 (vma->node.start + vma->node.size - 1) >> 32) in eb_vma_misplaced() 488 eb->vma[i] = vma; in eb_add_vma() 499 eb_unreserve_vma(vma, vma->exec_flags); in eb_add_vma() 622 struct i915_vma *vma = eb->vma[i]; in eb_reserve() local 743 GEM_BUG_ON(vma != eb->vma[i]); in eb_lookup_vmas() 802 struct i915_vma *vma = eb->vma[i]; in eb_release_vmas() local 805 if (!vma) in eb_release_vmas() 1764 struct i915_vma *vma = eb->vma[i]; in eb_move_to_gpu() local 1775 capture->vma = eb->vma[i]; in eb_move_to_gpu() 1807 struct i915_vma *vma = eb->vma[i]; in eb_move_to_gpu() local [all …]
|
H A D | i915_gem_tiling.c | 187 if (vma->node.size < size) in i915_vma_fence_prepare() 202 struct i915_vma *vma; in i915_gem_object_fence_prepare() local 209 if (!i915_vma_is_ggtt(vma)) in i915_gem_object_fence_prepare() 215 ret = i915_vma_unbind(vma); in i915_gem_object_fence_prepare() 228 struct i915_vma *vma; in i915_gem_object_set_tiling() local 289 if (!i915_vma_is_ggtt(vma)) in i915_gem_object_set_tiling() 292 vma->fence_size = in i915_gem_object_set_tiling() 294 vma->fence_alignment = in i915_gem_object_set_tiling() 296 vma->size, tiling, stride); in i915_gem_object_set_tiling() 298 if (vma->fence) in i915_gem_object_set_tiling() [all …]
|
H A D | i915_gem_render_state.c | 33 struct i915_vma *vma; member 77 struct drm_i915_gem_object *obj = so->vma->obj; in render_state_setup() 93 u64 r = s + so->vma->node.start; in render_state_setup() 115 so->batch_offset = so->vma->node.start; in render_state_setup() 204 if (IS_ERR(so->vma)) { in i915_gem_render_state_init() 205 ret = PTR_ERR(so->vma); in i915_gem_render_state_init() 232 if (!i915_gem_object_has_pages(so->vma->obj)) in i915_gem_render_state_emit() 263 i915_vma_move_to_active(so->vma, req, 0); in i915_gem_render_state_emit() 265 i915_vma_unpin(so->vma); in i915_gem_render_state_emit() 278 obj = so->vma->obj; in i915_gem_render_state_fini() [all …]
|
H A D | i915_gem_gtt.c | 200 ret = vma->vm->allocate_va_range(vma->vm, vma->node.start, in ppgtt_bind_vma() 211 vma->vm->insert_entries(vma->vm, vma, cache_level, pte_flags); in ppgtt_bind_vma() 218 vma->vm->clear_range(vma->vm, vma->node.start, vma->size); in ppgtt_unbind_vma() 225 vma->pages = vma->obj->mm.pages; in ppgtt_set_pages() 236 if (vma->pages != vma->obj->mm.pages) { in clear_pages() 2598 vma->vm->insert_entries(vma->vm, vma, cache_level, pte_flags); in ggtt_bind_vma() 2618 vma->vm->clear_range(vma->vm, vma->node.start, vma->size); in ggtt_unbind_vma() 2653 vma->vm->insert_entries(vma->vm, vma, cache_level, pte_flags); in aliasing_gtt_bind_vma() 2666 vma->vm->clear_range(vma->vm, vma->node.start, vma->size); in aliasing_gtt_unbind_vma() 3812 vma->pages = vma->obj->mm.pages; in i915_get_ggtt_vma_pages() [all …]
|
H A D | intel_guc_ct.c | 110 return ctch->vma != NULL; in ctch_is_open() 116 struct i915_vma *vma; in ctch_init() local 121 GEM_BUG_ON(ctch->vma); in ctch_init() 147 if (IS_ERR(vma)) { in ctch_init() 148 err = PTR_ERR(vma); in ctch_init() 151 ctch->vma = vma; in ctch_init() 171 i915_vma_unpin_and_release(&ctch->vma); in ctch_init() 181 GEM_BUG_ON(!ctch->vma); in ctch_fini() 184 i915_vma_unpin_and_release(&ctch->vma); in ctch_fini() 197 if (!ctch->vma) { in ctch_open() [all …]
|
H A D | intel_ringbuffer.c | 1213 struct i915_vma *vma = ring->vma; in intel_ring_pin() local 1300 if (IS_ERR(vma)) in intel_ring_create_vma() 1303 return vma; in intel_ring_create_vma() 1307 return vma; in intel_ring_create_vma() 1337 if (IS_ERR(vma)) { in intel_engine_create_ring() 1341 ring->vma = vma; in intel_engine_create_ring() 1408 if (IS_ERR(vma)) in alloc_context_vma() 1411 return vma; in alloc_context_vma() 1431 if (IS_ERR(vma)) { in intel_ring_context_pin() 1436 ce->state = vma; in intel_ring_context_pin() [all …]
|
H A D | intel_guc_log.c | 389 subbuf_size = guc->log.vma->obj->base.size; 440 i915_gem_object_unpin_map(guc->log.vma->obj); 456 i915_gem_object_unpin_map(guc->log.vma->obj); 531 struct i915_vma *vma; in intel_guc_log_create() local 537 GEM_BUG_ON(guc->log.vma); in intel_guc_log_create() 557 vma = intel_guc_allocate_vma(guc, size); in intel_guc_log_create() 558 if (IS_ERR(vma)) { in intel_guc_log_create() 559 ret = PTR_ERR(vma); in intel_guc_log_create() 563 guc->log.vma = vma; in intel_guc_log_create() 586 i915_vma_unpin_and_release(&guc->log.vma); in intel_guc_log_create() [all …]
|
H A D | intel_huc.c | 113 static int huc_ucode_xfer(struct intel_uc_fw *huc_fw, struct i915_vma *vma) in huc_ucode_xfer() argument 126 offset = guc_ggtt_offset(vma) + huc_fw->header_offset; in huc_ucode_xfer() 187 struct i915_vma *vma; in intel_huc_auth() local 193 vma = i915_gem_object_ggtt_pin(huc->fw.obj, NULL, 0, 0, in intel_huc_auth() 195 if (IS_ERR(vma)) { in intel_huc_auth() 197 (int)PTR_ERR(vma)); in intel_huc_auth() 202 guc_ggtt_offset(vma) + huc->fw.rsa_offset); in intel_huc_auth() 220 i915_vma_unpin(vma); in intel_huc_auth()
|
H A D | i915_guc_submission.c | 767 struct i915_vma *vma; in guc_client_alloc() local 791 if (IS_ERR(vma)) { in guc_client_alloc() 792 ret = PTR_ERR(vma); in guc_client_alloc() 797 client->vma = vma; in guc_client_alloc() 897 struct i915_vma *vma; in guc_ads_create() local 915 if (IS_ERR(vma)) in guc_ads_create() 916 return PTR_ERR(vma); in guc_ads_create() 918 guc->ads_vma = vma; in guc_ads_create() 977 struct i915_vma *vma; in i915_guc_submission_init() local 987 if (IS_ERR(vma)) in i915_guc_submission_init() [all …]
|
H A D | i915_gem.c | 1085 if (IS_ERR(vma)) { in i915_gem_gtt_pread() 1603 list_move_tail(&vma->vm_link, &vma->vm->inactive_list); in i915_gem_object_bump_inactive_ggtt() 1854 if (vma) in i915_gem_mmap_ioctl() 2383 if (!reg->vma) in i915_gem_runtime_suspend() 3673 if (!--vma->open_count && !i915_vma_is_ggtt(vma)) in i915_gem_close_object() 4261 if (IS_ERR(vma)) in i915_gem_object_pin_to_display_plane() 4275 return vma; in i915_gem_object_pin_to_display_plane() 4279 return vma; in i915_gem_object_pin_to_display_plane() 4442 return vma; in i915_gem_object_ggtt_pin() 4446 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma)) in i915_gem_object_ggtt_pin() [all …]
|
H A D | intel_overlay.c | 174 struct i915_vma *vma; member 307 i915_gem_track_fb(overlay->vma ? overlay->vma->obj : NULL, in intel_overlay_flip_prepare() 308 vma ? vma->obj : NULL, in intel_overlay_flip_prepare() 315 if (vma) in intel_overlay_flip_prepare() 316 overlay->vma = i915_vma_get(vma); in intel_overlay_flip_prepare() 318 overlay->vma = NULL; in intel_overlay_flip_prepare() 367 if (WARN_ON(!vma)) in intel_overlay_release_old_vma() 374 i915_vma_put(vma); in intel_overlay_release_old_vma() 807 if (IS_ERR(vma)) { in intel_overlay_do_put_image() 1429 if (IS_ERR(vma)) { in intel_setup_overlay() [all …]
|
H A D | i915_mm.c | 61 int remap_io_mapping(struct vm_area_struct *vma, 68 GEM_BUG_ON((vma->vm_flags & 73 r.mm = vma->vm_mm; 76 (pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK)); 80 zap_vma_ptes(vma, addr, (r.pfn - pfn) << PAGE_SHIFT);
|
H A D | intel_fbdev.c | 186 struct i915_vma *vma; in intelfb_create() local 223 if (IS_ERR(vma)) { in intelfb_create() 224 ret = PTR_ERR(vma); in intelfb_create() 264 vaddr = i915_vma_pin_iomap(vma); in intelfb_create() 294 ifbdev->vma = vma; in intelfb_create() 302 intel_unpin_fb_vma(vma); in intelfb_create() 538 if (ifbdev->vma) { in intel_fbdev_destroy() 540 intel_unpin_fb_vma(ifbdev->vma); in intel_fbdev_destroy() 778 if (!ifbdev || !ifbdev->vma) in intel_fbdev_set_suspend() 830 if (ifbdev->vma) in intel_fbdev_output_poll_changed() [all …]
|
/dragonfly/sys/dev/drm/ttm/ |
H A D | ttm_bo_vm.c | 117 vma->vm_private_data; in ttm_bo_vm_fault() 217 page_last = vma_pages(vma) + vma->vm_pgoff - in ttm_bo_vm_fault() 230 cvma = *vma; in ttm_bo_vm_fault() 444 bo = ttm_bo_vm_lookup(bdev, vma->vm_pgoff, vma_pages(vma)); in ttm_bo_mmap() 484 if (vma->vm_pgoff != 0) in ttm_fbdev_mmap() 530 memset(vma, 0, sizeof(*vma)); in ttm_bo_vm_fault_dfly() 534 vmf->vma = vma; in ttm_bo_vm_fault_dfly() 664 cvma = *vma; in ttm_bo_vm_fault_dfly() 791 bzero(&vma, sizeof(vma)); in ttm_bo_mmap_single() 793 vma.vm_end = vma.vm_start + size; in ttm_bo_mmap_single() [all …]
|
/dragonfly/sys/dev/drm/include/linux/ |
H A D | mm.h | 65 struct vm_area_struct *vma; member 80 int (*fault)(struct vm_area_struct *vma, struct vm_fault *vmf); 81 void (*open)(struct vm_area_struct *vma); 82 void (*close)(struct vm_area_struct *vma); 83 int (*access)(struct vm_area_struct *vma, unsigned long addr, 109 io_remap_pfn_range(struct vm_area_struct *vma, in io_remap_pfn_range() argument 113 vma->vm_page_prot = prot; in io_remap_pfn_range() 114 vma->vm_pfn = pfn; in io_remap_pfn_range() 120 vma_pages(struct vm_area_struct *vma) in vma_pages() argument 124 size = vma->vm_end - vma->vm_start; in vma_pages()
|
/dragonfly/contrib/gdb-7/bfd/ |
H A D | tekhex.c | 255 bfd_vma vma; member 267 #define enda(x) (x->vma + x->size) 320 vma &= ~CHUNK_MASK; in find_chunk() 321 while (d && (d->vma) != vma) in find_chunk() 334 d->vma = vma; in find_chunk() 652 bfd_vma vma; in tekhex_set_section_contents() local 658 for (vma = s->vma & ~(bfd_vma) CHUNK_MASK; in tekhex_set_section_contents() 659 vma < s->vma + s->size; in tekhex_set_section_contents() 660 vma += CHUNK_MASK) in tekhex_set_section_contents() 661 find_chunk (abfd, vma); in tekhex_set_section_contents() [all …]
|
/dragonfly/contrib/binutils-2.34/ld/ |
H A D | mri.c | 42 etree_type *vma; member 82 etree_type *vma, in mri_add_to_list() argument 90 (*ptr)->vma = vma; in mri_add_to_list() 99 mri_output_section (const char *name, etree_type *vma) in mri_output_section() argument 101 mri_add_to_list (&address, name, vma, 0, 0, 0); in mri_output_section() 149 olist->vma = alist->vma; in mri_draw_tree() 157 mri_add_to_list (&order, alist->name, alist->vma, 0, 0, 0); in mri_draw_tree() 210 base = p->vma ? p->vma : exp_nameop (NAME, "."); in mri_draw_tree()
|
/dragonfly/contrib/binutils-2.27/ld/ |
H A D | mri.c | 40 etree_type *vma; member 80 etree_type *vma, in mri_add_to_list() argument 88 (*ptr)->vma = vma; in mri_add_to_list() 97 mri_output_section (const char *name, etree_type *vma) in mri_output_section() argument 99 mri_add_to_list (&address, name, vma, 0, 0, 0); in mri_output_section() 147 olist->vma = alist->vma; in mri_draw_tree() 155 mri_add_to_list (&order, alist->name, alist->vma, 0, 0, 0); in mri_draw_tree() 208 base = p->vma ? p->vma : exp_nameop (NAME, "."); in mri_draw_tree()
|
/dragonfly/sys/dev/drm/ |
H A D | drm_gem.c | 887 void drm_gem_vm_open(struct vm_area_struct *vma) in drm_gem_vm_open() argument 902 void drm_gem_vm_close(struct vm_area_struct *vma) in drm_gem_vm_close() argument 936 struct vm_area_struct *vma) 941 if (obj_size < vma->vm_end - vma->vm_start) 948 vma->vm_ops = dev->driver->gem_vm_ops; 949 vma->vm_private_data = obj; 950 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags)); 951 vma->vm_page_prot = pgprot_decrypted(vma->vm_page_prot); 993 vma->vm_pgoff, 994 vma_pages(vma)); [all …]
|
/dragonfly/games/sail/ |
H A D | pl_5.c | 60 int vma, dir; in acceptmove() local 76 vma = ma; in acceptmove() 94 vma = min(ma, maxmove(ms, dir, 0)); in acceptmove() 95 if ((ta < 0 && moved) || (vma < 0 && moved)) in acceptmove() 100 vma--; in acceptmove() 102 if ((ta < 0 && moved) || (vma < 0 && moved)) in acceptmove() 121 vma -= *p - '0'; in acceptmove() 122 if ((ta < 0 && moved) || (vma < 0 && moved)) in acceptmove() 131 if ((ta < 0 && moved) || (vma < 0 && moved) in acceptmove()
|
/dragonfly/contrib/binutils-2.34/bfd/ |
H A D | tekhex.c | 255 bfd_vma vma; member 267 #define enda(x) (x->vma + x->size) 323 vma &= ~CHUNK_MASK; in find_chunk() 324 while (d && (d->vma) != vma) in find_chunk() 337 d->vma = vma; in find_chunk() 413 if (val < section->vma) in first_phase() 414 val = section->vma; in first_phase() 415 section->size = val - section->vma; in first_phase() 824 writevalue (&dst, addr + d->vma); in tekhex_write_object_contents() 842 writevalue (&dst, s->vma); in tekhex_write_object_contents() [all …]
|