Home
last modified time | relevance | path

Searched refs:vma (Results 1 – 25 of 205) sorted by relevance

123456789

/dragonfly/sys/dev/drm/i915/
H A Di915_vma.c49 list_move_tail(&vma->vm_link, &vma->vm->inactive_list); in i915_vma_retire()
227 if (!vma) in i915_vma_instance()
254 GEM_BUG_ON(vma->size > vma->node.size); in i915_vma_bind()
349 if (!vma) in i915_vma_unpin_and_release()
404 mappable = vma->node.start + vma->fence_size <= i915_vm_to_ggtt(vma->vm)->mappable_end; in __i915_vma_set_map_and_fenceable()
512 ret = vma->vm->set_pages(vma); in i915_vma_insert()
563 ret = i915_gem_gtt_insert(vma->vm, &vma->node, in i915_vma_insert()
587 vma->vm->clear_pages(vma); in i915_vma_insert()
602 vma->vm->clear_pages(vma); in i915_vma_remove()
691 rb_erase(&vma->obj_node, &vma->obj->vma_tree); in i915_vma_close()
[all …]
H A Di915_vma.h169 return vma->active; in i915_vma_get_active()
180 vma->active |= BIT(engine); in i915_vma_set_active()
200 GEM_BUG_ON(upper_32_bits(vma->node.start + vma->node.size - 1)); in i915_ggtt_offset()
207 return vma; in i915_vma_get()
234 cmp = vma->ggtt_view.type; in i915_vma_compare()
303 vma->flags++; in __i915_vma_pin()
309 vma->flags--; in __i915_vma_unpin()
316 __i915_vma_unpin(vma); in i915_vma_unpin()
348 GEM_BUG_ON(!vma->pages); in i915_vma_first_page()
373 vma->fence->pin_count--; in __i915_vma_unpin_fence()
[all …]
H A Di915_gem_fence_reg.c79 if (vma) { in i965_write_fence_reg()
122 if (vma) { in i915_write_fence_reg()
162 if (vma) { in i830_write_fence_reg()
215 if (vma) { in fence_update()
239 if (fence->vma && fence->vma != vma) { in fence_update()
262 if (fence->vma != vma) { in fence_update()
264 fence->vma = vma; in fence_update()
350 GEM_BUG_ON(fence->vma != vma); in i915_vma_pin_fence()
473 struct i915_vma *vma = reg->vma; in i915_gem_restore_fences() local
475 GEM_BUG_ON(vma && vma->fence != reg); in i915_gem_restore_fences()
[all …]
H A Di915_gem_evict.c81 struct i915_vma *vma, in mark_free() argument
85 if (i915_vma_is_pinned(vma)) in mark_free()
243 __i915_vma_pin(vma); in i915_gem_evict_something()
251 __i915_vma_unpin(vma); in i915_gem_evict_something()
320 vma = container_of(node, typeof(*vma), node); in i915_gem_evict_for_node()
340 (i915_vma_is_pinned(vma) || i915_vma_is_active(vma))) { in i915_gem_evict_for_node()
353 if (vma->exec_flags && in i915_gem_evict_for_node()
367 __i915_vma_pin(vma); in i915_gem_evict_for_node()
372 __i915_vma_unpin(vma); in i915_gem_evict_for_node()
424 __i915_vma_pin(vma); in i915_gem_evict_vm()
[all …]
H A Di915_gem_execbuffer.c343 (vma->node.start + vma->node.size - 1) >> 32) in eb_vma_misplaced()
488 eb->vma[i] = vma; in eb_add_vma()
499 eb_unreserve_vma(vma, vma->exec_flags); in eb_add_vma()
622 struct i915_vma *vma = eb->vma[i]; in eb_reserve() local
743 GEM_BUG_ON(vma != eb->vma[i]); in eb_lookup_vmas()
802 struct i915_vma *vma = eb->vma[i]; in eb_release_vmas() local
805 if (!vma) in eb_release_vmas()
1764 struct i915_vma *vma = eb->vma[i]; in eb_move_to_gpu() local
1775 capture->vma = eb->vma[i]; in eb_move_to_gpu()
1807 struct i915_vma *vma = eb->vma[i]; in eb_move_to_gpu() local
[all …]
H A Di915_gem_tiling.c187 if (vma->node.size < size) in i915_vma_fence_prepare()
202 struct i915_vma *vma; in i915_gem_object_fence_prepare() local
209 if (!i915_vma_is_ggtt(vma)) in i915_gem_object_fence_prepare()
215 ret = i915_vma_unbind(vma); in i915_gem_object_fence_prepare()
228 struct i915_vma *vma; in i915_gem_object_set_tiling() local
289 if (!i915_vma_is_ggtt(vma)) in i915_gem_object_set_tiling()
292 vma->fence_size = in i915_gem_object_set_tiling()
294 vma->fence_alignment = in i915_gem_object_set_tiling()
296 vma->size, tiling, stride); in i915_gem_object_set_tiling()
298 if (vma->fence) in i915_gem_object_set_tiling()
[all …]
H A Di915_gem_render_state.c33 struct i915_vma *vma; member
77 struct drm_i915_gem_object *obj = so->vma->obj; in render_state_setup()
93 u64 r = s + so->vma->node.start; in render_state_setup()
115 so->batch_offset = so->vma->node.start; in render_state_setup()
204 if (IS_ERR(so->vma)) { in i915_gem_render_state_init()
205 ret = PTR_ERR(so->vma); in i915_gem_render_state_init()
232 if (!i915_gem_object_has_pages(so->vma->obj)) in i915_gem_render_state_emit()
263 i915_vma_move_to_active(so->vma, req, 0); in i915_gem_render_state_emit()
265 i915_vma_unpin(so->vma); in i915_gem_render_state_emit()
278 obj = so->vma->obj; in i915_gem_render_state_fini()
[all …]
H A Di915_gem_gtt.c200 ret = vma->vm->allocate_va_range(vma->vm, vma->node.start, in ppgtt_bind_vma()
211 vma->vm->insert_entries(vma->vm, vma, cache_level, pte_flags); in ppgtt_bind_vma()
218 vma->vm->clear_range(vma->vm, vma->node.start, vma->size); in ppgtt_unbind_vma()
225 vma->pages = vma->obj->mm.pages; in ppgtt_set_pages()
236 if (vma->pages != vma->obj->mm.pages) { in clear_pages()
2598 vma->vm->insert_entries(vma->vm, vma, cache_level, pte_flags); in ggtt_bind_vma()
2618 vma->vm->clear_range(vma->vm, vma->node.start, vma->size); in ggtt_unbind_vma()
2653 vma->vm->insert_entries(vma->vm, vma, cache_level, pte_flags); in aliasing_gtt_bind_vma()
2666 vma->vm->clear_range(vma->vm, vma->node.start, vma->size); in aliasing_gtt_unbind_vma()
3812 vma->pages = vma->obj->mm.pages; in i915_get_ggtt_vma_pages()
[all …]
H A Dintel_guc_ct.c110 return ctch->vma != NULL; in ctch_is_open()
116 struct i915_vma *vma; in ctch_init() local
121 GEM_BUG_ON(ctch->vma); in ctch_init()
147 if (IS_ERR(vma)) { in ctch_init()
148 err = PTR_ERR(vma); in ctch_init()
151 ctch->vma = vma; in ctch_init()
171 i915_vma_unpin_and_release(&ctch->vma); in ctch_init()
181 GEM_BUG_ON(!ctch->vma); in ctch_fini()
184 i915_vma_unpin_and_release(&ctch->vma); in ctch_fini()
197 if (!ctch->vma) { in ctch_open()
[all …]
H A Dintel_ringbuffer.c1213 struct i915_vma *vma = ring->vma; in intel_ring_pin() local
1300 if (IS_ERR(vma)) in intel_ring_create_vma()
1303 return vma; in intel_ring_create_vma()
1307 return vma; in intel_ring_create_vma()
1337 if (IS_ERR(vma)) { in intel_engine_create_ring()
1341 ring->vma = vma; in intel_engine_create_ring()
1408 if (IS_ERR(vma)) in alloc_context_vma()
1411 return vma; in alloc_context_vma()
1431 if (IS_ERR(vma)) { in intel_ring_context_pin()
1436 ce->state = vma; in intel_ring_context_pin()
[all …]
H A Dintel_guc_log.c389 subbuf_size = guc->log.vma->obj->base.size;
440 i915_gem_object_unpin_map(guc->log.vma->obj);
456 i915_gem_object_unpin_map(guc->log.vma->obj);
531 struct i915_vma *vma; in intel_guc_log_create() local
537 GEM_BUG_ON(guc->log.vma); in intel_guc_log_create()
557 vma = intel_guc_allocate_vma(guc, size); in intel_guc_log_create()
558 if (IS_ERR(vma)) { in intel_guc_log_create()
559 ret = PTR_ERR(vma); in intel_guc_log_create()
563 guc->log.vma = vma; in intel_guc_log_create()
586 i915_vma_unpin_and_release(&guc->log.vma); in intel_guc_log_create()
[all …]
H A Dintel_huc.c113 static int huc_ucode_xfer(struct intel_uc_fw *huc_fw, struct i915_vma *vma) in huc_ucode_xfer() argument
126 offset = guc_ggtt_offset(vma) + huc_fw->header_offset; in huc_ucode_xfer()
187 struct i915_vma *vma; in intel_huc_auth() local
193 vma = i915_gem_object_ggtt_pin(huc->fw.obj, NULL, 0, 0, in intel_huc_auth()
195 if (IS_ERR(vma)) { in intel_huc_auth()
197 (int)PTR_ERR(vma)); in intel_huc_auth()
202 guc_ggtt_offset(vma) + huc->fw.rsa_offset); in intel_huc_auth()
220 i915_vma_unpin(vma); in intel_huc_auth()
H A Di915_guc_submission.c767 struct i915_vma *vma; in guc_client_alloc() local
791 if (IS_ERR(vma)) { in guc_client_alloc()
792 ret = PTR_ERR(vma); in guc_client_alloc()
797 client->vma = vma; in guc_client_alloc()
897 struct i915_vma *vma; in guc_ads_create() local
915 if (IS_ERR(vma)) in guc_ads_create()
916 return PTR_ERR(vma); in guc_ads_create()
918 guc->ads_vma = vma; in guc_ads_create()
977 struct i915_vma *vma; in i915_guc_submission_init() local
987 if (IS_ERR(vma)) in i915_guc_submission_init()
[all …]
H A Di915_gem.c1085 if (IS_ERR(vma)) { in i915_gem_gtt_pread()
1603 list_move_tail(&vma->vm_link, &vma->vm->inactive_list); in i915_gem_object_bump_inactive_ggtt()
1854 if (vma) in i915_gem_mmap_ioctl()
2383 if (!reg->vma) in i915_gem_runtime_suspend()
3673 if (!--vma->open_count && !i915_vma_is_ggtt(vma)) in i915_gem_close_object()
4261 if (IS_ERR(vma)) in i915_gem_object_pin_to_display_plane()
4275 return vma; in i915_gem_object_pin_to_display_plane()
4279 return vma; in i915_gem_object_pin_to_display_plane()
4442 return vma; in i915_gem_object_ggtt_pin()
4446 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma)) in i915_gem_object_ggtt_pin()
[all …]
H A Dintel_overlay.c174 struct i915_vma *vma; member
307 i915_gem_track_fb(overlay->vma ? overlay->vma->obj : NULL, in intel_overlay_flip_prepare()
308 vma ? vma->obj : NULL, in intel_overlay_flip_prepare()
315 if (vma) in intel_overlay_flip_prepare()
316 overlay->vma = i915_vma_get(vma); in intel_overlay_flip_prepare()
318 overlay->vma = NULL; in intel_overlay_flip_prepare()
367 if (WARN_ON(!vma)) in intel_overlay_release_old_vma()
374 i915_vma_put(vma); in intel_overlay_release_old_vma()
807 if (IS_ERR(vma)) { in intel_overlay_do_put_image()
1429 if (IS_ERR(vma)) { in intel_setup_overlay()
[all …]
H A Di915_mm.c61 int remap_io_mapping(struct vm_area_struct *vma,
68 GEM_BUG_ON((vma->vm_flags &
73 r.mm = vma->vm_mm;
76 (pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK));
80 zap_vma_ptes(vma, addr, (r.pfn - pfn) << PAGE_SHIFT);
H A Dintel_fbdev.c186 struct i915_vma *vma; in intelfb_create() local
223 if (IS_ERR(vma)) { in intelfb_create()
224 ret = PTR_ERR(vma); in intelfb_create()
264 vaddr = i915_vma_pin_iomap(vma); in intelfb_create()
294 ifbdev->vma = vma; in intelfb_create()
302 intel_unpin_fb_vma(vma); in intelfb_create()
538 if (ifbdev->vma) { in intel_fbdev_destroy()
540 intel_unpin_fb_vma(ifbdev->vma); in intel_fbdev_destroy()
778 if (!ifbdev || !ifbdev->vma) in intel_fbdev_set_suspend()
830 if (ifbdev->vma) in intel_fbdev_output_poll_changed()
[all …]
/dragonfly/sys/dev/drm/ttm/
H A Dttm_bo_vm.c117 vma->vm_private_data; in ttm_bo_vm_fault()
217 page_last = vma_pages(vma) + vma->vm_pgoff - in ttm_bo_vm_fault()
230 cvma = *vma; in ttm_bo_vm_fault()
444 bo = ttm_bo_vm_lookup(bdev, vma->vm_pgoff, vma_pages(vma)); in ttm_bo_mmap()
484 if (vma->vm_pgoff != 0) in ttm_fbdev_mmap()
530 memset(vma, 0, sizeof(*vma)); in ttm_bo_vm_fault_dfly()
534 vmf->vma = vma; in ttm_bo_vm_fault_dfly()
664 cvma = *vma; in ttm_bo_vm_fault_dfly()
791 bzero(&vma, sizeof(vma)); in ttm_bo_mmap_single()
793 vma.vm_end = vma.vm_start + size; in ttm_bo_mmap_single()
[all …]
/dragonfly/sys/dev/drm/include/linux/
H A Dmm.h65 struct vm_area_struct *vma; member
80 int (*fault)(struct vm_area_struct *vma, struct vm_fault *vmf);
81 void (*open)(struct vm_area_struct *vma);
82 void (*close)(struct vm_area_struct *vma);
83 int (*access)(struct vm_area_struct *vma, unsigned long addr,
109 io_remap_pfn_range(struct vm_area_struct *vma, in io_remap_pfn_range() argument
113 vma->vm_page_prot = prot; in io_remap_pfn_range()
114 vma->vm_pfn = pfn; in io_remap_pfn_range()
120 vma_pages(struct vm_area_struct *vma) in vma_pages() argument
124 size = vma->vm_end - vma->vm_start; in vma_pages()
/dragonfly/contrib/gdb-7/bfd/
H A Dtekhex.c255 bfd_vma vma; member
267 #define enda(x) (x->vma + x->size)
320 vma &= ~CHUNK_MASK; in find_chunk()
321 while (d && (d->vma) != vma) in find_chunk()
334 d->vma = vma; in find_chunk()
652 bfd_vma vma; in tekhex_set_section_contents() local
658 for (vma = s->vma & ~(bfd_vma) CHUNK_MASK; in tekhex_set_section_contents()
659 vma < s->vma + s->size; in tekhex_set_section_contents()
660 vma += CHUNK_MASK) in tekhex_set_section_contents()
661 find_chunk (abfd, vma); in tekhex_set_section_contents()
[all …]
/dragonfly/contrib/binutils-2.34/ld/
H A Dmri.c42 etree_type *vma; member
82 etree_type *vma, in mri_add_to_list() argument
90 (*ptr)->vma = vma; in mri_add_to_list()
99 mri_output_section (const char *name, etree_type *vma) in mri_output_section() argument
101 mri_add_to_list (&address, name, vma, 0, 0, 0); in mri_output_section()
149 olist->vma = alist->vma; in mri_draw_tree()
157 mri_add_to_list (&order, alist->name, alist->vma, 0, 0, 0); in mri_draw_tree()
210 base = p->vma ? p->vma : exp_nameop (NAME, "."); in mri_draw_tree()
/dragonfly/contrib/binutils-2.27/ld/
H A Dmri.c40 etree_type *vma; member
80 etree_type *vma, in mri_add_to_list() argument
88 (*ptr)->vma = vma; in mri_add_to_list()
97 mri_output_section (const char *name, etree_type *vma) in mri_output_section() argument
99 mri_add_to_list (&address, name, vma, 0, 0, 0); in mri_output_section()
147 olist->vma = alist->vma; in mri_draw_tree()
155 mri_add_to_list (&order, alist->name, alist->vma, 0, 0, 0); in mri_draw_tree()
208 base = p->vma ? p->vma : exp_nameop (NAME, "."); in mri_draw_tree()
/dragonfly/sys/dev/drm/
H A Ddrm_gem.c887 void drm_gem_vm_open(struct vm_area_struct *vma) in drm_gem_vm_open() argument
902 void drm_gem_vm_close(struct vm_area_struct *vma) in drm_gem_vm_close() argument
936 struct vm_area_struct *vma)
941 if (obj_size < vma->vm_end - vma->vm_start)
948 vma->vm_ops = dev->driver->gem_vm_ops;
949 vma->vm_private_data = obj;
950 vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
951 vma->vm_page_prot = pgprot_decrypted(vma->vm_page_prot);
993 vma->vm_pgoff,
994 vma_pages(vma));
[all …]
/dragonfly/games/sail/
H A Dpl_5.c60 int vma, dir; in acceptmove() local
76 vma = ma; in acceptmove()
94 vma = min(ma, maxmove(ms, dir, 0)); in acceptmove()
95 if ((ta < 0 && moved) || (vma < 0 && moved)) in acceptmove()
100 vma--; in acceptmove()
102 if ((ta < 0 && moved) || (vma < 0 && moved)) in acceptmove()
121 vma -= *p - '0'; in acceptmove()
122 if ((ta < 0 && moved) || (vma < 0 && moved)) in acceptmove()
131 if ((ta < 0 && moved) || (vma < 0 && moved) in acceptmove()
/dragonfly/contrib/binutils-2.34/bfd/
H A Dtekhex.c255 bfd_vma vma; member
267 #define enda(x) (x->vma + x->size)
323 vma &= ~CHUNK_MASK; in find_chunk()
324 while (d && (d->vma) != vma) in find_chunk()
337 d->vma = vma; in find_chunk()
413 if (val < section->vma) in first_phase()
414 val = section->vma; in first_phase()
415 section->size = val - section->vma; in first_phase()
824 writevalue (&dst, addr + d->vma); in tekhex_write_object_contents()
842 writevalue (&dst, s->vma); in tekhex_write_object_contents()
[all …]

123456789