/dragonfly/sys/dev/drm/ |
H A D | drm_gem.c | 192 struct vm_object *filp; in drm_gem_object_init() local 196 filp = default_pager_alloc(NULL, size, in drm_gem_object_init() 199 obj->filp = filp; in drm_gem_object_init() 221 obj->filp = NULL; in drm_gem_private_object_init() 237 mutex_lock(&filp->prime.lock); in drm_gem_remove_prime_handles() 242 mutex_unlock(&filp->prime.lock); in drm_gem_remove_prime_handles() 352 lockmgr(&filp->table_lock, LK_RELEASE); in drm_gem_handle_delete() 361 idr_remove(&filp->object_idr, handle); in drm_gem_handle_delete() 362 lockmgr(&filp->table_lock, LK_RELEASE); in drm_gem_handle_delete() 628 lockmgr(&filp->table_lock, LK_RELEASE); in drm_gem_object_lookup() [all …]
|
H A D | drm_file.c | 168 filp->f_mapping = dev->anon_inode->i_mapping; in drm_open() 202 struct file *filp = ap->a_fp; in drm_close() local 205 struct drm_file *file_priv = filp->private_data; in drm_close() 231 drm_legacy_lock_release(dev, filp); in drm_close() 325 filp->private_data = priv; in drm_open_helper() 326 priv->filp = filp; in drm_open_helper() 410 filp->private_data = NULL; in drm_open_helper() 521 drm_legacy_lock_release(dev, filp); in drm_release() 606 struct file *filp = ap->a_fp; in drm_read() local 635 if (filp->f_flag & O_NONBLOCK) { in drm_read() [all …]
|
H A D | drm_internal.h | 65 struct drm_file *filp); 76 struct drm_file *filp); 79 struct drm_file *filp);
|
H A D | drm_lock.c | 343 void drm_legacy_lock_release(struct drm_device *dev, struct file *filp) in drm_legacy_lock_release() argument 346 struct drm_file *file_priv = filp->private_data; in drm_legacy_lock_release() 354 filp, _DRM_LOCKING_CONTEXT(file_priv->master->lock.hw_lock->lock)); in drm_legacy_lock_release()
|
H A D | drm_legacy.h | 108 void drm_legacy_lock_release(struct drm_device *dev, struct file *filp); 116 struct drm_file *filp);
|
/dragonfly/sys/dev/drm/radeon/ |
H A D | radeon_gem.c | 222 struct drm_file *filp) in radeon_gem_info_ioctl() argument 240 struct drm_file *filp) in radeon_gem_pread_ioctl() argument 248 struct drm_file *filp) in radeon_gem_pwrite_ioctl() argument 256 struct drm_file *filp) in radeon_gem_create_ioctl() argument 381 struct drm_file *filp) in radeon_gem_set_domain_ioctl() argument 435 struct drm_file *filp) in radeon_gem_mmap_ioctl() argument 443 struct drm_file *filp) in radeon_gem_busy_ioctl() argument 505 struct drm_file *filp) in radeon_gem_set_tiling_ioctl() argument 523 struct drm_file *filp) in radeon_gem_get_tiling_ioctl() argument 608 struct drm_file *filp) in radeon_gem_va_ioctl() argument [all …]
|
H A D | radeon_vce.c | 168 rdev->vce.filp[i] = NULL; in radeon_vce_init() 316 void radeon_vce_free_handles(struct radeon_device *rdev, struct drm_file *filp) in radeon_vce_free_handles() argument 321 if (!handle || rdev->vce.filp[i] != filp) in radeon_vce_free_handles() 331 rdev->vce.filp[i] = NULL; in radeon_vce_free_handles() 529 if (p->rdev->vce.filp[i] != p->filp) { in radeon_vce_validate_handle() 540 p->rdev->vce.filp[i] = p->filp; in radeon_vce_validate_handle()
|
H A D | radeon_drv.c | 138 int radeon_mmap(struct file *filp, struct vm_area_struct *vma); 140 int radeon_mode_dumb_mmap(struct drm_file *filp, 528 long radeon_drm_ioctl(struct file *filp, in radeon_drm_ioctl() argument 531 struct drm_file *file_priv = filp->private_data; in radeon_drm_ioctl() 539 ret = drm_ioctl(filp, cmd, arg); in radeon_drm_ioctl() 547 static long radeon_kms_compat_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) in radeon_kms_compat_ioctl() argument 553 return drm_compat_ioctl(filp, cmd, arg); in radeon_kms_compat_ioctl() 555 ret = radeon_drm_ioctl(filp, cmd, arg); in radeon_kms_compat_ioctl()
|
H A D | radeon_uvd.c | 221 rdev->uvd.filp[i] = NULL; in radeon_uvd_init() 273 rdev->uvd.filp[i] = NULL; in radeon_uvd_suspend() 328 void radeon_uvd_free_handles(struct radeon_device *rdev, struct drm_file *filp) in radeon_uvd_free_handles() argument 333 if (handle != 0 && rdev->uvd.filp[i] == filp) { in radeon_uvd_free_handles() 348 rdev->uvd.filp[i] = NULL; in radeon_uvd_free_handles() 523 p->rdev->uvd.filp[i] = p->filp; in radeon_uvd_cs_msg() 544 if (p->rdev->uvd.filp[i] != p->filp) { in radeon_uvd_cs_msg()
|
/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_sched.c | 60 struct file *filp = fget(fd); in amdgpu_sched_process_priority_override() 66 if (!filp) in amdgpu_sched_process_priority_override() 69 file = filp->private_data; in amdgpu_sched_process_priority_override() 74 fput(filp); in amdgpu_sched_process_priority_override() 81 struct drm_file *filp); 83 struct drm_file *filp) in amdgpu_sched_ioctl() argument
|
H A D | amdgpu_cs.c | 44 gobj = drm_gem_object_lookup(p->filp, data->handle); in amdgpu_cs_user_fence_chunk() 103 struct amdgpu_fpriv *fpriv = p->filp->driver_priv; in amdgpu_cs_parser_init() 572 struct amdgpu_fpriv *fpriv = p->filp->driver_priv; in amdgpu_cs_parser_bos() 1233 job->owner = p->filp; in amdgpu_cs_submit() 1295 parser.filp = filp; in amdgpu_cs_ioctl() 1350 struct drm_file *filp) in amdgpu_cs_wait_ioctl() argument 1401 struct drm_file *filp, in amdgpu_cs_get_fence() argument 1427 struct drm_file *filp) in amdgpu_cs_fence_to_handle_ioctl() argument 1491 struct drm_file *filp, in amdgpu_cs_wait_all_fences() argument 1536 struct drm_file *filp, in amdgpu_cs_wait_any_fence() argument [all …]
|
H A D | amdgpu_gem.c | 209 struct drm_file *filp) in amdgpu_gem_create_ioctl() argument 212 struct amdgpu_fpriv *fpriv = filp->driver_priv; in amdgpu_gem_create_ioctl() 280 r = drm_gem_handle_create(filp, gobj, &handle); in amdgpu_gem_create_ioctl() 292 struct drm_file *filp) in amdgpu_gem_userptr_ioctl() argument 383 gobj = drm_gem_object_lookup(filp, handle); in amdgpu_mode_dumb_mmap() 399 struct drm_file *filp) in amdgpu_gem_mmap_ioctl() argument 436 struct drm_file *filp) in amdgpu_gem_wait_idle_ioctl() argument 446 gobj = drm_gem_object_lookup(filp, handle); in amdgpu_gem_wait_idle_ioctl() 469 struct drm_file *filp) in amdgpu_gem_metadata_ioctl() argument 551 struct drm_file *filp) in amdgpu_gem_va_ioctl() argument [all …]
|
H A D | amdgpu_ctx.c | 30 static int amdgpu_ctx_priority_permit(struct drm_file *filp, in amdgpu_ctx_priority_permit() argument 42 if (drm_is_current_master(filp)) in amdgpu_ctx_priority_permit() 50 struct drm_file *filp, in amdgpu_ctx_init() argument 59 r = amdgpu_ctx_priority_permit(filp, priority); in amdgpu_ctx_init() 139 struct drm_file *filp, in amdgpu_ctx_alloc() argument 160 r = amdgpu_ctx_init(adev, priority, filp, ctx); in amdgpu_ctx_alloc() 272 struct drm_file *filp) in amdgpu_ctx_ioctl() argument 280 struct amdgpu_fpriv *fpriv = filp->driver_priv; in amdgpu_ctx_ioctl() 293 r = amdgpu_ctx_alloc(adev, fpriv, filp, priority, &id); in amdgpu_ctx_ioctl()
|
H A D | amdgpu_bo_list.c | 58 int amdgpu_bo_list_create(struct amdgpu_device *adev, struct drm_file *filp, in amdgpu_bo_list_create() argument 94 gobj = drm_gem_object_lookup(filp, info[i].bo_handle); in amdgpu_bo_list_create() 255 struct drm_file *filp) in amdgpu_bo_list_ioctl() argument 258 struct amdgpu_fpriv *fpriv = filp->driver_priv; in amdgpu_bo_list_ioctl() 271 r = amdgpu_bo_list_create(adev, filp, info, args->in.bo_number, in amdgpu_bo_list_ioctl() 292 r = amdgpu_bo_list_create(adev, filp, info, args->in.bo_number, in amdgpu_bo_list_ioctl()
|
H A D | amdgpu.h | 675 struct drm_file *filp); 985 struct drm_file *filp; member 1175 struct drm_file *filp); 1177 struct drm_file *filp); 1180 struct drm_file *filp); 1182 struct drm_file *filp); 1184 struct drm_file *filp); 1188 struct drm_file *filp); 1190 struct drm_file *filp); 1196 struct drm_file *filp); [all …]
|
H A D | amdgpu_uvd.h | 63 struct drm_file *filp[AMDGPU_MAX_UVD_HANDLES]; member 80 struct drm_file *filp);
|
H A D | amdgpu_vce.c | 191 adev->vce.filp[i] = NULL; in amdgpu_vce_sw_init() 402 void amdgpu_vce_free_handles(struct amdgpu_device *adev, struct drm_file *filp) in amdgpu_vce_free_handles() argument 409 if (!handle || adev->vce.filp[i] != filp) in amdgpu_vce_free_handles() 416 adev->vce.filp[i] = NULL; in amdgpu_vce_free_handles() 680 if (p->adev->vce.filp[i] != p->filp) { in amdgpu_vce_validate_handle() 691 p->adev->vce.filp[i] = p->filp; in amdgpu_vce_validate_handle()
|
H A D | amdgpu_vce.h | 43 struct drm_file *filp[AMDGPU_MAX_VCE_HANDLES]; member 65 void amdgpu_vce_free_handles(struct amdgpu_device *adev, struct drm_file *filp);
|
H A D | amdgpu_uvd.c | 269 adev->uvd.filp[i] = NULL; in amdgpu_uvd_sw_init() 419 void amdgpu_uvd_free_handles(struct amdgpu_device *adev, struct drm_file *filp) in amdgpu_uvd_free_handles() argument 427 if (handle != 0 && adev->uvd.filp[i] == filp) { in amdgpu_uvd_free_handles() 440 adev->uvd.filp[i] = NULL; in amdgpu_uvd_free_handles() 752 adev->uvd.filp[i] = ctx->parser->filp; in amdgpu_uvd_cs_msg() 770 if (adev->uvd.filp[i] != ctx->parser->filp) { in amdgpu_uvd_cs_msg()
|
/dragonfly/sys/dev/drm/include/drm/ |
H A D | drm_file.h | 242 struct file *filp; member 374 int drm_open(struct inode *inode, struct file *filp); 375 ssize_t drm_read(struct file *filp, char __user *buffer, 378 int drm_release(struct inode *inode, struct file *filp); 379 unsigned int drm_poll(struct file *filp, struct poll_table_struct *wait);
|
H A D | drm_gem.h | 99 struct vm_object *filp; member 101 struct file *filp; member 221 int drm_gem_mmap(struct file *filp, struct vm_area_struct *vma); 316 int drm_gem_handle_delete(struct drm_file *filp, u32 handle); 327 struct drm_gem_object *drm_gem_object_lookup(struct drm_file *filp, u32 handle);
|
H A D | drm_ioctl.h | 69 typedef int drm_ioctl_compat_t(struct file *filp, unsigned int cmd, 178 long drm_compat_ioctl(struct file *filp, unsigned int cmd, unsigned long arg);
|
H A D | drm_gem_cma_helper.h | 77 int drm_gem_cma_mmap(struct file *filp, struct vm_area_struct *vma); 86 unsigned long drm_gem_cma_get_unmapped_area(struct file *filp,
|
/dragonfly/sys/dev/drm/i915/ |
H A D | i915_gem_dmabuf.c | 167 if (!obj->base.filp) in i915_gem_dmabuf_mmap() 171 ret = obj->base.filp->f_op->mmap(obj->base.filp, vma); in i915_gem_dmabuf_mmap() 176 vma->vm_file = get_file(obj->base.filp); in i915_gem_dmabuf_mmap()
|
H A D | i915_gem.c | 1464 vm_obj = obj->base.filp; in i915_gem_shmem_pwrite() 1490 vm_obj, obj->base.filp); in i915_gem_shmem_pwrite() 1776 if (!obj->base.filp) { in i915_gem_mmap_ioctl() 1839 vm_obj = obj->base.filp; in i915_gem_mmap_ioctl() 2475 if (obj->base.filp == NULL) in i915_gem_object_truncate() 2677 mapping = obj->base.filp; in i915_gem_object_get_pages_gtt() 4716 struct file *filp; in i915_gem_object_create_shmem() local 4732 if (IS_ERR(filp)) in i915_gem_object_create_shmem() 4733 return PTR_ERR(filp); in i915_gem_object_create_shmem() 4735 obj->filp = filp; in i915_gem_object_create_shmem() [all …]
|