xref: /linux/drivers/gpu/drm/panthor/panthor_drv.c (revision 1a9a7143)
1 // SPDX-License-Identifier: GPL-2.0 or MIT
2 /* Copyright 2018 Marty E. Plummer <hanetzer@startmail.com> */
3 /* Copyright 2019 Linaro, Ltd., Rob Herring <robh@kernel.org> */
4 /* Copyright 2019 Collabora ltd. */
5 
6 #include <linux/list.h>
7 #include <linux/module.h>
8 #include <linux/of_platform.h>
9 #include <linux/pagemap.h>
10 #include <linux/platform_device.h>
11 #include <linux/pm_runtime.h>
12 
13 #include <drm/drm_debugfs.h>
14 #include <drm/drm_drv.h>
15 #include <drm/drm_exec.h>
16 #include <drm/drm_ioctl.h>
17 #include <drm/drm_syncobj.h>
18 #include <drm/drm_utils.h>
19 #include <drm/gpu_scheduler.h>
20 #include <drm/panthor_drm.h>
21 
22 #include "panthor_device.h"
23 #include "panthor_fw.h"
24 #include "panthor_gem.h"
25 #include "panthor_gpu.h"
26 #include "panthor_heap.h"
27 #include "panthor_mmu.h"
28 #include "panthor_regs.h"
29 #include "panthor_sched.h"
30 
31 /**
32  * DOC: user <-> kernel object copy helpers.
33  */
34 
35 /**
36  * panthor_set_uobj() - Copy kernel object to user object.
37  * @usr_ptr: Users pointer.
38  * @usr_size: Size of the user object.
39  * @min_size: Minimum size for this object.
40  * @kern_size: Size of the kernel object.
41  * @in: Address of the kernel object to copy.
42  *
43  * Helper automating kernel -> user object copies.
44  *
45  * Don't use this function directly, use PANTHOR_UOBJ_SET() instead.
46  *
47  * Return: 0 on success, a negative error code otherwise.
48  */
49 static int
50 panthor_set_uobj(u64 usr_ptr, u32 usr_size, u32 min_size, u32 kern_size, const void *in)
51 {
52 	/* User size shouldn't be smaller than the minimal object size. */
53 	if (usr_size < min_size)
54 		return -EINVAL;
55 
56 	if (copy_to_user(u64_to_user_ptr(usr_ptr), in, min_t(u32, usr_size, kern_size)))
57 		return -EFAULT;
58 
59 	/* When the kernel object is smaller than the user object, we fill the gap with
60 	 * zeros.
61 	 */
62 	if (usr_size > kern_size &&
63 	    clear_user(u64_to_user_ptr(usr_ptr + kern_size), usr_size - kern_size)) {
64 		return -EFAULT;
65 	}
66 
67 	return 0;
68 }
69 
70 /**
71  * panthor_get_uobj_array() - Copy a user object array into a kernel accessible object array.
72  * @in: The object array to copy.
73  * @min_stride: Minimum array stride.
74  * @obj_size: Kernel object size.
75  *
76  * Helper automating user -> kernel object copies.
77  *
78  * Don't use this function directly, use PANTHOR_UOBJ_GET_ARRAY() instead.
79  *
80  * Return: newly allocated object array or an ERR_PTR on error.
81  */
82 static void *
83 panthor_get_uobj_array(const struct drm_panthor_obj_array *in, u32 min_stride,
84 		       u32 obj_size)
85 {
86 	int ret = 0;
87 	void *out_alloc;
88 
89 	if (!in->count)
90 		return NULL;
91 
92 	/* User stride must be at least the minimum object size, otherwise it might
93 	 * lack useful information.
94 	 */
95 	if (in->stride < min_stride)
96 		return ERR_PTR(-EINVAL);
97 
98 	out_alloc = kvmalloc_array(in->count, obj_size, GFP_KERNEL);
99 	if (!out_alloc)
100 		return ERR_PTR(-ENOMEM);
101 
102 	if (obj_size == in->stride) {
103 		/* Fast path when user/kernel have the same uAPI header version. */
104 		if (copy_from_user(out_alloc, u64_to_user_ptr(in->array),
105 				   (unsigned long)obj_size * in->count))
106 			ret = -EFAULT;
107 	} else {
108 		void __user *in_ptr = u64_to_user_ptr(in->array);
109 		void *out_ptr = out_alloc;
110 
111 		/* If the sizes differ, we need to copy elements one by one. */
112 		for (u32 i = 0; i < in->count; i++) {
113 			ret = copy_struct_from_user(out_ptr, obj_size, in_ptr, in->stride);
114 			if (ret)
115 				break;
116 
117 			out_ptr += obj_size;
118 			in_ptr += in->stride;
119 		}
120 	}
121 
122 	if (ret) {
123 		kvfree(out_alloc);
124 		return ERR_PTR(ret);
125 	}
126 
127 	return out_alloc;
128 }
129 
130 /**
131  * PANTHOR_UOBJ_MIN_SIZE_INTERNAL() - Get the minimum user object size
132  * @_typename: Object type.
133  * @_last_mandatory_field: Last mandatory field.
134  *
135  * Get the minimum user object size based on the last mandatory field name,
136  * A.K.A, the name of the last field of the structure at the time this
137  * structure was added to the uAPI.
138  *
139  * Don't use directly, use PANTHOR_UOBJ_DECL() instead.
140  */
141 #define PANTHOR_UOBJ_MIN_SIZE_INTERNAL(_typename, _last_mandatory_field) \
142 	(offsetof(_typename, _last_mandatory_field) + \
143 	 sizeof(((_typename *)NULL)->_last_mandatory_field))
144 
145 /**
146  * PANTHOR_UOBJ_DECL() - Declare a new uAPI object whose subject to
147  * evolutions.
148  * @_typename: Object type.
149  * @_last_mandatory_field: Last mandatory field.
150  *
151  * Should be used to extend the PANTHOR_UOBJ_MIN_SIZE() list.
152  */
153 #define PANTHOR_UOBJ_DECL(_typename, _last_mandatory_field) \
154 	_typename : PANTHOR_UOBJ_MIN_SIZE_INTERNAL(_typename, _last_mandatory_field)
155 
156 /**
157  * PANTHOR_UOBJ_MIN_SIZE() - Get the minimum size of a given uAPI object
158  * @_obj_name: Object to get the minimum size of.
159  *
160  * Don't use this macro directly, it's automatically called by
161  * PANTHOR_UOBJ_{SET,GET_ARRAY}().
162  */
163 #define PANTHOR_UOBJ_MIN_SIZE(_obj_name) \
164 	_Generic(_obj_name, \
165 		 PANTHOR_UOBJ_DECL(struct drm_panthor_gpu_info, tiler_present), \
166 		 PANTHOR_UOBJ_DECL(struct drm_panthor_csif_info, pad), \
167 		 PANTHOR_UOBJ_DECL(struct drm_panthor_sync_op, timeline_value), \
168 		 PANTHOR_UOBJ_DECL(struct drm_panthor_queue_submit, syncs), \
169 		 PANTHOR_UOBJ_DECL(struct drm_panthor_queue_create, ringbuf_size), \
170 		 PANTHOR_UOBJ_DECL(struct drm_panthor_vm_bind_op, syncs))
171 
172 /**
173  * PANTHOR_UOBJ_SET() - Copy a kernel object to a user object.
174  * @_dest_usr_ptr: User pointer to copy to.
175  * @_usr_size: Size of the user object.
176  * @_src_obj: Kernel object to copy (not a pointer).
177  *
178  * Return: 0 on success, a negative error code otherwise.
179  */
180 #define PANTHOR_UOBJ_SET(_dest_usr_ptr, _usr_size, _src_obj) \
181 	panthor_set_uobj(_dest_usr_ptr, _usr_size, \
182 			 PANTHOR_UOBJ_MIN_SIZE(_src_obj), \
183 			 sizeof(_src_obj), &(_src_obj))
184 
185 /**
186  * PANTHOR_UOBJ_GET_ARRAY() - Copy a user object array to a kernel accessible
187  * object array.
188  * @_dest_array: Local variable that will hold the newly allocated kernel
189  * object array.
190  * @_uobj_array: The drm_panthor_obj_array object describing the user object
191  * array.
192  *
193  * Return: 0 on success, a negative error code otherwise.
194  */
195 #define PANTHOR_UOBJ_GET_ARRAY(_dest_array, _uobj_array) \
196 	({ \
197 		typeof(_dest_array) _tmp; \
198 		_tmp = panthor_get_uobj_array(_uobj_array, \
199 					      PANTHOR_UOBJ_MIN_SIZE((_dest_array)[0]), \
200 					      sizeof((_dest_array)[0])); \
201 		if (!IS_ERR(_tmp)) \
202 			_dest_array = _tmp; \
203 		PTR_ERR_OR_ZERO(_tmp); \
204 	})
205 
206 /**
207  * struct panthor_sync_signal - Represent a synchronization object point to attach
208  * our job fence to.
209  *
210  * This structure is here to keep track of fences that are currently bound to
211  * a specific syncobj point.
212  *
213  * At the beginning of a job submission, the fence
214  * is retrieved from the syncobj itself, and can be NULL if no fence was attached
215  * to this point.
216  *
217  * At the end, it points to the fence of the last job that had a
218  * %DRM_PANTHOR_SYNC_OP_SIGNAL on this syncobj.
219  *
220  * With jobs being submitted in batches, the fence might change several times during
221  * the process, allowing one job to wait on a job that's part of the same submission
222  * but appears earlier in the drm_panthor_group_submit::queue_submits array.
223  */
224 struct panthor_sync_signal {
225 	/** @node: list_head to track signal ops within a submit operation */
226 	struct list_head node;
227 
228 	/** @handle: The syncobj handle. */
229 	u32 handle;
230 
231 	/**
232 	 * @point: The syncobj point.
233 	 *
234 	 * Zero for regular syncobjs, and non-zero for timeline syncobjs.
235 	 */
236 	u64 point;
237 
238 	/**
239 	 * @syncobj: The sync object pointed by @handle.
240 	 */
241 	struct drm_syncobj *syncobj;
242 
243 	/**
244 	 * @chain: Chain object used to link the new fence to an existing
245 	 * timeline syncobj.
246 	 *
247 	 * NULL for regular syncobj, non-NULL for timeline syncobjs.
248 	 */
249 	struct dma_fence_chain *chain;
250 
251 	/**
252 	 * @fence: The fence to assign to the syncobj or syncobj-point.
253 	 */
254 	struct dma_fence *fence;
255 };
256 
257 /**
258  * struct panthor_job_ctx - Job context
259  */
260 struct panthor_job_ctx {
261 	/** @job: The job that is about to be submitted to drm_sched. */
262 	struct drm_sched_job *job;
263 
264 	/** @syncops: Array of sync operations. */
265 	struct drm_panthor_sync_op *syncops;
266 
267 	/** @syncop_count: Number of sync operations. */
268 	u32 syncop_count;
269 };
270 
271 /**
272  * struct panthor_submit_ctx - Submission context
273  *
274  * Anything that's related to a submission (%DRM_IOCTL_PANTHOR_VM_BIND or
275  * %DRM_IOCTL_PANTHOR_GROUP_SUBMIT) is kept here, so we can automate the
276  * initialization and cleanup steps.
277  */
278 struct panthor_submit_ctx {
279 	/** @file: DRM file this submission happens on. */
280 	struct drm_file *file;
281 
282 	/**
283 	 * @signals: List of struct panthor_sync_signal.
284 	 *
285 	 * %DRM_PANTHOR_SYNC_OP_SIGNAL operations will be recorded here,
286 	 * and %DRM_PANTHOR_SYNC_OP_WAIT will first check if an entry
287 	 * matching the syncobj+point exists before calling
288 	 * drm_syncobj_find_fence(). This allows us to describe dependencies
289 	 * existing between jobs that are part of the same batch.
290 	 */
291 	struct list_head signals;
292 
293 	/** @jobs: Array of jobs. */
294 	struct panthor_job_ctx *jobs;
295 
296 	/** @job_count: Number of entries in the @jobs array. */
297 	u32 job_count;
298 
299 	/** @exec: drm_exec context used to acquire and prepare resv objects. */
300 	struct drm_exec exec;
301 };
302 
303 #define PANTHOR_SYNC_OP_FLAGS_MASK \
304 	(DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK | DRM_PANTHOR_SYNC_OP_SIGNAL)
305 
306 static bool sync_op_is_signal(const struct drm_panthor_sync_op *sync_op)
307 {
308 	return !!(sync_op->flags & DRM_PANTHOR_SYNC_OP_SIGNAL);
309 }
310 
311 static bool sync_op_is_wait(const struct drm_panthor_sync_op *sync_op)
312 {
313 	/* Note that DRM_PANTHOR_SYNC_OP_WAIT == 0 */
314 	return !(sync_op->flags & DRM_PANTHOR_SYNC_OP_SIGNAL);
315 }
316 
317 /**
318  * panthor_check_sync_op() - Check drm_panthor_sync_op fields
319  * @sync_op: The sync operation to check.
320  *
321  * Return: 0 on success, -EINVAL otherwise.
322  */
323 static int
324 panthor_check_sync_op(const struct drm_panthor_sync_op *sync_op)
325 {
326 	u8 handle_type;
327 
328 	if (sync_op->flags & ~PANTHOR_SYNC_OP_FLAGS_MASK)
329 		return -EINVAL;
330 
331 	handle_type = sync_op->flags & DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_MASK;
332 	if (handle_type != DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ &&
333 	    handle_type != DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_TIMELINE_SYNCOBJ)
334 		return -EINVAL;
335 
336 	if (handle_type == DRM_PANTHOR_SYNC_OP_HANDLE_TYPE_SYNCOBJ &&
337 	    sync_op->timeline_value != 0)
338 		return -EINVAL;
339 
340 	return 0;
341 }
342 
343 /**
344  * panthor_sync_signal_free() - Release resources and free a panthor_sync_signal object
345  * @sig_sync: Signal object to free.
346  */
347 static void
348 panthor_sync_signal_free(struct panthor_sync_signal *sig_sync)
349 {
350 	if (!sig_sync)
351 		return;
352 
353 	drm_syncobj_put(sig_sync->syncobj);
354 	dma_fence_chain_free(sig_sync->chain);
355 	dma_fence_put(sig_sync->fence);
356 	kfree(sig_sync);
357 }
358 
359 /**
360  * panthor_submit_ctx_add_sync_signal() - Add a signal operation to a submit context
361  * @ctx: Context to add the signal operation to.
362  * @handle: Syncobj handle.
363  * @point: Syncobj point.
364  *
365  * Return: 0 on success, otherwise negative error value.
366  */
367 static int
368 panthor_submit_ctx_add_sync_signal(struct panthor_submit_ctx *ctx, u32 handle, u64 point)
369 {
370 	struct panthor_sync_signal *sig_sync;
371 	struct dma_fence *cur_fence;
372 	int ret;
373 
374 	sig_sync = kzalloc(sizeof(*sig_sync), GFP_KERNEL);
375 	if (!sig_sync)
376 		return -ENOMEM;
377 
378 	sig_sync->handle = handle;
379 	sig_sync->point = point;
380 
381 	if (point > 0) {
382 		sig_sync->chain = dma_fence_chain_alloc();
383 		if (!sig_sync->chain) {
384 			ret = -ENOMEM;
385 			goto err_free_sig_sync;
386 		}
387 	}
388 
389 	sig_sync->syncobj = drm_syncobj_find(ctx->file, handle);
390 	if (!sig_sync->syncobj) {
391 		ret = -EINVAL;
392 		goto err_free_sig_sync;
393 	}
394 
395 	/* Retrieve the current fence attached to that point. It's
396 	 * perfectly fine to get a NULL fence here, it just means there's
397 	 * no fence attached to that point yet.
398 	 */
399 	if (!drm_syncobj_find_fence(ctx->file, handle, point, 0, &cur_fence))
400 		sig_sync->fence = cur_fence;
401 
402 	list_add_tail(&sig_sync->node, &ctx->signals);
403 
404 	return 0;
405 
406 err_free_sig_sync:
407 	panthor_sync_signal_free(sig_sync);
408 	return ret;
409 }
410 
411 /**
412  * panthor_submit_ctx_search_sync_signal() - Search an existing signal operation in a
413  * submit context.
414  * @ctx: Context to search the signal operation in.
415  * @handle: Syncobj handle.
416  * @point: Syncobj point.
417  *
418  * Return: A valid panthor_sync_signal object if found, NULL otherwise.
419  */
420 static struct panthor_sync_signal *
421 panthor_submit_ctx_search_sync_signal(struct panthor_submit_ctx *ctx, u32 handle, u64 point)
422 {
423 	struct panthor_sync_signal *sig_sync;
424 
425 	list_for_each_entry(sig_sync, &ctx->signals, node) {
426 		if (handle == sig_sync->handle && point == sig_sync->point)
427 			return sig_sync;
428 	}
429 
430 	return NULL;
431 }
432 
433 /**
434  * panthor_submit_ctx_add_job() - Add a job to a submit context
435  * @ctx: Context to search the signal operation in.
436  * @idx: Index of the job in the context.
437  * @job: Job to add.
438  * @syncs: Sync operations provided by userspace.
439  *
440  * Return: 0 on success, a negative error code otherwise.
441  */
442 static int
443 panthor_submit_ctx_add_job(struct panthor_submit_ctx *ctx, u32 idx,
444 			   struct drm_sched_job *job,
445 			   const struct drm_panthor_obj_array *syncs)
446 {
447 	int ret;
448 
449 	ctx->jobs[idx].job = job;
450 
451 	ret = PANTHOR_UOBJ_GET_ARRAY(ctx->jobs[idx].syncops, syncs);
452 	if (ret)
453 		return ret;
454 
455 	ctx->jobs[idx].syncop_count = syncs->count;
456 	return 0;
457 }
458 
459 /**
460  * panthor_submit_ctx_get_sync_signal() - Search signal operation and add one if none was found.
461  * @ctx: Context to search the signal operation in.
462  * @handle: Syncobj handle.
463  * @point: Syncobj point.
464  *
465  * Return: 0 on success, a negative error code otherwise.
466  */
467 static int
468 panthor_submit_ctx_get_sync_signal(struct panthor_submit_ctx *ctx, u32 handle, u64 point)
469 {
470 	struct panthor_sync_signal *sig_sync;
471 
472 	sig_sync = panthor_submit_ctx_search_sync_signal(ctx, handle, point);
473 	if (sig_sync)
474 		return 0;
475 
476 	return panthor_submit_ctx_add_sync_signal(ctx, handle, point);
477 }
478 
479 /**
480  * panthor_submit_ctx_update_job_sync_signal_fences() - Update fences
481  * on the signal operations specified by a job.
482  * @ctx: Context to search the signal operation in.
483  * @job_idx: Index of the job to operate on.
484  *
485  * Return: 0 on success, a negative error code otherwise.
486  */
487 static int
488 panthor_submit_ctx_update_job_sync_signal_fences(struct panthor_submit_ctx *ctx,
489 						 u32 job_idx)
490 {
491 	struct panthor_device *ptdev = container_of(ctx->file->minor->dev,
492 						    struct panthor_device,
493 						    base);
494 	struct dma_fence *done_fence = &ctx->jobs[job_idx].job->s_fence->finished;
495 	const struct drm_panthor_sync_op *sync_ops = ctx->jobs[job_idx].syncops;
496 	u32 sync_op_count = ctx->jobs[job_idx].syncop_count;
497 
498 	for (u32 i = 0; i < sync_op_count; i++) {
499 		struct dma_fence *old_fence;
500 		struct panthor_sync_signal *sig_sync;
501 
502 		if (!sync_op_is_signal(&sync_ops[i]))
503 			continue;
504 
505 		sig_sync = panthor_submit_ctx_search_sync_signal(ctx, sync_ops[i].handle,
506 								 sync_ops[i].timeline_value);
507 		if (drm_WARN_ON(&ptdev->base, !sig_sync))
508 			return -EINVAL;
509 
510 		old_fence = sig_sync->fence;
511 		sig_sync->fence = dma_fence_get(done_fence);
512 		dma_fence_put(old_fence);
513 
514 		if (drm_WARN_ON(&ptdev->base, !sig_sync->fence))
515 			return -EINVAL;
516 	}
517 
518 	return 0;
519 }
520 
521 /**
522  * panthor_submit_ctx_collect_job_signal_ops() - Iterate over all job signal operations
523  * and add them to the context.
524  * @ctx: Context to search the signal operation in.
525  * @job_idx: Index of the job to operate on.
526  *
527  * Return: 0 on success, a negative error code otherwise.
528  */
529 static int
530 panthor_submit_ctx_collect_job_signal_ops(struct panthor_submit_ctx *ctx,
531 					  u32 job_idx)
532 {
533 	const struct drm_panthor_sync_op *sync_ops = ctx->jobs[job_idx].syncops;
534 	u32 sync_op_count = ctx->jobs[job_idx].syncop_count;
535 
536 	for (u32 i = 0; i < sync_op_count; i++) {
537 		int ret;
538 
539 		if (!sync_op_is_signal(&sync_ops[i]))
540 			continue;
541 
542 		ret = panthor_check_sync_op(&sync_ops[i]);
543 		if (ret)
544 			return ret;
545 
546 		ret = panthor_submit_ctx_get_sync_signal(ctx,
547 							 sync_ops[i].handle,
548 							 sync_ops[i].timeline_value);
549 		if (ret)
550 			return ret;
551 	}
552 
553 	return 0;
554 }
555 
556 /**
557  * panthor_submit_ctx_push_fences() - Iterate over the signal array, and for each entry, push
558  * the currently assigned fence to the associated syncobj.
559  * @ctx: Context to push fences on.
560  *
561  * This is the last step of a submission procedure, and is done once we know the submission
562  * is effective and job fences are guaranteed to be signaled in finite time.
563  */
564 static void
565 panthor_submit_ctx_push_fences(struct panthor_submit_ctx *ctx)
566 {
567 	struct panthor_sync_signal *sig_sync;
568 
569 	list_for_each_entry(sig_sync, &ctx->signals, node) {
570 		if (sig_sync->chain) {
571 			drm_syncobj_add_point(sig_sync->syncobj, sig_sync->chain,
572 					      sig_sync->fence, sig_sync->point);
573 			sig_sync->chain = NULL;
574 		} else {
575 			drm_syncobj_replace_fence(sig_sync->syncobj, sig_sync->fence);
576 		}
577 	}
578 }
579 
580 /**
581  * panthor_submit_ctx_add_sync_deps_to_job() - Add sync wait operations as
582  * job dependencies.
583  * @ctx: Submit context.
584  * @job_idx: Index of the job to operate on.
585  *
586  * Return: 0 on success, a negative error code otherwise.
587  */
588 static int
589 panthor_submit_ctx_add_sync_deps_to_job(struct panthor_submit_ctx *ctx,
590 					u32 job_idx)
591 {
592 	struct panthor_device *ptdev = container_of(ctx->file->minor->dev,
593 						    struct panthor_device,
594 						    base);
595 	const struct drm_panthor_sync_op *sync_ops = ctx->jobs[job_idx].syncops;
596 	struct drm_sched_job *job = ctx->jobs[job_idx].job;
597 	u32 sync_op_count = ctx->jobs[job_idx].syncop_count;
598 	int ret = 0;
599 
600 	for (u32 i = 0; i < sync_op_count; i++) {
601 		struct panthor_sync_signal *sig_sync;
602 		struct dma_fence *fence;
603 
604 		if (!sync_op_is_wait(&sync_ops[i]))
605 			continue;
606 
607 		ret = panthor_check_sync_op(&sync_ops[i]);
608 		if (ret)
609 			return ret;
610 
611 		sig_sync = panthor_submit_ctx_search_sync_signal(ctx, sync_ops[i].handle,
612 								 sync_ops[i].timeline_value);
613 		if (sig_sync) {
614 			if (drm_WARN_ON(&ptdev->base, !sig_sync->fence))
615 				return -EINVAL;
616 
617 			fence = dma_fence_get(sig_sync->fence);
618 		} else {
619 			ret = drm_syncobj_find_fence(ctx->file, sync_ops[i].handle,
620 						     sync_ops[i].timeline_value,
621 						     0, &fence);
622 			if (ret)
623 				return ret;
624 		}
625 
626 		ret = drm_sched_job_add_dependency(job, fence);
627 		if (ret)
628 			return ret;
629 	}
630 
631 	return 0;
632 }
633 
634 /**
635  * panthor_submit_ctx_collect_jobs_signal_ops() - Collect all signal operations
636  * and add them to the submit context.
637  * @ctx: Submit context.
638  *
639  * Return: 0 on success, a negative error code otherwise.
640  */
641 static int
642 panthor_submit_ctx_collect_jobs_signal_ops(struct panthor_submit_ctx *ctx)
643 {
644 	for (u32 i = 0; i < ctx->job_count; i++) {
645 		int ret;
646 
647 		ret = panthor_submit_ctx_collect_job_signal_ops(ctx, i);
648 		if (ret)
649 			return ret;
650 	}
651 
652 	return 0;
653 }
654 
655 /**
656  * panthor_submit_ctx_add_deps_and_arm_jobs() - Add jobs dependencies and arm jobs
657  * @ctx: Submit context.
658  *
659  * Must be called after the resv preparation has been taken care of.
660  *
661  * Return: 0 on success, a negative error code otherwise.
662  */
663 static int
664 panthor_submit_ctx_add_deps_and_arm_jobs(struct panthor_submit_ctx *ctx)
665 {
666 	for (u32 i = 0; i < ctx->job_count; i++) {
667 		int ret;
668 
669 		ret = panthor_submit_ctx_add_sync_deps_to_job(ctx, i);
670 		if (ret)
671 			return ret;
672 
673 		drm_sched_job_arm(ctx->jobs[i].job);
674 
675 		ret = panthor_submit_ctx_update_job_sync_signal_fences(ctx, i);
676 		if (ret)
677 			return ret;
678 	}
679 
680 	return 0;
681 }
682 
683 /**
684  * panthor_submit_ctx_push_jobs() - Push jobs to their scheduling entities.
685  * @ctx: Submit context.
686  * @upd_resvs: Callback used to update reservation objects that were previously
687  * preapred.
688  */
689 static void
690 panthor_submit_ctx_push_jobs(struct panthor_submit_ctx *ctx,
691 			     void (*upd_resvs)(struct drm_exec *, struct drm_sched_job *))
692 {
693 	for (u32 i = 0; i < ctx->job_count; i++) {
694 		upd_resvs(&ctx->exec, ctx->jobs[i].job);
695 		drm_sched_entity_push_job(ctx->jobs[i].job);
696 
697 		/* Job is owned by the scheduler now. */
698 		ctx->jobs[i].job = NULL;
699 	}
700 
701 	panthor_submit_ctx_push_fences(ctx);
702 }
703 
704 /**
705  * panthor_submit_ctx_init() - Initializes a submission context
706  * @ctx: Submit context to initialize.
707  * @file: drm_file this submission happens on.
708  * @job_count: Number of jobs that will be submitted.
709  *
710  * Return: 0 on success, a negative error code otherwise.
711  */
712 static int panthor_submit_ctx_init(struct panthor_submit_ctx *ctx,
713 				   struct drm_file *file, u32 job_count)
714 {
715 	ctx->jobs = kvmalloc_array(job_count, sizeof(*ctx->jobs),
716 				   GFP_KERNEL | __GFP_ZERO);
717 	if (!ctx->jobs)
718 		return -ENOMEM;
719 
720 	ctx->file = file;
721 	ctx->job_count = job_count;
722 	INIT_LIST_HEAD(&ctx->signals);
723 	drm_exec_init(&ctx->exec,
724 		      DRM_EXEC_INTERRUPTIBLE_WAIT | DRM_EXEC_IGNORE_DUPLICATES,
725 		      0);
726 	return 0;
727 }
728 
729 /**
730  * panthor_submit_ctx_cleanup() - Cleanup a submission context
731  * @ctx: Submit context to cleanup.
732  * @job_put: Job put callback.
733  */
734 static void panthor_submit_ctx_cleanup(struct panthor_submit_ctx *ctx,
735 				       void (*job_put)(struct drm_sched_job *))
736 {
737 	struct panthor_sync_signal *sig_sync, *tmp;
738 	unsigned long i;
739 
740 	drm_exec_fini(&ctx->exec);
741 
742 	list_for_each_entry_safe(sig_sync, tmp, &ctx->signals, node)
743 		panthor_sync_signal_free(sig_sync);
744 
745 	for (i = 0; i < ctx->job_count; i++) {
746 		job_put(ctx->jobs[i].job);
747 		kvfree(ctx->jobs[i].syncops);
748 	}
749 
750 	kvfree(ctx->jobs);
751 }
752 
753 static int panthor_ioctl_dev_query(struct drm_device *ddev, void *data, struct drm_file *file)
754 {
755 	struct panthor_device *ptdev = container_of(ddev, struct panthor_device, base);
756 	struct drm_panthor_dev_query *args = data;
757 
758 	if (!args->pointer) {
759 		switch (args->type) {
760 		case DRM_PANTHOR_DEV_QUERY_GPU_INFO:
761 			args->size = sizeof(ptdev->gpu_info);
762 			return 0;
763 
764 		case DRM_PANTHOR_DEV_QUERY_CSIF_INFO:
765 			args->size = sizeof(ptdev->csif_info);
766 			return 0;
767 
768 		default:
769 			return -EINVAL;
770 		}
771 	}
772 
773 	switch (args->type) {
774 	case DRM_PANTHOR_DEV_QUERY_GPU_INFO:
775 		return PANTHOR_UOBJ_SET(args->pointer, args->size, ptdev->gpu_info);
776 
777 	case DRM_PANTHOR_DEV_QUERY_CSIF_INFO:
778 		return PANTHOR_UOBJ_SET(args->pointer, args->size, ptdev->csif_info);
779 
780 	default:
781 		return -EINVAL;
782 	}
783 }
784 
785 #define PANTHOR_VM_CREATE_FLAGS			0
786 
787 static int panthor_ioctl_vm_create(struct drm_device *ddev, void *data,
788 				   struct drm_file *file)
789 {
790 	struct panthor_device *ptdev = container_of(ddev, struct panthor_device, base);
791 	struct panthor_file *pfile = file->driver_priv;
792 	struct drm_panthor_vm_create *args = data;
793 	int cookie, ret;
794 
795 	if (!drm_dev_enter(ddev, &cookie))
796 		return -ENODEV;
797 
798 	ret = panthor_vm_pool_create_vm(ptdev, pfile->vms,  args);
799 	if (ret >= 0) {
800 		args->id = ret;
801 		ret = 0;
802 	}
803 
804 	drm_dev_exit(cookie);
805 	return ret;
806 }
807 
808 static int panthor_ioctl_vm_destroy(struct drm_device *ddev, void *data,
809 				    struct drm_file *file)
810 {
811 	struct panthor_file *pfile = file->driver_priv;
812 	struct drm_panthor_vm_destroy *args = data;
813 
814 	if (args->pad)
815 		return -EINVAL;
816 
817 	return panthor_vm_pool_destroy_vm(pfile->vms, args->id);
818 }
819 
820 #define PANTHOR_BO_FLAGS		DRM_PANTHOR_BO_NO_MMAP
821 
822 static int panthor_ioctl_bo_create(struct drm_device *ddev, void *data,
823 				   struct drm_file *file)
824 {
825 	struct panthor_file *pfile = file->driver_priv;
826 	struct drm_panthor_bo_create *args = data;
827 	struct panthor_vm *vm = NULL;
828 	int cookie, ret;
829 
830 	if (!drm_dev_enter(ddev, &cookie))
831 		return -ENODEV;
832 
833 	if (!args->size || args->pad ||
834 	    (args->flags & ~PANTHOR_BO_FLAGS)) {
835 		ret = -EINVAL;
836 		goto out_dev_exit;
837 	}
838 
839 	if (args->exclusive_vm_id) {
840 		vm = panthor_vm_pool_get_vm(pfile->vms, args->exclusive_vm_id);
841 		if (!vm) {
842 			ret = -EINVAL;
843 			goto out_dev_exit;
844 		}
845 	}
846 
847 	ret = panthor_gem_create_with_handle(file, ddev, vm, &args->size,
848 					     args->flags, &args->handle);
849 
850 	panthor_vm_put(vm);
851 
852 out_dev_exit:
853 	drm_dev_exit(cookie);
854 	return ret;
855 }
856 
857 static int panthor_ioctl_bo_mmap_offset(struct drm_device *ddev, void *data,
858 					struct drm_file *file)
859 {
860 	struct drm_panthor_bo_mmap_offset *args = data;
861 	struct drm_gem_object *obj;
862 	int ret;
863 
864 	if (args->pad)
865 		return -EINVAL;
866 
867 	obj = drm_gem_object_lookup(file, args->handle);
868 	if (!obj)
869 		return -ENOENT;
870 
871 	ret = drm_gem_create_mmap_offset(obj);
872 	if (ret)
873 		goto out;
874 
875 	args->offset = drm_vma_node_offset_addr(&obj->vma_node);
876 
877 out:
878 	drm_gem_object_put(obj);
879 	return ret;
880 }
881 
882 static int panthor_ioctl_group_submit(struct drm_device *ddev, void *data,
883 				      struct drm_file *file)
884 {
885 	struct panthor_file *pfile = file->driver_priv;
886 	struct drm_panthor_group_submit *args = data;
887 	struct drm_panthor_queue_submit *jobs_args;
888 	struct panthor_submit_ctx ctx;
889 	int ret = 0, cookie;
890 
891 	if (args->pad)
892 		return -EINVAL;
893 
894 	if (!drm_dev_enter(ddev, &cookie))
895 		return -ENODEV;
896 
897 	ret = PANTHOR_UOBJ_GET_ARRAY(jobs_args, &args->queue_submits);
898 	if (ret)
899 		goto out_dev_exit;
900 
901 	ret = panthor_submit_ctx_init(&ctx, file, args->queue_submits.count);
902 	if (ret)
903 		goto out_free_jobs_args;
904 
905 	/* Create jobs and attach sync operations */
906 	for (u32 i = 0; i < args->queue_submits.count; i++) {
907 		const struct drm_panthor_queue_submit *qsubmit = &jobs_args[i];
908 		struct drm_sched_job *job;
909 
910 		job = panthor_job_create(pfile, args->group_handle, qsubmit);
911 		if (IS_ERR(job)) {
912 			ret = PTR_ERR(job);
913 			goto out_cleanup_submit_ctx;
914 		}
915 
916 		ret = panthor_submit_ctx_add_job(&ctx, i, job, &qsubmit->syncs);
917 		if (ret)
918 			goto out_cleanup_submit_ctx;
919 	}
920 
921 	/*
922 	 * Collect signal operations on all jobs, such that each job can pick
923 	 * from it for its dependencies and update the fence to signal when the
924 	 * job is submitted.
925 	 */
926 	ret = panthor_submit_ctx_collect_jobs_signal_ops(&ctx);
927 	if (ret)
928 		goto out_cleanup_submit_ctx;
929 
930 	/*
931 	 * We acquire/prepare revs on all jobs before proceeding with the
932 	 * dependency registration.
933 	 *
934 	 * This is solving two problems:
935 	 * 1. drm_sched_job_arm() and drm_sched_entity_push_job() must be
936 	 *    protected by a lock to make sure no concurrent access to the same
937 	 *    entity get interleaved, which would mess up with the fence seqno
938 	 *    ordering. Luckily, one of the resv being acquired is the VM resv,
939 	 *    and a scheduling entity is only bound to a single VM. As soon as
940 	 *    we acquire the VM resv, we should be safe.
941 	 * 2. Jobs might depend on fences that were issued by previous jobs in
942 	 *    the same batch, so we can't add dependencies on all jobs before
943 	 *    arming previous jobs and registering the fence to the signal
944 	 *    array, otherwise we might miss dependencies, or point to an
945 	 *    outdated fence.
946 	 */
947 	if (args->queue_submits.count > 0) {
948 		/* All jobs target the same group, so they also point to the same VM. */
949 		struct panthor_vm *vm = panthor_job_vm(ctx.jobs[0].job);
950 
951 		drm_exec_until_all_locked(&ctx.exec) {
952 			ret = panthor_vm_prepare_mapped_bos_resvs(&ctx.exec, vm,
953 								  args->queue_submits.count);
954 		}
955 
956 		if (ret)
957 			goto out_cleanup_submit_ctx;
958 	}
959 
960 	/*
961 	 * Now that resvs are locked/prepared, we can iterate over each job to
962 	 * add the dependencies, arm the job fence, register the job fence to
963 	 * the signal array.
964 	 */
965 	ret = panthor_submit_ctx_add_deps_and_arm_jobs(&ctx);
966 	if (ret)
967 		goto out_cleanup_submit_ctx;
968 
969 	/* Nothing can fail after that point, so we can make our job fences
970 	 * visible to the outside world. Push jobs and set the job fences to
971 	 * the resv slots we reserved.  This also pushes the fences to the
972 	 * syncobjs that are part of the signal array.
973 	 */
974 	panthor_submit_ctx_push_jobs(&ctx, panthor_job_update_resvs);
975 
976 out_cleanup_submit_ctx:
977 	panthor_submit_ctx_cleanup(&ctx, panthor_job_put);
978 
979 out_free_jobs_args:
980 	kvfree(jobs_args);
981 
982 out_dev_exit:
983 	drm_dev_exit(cookie);
984 	return ret;
985 }
986 
987 static int panthor_ioctl_group_destroy(struct drm_device *ddev, void *data,
988 				       struct drm_file *file)
989 {
990 	struct panthor_file *pfile = file->driver_priv;
991 	struct drm_panthor_group_destroy *args = data;
992 
993 	if (args->pad)
994 		return -EINVAL;
995 
996 	return panthor_group_destroy(pfile, args->group_handle);
997 }
998 
999 static int panthor_ioctl_group_create(struct drm_device *ddev, void *data,
1000 				      struct drm_file *file)
1001 {
1002 	struct panthor_file *pfile = file->driver_priv;
1003 	struct drm_panthor_group_create *args = data;
1004 	struct drm_panthor_queue_create *queue_args;
1005 	int ret;
1006 
1007 	if (!args->queues.count)
1008 		return -EINVAL;
1009 
1010 	ret = PANTHOR_UOBJ_GET_ARRAY(queue_args, &args->queues);
1011 	if (ret)
1012 		return ret;
1013 
1014 	ret = panthor_group_create(pfile, args, queue_args);
1015 	if (ret >= 0) {
1016 		args->group_handle = ret;
1017 		ret = 0;
1018 	}
1019 
1020 	kvfree(queue_args);
1021 	return ret;
1022 }
1023 
1024 static int panthor_ioctl_group_get_state(struct drm_device *ddev, void *data,
1025 					 struct drm_file *file)
1026 {
1027 	struct panthor_file *pfile = file->driver_priv;
1028 	struct drm_panthor_group_get_state *args = data;
1029 
1030 	return panthor_group_get_state(pfile, args);
1031 }
1032 
1033 static int panthor_ioctl_tiler_heap_create(struct drm_device *ddev, void *data,
1034 					   struct drm_file *file)
1035 {
1036 	struct panthor_file *pfile = file->driver_priv;
1037 	struct drm_panthor_tiler_heap_create *args = data;
1038 	struct panthor_heap_pool *pool;
1039 	struct panthor_vm *vm;
1040 	int ret;
1041 
1042 	vm = panthor_vm_pool_get_vm(pfile->vms, args->vm_id);
1043 	if (!vm)
1044 		return -EINVAL;
1045 
1046 	pool = panthor_vm_get_heap_pool(vm, true);
1047 	if (IS_ERR(pool)) {
1048 		ret = PTR_ERR(pool);
1049 		goto out_put_vm;
1050 	}
1051 
1052 	ret = panthor_heap_create(pool,
1053 				  args->initial_chunk_count,
1054 				  args->chunk_size,
1055 				  args->max_chunks,
1056 				  args->target_in_flight,
1057 				  &args->tiler_heap_ctx_gpu_va,
1058 				  &args->first_heap_chunk_gpu_va);
1059 	if (ret < 0)
1060 		goto out_put_heap_pool;
1061 
1062 	/* Heap pools are per-VM. We combine the VM and HEAP id to make
1063 	 * a unique heap handle.
1064 	 */
1065 	args->handle = (args->vm_id << 16) | ret;
1066 	ret = 0;
1067 
1068 out_put_heap_pool:
1069 	panthor_heap_pool_put(pool);
1070 
1071 out_put_vm:
1072 	panthor_vm_put(vm);
1073 	return ret;
1074 }
1075 
1076 static int panthor_ioctl_tiler_heap_destroy(struct drm_device *ddev, void *data,
1077 					    struct drm_file *file)
1078 {
1079 	struct panthor_file *pfile = file->driver_priv;
1080 	struct drm_panthor_tiler_heap_destroy *args = data;
1081 	struct panthor_heap_pool *pool;
1082 	struct panthor_vm *vm;
1083 	int ret;
1084 
1085 	if (args->pad)
1086 		return -EINVAL;
1087 
1088 	vm = panthor_vm_pool_get_vm(pfile->vms, args->handle >> 16);
1089 	if (!vm)
1090 		return -EINVAL;
1091 
1092 	pool = panthor_vm_get_heap_pool(vm, false);
1093 	if (IS_ERR(pool)) {
1094 		ret = PTR_ERR(pool);
1095 		goto out_put_vm;
1096 	}
1097 
1098 	ret = panthor_heap_destroy(pool, args->handle & GENMASK(15, 0));
1099 	panthor_heap_pool_put(pool);
1100 
1101 out_put_vm:
1102 	panthor_vm_put(vm);
1103 	return ret;
1104 }
1105 
1106 static int panthor_ioctl_vm_bind_async(struct drm_device *ddev,
1107 				       struct drm_panthor_vm_bind *args,
1108 				       struct drm_file *file)
1109 {
1110 	struct panthor_file *pfile = file->driver_priv;
1111 	struct drm_panthor_vm_bind_op *jobs_args;
1112 	struct panthor_submit_ctx ctx;
1113 	struct panthor_vm *vm;
1114 	int ret = 0;
1115 
1116 	vm = panthor_vm_pool_get_vm(pfile->vms, args->vm_id);
1117 	if (!vm)
1118 		return -EINVAL;
1119 
1120 	ret = PANTHOR_UOBJ_GET_ARRAY(jobs_args, &args->ops);
1121 	if (ret)
1122 		goto out_put_vm;
1123 
1124 	ret = panthor_submit_ctx_init(&ctx, file, args->ops.count);
1125 	if (ret)
1126 		goto out_free_jobs_args;
1127 
1128 	for (u32 i = 0; i < args->ops.count; i++) {
1129 		struct drm_panthor_vm_bind_op *op = &jobs_args[i];
1130 		struct drm_sched_job *job;
1131 
1132 		job = panthor_vm_bind_job_create(file, vm, op);
1133 		if (IS_ERR(job)) {
1134 			ret = PTR_ERR(job);
1135 			goto out_cleanup_submit_ctx;
1136 		}
1137 
1138 		ret = panthor_submit_ctx_add_job(&ctx, i, job, &op->syncs);
1139 		if (ret)
1140 			goto out_cleanup_submit_ctx;
1141 	}
1142 
1143 	ret = panthor_submit_ctx_collect_jobs_signal_ops(&ctx);
1144 	if (ret)
1145 		goto out_cleanup_submit_ctx;
1146 
1147 	/* Prepare reservation objects for each VM_BIND job. */
1148 	drm_exec_until_all_locked(&ctx.exec) {
1149 		for (u32 i = 0; i < ctx.job_count; i++) {
1150 			ret = panthor_vm_bind_job_prepare_resvs(&ctx.exec, ctx.jobs[i].job);
1151 			drm_exec_retry_on_contention(&ctx.exec);
1152 			if (ret)
1153 				goto out_cleanup_submit_ctx;
1154 		}
1155 	}
1156 
1157 	ret = panthor_submit_ctx_add_deps_and_arm_jobs(&ctx);
1158 	if (ret)
1159 		goto out_cleanup_submit_ctx;
1160 
1161 	/* Nothing can fail after that point. */
1162 	panthor_submit_ctx_push_jobs(&ctx, panthor_vm_bind_job_update_resvs);
1163 
1164 out_cleanup_submit_ctx:
1165 	panthor_submit_ctx_cleanup(&ctx, panthor_vm_bind_job_put);
1166 
1167 out_free_jobs_args:
1168 	kvfree(jobs_args);
1169 
1170 out_put_vm:
1171 	panthor_vm_put(vm);
1172 	return ret;
1173 }
1174 
1175 static int panthor_ioctl_vm_bind_sync(struct drm_device *ddev,
1176 				      struct drm_panthor_vm_bind *args,
1177 				      struct drm_file *file)
1178 {
1179 	struct panthor_file *pfile = file->driver_priv;
1180 	struct drm_panthor_vm_bind_op *jobs_args;
1181 	struct panthor_vm *vm;
1182 	int ret;
1183 
1184 	vm = panthor_vm_pool_get_vm(pfile->vms, args->vm_id);
1185 	if (!vm)
1186 		return -EINVAL;
1187 
1188 	ret = PANTHOR_UOBJ_GET_ARRAY(jobs_args, &args->ops);
1189 	if (ret)
1190 		goto out_put_vm;
1191 
1192 	for (u32 i = 0; i < args->ops.count; i++) {
1193 		ret = panthor_vm_bind_exec_sync_op(file, vm, &jobs_args[i]);
1194 		if (ret) {
1195 			/* Update ops.count so the user knows where things failed. */
1196 			args->ops.count = i;
1197 			break;
1198 		}
1199 	}
1200 
1201 	kvfree(jobs_args);
1202 
1203 out_put_vm:
1204 	panthor_vm_put(vm);
1205 	return ret;
1206 }
1207 
1208 #define PANTHOR_VM_BIND_FLAGS DRM_PANTHOR_VM_BIND_ASYNC
1209 
1210 static int panthor_ioctl_vm_bind(struct drm_device *ddev, void *data,
1211 				 struct drm_file *file)
1212 {
1213 	struct drm_panthor_vm_bind *args = data;
1214 	int cookie, ret;
1215 
1216 	if (!drm_dev_enter(ddev, &cookie))
1217 		return -ENODEV;
1218 
1219 	if (args->flags & DRM_PANTHOR_VM_BIND_ASYNC)
1220 		ret = panthor_ioctl_vm_bind_async(ddev, args, file);
1221 	else
1222 		ret = panthor_ioctl_vm_bind_sync(ddev, args, file);
1223 
1224 	drm_dev_exit(cookie);
1225 	return ret;
1226 }
1227 
1228 static int panthor_ioctl_vm_get_state(struct drm_device *ddev, void *data,
1229 				      struct drm_file *file)
1230 {
1231 	struct panthor_file *pfile = file->driver_priv;
1232 	struct drm_panthor_vm_get_state *args = data;
1233 	struct panthor_vm *vm;
1234 
1235 	vm = panthor_vm_pool_get_vm(pfile->vms, args->vm_id);
1236 	if (!vm)
1237 		return -EINVAL;
1238 
1239 	if (panthor_vm_is_unusable(vm))
1240 		args->state = DRM_PANTHOR_VM_STATE_UNUSABLE;
1241 	else
1242 		args->state = DRM_PANTHOR_VM_STATE_USABLE;
1243 
1244 	panthor_vm_put(vm);
1245 	return 0;
1246 }
1247 
1248 static int
1249 panthor_open(struct drm_device *ddev, struct drm_file *file)
1250 {
1251 	struct panthor_device *ptdev = container_of(ddev, struct panthor_device, base);
1252 	struct panthor_file *pfile;
1253 	int ret;
1254 
1255 	if (!try_module_get(THIS_MODULE))
1256 		return -EINVAL;
1257 
1258 	pfile = kzalloc(sizeof(*pfile), GFP_KERNEL);
1259 	if (!pfile) {
1260 		ret = -ENOMEM;
1261 		goto err_put_mod;
1262 	}
1263 
1264 	pfile->ptdev = ptdev;
1265 
1266 	ret = panthor_vm_pool_create(pfile);
1267 	if (ret)
1268 		goto err_free_file;
1269 
1270 	ret = panthor_group_pool_create(pfile);
1271 	if (ret)
1272 		goto err_destroy_vm_pool;
1273 
1274 	file->driver_priv = pfile;
1275 	return 0;
1276 
1277 err_destroy_vm_pool:
1278 	panthor_vm_pool_destroy(pfile);
1279 
1280 err_free_file:
1281 	kfree(pfile);
1282 
1283 err_put_mod:
1284 	module_put(THIS_MODULE);
1285 	return ret;
1286 }
1287 
1288 static void
1289 panthor_postclose(struct drm_device *ddev, struct drm_file *file)
1290 {
1291 	struct panthor_file *pfile = file->driver_priv;
1292 
1293 	panthor_group_pool_destroy(pfile);
1294 	panthor_vm_pool_destroy(pfile);
1295 
1296 	kfree(pfile);
1297 	module_put(THIS_MODULE);
1298 }
1299 
1300 static const struct drm_ioctl_desc panthor_drm_driver_ioctls[] = {
1301 #define PANTHOR_IOCTL(n, func, flags) \
1302 	DRM_IOCTL_DEF_DRV(PANTHOR_##n, panthor_ioctl_##func, flags)
1303 
1304 	PANTHOR_IOCTL(DEV_QUERY, dev_query, DRM_RENDER_ALLOW),
1305 	PANTHOR_IOCTL(VM_CREATE, vm_create, DRM_RENDER_ALLOW),
1306 	PANTHOR_IOCTL(VM_DESTROY, vm_destroy, DRM_RENDER_ALLOW),
1307 	PANTHOR_IOCTL(VM_BIND, vm_bind, DRM_RENDER_ALLOW),
1308 	PANTHOR_IOCTL(VM_GET_STATE, vm_get_state, DRM_RENDER_ALLOW),
1309 	PANTHOR_IOCTL(BO_CREATE, bo_create, DRM_RENDER_ALLOW),
1310 	PANTHOR_IOCTL(BO_MMAP_OFFSET, bo_mmap_offset, DRM_RENDER_ALLOW),
1311 	PANTHOR_IOCTL(GROUP_CREATE, group_create, DRM_RENDER_ALLOW),
1312 	PANTHOR_IOCTL(GROUP_DESTROY, group_destroy, DRM_RENDER_ALLOW),
1313 	PANTHOR_IOCTL(GROUP_GET_STATE, group_get_state, DRM_RENDER_ALLOW),
1314 	PANTHOR_IOCTL(TILER_HEAP_CREATE, tiler_heap_create, DRM_RENDER_ALLOW),
1315 	PANTHOR_IOCTL(TILER_HEAP_DESTROY, tiler_heap_destroy, DRM_RENDER_ALLOW),
1316 	PANTHOR_IOCTL(GROUP_SUBMIT, group_submit, DRM_RENDER_ALLOW),
1317 };
1318 
1319 static int panthor_mmap(struct file *filp, struct vm_area_struct *vma)
1320 {
1321 	struct drm_file *file = filp->private_data;
1322 	struct panthor_file *pfile = file->driver_priv;
1323 	struct panthor_device *ptdev = pfile->ptdev;
1324 	u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT;
1325 	int ret, cookie;
1326 
1327 	if (!drm_dev_enter(file->minor->dev, &cookie))
1328 		return -ENODEV;
1329 
1330 #ifdef CONFIG_ARM64
1331 	/*
1332 	 * With 32-bit systems being limited by the 32-bit representation of
1333 	 * mmap2's pgoffset field, we need to make the MMIO offset arch
1334 	 * specific. This converts a user MMIO offset into something the kernel
1335 	 * driver understands.
1336 	 */
1337 	if (test_tsk_thread_flag(current, TIF_32BIT) &&
1338 	    offset >= DRM_PANTHOR_USER_MMIO_OFFSET_32BIT) {
1339 		offset += DRM_PANTHOR_USER_MMIO_OFFSET_64BIT -
1340 			  DRM_PANTHOR_USER_MMIO_OFFSET_32BIT;
1341 		vma->vm_pgoff = offset >> PAGE_SHIFT;
1342 	}
1343 #endif
1344 
1345 	if (offset >= DRM_PANTHOR_USER_MMIO_OFFSET)
1346 		ret = panthor_device_mmap_io(ptdev, vma);
1347 	else
1348 		ret = drm_gem_mmap(filp, vma);
1349 
1350 	drm_dev_exit(cookie);
1351 	return ret;
1352 }
1353 
1354 static const struct file_operations panthor_drm_driver_fops = {
1355 	.open = drm_open,
1356 	.release = drm_release,
1357 	.unlocked_ioctl = drm_ioctl,
1358 	.compat_ioctl = drm_compat_ioctl,
1359 	.poll = drm_poll,
1360 	.read = drm_read,
1361 	.llseek = noop_llseek,
1362 	.mmap = panthor_mmap,
1363 };
1364 
1365 #ifdef CONFIG_DEBUG_FS
1366 static void panthor_debugfs_init(struct drm_minor *minor)
1367 {
1368 	panthor_mmu_debugfs_init(minor);
1369 }
1370 #endif
1371 
1372 /*
1373  * PanCSF driver version:
1374  * - 1.0 - initial interface
1375  */
1376 static const struct drm_driver panthor_drm_driver = {
1377 	.driver_features = DRIVER_RENDER | DRIVER_GEM | DRIVER_SYNCOBJ |
1378 			   DRIVER_SYNCOBJ_TIMELINE | DRIVER_GEM_GPUVA,
1379 	.open = panthor_open,
1380 	.postclose = panthor_postclose,
1381 	.ioctls = panthor_drm_driver_ioctls,
1382 	.num_ioctls = ARRAY_SIZE(panthor_drm_driver_ioctls),
1383 	.fops = &panthor_drm_driver_fops,
1384 	.name = "panthor",
1385 	.desc = "Panthor DRM driver",
1386 	.date = "20230801",
1387 	.major = 1,
1388 	.minor = 0,
1389 
1390 	.gem_create_object = panthor_gem_create_object,
1391 	.gem_prime_import_sg_table = drm_gem_shmem_prime_import_sg_table,
1392 #ifdef CONFIG_DEBUG_FS
1393 	.debugfs_init = panthor_debugfs_init,
1394 #endif
1395 };
1396 
1397 static int panthor_probe(struct platform_device *pdev)
1398 {
1399 	struct panthor_device *ptdev;
1400 
1401 	ptdev = devm_drm_dev_alloc(&pdev->dev, &panthor_drm_driver,
1402 				   struct panthor_device, base);
1403 	if (IS_ERR(ptdev))
1404 		return -ENOMEM;
1405 
1406 	platform_set_drvdata(pdev, ptdev);
1407 
1408 	return panthor_device_init(ptdev);
1409 }
1410 
1411 static void panthor_remove(struct platform_device *pdev)
1412 {
1413 	struct panthor_device *ptdev = platform_get_drvdata(pdev);
1414 
1415 	panthor_device_unplug(ptdev);
1416 }
1417 
1418 static const struct of_device_id dt_match[] = {
1419 	{ .compatible = "rockchip,rk3588-mali" },
1420 	{ .compatible = "arm,mali-valhall-csf" },
1421 	{}
1422 };
1423 MODULE_DEVICE_TABLE(of, dt_match);
1424 
1425 static DEFINE_RUNTIME_DEV_PM_OPS(panthor_pm_ops,
1426 				 panthor_device_suspend,
1427 				 panthor_device_resume,
1428 				 NULL);
1429 
1430 static struct platform_driver panthor_driver = {
1431 	.probe = panthor_probe,
1432 	.remove_new = panthor_remove,
1433 	.driver = {
1434 		.name = "panthor",
1435 		.pm = pm_ptr(&panthor_pm_ops),
1436 		.of_match_table = dt_match,
1437 	},
1438 };
1439 
1440 /*
1441  * Workqueue used to cleanup stuff.
1442  *
1443  * We create a dedicated workqueue so we can drain on unplug and
1444  * make sure all resources are freed before the module is unloaded.
1445  */
1446 struct workqueue_struct *panthor_cleanup_wq;
1447 
1448 static int __init panthor_init(void)
1449 {
1450 	int ret;
1451 
1452 	ret = panthor_mmu_pt_cache_init();
1453 	if (ret)
1454 		return ret;
1455 
1456 	panthor_cleanup_wq = alloc_workqueue("panthor-cleanup", WQ_UNBOUND, 0);
1457 	if (!panthor_cleanup_wq) {
1458 		pr_err("panthor: Failed to allocate the workqueues");
1459 		ret = -ENOMEM;
1460 		goto err_mmu_pt_cache_fini;
1461 	}
1462 
1463 	ret = platform_driver_register(&panthor_driver);
1464 	if (ret)
1465 		goto err_destroy_cleanup_wq;
1466 
1467 	return 0;
1468 
1469 err_destroy_cleanup_wq:
1470 	destroy_workqueue(panthor_cleanup_wq);
1471 
1472 err_mmu_pt_cache_fini:
1473 	panthor_mmu_pt_cache_fini();
1474 	return ret;
1475 }
1476 module_init(panthor_init);
1477 
1478 static void __exit panthor_exit(void)
1479 {
1480 	platform_driver_unregister(&panthor_driver);
1481 	destroy_workqueue(panthor_cleanup_wq);
1482 	panthor_mmu_pt_cache_fini();
1483 }
1484 module_exit(panthor_exit);
1485 
1486 MODULE_AUTHOR("Panthor Project Developers");
1487 MODULE_DESCRIPTION("Panthor DRM Driver");
1488 MODULE_LICENSE("Dual MIT/GPL");
1489