1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Xilinx Video DMA
4  *
5  * Copyright (C) 2013-2015 Ideas on Board
6  * Copyright (C) 2013-2015 Xilinx, Inc.
7  *
8  * Contacts: Hyun Kwon <hyun.kwon@xilinx.com>
9  *           Laurent Pinchart <laurent.pinchart@ideasonboard.com>
10  */
11 
12 #include <linux/dma/xilinx_dma.h>
13 #include <linux/lcm.h>
14 #include <linux/list.h>
15 #include <linux/module.h>
16 #include <linux/of.h>
17 #include <linux/slab.h>
18 
19 #include <media/v4l2-dev.h>
20 #include <media/v4l2-fh.h>
21 #include <media/v4l2-ioctl.h>
22 #include <media/videobuf2-v4l2.h>
23 #include <media/videobuf2-dma-contig.h>
24 
25 #include "xilinx-dma.h"
26 #include "xilinx-vip.h"
27 #include "xilinx-vipp.h"
28 
29 #define XVIP_DMA_DEF_FORMAT		V4L2_PIX_FMT_YUYV
30 #define XVIP_DMA_DEF_WIDTH		1920
31 #define XVIP_DMA_DEF_HEIGHT		1080
32 
33 /* Minimum and maximum widths are expressed in bytes */
34 #define XVIP_DMA_MIN_WIDTH		1U
35 #define XVIP_DMA_MAX_WIDTH		65535U
36 #define XVIP_DMA_MIN_HEIGHT		1U
37 #define XVIP_DMA_MAX_HEIGHT		8191U
38 
39 /* -----------------------------------------------------------------------------
40  * Helper functions
41  */
42 
43 static struct v4l2_subdev *
xvip_dma_remote_subdev(struct media_pad * local,u32 * pad)44 xvip_dma_remote_subdev(struct media_pad *local, u32 *pad)
45 {
46 	struct media_pad *remote;
47 
48 	remote = media_entity_remote_pad(local);
49 	if (!remote || !is_media_entity_v4l2_subdev(remote->entity))
50 		return NULL;
51 
52 	if (pad)
53 		*pad = remote->index;
54 
55 	return media_entity_to_v4l2_subdev(remote->entity);
56 }
57 
xvip_dma_verify_format(struct xvip_dma * dma)58 static int xvip_dma_verify_format(struct xvip_dma *dma)
59 {
60 	struct v4l2_subdev_format fmt;
61 	struct v4l2_subdev *subdev;
62 	int ret;
63 
64 	subdev = xvip_dma_remote_subdev(&dma->pad, &fmt.pad);
65 	if (subdev == NULL)
66 		return -EPIPE;
67 
68 	fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE;
69 	ret = v4l2_subdev_call(subdev, pad, get_fmt, NULL, &fmt);
70 	if (ret < 0)
71 		return ret == -ENOIOCTLCMD ? -EINVAL : ret;
72 
73 	if (dma->fmtinfo->code != fmt.format.code ||
74 	    dma->format.height != fmt.format.height ||
75 	    dma->format.width != fmt.format.width ||
76 	    dma->format.colorspace != fmt.format.colorspace)
77 		return -EINVAL;
78 
79 	return 0;
80 }
81 
82 /* -----------------------------------------------------------------------------
83  * Pipeline Stream Management
84  */
85 
86 /**
87  * xvip_pipeline_start_stop - Start ot stop streaming on a pipeline
88  * @pipe: The pipeline
89  * @start: Start (when true) or stop (when false) the pipeline
90  *
91  * Walk the entities chain starting at the pipeline output video node and start
92  * or stop all of them.
93  *
94  * Return: 0 if successful, or the return value of the failed video::s_stream
95  * operation otherwise.
96  */
xvip_pipeline_start_stop(struct xvip_pipeline * pipe,bool start)97 static int xvip_pipeline_start_stop(struct xvip_pipeline *pipe, bool start)
98 {
99 	struct xvip_dma *dma = pipe->output;
100 	struct media_entity *entity;
101 	struct media_pad *pad;
102 	struct v4l2_subdev *subdev;
103 	int ret;
104 
105 	entity = &dma->video.entity;
106 	while (1) {
107 		pad = &entity->pads[0];
108 		if (!(pad->flags & MEDIA_PAD_FL_SINK))
109 			break;
110 
111 		pad = media_entity_remote_pad(pad);
112 		if (!pad || !is_media_entity_v4l2_subdev(pad->entity))
113 			break;
114 
115 		entity = pad->entity;
116 		subdev = media_entity_to_v4l2_subdev(entity);
117 
118 		ret = v4l2_subdev_call(subdev, video, s_stream, start);
119 		if (start && ret < 0 && ret != -ENOIOCTLCMD)
120 			return ret;
121 	}
122 
123 	return 0;
124 }
125 
126 /**
127  * xvip_pipeline_set_stream - Enable/disable streaming on a pipeline
128  * @pipe: The pipeline
129  * @on: Turn the stream on when true or off when false
130  *
131  * The pipeline is shared between all DMA engines connect at its input and
132  * output. While the stream state of DMA engines can be controlled
133  * independently, pipelines have a shared stream state that enable or disable
134  * all entities in the pipeline. For this reason the pipeline uses a streaming
135  * counter that tracks the number of DMA engines that have requested the stream
136  * to be enabled.
137  *
138  * When called with the @on argument set to true, this function will increment
139  * the pipeline streaming count. If the streaming count reaches the number of
140  * DMA engines in the pipeline it will enable all entities that belong to the
141  * pipeline.
142  *
143  * Similarly, when called with the @on argument set to false, this function will
144  * decrement the pipeline streaming count and disable all entities in the
145  * pipeline when the streaming count reaches zero.
146  *
147  * Return: 0 if successful, or the return value of the failed video::s_stream
148  * operation otherwise. Stopping the pipeline never fails. The pipeline state is
149  * not updated when the operation fails.
150  */
xvip_pipeline_set_stream(struct xvip_pipeline * pipe,bool on)151 static int xvip_pipeline_set_stream(struct xvip_pipeline *pipe, bool on)
152 {
153 	int ret = 0;
154 
155 	mutex_lock(&pipe->lock);
156 
157 	if (on) {
158 		if (pipe->stream_count == pipe->num_dmas - 1) {
159 			ret = xvip_pipeline_start_stop(pipe, true);
160 			if (ret < 0)
161 				goto done;
162 		}
163 		pipe->stream_count++;
164 	} else {
165 		if (--pipe->stream_count == 0)
166 			xvip_pipeline_start_stop(pipe, false);
167 	}
168 
169 done:
170 	mutex_unlock(&pipe->lock);
171 	return ret;
172 }
173 
xvip_pipeline_validate(struct xvip_pipeline * pipe,struct xvip_dma * start)174 static int xvip_pipeline_validate(struct xvip_pipeline *pipe,
175 				  struct xvip_dma *start)
176 {
177 	struct media_graph graph;
178 	struct media_entity *entity = &start->video.entity;
179 	struct media_device *mdev = entity->graph_obj.mdev;
180 	unsigned int num_inputs = 0;
181 	unsigned int num_outputs = 0;
182 	int ret;
183 
184 	mutex_lock(&mdev->graph_mutex);
185 
186 	/* Walk the graph to locate the video nodes. */
187 	ret = media_graph_walk_init(&graph, mdev);
188 	if (ret) {
189 		mutex_unlock(&mdev->graph_mutex);
190 		return ret;
191 	}
192 
193 	media_graph_walk_start(&graph, entity);
194 
195 	while ((entity = media_graph_walk_next(&graph))) {
196 		struct xvip_dma *dma;
197 
198 		if (entity->function != MEDIA_ENT_F_IO_V4L)
199 			continue;
200 
201 		dma = to_xvip_dma(media_entity_to_video_device(entity));
202 
203 		if (dma->pad.flags & MEDIA_PAD_FL_SINK) {
204 			pipe->output = dma;
205 			num_outputs++;
206 		} else {
207 			num_inputs++;
208 		}
209 	}
210 
211 	mutex_unlock(&mdev->graph_mutex);
212 
213 	media_graph_walk_cleanup(&graph);
214 
215 	/* We need exactly one output and zero or one input. */
216 	if (num_outputs != 1 || num_inputs > 1)
217 		return -EPIPE;
218 
219 	pipe->num_dmas = num_inputs + num_outputs;
220 
221 	return 0;
222 }
223 
__xvip_pipeline_cleanup(struct xvip_pipeline * pipe)224 static void __xvip_pipeline_cleanup(struct xvip_pipeline *pipe)
225 {
226 	pipe->num_dmas = 0;
227 	pipe->output = NULL;
228 }
229 
230 /**
231  * xvip_pipeline_cleanup - Cleanup the pipeline after streaming
232  * @pipe: the pipeline
233  *
234  * Decrease the pipeline use count and clean it up if we were the last user.
235  */
xvip_pipeline_cleanup(struct xvip_pipeline * pipe)236 static void xvip_pipeline_cleanup(struct xvip_pipeline *pipe)
237 {
238 	mutex_lock(&pipe->lock);
239 
240 	/* If we're the last user clean up the pipeline. */
241 	if (--pipe->use_count == 0)
242 		__xvip_pipeline_cleanup(pipe);
243 
244 	mutex_unlock(&pipe->lock);
245 }
246 
247 /**
248  * xvip_pipeline_prepare - Prepare the pipeline for streaming
249  * @pipe: the pipeline
250  * @dma: DMA engine at one end of the pipeline
251  *
252  * Validate the pipeline if no user exists yet, otherwise just increase the use
253  * count.
254  *
255  * Return: 0 if successful or -EPIPE if the pipeline is not valid.
256  */
xvip_pipeline_prepare(struct xvip_pipeline * pipe,struct xvip_dma * dma)257 static int xvip_pipeline_prepare(struct xvip_pipeline *pipe,
258 				 struct xvip_dma *dma)
259 {
260 	int ret;
261 
262 	mutex_lock(&pipe->lock);
263 
264 	/* If we're the first user validate and initialize the pipeline. */
265 	if (pipe->use_count == 0) {
266 		ret = xvip_pipeline_validate(pipe, dma);
267 		if (ret < 0) {
268 			__xvip_pipeline_cleanup(pipe);
269 			goto done;
270 		}
271 	}
272 
273 	pipe->use_count++;
274 	ret = 0;
275 
276 done:
277 	mutex_unlock(&pipe->lock);
278 	return ret;
279 }
280 
281 /* -----------------------------------------------------------------------------
282  * videobuf2 queue operations
283  */
284 
285 /**
286  * struct xvip_dma_buffer - Video DMA buffer
287  * @buf: vb2 buffer base object
288  * @queue: buffer list entry in the DMA engine queued buffers list
289  * @dma: DMA channel that uses the buffer
290  */
291 struct xvip_dma_buffer {
292 	struct vb2_v4l2_buffer buf;
293 	struct list_head queue;
294 	struct xvip_dma *dma;
295 };
296 
297 #define to_xvip_dma_buffer(vb)	container_of(vb, struct xvip_dma_buffer, buf)
298 
xvip_dma_complete(void * param)299 static void xvip_dma_complete(void *param)
300 {
301 	struct xvip_dma_buffer *buf = param;
302 	struct xvip_dma *dma = buf->dma;
303 
304 	spin_lock(&dma->queued_lock);
305 	list_del(&buf->queue);
306 	spin_unlock(&dma->queued_lock);
307 
308 	buf->buf.field = V4L2_FIELD_NONE;
309 	buf->buf.sequence = dma->sequence++;
310 	buf->buf.vb2_buf.timestamp = ktime_get_ns();
311 	vb2_set_plane_payload(&buf->buf.vb2_buf, 0, dma->format.sizeimage);
312 	vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_DONE);
313 }
314 
315 static int
xvip_dma_queue_setup(struct vb2_queue * vq,unsigned int * nbuffers,unsigned int * nplanes,unsigned int sizes[],struct device * alloc_devs[])316 xvip_dma_queue_setup(struct vb2_queue *vq,
317 		     unsigned int *nbuffers, unsigned int *nplanes,
318 		     unsigned int sizes[], struct device *alloc_devs[])
319 {
320 	struct xvip_dma *dma = vb2_get_drv_priv(vq);
321 
322 	/* Make sure the image size is large enough. */
323 	if (*nplanes)
324 		return sizes[0] < dma->format.sizeimage ? -EINVAL : 0;
325 
326 	*nplanes = 1;
327 	sizes[0] = dma->format.sizeimage;
328 
329 	return 0;
330 }
331 
xvip_dma_buffer_prepare(struct vb2_buffer * vb)332 static int xvip_dma_buffer_prepare(struct vb2_buffer *vb)
333 {
334 	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
335 	struct xvip_dma *dma = vb2_get_drv_priv(vb->vb2_queue);
336 	struct xvip_dma_buffer *buf = to_xvip_dma_buffer(vbuf);
337 
338 	buf->dma = dma;
339 
340 	return 0;
341 }
342 
xvip_dma_buffer_queue(struct vb2_buffer * vb)343 static void xvip_dma_buffer_queue(struct vb2_buffer *vb)
344 {
345 	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
346 	struct xvip_dma *dma = vb2_get_drv_priv(vb->vb2_queue);
347 	struct xvip_dma_buffer *buf = to_xvip_dma_buffer(vbuf);
348 	struct dma_async_tx_descriptor *desc;
349 	dma_addr_t addr = vb2_dma_contig_plane_dma_addr(vb, 0);
350 	u32 flags;
351 
352 	if (dma->queue.type == V4L2_BUF_TYPE_VIDEO_CAPTURE) {
353 		flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK;
354 		dma->xt.dir = DMA_DEV_TO_MEM;
355 		dma->xt.src_sgl = false;
356 		dma->xt.dst_sgl = true;
357 		dma->xt.dst_start = addr;
358 	} else {
359 		flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK;
360 		dma->xt.dir = DMA_MEM_TO_DEV;
361 		dma->xt.src_sgl = true;
362 		dma->xt.dst_sgl = false;
363 		dma->xt.src_start = addr;
364 	}
365 
366 	dma->xt.frame_size = 1;
367 	dma->sgl[0].size = dma->format.width * dma->fmtinfo->bpp;
368 	dma->sgl[0].icg = dma->format.bytesperline - dma->sgl[0].size;
369 	dma->xt.numf = dma->format.height;
370 
371 	desc = dmaengine_prep_interleaved_dma(dma->dma, &dma->xt, flags);
372 	if (!desc) {
373 		dev_err(dma->xdev->dev, "Failed to prepare DMA transfer\n");
374 		vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_ERROR);
375 		return;
376 	}
377 	desc->callback = xvip_dma_complete;
378 	desc->callback_param = buf;
379 
380 	spin_lock_irq(&dma->queued_lock);
381 	list_add_tail(&buf->queue, &dma->queued_bufs);
382 	spin_unlock_irq(&dma->queued_lock);
383 
384 	dmaengine_submit(desc);
385 
386 	if (vb2_is_streaming(&dma->queue))
387 		dma_async_issue_pending(dma->dma);
388 }
389 
xvip_dma_start_streaming(struct vb2_queue * vq,unsigned int count)390 static int xvip_dma_start_streaming(struct vb2_queue *vq, unsigned int count)
391 {
392 	struct xvip_dma *dma = vb2_get_drv_priv(vq);
393 	struct xvip_dma_buffer *buf, *nbuf;
394 	struct xvip_pipeline *pipe;
395 	int ret;
396 
397 	dma->sequence = 0;
398 
399 	/*
400 	 * Start streaming on the pipeline. No link touching an entity in the
401 	 * pipeline can be activated or deactivated once streaming is started.
402 	 *
403 	 * Use the pipeline object embedded in the first DMA object that starts
404 	 * streaming.
405 	 */
406 	pipe = dma->video.entity.pipe
407 	     ? to_xvip_pipeline(&dma->video.entity) : &dma->pipe;
408 
409 	ret = media_pipeline_start(&dma->video.entity, &pipe->pipe);
410 	if (ret < 0)
411 		goto error;
412 
413 	/* Verify that the configured format matches the output of the
414 	 * connected subdev.
415 	 */
416 	ret = xvip_dma_verify_format(dma);
417 	if (ret < 0)
418 		goto error_stop;
419 
420 	ret = xvip_pipeline_prepare(pipe, dma);
421 	if (ret < 0)
422 		goto error_stop;
423 
424 	/* Start the DMA engine. This must be done before starting the blocks
425 	 * in the pipeline to avoid DMA synchronization issues.
426 	 */
427 	dma_async_issue_pending(dma->dma);
428 
429 	/* Start the pipeline. */
430 	xvip_pipeline_set_stream(pipe, true);
431 
432 	return 0;
433 
434 error_stop:
435 	media_pipeline_stop(&dma->video.entity);
436 
437 error:
438 	/* Give back all queued buffers to videobuf2. */
439 	spin_lock_irq(&dma->queued_lock);
440 	list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) {
441 		vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_QUEUED);
442 		list_del(&buf->queue);
443 	}
444 	spin_unlock_irq(&dma->queued_lock);
445 
446 	return ret;
447 }
448 
xvip_dma_stop_streaming(struct vb2_queue * vq)449 static void xvip_dma_stop_streaming(struct vb2_queue *vq)
450 {
451 	struct xvip_dma *dma = vb2_get_drv_priv(vq);
452 	struct xvip_pipeline *pipe = to_xvip_pipeline(&dma->video.entity);
453 	struct xvip_dma_buffer *buf, *nbuf;
454 
455 	/* Stop the pipeline. */
456 	xvip_pipeline_set_stream(pipe, false);
457 
458 	/* Stop and reset the DMA engine. */
459 	dmaengine_terminate_all(dma->dma);
460 
461 	/* Cleanup the pipeline and mark it as being stopped. */
462 	xvip_pipeline_cleanup(pipe);
463 	media_pipeline_stop(&dma->video.entity);
464 
465 	/* Give back all queued buffers to videobuf2. */
466 	spin_lock_irq(&dma->queued_lock);
467 	list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) {
468 		vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_ERROR);
469 		list_del(&buf->queue);
470 	}
471 	spin_unlock_irq(&dma->queued_lock);
472 }
473 
474 static const struct vb2_ops xvip_dma_queue_qops = {
475 	.queue_setup = xvip_dma_queue_setup,
476 	.buf_prepare = xvip_dma_buffer_prepare,
477 	.buf_queue = xvip_dma_buffer_queue,
478 	.wait_prepare = vb2_ops_wait_prepare,
479 	.wait_finish = vb2_ops_wait_finish,
480 	.start_streaming = xvip_dma_start_streaming,
481 	.stop_streaming = xvip_dma_stop_streaming,
482 };
483 
484 /* -----------------------------------------------------------------------------
485  * V4L2 ioctls
486  */
487 
488 static int
xvip_dma_querycap(struct file * file,void * fh,struct v4l2_capability * cap)489 xvip_dma_querycap(struct file *file, void *fh, struct v4l2_capability *cap)
490 {
491 	struct v4l2_fh *vfh = file->private_data;
492 	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
493 
494 	cap->capabilities = dma->xdev->v4l2_caps | V4L2_CAP_STREAMING |
495 			    V4L2_CAP_DEVICE_CAPS;
496 
497 	strscpy(cap->driver, "xilinx-vipp", sizeof(cap->driver));
498 	strscpy(cap->card, dma->video.name, sizeof(cap->card));
499 	snprintf(cap->bus_info, sizeof(cap->bus_info), "platform:%pOFn:%u",
500 		 dma->xdev->dev->of_node, dma->port);
501 
502 	return 0;
503 }
504 
505 /* FIXME: without this callback function, some applications are not configured
506  * with correct formats, and it results in frames in wrong format. Whether this
507  * callback needs to be required is not clearly defined, so it should be
508  * clarified through the mailing list.
509  */
510 static int
xvip_dma_enum_format(struct file * file,void * fh,struct v4l2_fmtdesc * f)511 xvip_dma_enum_format(struct file *file, void *fh, struct v4l2_fmtdesc *f)
512 {
513 	struct v4l2_fh *vfh = file->private_data;
514 	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
515 
516 	if (f->index > 0)
517 		return -EINVAL;
518 
519 	f->pixelformat = dma->format.pixelformat;
520 
521 	return 0;
522 }
523 
524 static int
xvip_dma_get_format(struct file * file,void * fh,struct v4l2_format * format)525 xvip_dma_get_format(struct file *file, void *fh, struct v4l2_format *format)
526 {
527 	struct v4l2_fh *vfh = file->private_data;
528 	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
529 
530 	format->fmt.pix = dma->format;
531 
532 	return 0;
533 }
534 
535 static void
__xvip_dma_try_format(struct xvip_dma * dma,struct v4l2_pix_format * pix,const struct xvip_video_format ** fmtinfo)536 __xvip_dma_try_format(struct xvip_dma *dma, struct v4l2_pix_format *pix,
537 		      const struct xvip_video_format **fmtinfo)
538 {
539 	const struct xvip_video_format *info;
540 	unsigned int min_width;
541 	unsigned int max_width;
542 	unsigned int min_bpl;
543 	unsigned int max_bpl;
544 	unsigned int width;
545 	unsigned int align;
546 	unsigned int bpl;
547 
548 	/* Retrieve format information and select the default format if the
549 	 * requested format isn't supported.
550 	 */
551 	info = xvip_get_format_by_fourcc(pix->pixelformat);
552 	if (IS_ERR(info))
553 		info = xvip_get_format_by_fourcc(XVIP_DMA_DEF_FORMAT);
554 
555 	pix->pixelformat = info->fourcc;
556 	pix->field = V4L2_FIELD_NONE;
557 
558 	/* The transfer alignment requirements are expressed in bytes. Compute
559 	 * the minimum and maximum values, clamp the requested width and convert
560 	 * it back to pixels.
561 	 */
562 	align = lcm(dma->align, info->bpp);
563 	min_width = roundup(XVIP_DMA_MIN_WIDTH, align);
564 	max_width = rounddown(XVIP_DMA_MAX_WIDTH, align);
565 	width = rounddown(pix->width * info->bpp, align);
566 
567 	pix->width = clamp(width, min_width, max_width) / info->bpp;
568 	pix->height = clamp(pix->height, XVIP_DMA_MIN_HEIGHT,
569 			    XVIP_DMA_MAX_HEIGHT);
570 
571 	/* Clamp the requested bytes per line value. If the maximum bytes per
572 	 * line value is zero, the module doesn't support user configurable line
573 	 * sizes. Override the requested value with the minimum in that case.
574 	 */
575 	min_bpl = pix->width * info->bpp;
576 	max_bpl = rounddown(XVIP_DMA_MAX_WIDTH, dma->align);
577 	bpl = rounddown(pix->bytesperline, dma->align);
578 
579 	pix->bytesperline = clamp(bpl, min_bpl, max_bpl);
580 	pix->sizeimage = pix->bytesperline * pix->height;
581 
582 	if (fmtinfo)
583 		*fmtinfo = info;
584 }
585 
586 static int
xvip_dma_try_format(struct file * file,void * fh,struct v4l2_format * format)587 xvip_dma_try_format(struct file *file, void *fh, struct v4l2_format *format)
588 {
589 	struct v4l2_fh *vfh = file->private_data;
590 	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
591 
592 	__xvip_dma_try_format(dma, &format->fmt.pix, NULL);
593 	return 0;
594 }
595 
596 static int
xvip_dma_set_format(struct file * file,void * fh,struct v4l2_format * format)597 xvip_dma_set_format(struct file *file, void *fh, struct v4l2_format *format)
598 {
599 	struct v4l2_fh *vfh = file->private_data;
600 	struct xvip_dma *dma = to_xvip_dma(vfh->vdev);
601 	const struct xvip_video_format *info;
602 
603 	__xvip_dma_try_format(dma, &format->fmt.pix, &info);
604 
605 	if (vb2_is_busy(&dma->queue))
606 		return -EBUSY;
607 
608 	dma->format = format->fmt.pix;
609 	dma->fmtinfo = info;
610 
611 	return 0;
612 }
613 
614 static const struct v4l2_ioctl_ops xvip_dma_ioctl_ops = {
615 	.vidioc_querycap		= xvip_dma_querycap,
616 	.vidioc_enum_fmt_vid_cap	= xvip_dma_enum_format,
617 	.vidioc_g_fmt_vid_cap		= xvip_dma_get_format,
618 	.vidioc_g_fmt_vid_out		= xvip_dma_get_format,
619 	.vidioc_s_fmt_vid_cap		= xvip_dma_set_format,
620 	.vidioc_s_fmt_vid_out		= xvip_dma_set_format,
621 	.vidioc_try_fmt_vid_cap		= xvip_dma_try_format,
622 	.vidioc_try_fmt_vid_out		= xvip_dma_try_format,
623 	.vidioc_reqbufs			= vb2_ioctl_reqbufs,
624 	.vidioc_querybuf		= vb2_ioctl_querybuf,
625 	.vidioc_qbuf			= vb2_ioctl_qbuf,
626 	.vidioc_dqbuf			= vb2_ioctl_dqbuf,
627 	.vidioc_create_bufs		= vb2_ioctl_create_bufs,
628 	.vidioc_expbuf			= vb2_ioctl_expbuf,
629 	.vidioc_streamon		= vb2_ioctl_streamon,
630 	.vidioc_streamoff		= vb2_ioctl_streamoff,
631 };
632 
633 /* -----------------------------------------------------------------------------
634  * V4L2 file operations
635  */
636 
637 static const struct v4l2_file_operations xvip_dma_fops = {
638 	.owner		= THIS_MODULE,
639 	.unlocked_ioctl	= video_ioctl2,
640 	.open		= v4l2_fh_open,
641 	.release	= vb2_fop_release,
642 	.poll		= vb2_fop_poll,
643 	.mmap		= vb2_fop_mmap,
644 };
645 
646 /* -----------------------------------------------------------------------------
647  * Xilinx Video DMA Core
648  */
649 
xvip_dma_init(struct xvip_composite_device * xdev,struct xvip_dma * dma,enum v4l2_buf_type type,unsigned int port)650 int xvip_dma_init(struct xvip_composite_device *xdev, struct xvip_dma *dma,
651 		  enum v4l2_buf_type type, unsigned int port)
652 {
653 	char name[16];
654 	int ret;
655 
656 	dma->xdev = xdev;
657 	dma->port = port;
658 	mutex_init(&dma->lock);
659 	mutex_init(&dma->pipe.lock);
660 	INIT_LIST_HEAD(&dma->queued_bufs);
661 	spin_lock_init(&dma->queued_lock);
662 
663 	dma->fmtinfo = xvip_get_format_by_fourcc(XVIP_DMA_DEF_FORMAT);
664 	dma->format.pixelformat = dma->fmtinfo->fourcc;
665 	dma->format.colorspace = V4L2_COLORSPACE_SRGB;
666 	dma->format.field = V4L2_FIELD_NONE;
667 	dma->format.width = XVIP_DMA_DEF_WIDTH;
668 	dma->format.height = XVIP_DMA_DEF_HEIGHT;
669 	dma->format.bytesperline = dma->format.width * dma->fmtinfo->bpp;
670 	dma->format.sizeimage = dma->format.bytesperline * dma->format.height;
671 
672 	/* Initialize the media entity... */
673 	dma->pad.flags = type == V4L2_BUF_TYPE_VIDEO_CAPTURE
674 		       ? MEDIA_PAD_FL_SINK : MEDIA_PAD_FL_SOURCE;
675 
676 	ret = media_entity_pads_init(&dma->video.entity, 1, &dma->pad);
677 	if (ret < 0)
678 		goto error;
679 
680 	/* ... and the video node... */
681 	dma->video.fops = &xvip_dma_fops;
682 	dma->video.v4l2_dev = &xdev->v4l2_dev;
683 	dma->video.queue = &dma->queue;
684 	snprintf(dma->video.name, sizeof(dma->video.name), "%pOFn %s %u",
685 		 xdev->dev->of_node,
686 		 type == V4L2_BUF_TYPE_VIDEO_CAPTURE ? "output" : "input",
687 		 port);
688 	dma->video.vfl_type = VFL_TYPE_VIDEO;
689 	dma->video.vfl_dir = type == V4L2_BUF_TYPE_VIDEO_CAPTURE
690 			   ? VFL_DIR_RX : VFL_DIR_TX;
691 	dma->video.release = video_device_release_empty;
692 	dma->video.ioctl_ops = &xvip_dma_ioctl_ops;
693 	dma->video.lock = &dma->lock;
694 	dma->video.device_caps = V4L2_CAP_STREAMING;
695 	if (type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
696 		dma->video.device_caps |= V4L2_CAP_VIDEO_CAPTURE;
697 	else
698 		dma->video.device_caps |= V4L2_CAP_VIDEO_OUTPUT;
699 
700 	video_set_drvdata(&dma->video, dma);
701 
702 	/* ... and the buffers queue... */
703 	/* Don't enable VB2_READ and VB2_WRITE, as using the read() and write()
704 	 * V4L2 APIs would be inefficient. Testing on the command line with a
705 	 * 'cat /dev/video?' thus won't be possible, but given that the driver
706 	 * anyway requires a test tool to setup the pipeline before any video
707 	 * stream can be started, requiring a specific V4L2 test tool as well
708 	 * instead of 'cat' isn't really a drawback.
709 	 */
710 	dma->queue.type = type;
711 	dma->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
712 	dma->queue.lock = &dma->lock;
713 	dma->queue.drv_priv = dma;
714 	dma->queue.buf_struct_size = sizeof(struct xvip_dma_buffer);
715 	dma->queue.ops = &xvip_dma_queue_qops;
716 	dma->queue.mem_ops = &vb2_dma_contig_memops;
717 	dma->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC
718 				   | V4L2_BUF_FLAG_TSTAMP_SRC_EOF;
719 	dma->queue.dev = dma->xdev->dev;
720 	ret = vb2_queue_init(&dma->queue);
721 	if (ret < 0) {
722 		dev_err(dma->xdev->dev, "failed to initialize VB2 queue\n");
723 		goto error;
724 	}
725 
726 	/* ... and the DMA channel. */
727 	snprintf(name, sizeof(name), "port%u", port);
728 	dma->dma = dma_request_chan(dma->xdev->dev, name);
729 	if (IS_ERR(dma->dma)) {
730 		ret = PTR_ERR(dma->dma);
731 		if (ret != -EPROBE_DEFER)
732 			dev_err(dma->xdev->dev, "no VDMA channel found\n");
733 		goto error;
734 	}
735 
736 	dma->align = 1 << dma->dma->device->copy_align;
737 
738 	ret = video_register_device(&dma->video, VFL_TYPE_VIDEO, -1);
739 	if (ret < 0) {
740 		dev_err(dma->xdev->dev, "failed to register video device\n");
741 		goto error;
742 	}
743 
744 	return 0;
745 
746 error:
747 	xvip_dma_cleanup(dma);
748 	return ret;
749 }
750 
xvip_dma_cleanup(struct xvip_dma * dma)751 void xvip_dma_cleanup(struct xvip_dma *dma)
752 {
753 	if (video_is_registered(&dma->video))
754 		video_unregister_device(&dma->video);
755 
756 	if (!IS_ERR_OR_NULL(dma->dma))
757 		dma_release_channel(dma->dma);
758 
759 	media_entity_cleanup(&dma->video.entity);
760 
761 	mutex_destroy(&dma->lock);
762 	mutex_destroy(&dma->pipe.lock);
763 }
764