1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Xilinx Video DMA 4 * 5 * Copyright (C) 2013-2015 Ideas on Board 6 * Copyright (C) 2013-2015 Xilinx, Inc. 7 * 8 * Contacts: Hyun Kwon <hyun.kwon@xilinx.com> 9 * Laurent Pinchart <laurent.pinchart@ideasonboard.com> 10 */ 11 12 #include <linux/dma/xilinx_dma.h> 13 #include <linux/lcm.h> 14 #include <linux/list.h> 15 #include <linux/module.h> 16 #include <linux/of.h> 17 #include <linux/slab.h> 18 19 #include <media/v4l2-dev.h> 20 #include <media/v4l2-fh.h> 21 #include <media/v4l2-ioctl.h> 22 #include <media/videobuf2-v4l2.h> 23 #include <media/videobuf2-dma-contig.h> 24 25 #include "xilinx-dma.h" 26 #include "xilinx-vip.h" 27 #include "xilinx-vipp.h" 28 29 #define XVIP_DMA_DEF_WIDTH 1920 30 #define XVIP_DMA_DEF_HEIGHT 1080 31 32 /* Minimum and maximum widths are expressed in bytes */ 33 #define XVIP_DMA_MIN_WIDTH 1U 34 #define XVIP_DMA_MAX_WIDTH 65535U 35 #define XVIP_DMA_MIN_HEIGHT 1U 36 #define XVIP_DMA_MAX_HEIGHT 8191U 37 38 /* ----------------------------------------------------------------------------- 39 * Helper functions 40 */ 41 42 static struct v4l2_subdev * 43 xvip_dma_remote_subdev(struct media_pad *local, u32 *pad) 44 { 45 struct media_pad *remote; 46 47 remote = media_pad_remote_pad_first(local); 48 if (!remote || !is_media_entity_v4l2_subdev(remote->entity)) 49 return NULL; 50 51 if (pad) 52 *pad = remote->index; 53 54 return media_entity_to_v4l2_subdev(remote->entity); 55 } 56 57 static int xvip_dma_verify_format(struct xvip_dma *dma) 58 { 59 struct v4l2_subdev_format fmt; 60 struct v4l2_subdev *subdev; 61 int ret; 62 63 subdev = xvip_dma_remote_subdev(&dma->pad, &fmt.pad); 64 if (subdev == NULL) 65 return -EPIPE; 66 67 fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE; 68 ret = v4l2_subdev_call(subdev, pad, get_fmt, NULL, &fmt); 69 if (ret < 0) 70 return ret == -ENOIOCTLCMD ? -EINVAL : ret; 71 72 if (dma->fmtinfo->code != fmt.format.code || 73 dma->format.height != fmt.format.height || 74 dma->format.width != fmt.format.width || 75 dma->format.colorspace != fmt.format.colorspace) 76 return -EINVAL; 77 78 return 0; 79 } 80 81 /* ----------------------------------------------------------------------------- 82 * Pipeline Stream Management 83 */ 84 85 /** 86 * xvip_pipeline_start_stop - Start ot stop streaming on a pipeline 87 * @pipe: The pipeline 88 * @start: Start (when true) or stop (when false) the pipeline 89 * 90 * Walk the entities chain starting at the pipeline output video node and start 91 * or stop all of them. 92 * 93 * Return: 0 if successful, or the return value of the failed video::s_stream 94 * operation otherwise. 95 */ 96 static int xvip_pipeline_start_stop(struct xvip_pipeline *pipe, bool start) 97 { 98 struct xvip_dma *dma = pipe->output; 99 struct media_entity *entity; 100 struct media_pad *pad; 101 struct v4l2_subdev *subdev; 102 int ret; 103 104 entity = &dma->video.entity; 105 while (1) { 106 pad = &entity->pads[0]; 107 if (!(pad->flags & MEDIA_PAD_FL_SINK)) 108 break; 109 110 pad = media_pad_remote_pad_first(pad); 111 if (!pad || !is_media_entity_v4l2_subdev(pad->entity)) 112 break; 113 114 entity = pad->entity; 115 subdev = media_entity_to_v4l2_subdev(entity); 116 117 ret = v4l2_subdev_call(subdev, video, s_stream, start); 118 if (start && ret < 0 && ret != -ENOIOCTLCMD) 119 return ret; 120 } 121 122 return 0; 123 } 124 125 /** 126 * xvip_pipeline_set_stream - Enable/disable streaming on a pipeline 127 * @pipe: The pipeline 128 * @on: Turn the stream on when true or off when false 129 * 130 * The pipeline is shared between all DMA engines connect at its input and 131 * output. While the stream state of DMA engines can be controlled 132 * independently, pipelines have a shared stream state that enable or disable 133 * all entities in the pipeline. For this reason the pipeline uses a streaming 134 * counter that tracks the number of DMA engines that have requested the stream 135 * to be enabled. 136 * 137 * When called with the @on argument set to true, this function will increment 138 * the pipeline streaming count. If the streaming count reaches the number of 139 * DMA engines in the pipeline it will enable all entities that belong to the 140 * pipeline. 141 * 142 * Similarly, when called with the @on argument set to false, this function will 143 * decrement the pipeline streaming count and disable all entities in the 144 * pipeline when the streaming count reaches zero. 145 * 146 * Return: 0 if successful, or the return value of the failed video::s_stream 147 * operation otherwise. Stopping the pipeline never fails. The pipeline state is 148 * not updated when the operation fails. 149 */ 150 static int xvip_pipeline_set_stream(struct xvip_pipeline *pipe, bool on) 151 { 152 int ret = 0; 153 154 mutex_lock(&pipe->lock); 155 156 if (on) { 157 if (pipe->stream_count == pipe->num_dmas - 1) { 158 ret = xvip_pipeline_start_stop(pipe, true); 159 if (ret < 0) 160 goto done; 161 } 162 pipe->stream_count++; 163 } else { 164 if (--pipe->stream_count == 0) 165 xvip_pipeline_start_stop(pipe, false); 166 } 167 168 done: 169 mutex_unlock(&pipe->lock); 170 return ret; 171 } 172 173 static int xvip_pipeline_validate(struct xvip_pipeline *pipe, 174 struct xvip_dma *start) 175 { 176 struct media_graph graph; 177 struct media_entity *entity = &start->video.entity; 178 struct media_device *mdev = entity->graph_obj.mdev; 179 unsigned int num_inputs = 0; 180 unsigned int num_outputs = 0; 181 int ret; 182 183 mutex_lock(&mdev->graph_mutex); 184 185 /* Walk the graph to locate the video nodes. */ 186 ret = media_graph_walk_init(&graph, mdev); 187 if (ret) { 188 mutex_unlock(&mdev->graph_mutex); 189 return ret; 190 } 191 192 media_graph_walk_start(&graph, entity); 193 194 while ((entity = media_graph_walk_next(&graph))) { 195 struct xvip_dma *dma; 196 197 if (entity->function != MEDIA_ENT_F_IO_V4L) 198 continue; 199 200 dma = to_xvip_dma(media_entity_to_video_device(entity)); 201 202 if (dma->pad.flags & MEDIA_PAD_FL_SINK) { 203 pipe->output = dma; 204 num_outputs++; 205 } else { 206 num_inputs++; 207 } 208 } 209 210 mutex_unlock(&mdev->graph_mutex); 211 212 media_graph_walk_cleanup(&graph); 213 214 /* We need exactly one output and zero or one input. */ 215 if (num_outputs != 1 || num_inputs > 1) 216 return -EPIPE; 217 218 pipe->num_dmas = num_inputs + num_outputs; 219 220 return 0; 221 } 222 223 static void __xvip_pipeline_cleanup(struct xvip_pipeline *pipe) 224 { 225 pipe->num_dmas = 0; 226 pipe->output = NULL; 227 } 228 229 /** 230 * xvip_pipeline_cleanup - Cleanup the pipeline after streaming 231 * @pipe: the pipeline 232 * 233 * Decrease the pipeline use count and clean it up if we were the last user. 234 */ 235 static void xvip_pipeline_cleanup(struct xvip_pipeline *pipe) 236 { 237 mutex_lock(&pipe->lock); 238 239 /* If we're the last user clean up the pipeline. */ 240 if (--pipe->use_count == 0) 241 __xvip_pipeline_cleanup(pipe); 242 243 mutex_unlock(&pipe->lock); 244 } 245 246 /** 247 * xvip_pipeline_prepare - Prepare the pipeline for streaming 248 * @pipe: the pipeline 249 * @dma: DMA engine at one end of the pipeline 250 * 251 * Validate the pipeline if no user exists yet, otherwise just increase the use 252 * count. 253 * 254 * Return: 0 if successful or -EPIPE if the pipeline is not valid. 255 */ 256 static int xvip_pipeline_prepare(struct xvip_pipeline *pipe, 257 struct xvip_dma *dma) 258 { 259 int ret; 260 261 mutex_lock(&pipe->lock); 262 263 /* If we're the first user validate and initialize the pipeline. */ 264 if (pipe->use_count == 0) { 265 ret = xvip_pipeline_validate(pipe, dma); 266 if (ret < 0) { 267 __xvip_pipeline_cleanup(pipe); 268 goto done; 269 } 270 } 271 272 pipe->use_count++; 273 ret = 0; 274 275 done: 276 mutex_unlock(&pipe->lock); 277 return ret; 278 } 279 280 /* ----------------------------------------------------------------------------- 281 * videobuf2 queue operations 282 */ 283 284 /** 285 * struct xvip_dma_buffer - Video DMA buffer 286 * @buf: vb2 buffer base object 287 * @queue: buffer list entry in the DMA engine queued buffers list 288 * @dma: DMA channel that uses the buffer 289 */ 290 struct xvip_dma_buffer { 291 struct vb2_v4l2_buffer buf; 292 struct list_head queue; 293 struct xvip_dma *dma; 294 }; 295 296 #define to_xvip_dma_buffer(vb) container_of(vb, struct xvip_dma_buffer, buf) 297 298 static void xvip_dma_complete(void *param) 299 { 300 struct xvip_dma_buffer *buf = param; 301 struct xvip_dma *dma = buf->dma; 302 303 spin_lock(&dma->queued_lock); 304 list_del(&buf->queue); 305 spin_unlock(&dma->queued_lock); 306 307 buf->buf.field = V4L2_FIELD_NONE; 308 buf->buf.sequence = dma->sequence++; 309 buf->buf.vb2_buf.timestamp = ktime_get_ns(); 310 vb2_set_plane_payload(&buf->buf.vb2_buf, 0, dma->format.sizeimage); 311 vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_DONE); 312 } 313 314 static int 315 xvip_dma_queue_setup(struct vb2_queue *vq, 316 unsigned int *nbuffers, unsigned int *nplanes, 317 unsigned int sizes[], struct device *alloc_devs[]) 318 { 319 struct xvip_dma *dma = vb2_get_drv_priv(vq); 320 321 /* Make sure the image size is large enough. */ 322 if (*nplanes) 323 return sizes[0] < dma->format.sizeimage ? -EINVAL : 0; 324 325 *nplanes = 1; 326 sizes[0] = dma->format.sizeimage; 327 328 return 0; 329 } 330 331 static int xvip_dma_buffer_prepare(struct vb2_buffer *vb) 332 { 333 struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); 334 struct xvip_dma *dma = vb2_get_drv_priv(vb->vb2_queue); 335 struct xvip_dma_buffer *buf = to_xvip_dma_buffer(vbuf); 336 337 buf->dma = dma; 338 339 return 0; 340 } 341 342 static void xvip_dma_buffer_queue(struct vb2_buffer *vb) 343 { 344 struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); 345 struct xvip_dma *dma = vb2_get_drv_priv(vb->vb2_queue); 346 struct xvip_dma_buffer *buf = to_xvip_dma_buffer(vbuf); 347 struct dma_async_tx_descriptor *desc; 348 dma_addr_t addr = vb2_dma_contig_plane_dma_addr(vb, 0); 349 u32 flags; 350 351 if (dma->queue.type == V4L2_BUF_TYPE_VIDEO_CAPTURE) { 352 flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK; 353 dma->xt.dir = DMA_DEV_TO_MEM; 354 dma->xt.src_sgl = false; 355 dma->xt.dst_sgl = true; 356 dma->xt.dst_start = addr; 357 } else { 358 flags = DMA_PREP_INTERRUPT | DMA_CTRL_ACK; 359 dma->xt.dir = DMA_MEM_TO_DEV; 360 dma->xt.src_sgl = true; 361 dma->xt.dst_sgl = false; 362 dma->xt.src_start = addr; 363 } 364 365 dma->xt.frame_size = 1; 366 dma->sgl[0].size = dma->format.width * dma->fmtinfo->bpp; 367 dma->sgl[0].icg = dma->format.bytesperline - dma->sgl[0].size; 368 dma->xt.numf = dma->format.height; 369 370 desc = dmaengine_prep_interleaved_dma(dma->dma, &dma->xt, flags); 371 if (!desc) { 372 dev_err(dma->xdev->dev, "Failed to prepare DMA transfer\n"); 373 vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_ERROR); 374 return; 375 } 376 desc->callback = xvip_dma_complete; 377 desc->callback_param = buf; 378 379 spin_lock_irq(&dma->queued_lock); 380 list_add_tail(&buf->queue, &dma->queued_bufs); 381 spin_unlock_irq(&dma->queued_lock); 382 383 dmaengine_submit(desc); 384 385 if (vb2_is_streaming(&dma->queue)) 386 dma_async_issue_pending(dma->dma); 387 } 388 389 static int xvip_dma_start_streaming(struct vb2_queue *vq, unsigned int count) 390 { 391 struct xvip_dma *dma = vb2_get_drv_priv(vq); 392 struct xvip_dma_buffer *buf, *nbuf; 393 struct xvip_pipeline *pipe; 394 int ret; 395 396 dma->sequence = 0; 397 398 /* 399 * Start streaming on the pipeline. No link touching an entity in the 400 * pipeline can be activated or deactivated once streaming is started. 401 * 402 * Use the pipeline object embedded in the first DMA object that starts 403 * streaming. 404 */ 405 pipe = dma->video.entity.pipe 406 ? to_xvip_pipeline(&dma->video.entity) : &dma->pipe; 407 408 ret = media_pipeline_start(&dma->video.entity, &pipe->pipe); 409 if (ret < 0) 410 goto error; 411 412 /* Verify that the configured format matches the output of the 413 * connected subdev. 414 */ 415 ret = xvip_dma_verify_format(dma); 416 if (ret < 0) 417 goto error_stop; 418 419 ret = xvip_pipeline_prepare(pipe, dma); 420 if (ret < 0) 421 goto error_stop; 422 423 /* Start the DMA engine. This must be done before starting the blocks 424 * in the pipeline to avoid DMA synchronization issues. 425 */ 426 dma_async_issue_pending(dma->dma); 427 428 /* Start the pipeline. */ 429 xvip_pipeline_set_stream(pipe, true); 430 431 return 0; 432 433 error_stop: 434 media_pipeline_stop(&dma->video.entity); 435 436 error: 437 /* Give back all queued buffers to videobuf2. */ 438 spin_lock_irq(&dma->queued_lock); 439 list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) { 440 vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_QUEUED); 441 list_del(&buf->queue); 442 } 443 spin_unlock_irq(&dma->queued_lock); 444 445 return ret; 446 } 447 448 static void xvip_dma_stop_streaming(struct vb2_queue *vq) 449 { 450 struct xvip_dma *dma = vb2_get_drv_priv(vq); 451 struct xvip_pipeline *pipe = to_xvip_pipeline(&dma->video.entity); 452 struct xvip_dma_buffer *buf, *nbuf; 453 454 /* Stop the pipeline. */ 455 xvip_pipeline_set_stream(pipe, false); 456 457 /* Stop and reset the DMA engine. */ 458 dmaengine_terminate_all(dma->dma); 459 460 /* Cleanup the pipeline and mark it as being stopped. */ 461 xvip_pipeline_cleanup(pipe); 462 media_pipeline_stop(&dma->video.entity); 463 464 /* Give back all queued buffers to videobuf2. */ 465 spin_lock_irq(&dma->queued_lock); 466 list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) { 467 vb2_buffer_done(&buf->buf.vb2_buf, VB2_BUF_STATE_ERROR); 468 list_del(&buf->queue); 469 } 470 spin_unlock_irq(&dma->queued_lock); 471 } 472 473 static const struct vb2_ops xvip_dma_queue_qops = { 474 .queue_setup = xvip_dma_queue_setup, 475 .buf_prepare = xvip_dma_buffer_prepare, 476 .buf_queue = xvip_dma_buffer_queue, 477 .wait_prepare = vb2_ops_wait_prepare, 478 .wait_finish = vb2_ops_wait_finish, 479 .start_streaming = xvip_dma_start_streaming, 480 .stop_streaming = xvip_dma_stop_streaming, 481 }; 482 483 /* ----------------------------------------------------------------------------- 484 * V4L2 ioctls 485 */ 486 487 static int 488 xvip_dma_querycap(struct file *file, void *fh, struct v4l2_capability *cap) 489 { 490 struct v4l2_fh *vfh = file->private_data; 491 struct xvip_dma *dma = to_xvip_dma(vfh->vdev); 492 493 cap->capabilities = dma->xdev->v4l2_caps | V4L2_CAP_STREAMING | 494 V4L2_CAP_DEVICE_CAPS; 495 496 strscpy(cap->driver, "xilinx-vipp", sizeof(cap->driver)); 497 strscpy(cap->card, dma->video.name, sizeof(cap->card)); 498 snprintf(cap->bus_info, sizeof(cap->bus_info), "platform:%pOFn:%u", 499 dma->xdev->dev->of_node, dma->port); 500 501 return 0; 502 } 503 504 /* FIXME: without this callback function, some applications are not configured 505 * with correct formats, and it results in frames in wrong format. Whether this 506 * callback needs to be required is not clearly defined, so it should be 507 * clarified through the mailing list. 508 */ 509 static int 510 xvip_dma_enum_format(struct file *file, void *fh, struct v4l2_fmtdesc *f) 511 { 512 struct v4l2_fh *vfh = file->private_data; 513 struct xvip_dma *dma = to_xvip_dma(vfh->vdev); 514 515 if (f->index > 0) 516 return -EINVAL; 517 518 f->pixelformat = dma->format.pixelformat; 519 520 return 0; 521 } 522 523 static int 524 xvip_dma_get_format(struct file *file, void *fh, struct v4l2_format *format) 525 { 526 struct v4l2_fh *vfh = file->private_data; 527 struct xvip_dma *dma = to_xvip_dma(vfh->vdev); 528 529 format->fmt.pix = dma->format; 530 531 return 0; 532 } 533 534 static void 535 __xvip_dma_try_format(struct xvip_dma *dma, struct v4l2_pix_format *pix, 536 const struct xvip_video_format **fmtinfo) 537 { 538 const struct xvip_video_format *info; 539 unsigned int min_width; 540 unsigned int max_width; 541 unsigned int min_bpl; 542 unsigned int max_bpl; 543 unsigned int width; 544 unsigned int align; 545 unsigned int bpl; 546 547 /* Retrieve format information and select the default format if the 548 * requested format isn't supported. 549 */ 550 info = xvip_get_format_by_fourcc(pix->pixelformat); 551 552 pix->pixelformat = info->fourcc; 553 pix->field = V4L2_FIELD_NONE; 554 555 /* The transfer alignment requirements are expressed in bytes. Compute 556 * the minimum and maximum values, clamp the requested width and convert 557 * it back to pixels. 558 */ 559 align = lcm(dma->align, info->bpp); 560 min_width = roundup(XVIP_DMA_MIN_WIDTH, align); 561 max_width = rounddown(XVIP_DMA_MAX_WIDTH, align); 562 width = rounddown(pix->width * info->bpp, align); 563 564 pix->width = clamp(width, min_width, max_width) / info->bpp; 565 pix->height = clamp(pix->height, XVIP_DMA_MIN_HEIGHT, 566 XVIP_DMA_MAX_HEIGHT); 567 568 /* Clamp the requested bytes per line value. If the maximum bytes per 569 * line value is zero, the module doesn't support user configurable line 570 * sizes. Override the requested value with the minimum in that case. 571 */ 572 min_bpl = pix->width * info->bpp; 573 max_bpl = rounddown(XVIP_DMA_MAX_WIDTH, dma->align); 574 bpl = rounddown(pix->bytesperline, dma->align); 575 576 pix->bytesperline = clamp(bpl, min_bpl, max_bpl); 577 pix->sizeimage = pix->bytesperline * pix->height; 578 579 if (fmtinfo) 580 *fmtinfo = info; 581 } 582 583 static int 584 xvip_dma_try_format(struct file *file, void *fh, struct v4l2_format *format) 585 { 586 struct v4l2_fh *vfh = file->private_data; 587 struct xvip_dma *dma = to_xvip_dma(vfh->vdev); 588 589 __xvip_dma_try_format(dma, &format->fmt.pix, NULL); 590 return 0; 591 } 592 593 static int 594 xvip_dma_set_format(struct file *file, void *fh, struct v4l2_format *format) 595 { 596 struct v4l2_fh *vfh = file->private_data; 597 struct xvip_dma *dma = to_xvip_dma(vfh->vdev); 598 const struct xvip_video_format *info; 599 600 __xvip_dma_try_format(dma, &format->fmt.pix, &info); 601 602 if (vb2_is_busy(&dma->queue)) 603 return -EBUSY; 604 605 dma->format = format->fmt.pix; 606 dma->fmtinfo = info; 607 608 return 0; 609 } 610 611 static const struct v4l2_ioctl_ops xvip_dma_ioctl_ops = { 612 .vidioc_querycap = xvip_dma_querycap, 613 .vidioc_enum_fmt_vid_cap = xvip_dma_enum_format, 614 .vidioc_g_fmt_vid_cap = xvip_dma_get_format, 615 .vidioc_g_fmt_vid_out = xvip_dma_get_format, 616 .vidioc_s_fmt_vid_cap = xvip_dma_set_format, 617 .vidioc_s_fmt_vid_out = xvip_dma_set_format, 618 .vidioc_try_fmt_vid_cap = xvip_dma_try_format, 619 .vidioc_try_fmt_vid_out = xvip_dma_try_format, 620 .vidioc_reqbufs = vb2_ioctl_reqbufs, 621 .vidioc_querybuf = vb2_ioctl_querybuf, 622 .vidioc_qbuf = vb2_ioctl_qbuf, 623 .vidioc_dqbuf = vb2_ioctl_dqbuf, 624 .vidioc_create_bufs = vb2_ioctl_create_bufs, 625 .vidioc_expbuf = vb2_ioctl_expbuf, 626 .vidioc_streamon = vb2_ioctl_streamon, 627 .vidioc_streamoff = vb2_ioctl_streamoff, 628 }; 629 630 /* ----------------------------------------------------------------------------- 631 * V4L2 file operations 632 */ 633 634 static const struct v4l2_file_operations xvip_dma_fops = { 635 .owner = THIS_MODULE, 636 .unlocked_ioctl = video_ioctl2, 637 .open = v4l2_fh_open, 638 .release = vb2_fop_release, 639 .poll = vb2_fop_poll, 640 .mmap = vb2_fop_mmap, 641 }; 642 643 /* ----------------------------------------------------------------------------- 644 * Xilinx Video DMA Core 645 */ 646 647 int xvip_dma_init(struct xvip_composite_device *xdev, struct xvip_dma *dma, 648 enum v4l2_buf_type type, unsigned int port) 649 { 650 char name[16]; 651 int ret; 652 653 dma->xdev = xdev; 654 dma->port = port; 655 mutex_init(&dma->lock); 656 mutex_init(&dma->pipe.lock); 657 INIT_LIST_HEAD(&dma->queued_bufs); 658 spin_lock_init(&dma->queued_lock); 659 660 dma->fmtinfo = xvip_get_format_by_fourcc(V4L2_PIX_FMT_YUYV); 661 dma->format.pixelformat = dma->fmtinfo->fourcc; 662 dma->format.colorspace = V4L2_COLORSPACE_SRGB; 663 dma->format.field = V4L2_FIELD_NONE; 664 dma->format.width = XVIP_DMA_DEF_WIDTH; 665 dma->format.height = XVIP_DMA_DEF_HEIGHT; 666 dma->format.bytesperline = dma->format.width * dma->fmtinfo->bpp; 667 dma->format.sizeimage = dma->format.bytesperline * dma->format.height; 668 669 /* Initialize the media entity... */ 670 dma->pad.flags = type == V4L2_BUF_TYPE_VIDEO_CAPTURE 671 ? MEDIA_PAD_FL_SINK : MEDIA_PAD_FL_SOURCE; 672 673 ret = media_entity_pads_init(&dma->video.entity, 1, &dma->pad); 674 if (ret < 0) 675 goto error; 676 677 /* ... and the video node... */ 678 dma->video.fops = &xvip_dma_fops; 679 dma->video.v4l2_dev = &xdev->v4l2_dev; 680 dma->video.queue = &dma->queue; 681 snprintf(dma->video.name, sizeof(dma->video.name), "%pOFn %s %u", 682 xdev->dev->of_node, 683 type == V4L2_BUF_TYPE_VIDEO_CAPTURE ? "output" : "input", 684 port); 685 dma->video.vfl_type = VFL_TYPE_VIDEO; 686 dma->video.vfl_dir = type == V4L2_BUF_TYPE_VIDEO_CAPTURE 687 ? VFL_DIR_RX : VFL_DIR_TX; 688 dma->video.release = video_device_release_empty; 689 dma->video.ioctl_ops = &xvip_dma_ioctl_ops; 690 dma->video.lock = &dma->lock; 691 dma->video.device_caps = V4L2_CAP_STREAMING; 692 if (type == V4L2_BUF_TYPE_VIDEO_CAPTURE) 693 dma->video.device_caps |= V4L2_CAP_VIDEO_CAPTURE; 694 else 695 dma->video.device_caps |= V4L2_CAP_VIDEO_OUTPUT; 696 697 video_set_drvdata(&dma->video, dma); 698 699 /* ... and the buffers queue... */ 700 /* Don't enable VB2_READ and VB2_WRITE, as using the read() and write() 701 * V4L2 APIs would be inefficient. Testing on the command line with a 702 * 'cat /dev/video?' thus won't be possible, but given that the driver 703 * anyway requires a test tool to setup the pipeline before any video 704 * stream can be started, requiring a specific V4L2 test tool as well 705 * instead of 'cat' isn't really a drawback. 706 */ 707 dma->queue.type = type; 708 dma->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF; 709 dma->queue.lock = &dma->lock; 710 dma->queue.drv_priv = dma; 711 dma->queue.buf_struct_size = sizeof(struct xvip_dma_buffer); 712 dma->queue.ops = &xvip_dma_queue_qops; 713 dma->queue.mem_ops = &vb2_dma_contig_memops; 714 dma->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC 715 | V4L2_BUF_FLAG_TSTAMP_SRC_EOF; 716 dma->queue.dev = dma->xdev->dev; 717 ret = vb2_queue_init(&dma->queue); 718 if (ret < 0) { 719 dev_err(dma->xdev->dev, "failed to initialize VB2 queue\n"); 720 goto error; 721 } 722 723 /* ... and the DMA channel. */ 724 snprintf(name, sizeof(name), "port%u", port); 725 dma->dma = dma_request_chan(dma->xdev->dev, name); 726 if (IS_ERR(dma->dma)) { 727 ret = PTR_ERR(dma->dma); 728 if (ret != -EPROBE_DEFER) 729 dev_err(dma->xdev->dev, "no VDMA channel found\n"); 730 goto error; 731 } 732 733 dma->align = 1 << dma->dma->device->copy_align; 734 735 ret = video_register_device(&dma->video, VFL_TYPE_VIDEO, -1); 736 if (ret < 0) { 737 dev_err(dma->xdev->dev, "failed to register video device\n"); 738 goto error; 739 } 740 741 return 0; 742 743 error: 744 xvip_dma_cleanup(dma); 745 return ret; 746 } 747 748 void xvip_dma_cleanup(struct xvip_dma *dma) 749 { 750 if (video_is_registered(&dma->video)) 751 video_unregister_device(&dma->video); 752 753 if (!IS_ERR_OR_NULL(dma->dma)) 754 dma_release_channel(dma->dma); 755 756 media_entity_cleanup(&dma->video.entity); 757 758 mutex_destroy(&dma->lock); 759 mutex_destroy(&dma->pipe.lock); 760 } 761