1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * TI VPFE capture Driver 4 * 5 * Copyright (C) 2013 - 2014 Texas Instruments, Inc. 6 * 7 * Benoit Parrot <bparrot@ti.com> 8 * Lad, Prabhakar <prabhakar.csengg@gmail.com> 9 */ 10 11 #include <linux/delay.h> 12 #include <linux/err.h> 13 #include <linux/init.h> 14 #include <linux/interrupt.h> 15 #include <linux/io.h> 16 #include <linux/module.h> 17 #include <linux/of_graph.h> 18 #include <linux/pinctrl/consumer.h> 19 #include <linux/platform_device.h> 20 #include <linux/pm_runtime.h> 21 #include <linux/slab.h> 22 #include <linux/uaccess.h> 23 #include <linux/videodev2.h> 24 25 #include <media/v4l2-common.h> 26 #include <media/v4l2-ctrls.h> 27 #include <media/v4l2-event.h> 28 #include <media/v4l2-fwnode.h> 29 #include <media/v4l2-rect.h> 30 31 #include "am437x-vpfe.h" 32 33 #define VPFE_MODULE_NAME "vpfe" 34 #define VPFE_VERSION "0.1.0" 35 36 static int debug; 37 module_param(debug, int, 0644); 38 MODULE_PARM_DESC(debug, "Debug level 0-8"); 39 40 #define vpfe_dbg(level, dev, fmt, arg...) \ 41 v4l2_dbg(level, debug, &dev->v4l2_dev, fmt, ##arg) 42 #define vpfe_info(dev, fmt, arg...) \ 43 v4l2_info(&dev->v4l2_dev, fmt, ##arg) 44 #define vpfe_err(dev, fmt, arg...) \ 45 v4l2_err(&dev->v4l2_dev, fmt, ##arg) 46 47 /* standard information */ 48 struct vpfe_standard { 49 v4l2_std_id std_id; 50 unsigned int width; 51 unsigned int height; 52 struct v4l2_fract pixelaspect; 53 int frame_format; 54 }; 55 56 static const struct vpfe_standard vpfe_standards[] = { 57 {V4L2_STD_525_60, 720, 480, {11, 10}, 1}, 58 {V4L2_STD_625_50, 720, 576, {54, 59}, 1}, 59 }; 60 61 static struct vpfe_fmt formats[VPFE_NUM_FORMATS] = { 62 { 63 .fourcc = V4L2_PIX_FMT_YUYV, 64 .code = MEDIA_BUS_FMT_YUYV8_2X8, 65 .bitsperpixel = 16, 66 }, { 67 .fourcc = V4L2_PIX_FMT_UYVY, 68 .code = MEDIA_BUS_FMT_UYVY8_2X8, 69 .bitsperpixel = 16, 70 }, { 71 .fourcc = V4L2_PIX_FMT_YVYU, 72 .code = MEDIA_BUS_FMT_YVYU8_2X8, 73 .bitsperpixel = 16, 74 }, { 75 .fourcc = V4L2_PIX_FMT_VYUY, 76 .code = MEDIA_BUS_FMT_VYUY8_2X8, 77 .bitsperpixel = 16, 78 }, { 79 .fourcc = V4L2_PIX_FMT_SBGGR8, 80 .code = MEDIA_BUS_FMT_SBGGR8_1X8, 81 .bitsperpixel = 8, 82 }, { 83 .fourcc = V4L2_PIX_FMT_SGBRG8, 84 .code = MEDIA_BUS_FMT_SGBRG8_1X8, 85 .bitsperpixel = 8, 86 }, { 87 .fourcc = V4L2_PIX_FMT_SGRBG8, 88 .code = MEDIA_BUS_FMT_SGRBG8_1X8, 89 .bitsperpixel = 8, 90 }, { 91 .fourcc = V4L2_PIX_FMT_SRGGB8, 92 .code = MEDIA_BUS_FMT_SRGGB8_1X8, 93 .bitsperpixel = 8, 94 }, { 95 .fourcc = V4L2_PIX_FMT_RGB565, 96 .code = MEDIA_BUS_FMT_RGB565_2X8_LE, 97 .bitsperpixel = 16, 98 }, { 99 .fourcc = V4L2_PIX_FMT_RGB565X, 100 .code = MEDIA_BUS_FMT_RGB565_2X8_BE, 101 .bitsperpixel = 16, 102 }, 103 }; 104 105 static int __subdev_get_format(struct vpfe_device *vpfe, 106 struct v4l2_mbus_framefmt *fmt); 107 static int vpfe_calc_format_size(struct vpfe_device *vpfe, 108 const struct vpfe_fmt *fmt, 109 struct v4l2_format *f); 110 111 static struct vpfe_fmt *find_format_by_code(struct vpfe_device *vpfe, 112 unsigned int code) 113 { 114 struct vpfe_fmt *fmt; 115 unsigned int k; 116 117 for (k = 0; k < vpfe->num_active_fmt; k++) { 118 fmt = vpfe->active_fmt[k]; 119 if (fmt->code == code) 120 return fmt; 121 } 122 123 return NULL; 124 } 125 126 static struct vpfe_fmt *find_format_by_pix(struct vpfe_device *vpfe, 127 unsigned int pixelformat) 128 { 129 struct vpfe_fmt *fmt; 130 unsigned int k; 131 132 for (k = 0; k < vpfe->num_active_fmt; k++) { 133 fmt = vpfe->active_fmt[k]; 134 if (fmt->fourcc == pixelformat) 135 return fmt; 136 } 137 138 return NULL; 139 } 140 141 static unsigned int __get_bytesperpixel(struct vpfe_device *vpfe, 142 const struct vpfe_fmt *fmt) 143 { 144 struct vpfe_subdev_info *sdinfo = vpfe->current_subdev; 145 unsigned int bus_width = sdinfo->vpfe_param.bus_width; 146 u32 bpp, bus_width_bytes, clocksperpixel; 147 148 bus_width_bytes = ALIGN(bus_width, 8) >> 3; 149 clocksperpixel = DIV_ROUND_UP(fmt->bitsperpixel, bus_width); 150 bpp = clocksperpixel * bus_width_bytes; 151 152 return bpp; 153 } 154 155 /* Print Four-character-code (FOURCC) */ 156 static char *print_fourcc(u32 fmt) 157 { 158 static char code[5]; 159 160 code[0] = (unsigned char)(fmt & 0xff); 161 code[1] = (unsigned char)((fmt >> 8) & 0xff); 162 code[2] = (unsigned char)((fmt >> 16) & 0xff); 163 code[3] = (unsigned char)((fmt >> 24) & 0xff); 164 code[4] = '\0'; 165 166 return code; 167 } 168 169 static inline u32 vpfe_reg_read(struct vpfe_ccdc *ccdc, u32 offset) 170 { 171 return ioread32(ccdc->ccdc_cfg.base_addr + offset); 172 } 173 174 static inline void vpfe_reg_write(struct vpfe_ccdc *ccdc, u32 val, u32 offset) 175 { 176 iowrite32(val, ccdc->ccdc_cfg.base_addr + offset); 177 } 178 179 static inline struct vpfe_device *to_vpfe(struct vpfe_ccdc *ccdc) 180 { 181 return container_of(ccdc, struct vpfe_device, ccdc); 182 } 183 184 static inline 185 struct vpfe_cap_buffer *to_vpfe_buffer(struct vb2_v4l2_buffer *vb) 186 { 187 return container_of(vb, struct vpfe_cap_buffer, vb); 188 } 189 190 static inline void vpfe_pcr_enable(struct vpfe_ccdc *ccdc, int flag) 191 { 192 vpfe_reg_write(ccdc, !!flag, VPFE_PCR); 193 } 194 195 static void vpfe_config_enable(struct vpfe_ccdc *ccdc, int flag) 196 { 197 unsigned int cfg; 198 199 if (!flag) { 200 cfg = vpfe_reg_read(ccdc, VPFE_CONFIG); 201 cfg &= ~(VPFE_CONFIG_EN_ENABLE << VPFE_CONFIG_EN_SHIFT); 202 } else { 203 cfg = VPFE_CONFIG_EN_ENABLE << VPFE_CONFIG_EN_SHIFT; 204 } 205 206 vpfe_reg_write(ccdc, cfg, VPFE_CONFIG); 207 } 208 209 static void vpfe_ccdc_setwin(struct vpfe_ccdc *ccdc, 210 struct v4l2_rect *image_win, 211 enum ccdc_frmfmt frm_fmt, 212 int bpp) 213 { 214 int horz_start, horz_nr_pixels; 215 int vert_start, vert_nr_lines; 216 int val, mid_img; 217 218 /* 219 * ppc - per pixel count. indicates how many pixels per cell 220 * output to SDRAM. example, for ycbcr, it is one y and one c, so 2. 221 * raw capture this is 1 222 */ 223 horz_start = image_win->left * bpp; 224 horz_nr_pixels = (image_win->width * bpp) - 1; 225 vpfe_reg_write(ccdc, (horz_start << VPFE_HORZ_INFO_SPH_SHIFT) | 226 horz_nr_pixels, VPFE_HORZ_INFO); 227 228 vert_start = image_win->top; 229 230 if (frm_fmt == CCDC_FRMFMT_INTERLACED) { 231 vert_nr_lines = (image_win->height >> 1) - 1; 232 vert_start >>= 1; 233 /* configure VDINT0 */ 234 val = (vert_start << VPFE_VDINT_VDINT0_SHIFT); 235 } else { 236 vert_nr_lines = image_win->height - 1; 237 /* 238 * configure VDINT0 and VDINT1. VDINT1 will be at half 239 * of image height 240 */ 241 mid_img = vert_start + (image_win->height / 2); 242 val = (vert_start << VPFE_VDINT_VDINT0_SHIFT) | 243 (mid_img & VPFE_VDINT_VDINT1_MASK); 244 } 245 246 vpfe_reg_write(ccdc, val, VPFE_VDINT); 247 248 vpfe_reg_write(ccdc, (vert_start << VPFE_VERT_START_SLV0_SHIFT) | 249 vert_start, VPFE_VERT_START); 250 vpfe_reg_write(ccdc, vert_nr_lines, VPFE_VERT_LINES); 251 } 252 253 static void vpfe_reg_dump(struct vpfe_ccdc *ccdc) 254 { 255 struct vpfe_device *vpfe = to_vpfe(ccdc); 256 257 vpfe_dbg(3, vpfe, "ALAW: 0x%x\n", vpfe_reg_read(ccdc, VPFE_ALAW)); 258 vpfe_dbg(3, vpfe, "CLAMP: 0x%x\n", vpfe_reg_read(ccdc, VPFE_CLAMP)); 259 vpfe_dbg(3, vpfe, "DCSUB: 0x%x\n", vpfe_reg_read(ccdc, VPFE_DCSUB)); 260 vpfe_dbg(3, vpfe, "BLKCMP: 0x%x\n", vpfe_reg_read(ccdc, VPFE_BLKCMP)); 261 vpfe_dbg(3, vpfe, "COLPTN: 0x%x\n", vpfe_reg_read(ccdc, VPFE_COLPTN)); 262 vpfe_dbg(3, vpfe, "SDOFST: 0x%x\n", vpfe_reg_read(ccdc, VPFE_SDOFST)); 263 vpfe_dbg(3, vpfe, "SYN_MODE: 0x%x\n", 264 vpfe_reg_read(ccdc, VPFE_SYNMODE)); 265 vpfe_dbg(3, vpfe, "HSIZE_OFF: 0x%x\n", 266 vpfe_reg_read(ccdc, VPFE_HSIZE_OFF)); 267 vpfe_dbg(3, vpfe, "HORZ_INFO: 0x%x\n", 268 vpfe_reg_read(ccdc, VPFE_HORZ_INFO)); 269 vpfe_dbg(3, vpfe, "VERT_START: 0x%x\n", 270 vpfe_reg_read(ccdc, VPFE_VERT_START)); 271 vpfe_dbg(3, vpfe, "VERT_LINES: 0x%x\n", 272 vpfe_reg_read(ccdc, VPFE_VERT_LINES)); 273 } 274 275 static int 276 vpfe_ccdc_validate_param(struct vpfe_ccdc *ccdc, 277 struct vpfe_ccdc_config_params_raw *ccdcparam) 278 { 279 struct vpfe_device *vpfe = to_vpfe(ccdc); 280 u8 max_gamma, max_data; 281 282 if (!ccdcparam->alaw.enable) 283 return 0; 284 285 max_gamma = ccdc_gamma_width_max_bit(ccdcparam->alaw.gamma_wd); 286 max_data = ccdc_data_size_max_bit(ccdcparam->data_sz); 287 288 if (ccdcparam->alaw.gamma_wd > VPFE_CCDC_GAMMA_BITS_09_0 || 289 ccdcparam->data_sz > VPFE_CCDC_DATA_8BITS || 290 max_gamma > max_data) { 291 vpfe_dbg(1, vpfe, "Invalid data line select\n"); 292 return -EINVAL; 293 } 294 295 return 0; 296 } 297 298 static void 299 vpfe_ccdc_update_raw_params(struct vpfe_ccdc *ccdc, 300 struct vpfe_ccdc_config_params_raw *raw_params) 301 { 302 struct vpfe_ccdc_config_params_raw *config_params = 303 &ccdc->ccdc_cfg.bayer.config_params; 304 305 *config_params = *raw_params; 306 } 307 308 /* 309 * vpfe_ccdc_restore_defaults() 310 * This function will write defaults to all CCDC registers 311 */ 312 static void vpfe_ccdc_restore_defaults(struct vpfe_ccdc *ccdc) 313 { 314 int i; 315 316 /* Disable CCDC */ 317 vpfe_pcr_enable(ccdc, 0); 318 319 /* set all registers to default value */ 320 for (i = 4; i <= 0x94; i += 4) 321 vpfe_reg_write(ccdc, 0, i); 322 323 vpfe_reg_write(ccdc, VPFE_NO_CULLING, VPFE_CULLING); 324 vpfe_reg_write(ccdc, VPFE_CCDC_GAMMA_BITS_11_2, VPFE_ALAW); 325 } 326 327 static int vpfe_ccdc_close(struct vpfe_ccdc *ccdc, struct device *dev) 328 { 329 struct vpfe_device *vpfe = to_vpfe(ccdc); 330 u32 dma_cntl, pcr; 331 332 pcr = vpfe_reg_read(ccdc, VPFE_PCR); 333 if (pcr) 334 vpfe_dbg(1, vpfe, "VPFE_PCR is still set (%x)", pcr); 335 336 dma_cntl = vpfe_reg_read(ccdc, VPFE_DMA_CNTL); 337 if ((dma_cntl & VPFE_DMA_CNTL_OVERFLOW)) 338 vpfe_dbg(1, vpfe, "VPFE_DMA_CNTL_OVERFLOW is still set (%x)", 339 dma_cntl); 340 341 /* Disable CCDC by resetting all register to default POR values */ 342 vpfe_ccdc_restore_defaults(ccdc); 343 344 /* Disabled the module at the CONFIG level */ 345 vpfe_config_enable(ccdc, 0); 346 347 pm_runtime_put_sync(dev); 348 return 0; 349 } 350 351 static int vpfe_ccdc_set_params(struct vpfe_ccdc *ccdc, void __user *params) 352 { 353 struct vpfe_device *vpfe = to_vpfe(ccdc); 354 struct vpfe_ccdc_config_params_raw raw_params; 355 int x; 356 357 if (ccdc->ccdc_cfg.if_type != VPFE_RAW_BAYER) 358 return -EINVAL; 359 360 x = copy_from_user(&raw_params, params, sizeof(raw_params)); 361 if (x) { 362 vpfe_dbg(1, vpfe, 363 "%s: error in copying ccdc params, %d\n", 364 __func__, x); 365 return -EFAULT; 366 } 367 368 if (!vpfe_ccdc_validate_param(ccdc, &raw_params)) { 369 vpfe_ccdc_update_raw_params(ccdc, &raw_params); 370 return 0; 371 } 372 373 return -EINVAL; 374 } 375 376 /* 377 * vpfe_ccdc_config_ycbcr() 378 * This function will configure CCDC for YCbCr video capture 379 */ 380 static void vpfe_ccdc_config_ycbcr(struct vpfe_ccdc *ccdc) 381 { 382 struct ccdc_params_ycbcr *params = &ccdc->ccdc_cfg.ycbcr; 383 u32 syn_mode; 384 385 /* 386 * first restore the CCDC registers to default values 387 * This is important since we assume default values to be set in 388 * a lot of registers that we didn't touch 389 */ 390 vpfe_ccdc_restore_defaults(ccdc); 391 392 /* 393 * configure pixel format, frame format, configure video frame 394 * format, enable output to SDRAM, enable internal timing generator 395 * and 8bit pack mode 396 */ 397 syn_mode = (((params->pix_fmt & VPFE_SYN_MODE_INPMOD_MASK) << 398 VPFE_SYN_MODE_INPMOD_SHIFT) | 399 ((params->frm_fmt & VPFE_SYN_FLDMODE_MASK) << 400 VPFE_SYN_FLDMODE_SHIFT) | VPFE_VDHDEN_ENABLE | 401 VPFE_WEN_ENABLE | VPFE_DATA_PACK_ENABLE); 402 403 /* setup BT.656 sync mode */ 404 if (params->bt656_enable) { 405 vpfe_reg_write(ccdc, VPFE_REC656IF_BT656_EN, VPFE_REC656IF); 406 407 /* 408 * configure the FID, VD, HD pin polarity, 409 * fld,hd pol positive, vd negative, 8-bit data 410 */ 411 syn_mode |= VPFE_SYN_MODE_VD_POL_NEGATIVE; 412 if (ccdc->ccdc_cfg.if_type == VPFE_BT656_10BIT) 413 syn_mode |= VPFE_SYN_MODE_10BITS; 414 else 415 syn_mode |= VPFE_SYN_MODE_8BITS; 416 } else { 417 /* y/c external sync mode */ 418 syn_mode |= (((params->fid_pol & VPFE_FID_POL_MASK) << 419 VPFE_FID_POL_SHIFT) | 420 ((params->hd_pol & VPFE_HD_POL_MASK) << 421 VPFE_HD_POL_SHIFT) | 422 ((params->vd_pol & VPFE_VD_POL_MASK) << 423 VPFE_VD_POL_SHIFT)); 424 } 425 vpfe_reg_write(ccdc, syn_mode, VPFE_SYNMODE); 426 427 /* configure video window */ 428 vpfe_ccdc_setwin(ccdc, ¶ms->win, 429 params->frm_fmt, params->bytesperpixel); 430 431 /* 432 * configure the order of y cb cr in SDRAM, and disable latch 433 * internal register on vsync 434 */ 435 if (ccdc->ccdc_cfg.if_type == VPFE_BT656_10BIT) 436 vpfe_reg_write(ccdc, 437 (params->pix_order << VPFE_CCDCFG_Y8POS_SHIFT) | 438 VPFE_LATCH_ON_VSYNC_DISABLE | 439 VPFE_CCDCFG_BW656_10BIT, VPFE_CCDCFG); 440 else 441 vpfe_reg_write(ccdc, 442 (params->pix_order << VPFE_CCDCFG_Y8POS_SHIFT) | 443 VPFE_LATCH_ON_VSYNC_DISABLE, VPFE_CCDCFG); 444 445 /* 446 * configure the horizontal line offset. This should be a 447 * on 32 byte boundary. So clear LSB 5 bits 448 */ 449 vpfe_reg_write(ccdc, params->bytesperline, VPFE_HSIZE_OFF); 450 451 /* configure the memory line offset */ 452 if (params->buf_type == CCDC_BUFTYPE_FLD_INTERLEAVED) 453 /* two fields are interleaved in memory */ 454 vpfe_reg_write(ccdc, VPFE_SDOFST_FIELD_INTERLEAVED, 455 VPFE_SDOFST); 456 } 457 458 static void 459 vpfe_ccdc_config_black_clamp(struct vpfe_ccdc *ccdc, 460 struct vpfe_ccdc_black_clamp *bclamp) 461 { 462 u32 val; 463 464 if (!bclamp->enable) { 465 /* configure DCSub */ 466 val = (bclamp->dc_sub) & VPFE_BLK_DC_SUB_MASK; 467 vpfe_reg_write(ccdc, val, VPFE_DCSUB); 468 vpfe_reg_write(ccdc, VPFE_CLAMP_DEFAULT_VAL, VPFE_CLAMP); 469 return; 470 } 471 /* 472 * Configure gain, Start pixel, No of line to be avg, 473 * No of pixel/line to be avg, & Enable the Black clamping 474 */ 475 val = ((bclamp->sgain & VPFE_BLK_SGAIN_MASK) | 476 ((bclamp->start_pixel & VPFE_BLK_ST_PXL_MASK) << 477 VPFE_BLK_ST_PXL_SHIFT) | 478 ((bclamp->sample_ln & VPFE_BLK_SAMPLE_LINE_MASK) << 479 VPFE_BLK_SAMPLE_LINE_SHIFT) | 480 ((bclamp->sample_pixel & VPFE_BLK_SAMPLE_LN_MASK) << 481 VPFE_BLK_SAMPLE_LN_SHIFT) | VPFE_BLK_CLAMP_ENABLE); 482 vpfe_reg_write(ccdc, val, VPFE_CLAMP); 483 /* If Black clamping is enable then make dcsub 0 */ 484 vpfe_reg_write(ccdc, VPFE_DCSUB_DEFAULT_VAL, VPFE_DCSUB); 485 } 486 487 static void 488 vpfe_ccdc_config_black_compense(struct vpfe_ccdc *ccdc, 489 struct vpfe_ccdc_black_compensation *bcomp) 490 { 491 u32 val; 492 493 val = ((bcomp->b & VPFE_BLK_COMP_MASK) | 494 ((bcomp->gb & VPFE_BLK_COMP_MASK) << 495 VPFE_BLK_COMP_GB_COMP_SHIFT) | 496 ((bcomp->gr & VPFE_BLK_COMP_MASK) << 497 VPFE_BLK_COMP_GR_COMP_SHIFT) | 498 ((bcomp->r & VPFE_BLK_COMP_MASK) << 499 VPFE_BLK_COMP_R_COMP_SHIFT)); 500 vpfe_reg_write(ccdc, val, VPFE_BLKCMP); 501 } 502 503 /* 504 * vpfe_ccdc_config_raw() 505 * This function will configure CCDC for Raw capture mode 506 */ 507 static void vpfe_ccdc_config_raw(struct vpfe_ccdc *ccdc) 508 { 509 struct vpfe_device *vpfe = to_vpfe(ccdc); 510 struct vpfe_ccdc_config_params_raw *config_params = 511 &ccdc->ccdc_cfg.bayer.config_params; 512 struct ccdc_params_raw *params = &ccdc->ccdc_cfg.bayer; 513 unsigned int syn_mode; 514 unsigned int val; 515 516 /* Reset CCDC */ 517 vpfe_ccdc_restore_defaults(ccdc); 518 519 /* Disable latching function registers on VSYNC */ 520 vpfe_reg_write(ccdc, VPFE_LATCH_ON_VSYNC_DISABLE, VPFE_CCDCFG); 521 522 /* 523 * Configure the vertical sync polarity(SYN_MODE.VDPOL), 524 * horizontal sync polarity (SYN_MODE.HDPOL), frame id polarity 525 * (SYN_MODE.FLDPOL), frame format(progressive or interlace), 526 * data size(SYNMODE.DATSIZ), &pixel format (Input mode), output 527 * SDRAM, enable internal timing generator 528 */ 529 syn_mode = (((params->vd_pol & VPFE_VD_POL_MASK) << VPFE_VD_POL_SHIFT) | 530 ((params->hd_pol & VPFE_HD_POL_MASK) << VPFE_HD_POL_SHIFT) | 531 ((params->fid_pol & VPFE_FID_POL_MASK) << 532 VPFE_FID_POL_SHIFT) | ((params->frm_fmt & 533 VPFE_FRM_FMT_MASK) << VPFE_FRM_FMT_SHIFT) | 534 ((config_params->data_sz & VPFE_DATA_SZ_MASK) << 535 VPFE_DATA_SZ_SHIFT) | ((params->pix_fmt & 536 VPFE_PIX_FMT_MASK) << VPFE_PIX_FMT_SHIFT) | 537 VPFE_WEN_ENABLE | VPFE_VDHDEN_ENABLE); 538 539 /* Enable and configure aLaw register if needed */ 540 if (config_params->alaw.enable) { 541 val = ((config_params->alaw.gamma_wd & 542 VPFE_ALAW_GAMMA_WD_MASK) | VPFE_ALAW_ENABLE); 543 vpfe_reg_write(ccdc, val, VPFE_ALAW); 544 vpfe_dbg(3, vpfe, "\nWriting 0x%x to ALAW...\n", val); 545 } 546 547 /* Configure video window */ 548 vpfe_ccdc_setwin(ccdc, ¶ms->win, params->frm_fmt, 549 params->bytesperpixel); 550 551 /* Configure Black Clamp */ 552 vpfe_ccdc_config_black_clamp(ccdc, &config_params->blk_clamp); 553 554 /* Configure Black level compensation */ 555 vpfe_ccdc_config_black_compense(ccdc, &config_params->blk_comp); 556 557 /* If data size is 8 bit then pack the data */ 558 if ((config_params->data_sz == VPFE_CCDC_DATA_8BITS) || 559 config_params->alaw.enable) 560 syn_mode |= VPFE_DATA_PACK_ENABLE; 561 562 /* 563 * Configure Horizontal offset register. If pack 8 is enabled then 564 * 1 pixel will take 1 byte 565 */ 566 vpfe_reg_write(ccdc, params->bytesperline, VPFE_HSIZE_OFF); 567 568 vpfe_dbg(3, vpfe, "Writing %d (%x) to HSIZE_OFF\n", 569 params->bytesperline, params->bytesperline); 570 571 /* Set value for SDOFST */ 572 if (params->frm_fmt == CCDC_FRMFMT_INTERLACED) { 573 if (params->image_invert_enable) { 574 /* For interlace inverse mode */ 575 vpfe_reg_write(ccdc, VPFE_INTERLACED_IMAGE_INVERT, 576 VPFE_SDOFST); 577 } else { 578 /* For interlace non inverse mode */ 579 vpfe_reg_write(ccdc, VPFE_INTERLACED_NO_IMAGE_INVERT, 580 VPFE_SDOFST); 581 } 582 } else if (params->frm_fmt == CCDC_FRMFMT_PROGRESSIVE) { 583 vpfe_reg_write(ccdc, VPFE_PROGRESSIVE_NO_IMAGE_INVERT, 584 VPFE_SDOFST); 585 } 586 587 vpfe_reg_write(ccdc, syn_mode, VPFE_SYNMODE); 588 589 vpfe_reg_dump(ccdc); 590 } 591 592 static inline int 593 vpfe_ccdc_set_buftype(struct vpfe_ccdc *ccdc, 594 enum ccdc_buftype buf_type) 595 { 596 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) 597 ccdc->ccdc_cfg.bayer.buf_type = buf_type; 598 else 599 ccdc->ccdc_cfg.ycbcr.buf_type = buf_type; 600 601 return 0; 602 } 603 604 static inline enum ccdc_buftype vpfe_ccdc_get_buftype(struct vpfe_ccdc *ccdc) 605 { 606 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) 607 return ccdc->ccdc_cfg.bayer.buf_type; 608 609 return ccdc->ccdc_cfg.ycbcr.buf_type; 610 } 611 612 static int vpfe_ccdc_set_pixel_format(struct vpfe_ccdc *ccdc, u32 pixfmt) 613 { 614 struct vpfe_device *vpfe = to_vpfe(ccdc); 615 616 vpfe_dbg(1, vpfe, "%s: if_type: %d, pixfmt:%s\n", 617 __func__, ccdc->ccdc_cfg.if_type, print_fourcc(pixfmt)); 618 619 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) { 620 ccdc->ccdc_cfg.bayer.pix_fmt = CCDC_PIXFMT_RAW; 621 /* 622 * Need to clear it in case it was left on 623 * after the last capture. 624 */ 625 ccdc->ccdc_cfg.bayer.config_params.alaw.enable = 0; 626 627 switch (pixfmt) { 628 case V4L2_PIX_FMT_SBGGR8: 629 ccdc->ccdc_cfg.bayer.config_params.alaw.enable = 1; 630 break; 631 632 case V4L2_PIX_FMT_YUYV: 633 case V4L2_PIX_FMT_UYVY: 634 case V4L2_PIX_FMT_YUV420: 635 case V4L2_PIX_FMT_NV12: 636 case V4L2_PIX_FMT_RGB565X: 637 break; 638 639 case V4L2_PIX_FMT_SBGGR16: 640 default: 641 return -EINVAL; 642 } 643 } else { 644 switch (pixfmt) { 645 case V4L2_PIX_FMT_YUYV: 646 ccdc->ccdc_cfg.ycbcr.pix_order = CCDC_PIXORDER_YCBYCR; 647 break; 648 649 case V4L2_PIX_FMT_UYVY: 650 ccdc->ccdc_cfg.ycbcr.pix_order = CCDC_PIXORDER_CBYCRY; 651 break; 652 653 default: 654 return -EINVAL; 655 } 656 } 657 658 return 0; 659 } 660 661 static u32 vpfe_ccdc_get_pixel_format(struct vpfe_ccdc *ccdc) 662 { 663 u32 pixfmt; 664 665 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) { 666 pixfmt = V4L2_PIX_FMT_YUYV; 667 } else { 668 if (ccdc->ccdc_cfg.ycbcr.pix_order == CCDC_PIXORDER_YCBYCR) 669 pixfmt = V4L2_PIX_FMT_YUYV; 670 else 671 pixfmt = V4L2_PIX_FMT_UYVY; 672 } 673 674 return pixfmt; 675 } 676 677 static int 678 vpfe_ccdc_set_image_window(struct vpfe_ccdc *ccdc, 679 struct v4l2_rect *win, unsigned int bpp) 680 { 681 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) { 682 ccdc->ccdc_cfg.bayer.win = *win; 683 ccdc->ccdc_cfg.bayer.bytesperpixel = bpp; 684 ccdc->ccdc_cfg.bayer.bytesperline = ALIGN(win->width * bpp, 32); 685 } else { 686 ccdc->ccdc_cfg.ycbcr.win = *win; 687 ccdc->ccdc_cfg.ycbcr.bytesperpixel = bpp; 688 ccdc->ccdc_cfg.ycbcr.bytesperline = ALIGN(win->width * bpp, 32); 689 } 690 691 return 0; 692 } 693 694 static inline void 695 vpfe_ccdc_get_image_window(struct vpfe_ccdc *ccdc, 696 struct v4l2_rect *win) 697 { 698 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) 699 *win = ccdc->ccdc_cfg.bayer.win; 700 else 701 *win = ccdc->ccdc_cfg.ycbcr.win; 702 } 703 704 static inline unsigned int vpfe_ccdc_get_line_length(struct vpfe_ccdc *ccdc) 705 { 706 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) 707 return ccdc->ccdc_cfg.bayer.bytesperline; 708 709 return ccdc->ccdc_cfg.ycbcr.bytesperline; 710 } 711 712 static inline int 713 vpfe_ccdc_set_frame_format(struct vpfe_ccdc *ccdc, 714 enum ccdc_frmfmt frm_fmt) 715 { 716 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) 717 ccdc->ccdc_cfg.bayer.frm_fmt = frm_fmt; 718 else 719 ccdc->ccdc_cfg.ycbcr.frm_fmt = frm_fmt; 720 721 return 0; 722 } 723 724 static inline enum ccdc_frmfmt 725 vpfe_ccdc_get_frame_format(struct vpfe_ccdc *ccdc) 726 { 727 if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) 728 return ccdc->ccdc_cfg.bayer.frm_fmt; 729 730 return ccdc->ccdc_cfg.ycbcr.frm_fmt; 731 } 732 733 static inline int vpfe_ccdc_getfid(struct vpfe_ccdc *ccdc) 734 { 735 return (vpfe_reg_read(ccdc, VPFE_SYNMODE) >> 15) & 1; 736 } 737 738 static inline void vpfe_set_sdr_addr(struct vpfe_ccdc *ccdc, unsigned long addr) 739 { 740 vpfe_reg_write(ccdc, addr & 0xffffffe0, VPFE_SDR_ADDR); 741 } 742 743 static int vpfe_ccdc_set_hw_if_params(struct vpfe_ccdc *ccdc, 744 struct vpfe_hw_if_param *params) 745 { 746 struct vpfe_device *vpfe = to_vpfe(ccdc); 747 748 ccdc->ccdc_cfg.if_type = params->if_type; 749 750 switch (params->if_type) { 751 case VPFE_BT656: 752 case VPFE_YCBCR_SYNC_16: 753 case VPFE_YCBCR_SYNC_8: 754 case VPFE_BT656_10BIT: 755 ccdc->ccdc_cfg.ycbcr.vd_pol = params->vdpol; 756 ccdc->ccdc_cfg.ycbcr.hd_pol = params->hdpol; 757 break; 758 759 case VPFE_RAW_BAYER: 760 ccdc->ccdc_cfg.bayer.vd_pol = params->vdpol; 761 ccdc->ccdc_cfg.bayer.hd_pol = params->hdpol; 762 if (params->bus_width == 10) 763 ccdc->ccdc_cfg.bayer.config_params.data_sz = 764 VPFE_CCDC_DATA_10BITS; 765 else 766 ccdc->ccdc_cfg.bayer.config_params.data_sz = 767 VPFE_CCDC_DATA_8BITS; 768 vpfe_dbg(1, vpfe, "params.bus_width: %d\n", 769 params->bus_width); 770 vpfe_dbg(1, vpfe, "config_params.data_sz: %d\n", 771 ccdc->ccdc_cfg.bayer.config_params.data_sz); 772 break; 773 774 default: 775 return -EINVAL; 776 } 777 778 return 0; 779 } 780 781 static void vpfe_clear_intr(struct vpfe_ccdc *ccdc, int vdint) 782 { 783 unsigned int vpfe_int_status; 784 785 vpfe_int_status = vpfe_reg_read(ccdc, VPFE_IRQ_STS); 786 787 switch (vdint) { 788 /* VD0 interrupt */ 789 case VPFE_VDINT0: 790 vpfe_int_status &= ~VPFE_VDINT0; 791 vpfe_int_status |= VPFE_VDINT0; 792 break; 793 794 /* VD1 interrupt */ 795 case VPFE_VDINT1: 796 vpfe_int_status &= ~VPFE_VDINT1; 797 vpfe_int_status |= VPFE_VDINT1; 798 break; 799 800 /* VD2 interrupt */ 801 case VPFE_VDINT2: 802 vpfe_int_status &= ~VPFE_VDINT2; 803 vpfe_int_status |= VPFE_VDINT2; 804 break; 805 806 /* Clear all interrupts */ 807 default: 808 vpfe_int_status &= ~(VPFE_VDINT0 | 809 VPFE_VDINT1 | 810 VPFE_VDINT2); 811 vpfe_int_status |= (VPFE_VDINT0 | 812 VPFE_VDINT1 | 813 VPFE_VDINT2); 814 break; 815 } 816 /* Clear specific VDINT from the status register */ 817 vpfe_reg_write(ccdc, vpfe_int_status, VPFE_IRQ_STS); 818 819 vpfe_int_status = vpfe_reg_read(ccdc, VPFE_IRQ_STS); 820 821 /* Acknowledge that we are done with all interrupts */ 822 vpfe_reg_write(ccdc, 1, VPFE_IRQ_EOI); 823 } 824 825 static void vpfe_ccdc_config_defaults(struct vpfe_ccdc *ccdc) 826 { 827 ccdc->ccdc_cfg.if_type = VPFE_RAW_BAYER; 828 829 ccdc->ccdc_cfg.ycbcr.pix_fmt = CCDC_PIXFMT_YCBCR_8BIT; 830 ccdc->ccdc_cfg.ycbcr.frm_fmt = CCDC_FRMFMT_INTERLACED; 831 ccdc->ccdc_cfg.ycbcr.fid_pol = VPFE_PINPOL_POSITIVE; 832 ccdc->ccdc_cfg.ycbcr.vd_pol = VPFE_PINPOL_POSITIVE; 833 ccdc->ccdc_cfg.ycbcr.hd_pol = VPFE_PINPOL_POSITIVE; 834 ccdc->ccdc_cfg.ycbcr.pix_order = CCDC_PIXORDER_CBYCRY; 835 ccdc->ccdc_cfg.ycbcr.buf_type = CCDC_BUFTYPE_FLD_INTERLEAVED; 836 837 ccdc->ccdc_cfg.ycbcr.win.left = 0; 838 ccdc->ccdc_cfg.ycbcr.win.top = 0; 839 ccdc->ccdc_cfg.ycbcr.win.width = 720; 840 ccdc->ccdc_cfg.ycbcr.win.height = 576; 841 ccdc->ccdc_cfg.ycbcr.bt656_enable = 1; 842 843 ccdc->ccdc_cfg.bayer.pix_fmt = CCDC_PIXFMT_RAW; 844 ccdc->ccdc_cfg.bayer.frm_fmt = CCDC_FRMFMT_PROGRESSIVE; 845 ccdc->ccdc_cfg.bayer.fid_pol = VPFE_PINPOL_POSITIVE; 846 ccdc->ccdc_cfg.bayer.vd_pol = VPFE_PINPOL_POSITIVE; 847 ccdc->ccdc_cfg.bayer.hd_pol = VPFE_PINPOL_POSITIVE; 848 849 ccdc->ccdc_cfg.bayer.win.left = 0; 850 ccdc->ccdc_cfg.bayer.win.top = 0; 851 ccdc->ccdc_cfg.bayer.win.width = 800; 852 ccdc->ccdc_cfg.bayer.win.height = 600; 853 ccdc->ccdc_cfg.bayer.config_params.data_sz = VPFE_CCDC_DATA_8BITS; 854 ccdc->ccdc_cfg.bayer.config_params.alaw.gamma_wd = 855 VPFE_CCDC_GAMMA_BITS_09_0; 856 } 857 858 /* 859 * vpfe_get_ccdc_image_format - Get image parameters based on CCDC settings 860 */ 861 static int vpfe_get_ccdc_image_format(struct vpfe_device *vpfe, 862 struct v4l2_format *f) 863 { 864 struct v4l2_rect image_win; 865 enum ccdc_buftype buf_type; 866 enum ccdc_frmfmt frm_fmt; 867 868 memset(f, 0, sizeof(*f)); 869 f->type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 870 vpfe_ccdc_get_image_window(&vpfe->ccdc, &image_win); 871 f->fmt.pix.width = image_win.width; 872 f->fmt.pix.height = image_win.height; 873 f->fmt.pix.bytesperline = vpfe_ccdc_get_line_length(&vpfe->ccdc); 874 f->fmt.pix.sizeimage = f->fmt.pix.bytesperline * 875 f->fmt.pix.height; 876 buf_type = vpfe_ccdc_get_buftype(&vpfe->ccdc); 877 f->fmt.pix.pixelformat = vpfe_ccdc_get_pixel_format(&vpfe->ccdc); 878 frm_fmt = vpfe_ccdc_get_frame_format(&vpfe->ccdc); 879 880 if (frm_fmt == CCDC_FRMFMT_PROGRESSIVE) { 881 f->fmt.pix.field = V4L2_FIELD_NONE; 882 } else if (frm_fmt == CCDC_FRMFMT_INTERLACED) { 883 if (buf_type == CCDC_BUFTYPE_FLD_INTERLEAVED) { 884 f->fmt.pix.field = V4L2_FIELD_INTERLACED; 885 } else if (buf_type == CCDC_BUFTYPE_FLD_SEPARATED) { 886 f->fmt.pix.field = V4L2_FIELD_SEQ_TB; 887 } else { 888 vpfe_err(vpfe, "Invalid buf_type\n"); 889 return -EINVAL; 890 } 891 } else { 892 vpfe_err(vpfe, "Invalid frm_fmt\n"); 893 return -EINVAL; 894 } 895 return 0; 896 } 897 898 static int vpfe_config_ccdc_image_format(struct vpfe_device *vpfe) 899 { 900 enum ccdc_frmfmt frm_fmt = CCDC_FRMFMT_INTERLACED; 901 u32 bpp; 902 int ret = 0; 903 904 vpfe_dbg(1, vpfe, "pixelformat: %s\n", 905 print_fourcc(vpfe->fmt.fmt.pix.pixelformat)); 906 907 if (vpfe_ccdc_set_pixel_format(&vpfe->ccdc, 908 vpfe->fmt.fmt.pix.pixelformat) < 0) { 909 vpfe_err(vpfe, "couldn't set pix format in ccdc\n"); 910 return -EINVAL; 911 } 912 913 /* configure the image window */ 914 bpp = __get_bytesperpixel(vpfe, vpfe->current_vpfe_fmt); 915 vpfe_ccdc_set_image_window(&vpfe->ccdc, &vpfe->crop, bpp); 916 917 switch (vpfe->fmt.fmt.pix.field) { 918 case V4L2_FIELD_INTERLACED: 919 /* do nothing, since it is default */ 920 ret = vpfe_ccdc_set_buftype( 921 &vpfe->ccdc, 922 CCDC_BUFTYPE_FLD_INTERLEAVED); 923 break; 924 925 case V4L2_FIELD_NONE: 926 frm_fmt = CCDC_FRMFMT_PROGRESSIVE; 927 /* buffer type only applicable for interlaced scan */ 928 break; 929 930 case V4L2_FIELD_SEQ_TB: 931 ret = vpfe_ccdc_set_buftype( 932 &vpfe->ccdc, 933 CCDC_BUFTYPE_FLD_SEPARATED); 934 break; 935 936 default: 937 return -EINVAL; 938 } 939 940 if (ret) 941 return ret; 942 943 return vpfe_ccdc_set_frame_format(&vpfe->ccdc, frm_fmt); 944 } 945 946 /* 947 * vpfe_config_image_format() 948 * For a given standard, this functions sets up the default 949 * pix format & crop values in the vpfe device and ccdc. It first 950 * starts with defaults based values from the standard table. 951 * It then checks if sub device supports get_fmt and then override the 952 * values based on that.Sets crop values to match with scan resolution 953 * starting at 0,0. It calls vpfe_config_ccdc_image_format() set the 954 * values in ccdc 955 */ 956 static int vpfe_config_image_format(struct vpfe_device *vpfe, 957 v4l2_std_id std_id) 958 { 959 struct vpfe_fmt *fmt; 960 struct v4l2_mbus_framefmt mbus_fmt; 961 int i, ret; 962 963 for (i = 0; i < ARRAY_SIZE(vpfe_standards); i++) { 964 if (vpfe_standards[i].std_id & std_id) { 965 vpfe->std_info.active_pixels = 966 vpfe_standards[i].width; 967 vpfe->std_info.active_lines = 968 vpfe_standards[i].height; 969 vpfe->std_info.frame_format = 970 vpfe_standards[i].frame_format; 971 vpfe->std_index = i; 972 973 break; 974 } 975 } 976 977 if (i == ARRAY_SIZE(vpfe_standards)) { 978 vpfe_err(vpfe, "standard not supported\n"); 979 return -EINVAL; 980 } 981 982 ret = __subdev_get_format(vpfe, &mbus_fmt); 983 if (ret) 984 return ret; 985 986 fmt = find_format_by_code(vpfe, mbus_fmt.code); 987 if (!fmt) { 988 vpfe_dbg(3, vpfe, "mbus code format (0x%08x) not found.\n", 989 mbus_fmt.code); 990 return -EINVAL; 991 } 992 993 /* Save current subdev format */ 994 v4l2_fill_pix_format(&vpfe->fmt.fmt.pix, &mbus_fmt); 995 vpfe->fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 996 vpfe->fmt.fmt.pix.pixelformat = fmt->fourcc; 997 vpfe_calc_format_size(vpfe, fmt, &vpfe->fmt); 998 vpfe->current_vpfe_fmt = fmt; 999 1000 /* Update the crop window based on found values */ 1001 vpfe->crop.top = 0; 1002 vpfe->crop.left = 0; 1003 vpfe->crop.width = mbus_fmt.width; 1004 vpfe->crop.height = mbus_fmt.height; 1005 1006 return vpfe_config_ccdc_image_format(vpfe); 1007 } 1008 1009 static int vpfe_initialize_device(struct vpfe_device *vpfe) 1010 { 1011 struct vpfe_subdev_info *sdinfo; 1012 int ret; 1013 1014 sdinfo = &vpfe->cfg->sub_devs[0]; 1015 sdinfo->sd = vpfe->sd[0]; 1016 vpfe->current_input = 0; 1017 vpfe->std_index = 0; 1018 /* Configure the default format information */ 1019 ret = vpfe_config_image_format(vpfe, 1020 vpfe_standards[vpfe->std_index].std_id); 1021 if (ret) 1022 return ret; 1023 1024 ret = pm_runtime_resume_and_get(vpfe->pdev); 1025 if (ret < 0) 1026 return ret; 1027 1028 vpfe_config_enable(&vpfe->ccdc, 1); 1029 1030 vpfe_ccdc_restore_defaults(&vpfe->ccdc); 1031 1032 /* Clear all VPFE interrupts */ 1033 vpfe_clear_intr(&vpfe->ccdc, -1); 1034 1035 return ret; 1036 } 1037 1038 /* 1039 * vpfe_release : This function is based on the vb2_fop_release 1040 * helper function. 1041 * It has been augmented to handle module power management, 1042 * by disabling/enabling h/w module fcntl clock when necessary. 1043 */ 1044 static int vpfe_release(struct file *file) 1045 { 1046 struct vpfe_device *vpfe = video_drvdata(file); 1047 bool fh_singular; 1048 int ret; 1049 1050 mutex_lock(&vpfe->lock); 1051 1052 /* Save the singular status before we call the clean-up helper */ 1053 fh_singular = v4l2_fh_is_singular_file(file); 1054 1055 /* the release helper will cleanup any on-going streaming */ 1056 ret = _vb2_fop_release(file, NULL); 1057 1058 /* 1059 * If this was the last open file. 1060 * Then de-initialize hw module. 1061 */ 1062 if (fh_singular) 1063 vpfe_ccdc_close(&vpfe->ccdc, vpfe->pdev); 1064 1065 mutex_unlock(&vpfe->lock); 1066 1067 return ret; 1068 } 1069 1070 /* 1071 * vpfe_open : This function is based on the v4l2_fh_open helper function. 1072 * It has been augmented to handle module power management, 1073 * by disabling/enabling h/w module fcntl clock when necessary. 1074 */ 1075 static int vpfe_open(struct file *file) 1076 { 1077 struct vpfe_device *vpfe = video_drvdata(file); 1078 int ret; 1079 1080 mutex_lock(&vpfe->lock); 1081 1082 ret = v4l2_fh_open(file); 1083 if (ret) { 1084 vpfe_err(vpfe, "v4l2_fh_open failed\n"); 1085 goto unlock; 1086 } 1087 1088 if (!v4l2_fh_is_singular_file(file)) 1089 goto unlock; 1090 1091 if (vpfe_initialize_device(vpfe)) { 1092 v4l2_fh_release(file); 1093 ret = -ENODEV; 1094 } 1095 1096 unlock: 1097 mutex_unlock(&vpfe->lock); 1098 return ret; 1099 } 1100 1101 /** 1102 * vpfe_schedule_next_buffer: set next buffer address for capture 1103 * @vpfe : ptr to vpfe device 1104 * 1105 * This function will get next buffer from the dma queue and 1106 * set the buffer address in the vpfe register for capture. 1107 * the buffer is marked active 1108 */ 1109 static void vpfe_schedule_next_buffer(struct vpfe_device *vpfe) 1110 { 1111 dma_addr_t addr; 1112 1113 spin_lock(&vpfe->dma_queue_lock); 1114 if (list_empty(&vpfe->dma_queue)) { 1115 spin_unlock(&vpfe->dma_queue_lock); 1116 return; 1117 } 1118 1119 vpfe->next_frm = list_entry(vpfe->dma_queue.next, 1120 struct vpfe_cap_buffer, list); 1121 list_del(&vpfe->next_frm->list); 1122 spin_unlock(&vpfe->dma_queue_lock); 1123 1124 addr = vb2_dma_contig_plane_dma_addr(&vpfe->next_frm->vb.vb2_buf, 0); 1125 vpfe_set_sdr_addr(&vpfe->ccdc, addr); 1126 } 1127 1128 static inline void vpfe_schedule_bottom_field(struct vpfe_device *vpfe) 1129 { 1130 dma_addr_t addr; 1131 1132 addr = vb2_dma_contig_plane_dma_addr(&vpfe->next_frm->vb.vb2_buf, 0) + 1133 vpfe->field_off; 1134 1135 vpfe_set_sdr_addr(&vpfe->ccdc, addr); 1136 } 1137 1138 /* 1139 * vpfe_process_buffer_complete: process a completed buffer 1140 * @vpfe : ptr to vpfe device 1141 * 1142 * This function time stamp the buffer and mark it as DONE. It also 1143 * wake up any process waiting on the QUEUE and set the next buffer 1144 * as current 1145 */ 1146 static inline void vpfe_process_buffer_complete(struct vpfe_device *vpfe) 1147 { 1148 vpfe->cur_frm->vb.vb2_buf.timestamp = ktime_get_ns(); 1149 vpfe->cur_frm->vb.field = vpfe->fmt.fmt.pix.field; 1150 vpfe->cur_frm->vb.sequence = vpfe->sequence++; 1151 vb2_buffer_done(&vpfe->cur_frm->vb.vb2_buf, VB2_BUF_STATE_DONE); 1152 vpfe->cur_frm = vpfe->next_frm; 1153 } 1154 1155 static void vpfe_handle_interlaced_irq(struct vpfe_device *vpfe, 1156 enum v4l2_field field) 1157 { 1158 int fid; 1159 1160 /* interlaced or TB capture check which field 1161 * we are in hardware 1162 */ 1163 fid = vpfe_ccdc_getfid(&vpfe->ccdc); 1164 1165 /* switch the software maintained field id */ 1166 vpfe->field ^= 1; 1167 if (fid == vpfe->field) { 1168 /* we are in-sync here,continue */ 1169 if (fid == 0) { 1170 /* 1171 * One frame is just being captured. If the 1172 * next frame is available, release the 1173 * current frame and move on 1174 */ 1175 if (vpfe->cur_frm != vpfe->next_frm) 1176 vpfe_process_buffer_complete(vpfe); 1177 1178 if (vpfe->stopping) 1179 return; 1180 1181 /* 1182 * based on whether the two fields are stored 1183 * interleave or separately in memory, 1184 * reconfigure the CCDC memory address 1185 */ 1186 if (field == V4L2_FIELD_SEQ_TB) 1187 vpfe_schedule_bottom_field(vpfe); 1188 } else { 1189 /* 1190 * if one field is just being captured configure 1191 * the next frame get the next frame from the empty 1192 * queue if no frame is available hold on to the 1193 * current buffer 1194 */ 1195 if (vpfe->cur_frm == vpfe->next_frm) 1196 vpfe_schedule_next_buffer(vpfe); 1197 } 1198 } else if (fid == 0) { 1199 /* 1200 * out of sync. Recover from any hardware out-of-sync. 1201 * May loose one frame 1202 */ 1203 vpfe->field = fid; 1204 } 1205 } 1206 1207 /* 1208 * vpfe_isr : ISR handler for vpfe capture (VINT0) 1209 * @irq: irq number 1210 * @dev_id: dev_id ptr 1211 * 1212 * It changes status of the captured buffer, takes next buffer from the queue 1213 * and sets its address in VPFE registers 1214 */ 1215 static irqreturn_t vpfe_isr(int irq, void *dev) 1216 { 1217 struct vpfe_device *vpfe = (struct vpfe_device *)dev; 1218 enum v4l2_field field = vpfe->fmt.fmt.pix.field; 1219 int intr_status, stopping = vpfe->stopping; 1220 1221 intr_status = vpfe_reg_read(&vpfe->ccdc, VPFE_IRQ_STS); 1222 1223 if (intr_status & VPFE_VDINT0) { 1224 if (field == V4L2_FIELD_NONE) { 1225 if (vpfe->cur_frm != vpfe->next_frm) 1226 vpfe_process_buffer_complete(vpfe); 1227 } else { 1228 vpfe_handle_interlaced_irq(vpfe, field); 1229 } 1230 if (stopping) { 1231 vpfe->stopping = false; 1232 complete(&vpfe->capture_stop); 1233 } 1234 } 1235 1236 if (intr_status & VPFE_VDINT1 && !stopping) { 1237 if (field == V4L2_FIELD_NONE && 1238 vpfe->cur_frm == vpfe->next_frm) 1239 vpfe_schedule_next_buffer(vpfe); 1240 } 1241 1242 vpfe_clear_intr(&vpfe->ccdc, intr_status); 1243 1244 return IRQ_HANDLED; 1245 } 1246 1247 static inline void vpfe_detach_irq(struct vpfe_device *vpfe) 1248 { 1249 unsigned int intr = VPFE_VDINT0; 1250 enum ccdc_frmfmt frame_format; 1251 1252 frame_format = vpfe_ccdc_get_frame_format(&vpfe->ccdc); 1253 if (frame_format == CCDC_FRMFMT_PROGRESSIVE) 1254 intr |= VPFE_VDINT1; 1255 1256 vpfe_reg_write(&vpfe->ccdc, intr, VPFE_IRQ_EN_CLR); 1257 } 1258 1259 static inline void vpfe_attach_irq(struct vpfe_device *vpfe) 1260 { 1261 unsigned int intr = VPFE_VDINT0; 1262 enum ccdc_frmfmt frame_format; 1263 1264 frame_format = vpfe_ccdc_get_frame_format(&vpfe->ccdc); 1265 if (frame_format == CCDC_FRMFMT_PROGRESSIVE) 1266 intr |= VPFE_VDINT1; 1267 1268 vpfe_reg_write(&vpfe->ccdc, intr, VPFE_IRQ_EN_SET); 1269 } 1270 1271 static int vpfe_querycap(struct file *file, void *priv, 1272 struct v4l2_capability *cap) 1273 { 1274 struct vpfe_device *vpfe = video_drvdata(file); 1275 1276 strscpy(cap->driver, VPFE_MODULE_NAME, sizeof(cap->driver)); 1277 strscpy(cap->card, "TI AM437x VPFE", sizeof(cap->card)); 1278 snprintf(cap->bus_info, sizeof(cap->bus_info), 1279 "platform:%s", vpfe->v4l2_dev.name); 1280 return 0; 1281 } 1282 1283 /* get the format set at output pad of the adjacent subdev */ 1284 static int __subdev_get_format(struct vpfe_device *vpfe, 1285 struct v4l2_mbus_framefmt *fmt) 1286 { 1287 struct v4l2_subdev *sd = vpfe->current_subdev->sd; 1288 struct v4l2_subdev_format sd_fmt; 1289 struct v4l2_mbus_framefmt *mbus_fmt = &sd_fmt.format; 1290 int ret; 1291 1292 sd_fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE; 1293 sd_fmt.pad = 0; 1294 1295 ret = v4l2_subdev_call(sd, pad, get_fmt, NULL, &sd_fmt); 1296 if (ret) 1297 return ret; 1298 1299 *fmt = *mbus_fmt; 1300 1301 vpfe_dbg(1, vpfe, "%s: %dx%d code:%04X\n", __func__, 1302 fmt->width, fmt->height, fmt->code); 1303 1304 return 0; 1305 } 1306 1307 /* set the format at output pad of the adjacent subdev */ 1308 static int __subdev_set_format(struct vpfe_device *vpfe, 1309 struct v4l2_mbus_framefmt *fmt) 1310 { 1311 struct v4l2_subdev *sd = vpfe->current_subdev->sd; 1312 struct v4l2_subdev_format sd_fmt; 1313 struct v4l2_mbus_framefmt *mbus_fmt = &sd_fmt.format; 1314 int ret; 1315 1316 sd_fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE; 1317 sd_fmt.pad = 0; 1318 *mbus_fmt = *fmt; 1319 1320 ret = v4l2_subdev_call(sd, pad, set_fmt, NULL, &sd_fmt); 1321 if (ret) 1322 return ret; 1323 1324 vpfe_dbg(1, vpfe, "%s %dx%d code:%04X\n", __func__, 1325 fmt->width, fmt->height, fmt->code); 1326 1327 return 0; 1328 } 1329 1330 static int vpfe_calc_format_size(struct vpfe_device *vpfe, 1331 const struct vpfe_fmt *fmt, 1332 struct v4l2_format *f) 1333 { 1334 u32 bpp; 1335 1336 if (!fmt) { 1337 vpfe_dbg(3, vpfe, "No vpfe_fmt provided!\n"); 1338 return -EINVAL; 1339 } 1340 1341 bpp = __get_bytesperpixel(vpfe, fmt); 1342 1343 /* pitch should be 32 bytes aligned */ 1344 f->fmt.pix.bytesperline = ALIGN(f->fmt.pix.width * bpp, 32); 1345 f->fmt.pix.sizeimage = f->fmt.pix.bytesperline * 1346 f->fmt.pix.height; 1347 1348 vpfe_dbg(3, vpfe, "%s: fourcc: %s size: %dx%d bpl:%d img_size:%d\n", 1349 __func__, print_fourcc(f->fmt.pix.pixelformat), 1350 f->fmt.pix.width, f->fmt.pix.height, 1351 f->fmt.pix.bytesperline, f->fmt.pix.sizeimage); 1352 1353 return 0; 1354 } 1355 1356 static int vpfe_g_fmt(struct file *file, void *priv, 1357 struct v4l2_format *fmt) 1358 { 1359 struct vpfe_device *vpfe = video_drvdata(file); 1360 1361 *fmt = vpfe->fmt; 1362 1363 return 0; 1364 } 1365 1366 static int vpfe_enum_fmt(struct file *file, void *priv, 1367 struct v4l2_fmtdesc *f) 1368 { 1369 struct vpfe_device *vpfe = video_drvdata(file); 1370 struct vpfe_subdev_info *sdinfo; 1371 struct vpfe_fmt *fmt; 1372 1373 sdinfo = vpfe->current_subdev; 1374 if (!sdinfo->sd) 1375 return -EINVAL; 1376 1377 if (f->index >= vpfe->num_active_fmt) 1378 return -EINVAL; 1379 1380 fmt = vpfe->active_fmt[f->index]; 1381 1382 f->pixelformat = fmt->fourcc; 1383 1384 vpfe_dbg(1, vpfe, "%s: mbus index: %d code: %x pixelformat: %s\n", 1385 __func__, f->index, fmt->code, print_fourcc(fmt->fourcc)); 1386 1387 return 0; 1388 } 1389 1390 static int vpfe_try_fmt(struct file *file, void *priv, 1391 struct v4l2_format *f) 1392 { 1393 struct vpfe_device *vpfe = video_drvdata(file); 1394 struct v4l2_subdev *sd = vpfe->current_subdev->sd; 1395 const struct vpfe_fmt *fmt; 1396 struct v4l2_subdev_frame_size_enum fse; 1397 int ret, found; 1398 1399 fmt = find_format_by_pix(vpfe, f->fmt.pix.pixelformat); 1400 if (!fmt) { 1401 /* default to first entry */ 1402 vpfe_dbg(3, vpfe, "Invalid pixel code: %x, default used instead\n", 1403 f->fmt.pix.pixelformat); 1404 fmt = vpfe->active_fmt[0]; 1405 f->fmt.pix.pixelformat = fmt->fourcc; 1406 } 1407 1408 f->fmt.pix.field = vpfe->fmt.fmt.pix.field; 1409 1410 /* check for/find a valid width/height */ 1411 ret = 0; 1412 found = false; 1413 fse.pad = 0; 1414 fse.code = fmt->code; 1415 fse.which = V4L2_SUBDEV_FORMAT_ACTIVE; 1416 for (fse.index = 0; ; fse.index++) { 1417 ret = v4l2_subdev_call(sd, pad, enum_frame_size, 1418 NULL, &fse); 1419 if (ret) 1420 break; 1421 1422 if (f->fmt.pix.width == fse.max_width && 1423 f->fmt.pix.height == fse.max_height) { 1424 found = true; 1425 break; 1426 } else if (f->fmt.pix.width >= fse.min_width && 1427 f->fmt.pix.width <= fse.max_width && 1428 f->fmt.pix.height >= fse.min_height && 1429 f->fmt.pix.height <= fse.max_height) { 1430 found = true; 1431 break; 1432 } 1433 } 1434 1435 if (!found) { 1436 /* use existing values as default */ 1437 f->fmt.pix.width = vpfe->fmt.fmt.pix.width; 1438 f->fmt.pix.height = vpfe->fmt.fmt.pix.height; 1439 } 1440 1441 /* 1442 * Use current colorspace for now, it will get 1443 * updated properly during s_fmt 1444 */ 1445 f->fmt.pix.colorspace = vpfe->fmt.fmt.pix.colorspace; 1446 return vpfe_calc_format_size(vpfe, fmt, f); 1447 } 1448 1449 static int vpfe_s_fmt(struct file *file, void *priv, 1450 struct v4l2_format *fmt) 1451 { 1452 struct vpfe_device *vpfe = video_drvdata(file); 1453 struct vpfe_fmt *f; 1454 struct v4l2_mbus_framefmt mbus_fmt; 1455 int ret; 1456 1457 /* If streaming is started, return error */ 1458 if (vb2_is_busy(&vpfe->buffer_queue)) { 1459 vpfe_err(vpfe, "%s device busy\n", __func__); 1460 return -EBUSY; 1461 } 1462 1463 ret = vpfe_try_fmt(file, priv, fmt); 1464 if (ret < 0) 1465 return ret; 1466 1467 f = find_format_by_pix(vpfe, fmt->fmt.pix.pixelformat); 1468 1469 v4l2_fill_mbus_format(&mbus_fmt, &fmt->fmt.pix, f->code); 1470 1471 ret = __subdev_set_format(vpfe, &mbus_fmt); 1472 if (ret) 1473 return ret; 1474 1475 /* Just double check nothing has gone wrong */ 1476 if (mbus_fmt.code != f->code) { 1477 vpfe_dbg(3, vpfe, 1478 "%s subdev changed format on us, this should not happen\n", 1479 __func__); 1480 return -EINVAL; 1481 } 1482 1483 v4l2_fill_pix_format(&vpfe->fmt.fmt.pix, &mbus_fmt); 1484 vpfe->fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 1485 vpfe->fmt.fmt.pix.pixelformat = f->fourcc; 1486 vpfe_calc_format_size(vpfe, f, &vpfe->fmt); 1487 *fmt = vpfe->fmt; 1488 vpfe->current_vpfe_fmt = f; 1489 1490 /* Update the crop window based on found values */ 1491 vpfe->crop.width = fmt->fmt.pix.width; 1492 vpfe->crop.height = fmt->fmt.pix.height; 1493 1494 /* set image capture parameters in the ccdc */ 1495 return vpfe_config_ccdc_image_format(vpfe); 1496 } 1497 1498 static int vpfe_enum_size(struct file *file, void *priv, 1499 struct v4l2_frmsizeenum *fsize) 1500 { 1501 struct vpfe_device *vpfe = video_drvdata(file); 1502 struct v4l2_subdev_frame_size_enum fse; 1503 struct v4l2_subdev *sd = vpfe->current_subdev->sd; 1504 struct vpfe_fmt *fmt; 1505 int ret; 1506 1507 /* check for valid format */ 1508 fmt = find_format_by_pix(vpfe, fsize->pixel_format); 1509 if (!fmt) { 1510 vpfe_dbg(3, vpfe, "Invalid pixel code: %x\n", 1511 fsize->pixel_format); 1512 return -EINVAL; 1513 } 1514 1515 memset(fsize->reserved, 0x0, sizeof(fsize->reserved)); 1516 1517 memset(&fse, 0x0, sizeof(fse)); 1518 fse.index = fsize->index; 1519 fse.pad = 0; 1520 fse.code = fmt->code; 1521 fse.which = V4L2_SUBDEV_FORMAT_ACTIVE; 1522 ret = v4l2_subdev_call(sd, pad, enum_frame_size, NULL, &fse); 1523 if (ret) 1524 return ret; 1525 1526 vpfe_dbg(1, vpfe, "%s: index: %d code: %x W:[%d,%d] H:[%d,%d]\n", 1527 __func__, fse.index, fse.code, fse.min_width, fse.max_width, 1528 fse.min_height, fse.max_height); 1529 1530 fsize->type = V4L2_FRMSIZE_TYPE_DISCRETE; 1531 fsize->discrete.width = fse.max_width; 1532 fsize->discrete.height = fse.max_height; 1533 1534 vpfe_dbg(1, vpfe, "%s: index: %d pixformat: %s size: %dx%d\n", 1535 __func__, fsize->index, print_fourcc(fsize->pixel_format), 1536 fsize->discrete.width, fsize->discrete.height); 1537 1538 return 0; 1539 } 1540 1541 /* 1542 * vpfe_get_subdev_input_index - Get subdev index and subdev input index for a 1543 * given app input index 1544 */ 1545 static int 1546 vpfe_get_subdev_input_index(struct vpfe_device *vpfe, 1547 int *subdev_index, 1548 int *subdev_input_index, 1549 int app_input_index) 1550 { 1551 int i, j = 0; 1552 1553 for (i = 0; i < ARRAY_SIZE(vpfe->cfg->asd); i++) { 1554 if (app_input_index < (j + 1)) { 1555 *subdev_index = i; 1556 *subdev_input_index = app_input_index - j; 1557 return 0; 1558 } 1559 j++; 1560 } 1561 return -EINVAL; 1562 } 1563 1564 /* 1565 * vpfe_get_app_input - Get app input index for a given subdev input index 1566 * driver stores the input index of the current sub device and translate it 1567 * when application request the current input 1568 */ 1569 static int vpfe_get_app_input_index(struct vpfe_device *vpfe, 1570 int *app_input_index) 1571 { 1572 struct vpfe_config *cfg = vpfe->cfg; 1573 struct vpfe_subdev_info *sdinfo; 1574 struct i2c_client *client; 1575 struct i2c_client *curr_client; 1576 int i, j = 0; 1577 1578 curr_client = v4l2_get_subdevdata(vpfe->current_subdev->sd); 1579 for (i = 0; i < ARRAY_SIZE(vpfe->cfg->asd); i++) { 1580 sdinfo = &cfg->sub_devs[i]; 1581 client = v4l2_get_subdevdata(sdinfo->sd); 1582 if (client->addr == curr_client->addr && 1583 client->adapter->nr == curr_client->adapter->nr) { 1584 if (vpfe->current_input >= 1) 1585 return -1; 1586 *app_input_index = j + vpfe->current_input; 1587 return 0; 1588 } 1589 j++; 1590 } 1591 return -EINVAL; 1592 } 1593 1594 static int vpfe_enum_input(struct file *file, void *priv, 1595 struct v4l2_input *inp) 1596 { 1597 struct vpfe_device *vpfe = video_drvdata(file); 1598 struct vpfe_subdev_info *sdinfo; 1599 int subdev, index; 1600 1601 if (vpfe_get_subdev_input_index(vpfe, &subdev, &index, 1602 inp->index) < 0) { 1603 vpfe_dbg(1, vpfe, 1604 "input information not found for the subdev\n"); 1605 return -EINVAL; 1606 } 1607 sdinfo = &vpfe->cfg->sub_devs[subdev]; 1608 *inp = sdinfo->inputs[index]; 1609 1610 return 0; 1611 } 1612 1613 static int vpfe_g_input(struct file *file, void *priv, unsigned int *index) 1614 { 1615 struct vpfe_device *vpfe = video_drvdata(file); 1616 1617 return vpfe_get_app_input_index(vpfe, index); 1618 } 1619 1620 /* Assumes caller is holding vpfe_dev->lock */ 1621 static int vpfe_set_input(struct vpfe_device *vpfe, unsigned int index) 1622 { 1623 int subdev_index = 0, inp_index = 0; 1624 struct vpfe_subdev_info *sdinfo; 1625 struct vpfe_route *route; 1626 u32 input, output; 1627 int ret; 1628 1629 /* If streaming is started, return error */ 1630 if (vb2_is_busy(&vpfe->buffer_queue)) { 1631 vpfe_err(vpfe, "%s device busy\n", __func__); 1632 return -EBUSY; 1633 } 1634 ret = vpfe_get_subdev_input_index(vpfe, 1635 &subdev_index, 1636 &inp_index, 1637 index); 1638 if (ret < 0) { 1639 vpfe_err(vpfe, "invalid input index: %d\n", index); 1640 goto get_out; 1641 } 1642 1643 sdinfo = &vpfe->cfg->sub_devs[subdev_index]; 1644 sdinfo->sd = vpfe->sd[subdev_index]; 1645 route = &sdinfo->routes[inp_index]; 1646 if (route && sdinfo->can_route) { 1647 input = route->input; 1648 output = route->output; 1649 if (sdinfo->sd) { 1650 ret = v4l2_subdev_call(sdinfo->sd, video, 1651 s_routing, input, output, 0); 1652 if (ret) { 1653 vpfe_err(vpfe, "s_routing failed\n"); 1654 ret = -EINVAL; 1655 goto get_out; 1656 } 1657 } 1658 1659 } 1660 1661 vpfe->current_subdev = sdinfo; 1662 if (sdinfo->sd) 1663 vpfe->v4l2_dev.ctrl_handler = sdinfo->sd->ctrl_handler; 1664 vpfe->current_input = index; 1665 vpfe->std_index = 0; 1666 1667 /* set the bus/interface parameter for the sub device in ccdc */ 1668 ret = vpfe_ccdc_set_hw_if_params(&vpfe->ccdc, &sdinfo->vpfe_param); 1669 if (ret) 1670 return ret; 1671 1672 /* set the default image parameters in the device */ 1673 return vpfe_config_image_format(vpfe, 1674 vpfe_standards[vpfe->std_index].std_id); 1675 1676 get_out: 1677 return ret; 1678 } 1679 1680 static int vpfe_s_input(struct file *file, void *priv, unsigned int index) 1681 { 1682 struct vpfe_device *vpfe = video_drvdata(file); 1683 1684 return vpfe_set_input(vpfe, index); 1685 } 1686 1687 static int vpfe_querystd(struct file *file, void *priv, v4l2_std_id *std_id) 1688 { 1689 struct vpfe_device *vpfe = video_drvdata(file); 1690 struct vpfe_subdev_info *sdinfo; 1691 1692 sdinfo = vpfe->current_subdev; 1693 if (!(sdinfo->inputs[0].capabilities & V4L2_IN_CAP_STD)) 1694 return -ENODATA; 1695 1696 /* Call querystd function of decoder device */ 1697 return v4l2_device_call_until_err(&vpfe->v4l2_dev, sdinfo->grp_id, 1698 video, querystd, std_id); 1699 } 1700 1701 static int vpfe_s_std(struct file *file, void *priv, v4l2_std_id std_id) 1702 { 1703 struct vpfe_device *vpfe = video_drvdata(file); 1704 struct vpfe_subdev_info *sdinfo; 1705 int ret; 1706 1707 sdinfo = vpfe->current_subdev; 1708 if (!(sdinfo->inputs[0].capabilities & V4L2_IN_CAP_STD)) 1709 return -ENODATA; 1710 1711 /* if trying to set the same std then nothing to do */ 1712 if (vpfe_standards[vpfe->std_index].std_id == std_id) 1713 return 0; 1714 1715 /* If streaming is started, return error */ 1716 if (vb2_is_busy(&vpfe->buffer_queue)) { 1717 vpfe_err(vpfe, "%s device busy\n", __func__); 1718 ret = -EBUSY; 1719 return ret; 1720 } 1721 1722 ret = v4l2_device_call_until_err(&vpfe->v4l2_dev, sdinfo->grp_id, 1723 video, s_std, std_id); 1724 if (ret < 0) { 1725 vpfe_err(vpfe, "Failed to set standard\n"); 1726 return ret; 1727 } 1728 ret = vpfe_config_image_format(vpfe, std_id); 1729 1730 return ret; 1731 } 1732 1733 static int vpfe_g_std(struct file *file, void *priv, v4l2_std_id *std_id) 1734 { 1735 struct vpfe_device *vpfe = video_drvdata(file); 1736 struct vpfe_subdev_info *sdinfo; 1737 1738 sdinfo = vpfe->current_subdev; 1739 if (sdinfo->inputs[0].capabilities != V4L2_IN_CAP_STD) 1740 return -ENODATA; 1741 1742 *std_id = vpfe_standards[vpfe->std_index].std_id; 1743 1744 return 0; 1745 } 1746 1747 /* 1748 * vpfe_calculate_offsets : This function calculates buffers offset 1749 * for top and bottom field 1750 */ 1751 static void vpfe_calculate_offsets(struct vpfe_device *vpfe) 1752 { 1753 struct v4l2_rect image_win; 1754 1755 vpfe_ccdc_get_image_window(&vpfe->ccdc, &image_win); 1756 vpfe->field_off = image_win.height * image_win.width; 1757 } 1758 1759 /* 1760 * vpfe_queue_setup - Callback function for buffer setup. 1761 * @vq: vb2_queue ptr 1762 * @nbuffers: ptr to number of buffers requested by application 1763 * @nplanes:: contains number of distinct video planes needed to hold a frame 1764 * @sizes[]: contains the size (in bytes) of each plane. 1765 * @alloc_devs: ptr to allocation context 1766 * 1767 * This callback function is called when reqbuf() is called to adjust 1768 * the buffer count and buffer size 1769 */ 1770 static int vpfe_queue_setup(struct vb2_queue *vq, 1771 unsigned int *nbuffers, unsigned int *nplanes, 1772 unsigned int sizes[], struct device *alloc_devs[]) 1773 { 1774 struct vpfe_device *vpfe = vb2_get_drv_priv(vq); 1775 unsigned size = vpfe->fmt.fmt.pix.sizeimage; 1776 1777 if (vq->num_buffers + *nbuffers < 3) 1778 *nbuffers = 3 - vq->num_buffers; 1779 1780 if (*nplanes) { 1781 if (sizes[0] < size) 1782 return -EINVAL; 1783 size = sizes[0]; 1784 } 1785 1786 *nplanes = 1; 1787 sizes[0] = size; 1788 1789 vpfe_dbg(1, vpfe, 1790 "nbuffers=%d, size=%u\n", *nbuffers, sizes[0]); 1791 1792 /* Calculate field offset */ 1793 vpfe_calculate_offsets(vpfe); 1794 1795 return 0; 1796 } 1797 1798 /* 1799 * vpfe_buffer_prepare : callback function for buffer prepare 1800 * @vb: ptr to vb2_buffer 1801 * 1802 * This is the callback function for buffer prepare when vb2_qbuf() 1803 * function is called. The buffer is prepared and user space virtual address 1804 * or user address is converted into physical address 1805 */ 1806 static int vpfe_buffer_prepare(struct vb2_buffer *vb) 1807 { 1808 struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); 1809 struct vpfe_device *vpfe = vb2_get_drv_priv(vb->vb2_queue); 1810 1811 vb2_set_plane_payload(vb, 0, vpfe->fmt.fmt.pix.sizeimage); 1812 1813 if (vb2_get_plane_payload(vb, 0) > vb2_plane_size(vb, 0)) 1814 return -EINVAL; 1815 1816 vbuf->field = vpfe->fmt.fmt.pix.field; 1817 1818 return 0; 1819 } 1820 1821 /* 1822 * vpfe_buffer_queue : Callback function to add buffer to DMA queue 1823 * @vb: ptr to vb2_buffer 1824 */ 1825 static void vpfe_buffer_queue(struct vb2_buffer *vb) 1826 { 1827 struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); 1828 struct vpfe_device *vpfe = vb2_get_drv_priv(vb->vb2_queue); 1829 struct vpfe_cap_buffer *buf = to_vpfe_buffer(vbuf); 1830 unsigned long flags = 0; 1831 1832 /* add the buffer to the DMA queue */ 1833 spin_lock_irqsave(&vpfe->dma_queue_lock, flags); 1834 list_add_tail(&buf->list, &vpfe->dma_queue); 1835 spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags); 1836 } 1837 1838 static void vpfe_return_all_buffers(struct vpfe_device *vpfe, 1839 enum vb2_buffer_state state) 1840 { 1841 struct vpfe_cap_buffer *buf, *node; 1842 unsigned long flags; 1843 1844 spin_lock_irqsave(&vpfe->dma_queue_lock, flags); 1845 list_for_each_entry_safe(buf, node, &vpfe->dma_queue, list) { 1846 vb2_buffer_done(&buf->vb.vb2_buf, state); 1847 list_del(&buf->list); 1848 } 1849 1850 if (vpfe->cur_frm) 1851 vb2_buffer_done(&vpfe->cur_frm->vb.vb2_buf, state); 1852 1853 if (vpfe->next_frm && vpfe->next_frm != vpfe->cur_frm) 1854 vb2_buffer_done(&vpfe->next_frm->vb.vb2_buf, state); 1855 1856 vpfe->cur_frm = NULL; 1857 vpfe->next_frm = NULL; 1858 spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags); 1859 } 1860 1861 /* 1862 * vpfe_start_streaming : Starts the DMA engine for streaming 1863 * @vb: ptr to vb2_buffer 1864 * @count: number of buffers 1865 */ 1866 static int vpfe_start_streaming(struct vb2_queue *vq, unsigned int count) 1867 { 1868 struct vpfe_device *vpfe = vb2_get_drv_priv(vq); 1869 struct vpfe_subdev_info *sdinfo; 1870 unsigned long flags; 1871 unsigned long addr; 1872 int ret; 1873 1874 spin_lock_irqsave(&vpfe->dma_queue_lock, flags); 1875 1876 vpfe->field = 0; 1877 vpfe->sequence = 0; 1878 1879 sdinfo = vpfe->current_subdev; 1880 1881 vpfe_attach_irq(vpfe); 1882 1883 vpfe->stopping = false; 1884 init_completion(&vpfe->capture_stop); 1885 1886 if (vpfe->ccdc.ccdc_cfg.if_type == VPFE_RAW_BAYER) 1887 vpfe_ccdc_config_raw(&vpfe->ccdc); 1888 else 1889 vpfe_ccdc_config_ycbcr(&vpfe->ccdc); 1890 1891 /* Get the next frame from the buffer queue */ 1892 vpfe->next_frm = list_entry(vpfe->dma_queue.next, 1893 struct vpfe_cap_buffer, list); 1894 vpfe->cur_frm = vpfe->next_frm; 1895 /* Remove buffer from the buffer queue */ 1896 list_del(&vpfe->cur_frm->list); 1897 spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags); 1898 1899 addr = vb2_dma_contig_plane_dma_addr(&vpfe->cur_frm->vb.vb2_buf, 0); 1900 1901 vpfe_set_sdr_addr(&vpfe->ccdc, (unsigned long)(addr)); 1902 1903 vpfe_pcr_enable(&vpfe->ccdc, 1); 1904 1905 ret = v4l2_subdev_call(sdinfo->sd, video, s_stream, 1); 1906 if (ret < 0) { 1907 vpfe_err(vpfe, "Error in attaching interrupt handle\n"); 1908 goto err; 1909 } 1910 1911 return 0; 1912 1913 err: 1914 vpfe_return_all_buffers(vpfe, VB2_BUF_STATE_QUEUED); 1915 vpfe_pcr_enable(&vpfe->ccdc, 0); 1916 return ret; 1917 } 1918 1919 /* 1920 * vpfe_stop_streaming : Stop the DMA engine 1921 * @vq: ptr to vb2_queue 1922 * 1923 * This callback stops the DMA engine and any remaining buffers 1924 * in the DMA queue are released. 1925 */ 1926 static void vpfe_stop_streaming(struct vb2_queue *vq) 1927 { 1928 struct vpfe_device *vpfe = vb2_get_drv_priv(vq); 1929 struct vpfe_subdev_info *sdinfo; 1930 int ret; 1931 1932 vpfe_pcr_enable(&vpfe->ccdc, 0); 1933 1934 /* Wait for the last frame to be captured */ 1935 vpfe->stopping = true; 1936 wait_for_completion_timeout(&vpfe->capture_stop, 1937 msecs_to_jiffies(250)); 1938 1939 vpfe_detach_irq(vpfe); 1940 1941 sdinfo = vpfe->current_subdev; 1942 ret = v4l2_subdev_call(sdinfo->sd, video, s_stream, 0); 1943 if (ret && ret != -ENOIOCTLCMD && ret != -ENODEV) 1944 vpfe_dbg(1, vpfe, "stream off failed in subdev\n"); 1945 1946 /* release all active buffers */ 1947 vpfe_return_all_buffers(vpfe, VB2_BUF_STATE_ERROR); 1948 } 1949 1950 static int vpfe_g_pixelaspect(struct file *file, void *priv, 1951 int type, struct v4l2_fract *f) 1952 { 1953 struct vpfe_device *vpfe = video_drvdata(file); 1954 1955 if (type != V4L2_BUF_TYPE_VIDEO_CAPTURE || 1956 vpfe->std_index >= ARRAY_SIZE(vpfe_standards)) 1957 return -EINVAL; 1958 1959 *f = vpfe_standards[vpfe->std_index].pixelaspect; 1960 1961 return 0; 1962 } 1963 1964 static int 1965 vpfe_g_selection(struct file *file, void *fh, struct v4l2_selection *s) 1966 { 1967 struct vpfe_device *vpfe = video_drvdata(file); 1968 1969 if (s->type != V4L2_BUF_TYPE_VIDEO_CAPTURE || 1970 vpfe->std_index >= ARRAY_SIZE(vpfe_standards)) 1971 return -EINVAL; 1972 1973 switch (s->target) { 1974 case V4L2_SEL_TGT_CROP_BOUNDS: 1975 case V4L2_SEL_TGT_CROP_DEFAULT: 1976 s->r.left = 0; 1977 s->r.top = 0; 1978 s->r.width = vpfe_standards[vpfe->std_index].width; 1979 s->r.height = vpfe_standards[vpfe->std_index].height; 1980 break; 1981 1982 case V4L2_SEL_TGT_CROP: 1983 s->r = vpfe->crop; 1984 break; 1985 1986 default: 1987 return -EINVAL; 1988 } 1989 1990 return 0; 1991 } 1992 1993 static int 1994 vpfe_s_selection(struct file *file, void *fh, struct v4l2_selection *s) 1995 { 1996 struct vpfe_device *vpfe = video_drvdata(file); 1997 struct v4l2_rect cr = vpfe->crop; 1998 struct v4l2_rect r = s->r; 1999 u32 bpp; 2000 2001 /* If streaming is started, return error */ 2002 if (vb2_is_busy(&vpfe->buffer_queue)) { 2003 vpfe_err(vpfe, "%s device busy\n", __func__); 2004 return -EBUSY; 2005 } 2006 2007 if (s->type != V4L2_BUF_TYPE_VIDEO_CAPTURE || 2008 s->target != V4L2_SEL_TGT_CROP) 2009 return -EINVAL; 2010 2011 v4l_bound_align_image(&r.width, 0, cr.width, 0, 2012 &r.height, 0, cr.height, 0, 0); 2013 2014 r.left = clamp_t(unsigned int, r.left, 0, cr.width - r.width); 2015 r.top = clamp_t(unsigned int, r.top, 0, cr.height - r.height); 2016 2017 if (s->flags & V4L2_SEL_FLAG_LE && !v4l2_rect_enclosed(&r, &s->r)) 2018 return -ERANGE; 2019 2020 if (s->flags & V4L2_SEL_FLAG_GE && !v4l2_rect_enclosed(&s->r, &r)) 2021 return -ERANGE; 2022 2023 s->r = vpfe->crop = r; 2024 2025 bpp = __get_bytesperpixel(vpfe, vpfe->current_vpfe_fmt); 2026 vpfe_ccdc_set_image_window(&vpfe->ccdc, &r, bpp); 2027 vpfe->fmt.fmt.pix.width = r.width; 2028 vpfe->fmt.fmt.pix.height = r.height; 2029 vpfe->fmt.fmt.pix.bytesperline = 2030 vpfe_ccdc_get_line_length(&vpfe->ccdc); 2031 vpfe->fmt.fmt.pix.sizeimage = vpfe->fmt.fmt.pix.bytesperline * 2032 vpfe->fmt.fmt.pix.height; 2033 2034 vpfe_dbg(1, vpfe, "cropped (%d,%d)/%dx%d of %dx%d\n", 2035 r.left, r.top, r.width, r.height, cr.width, cr.height); 2036 2037 return 0; 2038 } 2039 2040 static long vpfe_ioctl_default(struct file *file, void *priv, 2041 bool valid_prio, unsigned int cmd, void *param) 2042 { 2043 struct vpfe_device *vpfe = video_drvdata(file); 2044 int ret; 2045 2046 if (!valid_prio) { 2047 vpfe_err(vpfe, "%s device busy\n", __func__); 2048 return -EBUSY; 2049 } 2050 2051 /* If streaming is started, return error */ 2052 if (vb2_is_busy(&vpfe->buffer_queue)) { 2053 vpfe_err(vpfe, "%s device busy\n", __func__); 2054 return -EBUSY; 2055 } 2056 2057 switch (cmd) { 2058 case VIDIOC_AM437X_CCDC_CFG: 2059 ret = vpfe_ccdc_set_params(&vpfe->ccdc, (void __user *)param); 2060 if (ret) { 2061 vpfe_dbg(2, vpfe, 2062 "Error setting parameters in CCDC\n"); 2063 return ret; 2064 } 2065 ret = vpfe_get_ccdc_image_format(vpfe, 2066 &vpfe->fmt); 2067 if (ret < 0) { 2068 vpfe_dbg(2, vpfe, 2069 "Invalid image format at CCDC\n"); 2070 return ret; 2071 } 2072 break; 2073 2074 default: 2075 ret = -ENOTTY; 2076 break; 2077 } 2078 2079 return ret; 2080 } 2081 2082 static const struct vb2_ops vpfe_video_qops = { 2083 .wait_prepare = vb2_ops_wait_prepare, 2084 .wait_finish = vb2_ops_wait_finish, 2085 .queue_setup = vpfe_queue_setup, 2086 .buf_prepare = vpfe_buffer_prepare, 2087 .buf_queue = vpfe_buffer_queue, 2088 .start_streaming = vpfe_start_streaming, 2089 .stop_streaming = vpfe_stop_streaming, 2090 }; 2091 2092 /* vpfe capture driver file operations */ 2093 static const struct v4l2_file_operations vpfe_fops = { 2094 .owner = THIS_MODULE, 2095 .open = vpfe_open, 2096 .release = vpfe_release, 2097 .read = vb2_fop_read, 2098 .poll = vb2_fop_poll, 2099 .unlocked_ioctl = video_ioctl2, 2100 .mmap = vb2_fop_mmap, 2101 }; 2102 2103 /* vpfe capture ioctl operations */ 2104 static const struct v4l2_ioctl_ops vpfe_ioctl_ops = { 2105 .vidioc_querycap = vpfe_querycap, 2106 .vidioc_enum_fmt_vid_cap = vpfe_enum_fmt, 2107 .vidioc_g_fmt_vid_cap = vpfe_g_fmt, 2108 .vidioc_s_fmt_vid_cap = vpfe_s_fmt, 2109 .vidioc_try_fmt_vid_cap = vpfe_try_fmt, 2110 2111 .vidioc_enum_framesizes = vpfe_enum_size, 2112 2113 .vidioc_enum_input = vpfe_enum_input, 2114 .vidioc_g_input = vpfe_g_input, 2115 .vidioc_s_input = vpfe_s_input, 2116 2117 .vidioc_querystd = vpfe_querystd, 2118 .vidioc_s_std = vpfe_s_std, 2119 .vidioc_g_std = vpfe_g_std, 2120 2121 .vidioc_reqbufs = vb2_ioctl_reqbufs, 2122 .vidioc_create_bufs = vb2_ioctl_create_bufs, 2123 .vidioc_prepare_buf = vb2_ioctl_prepare_buf, 2124 .vidioc_querybuf = vb2_ioctl_querybuf, 2125 .vidioc_qbuf = vb2_ioctl_qbuf, 2126 .vidioc_dqbuf = vb2_ioctl_dqbuf, 2127 .vidioc_expbuf = vb2_ioctl_expbuf, 2128 .vidioc_streamon = vb2_ioctl_streamon, 2129 .vidioc_streamoff = vb2_ioctl_streamoff, 2130 2131 .vidioc_log_status = v4l2_ctrl_log_status, 2132 .vidioc_subscribe_event = v4l2_ctrl_subscribe_event, 2133 .vidioc_unsubscribe_event = v4l2_event_unsubscribe, 2134 2135 .vidioc_g_pixelaspect = vpfe_g_pixelaspect, 2136 .vidioc_g_selection = vpfe_g_selection, 2137 .vidioc_s_selection = vpfe_s_selection, 2138 2139 .vidioc_default = vpfe_ioctl_default, 2140 }; 2141 2142 static int 2143 vpfe_async_bound(struct v4l2_async_notifier *notifier, 2144 struct v4l2_subdev *subdev, 2145 struct v4l2_async_subdev *asd) 2146 { 2147 struct vpfe_device *vpfe = container_of(notifier->v4l2_dev, 2148 struct vpfe_device, v4l2_dev); 2149 struct v4l2_subdev_mbus_code_enum mbus_code; 2150 struct vpfe_subdev_info *sdinfo; 2151 struct vpfe_fmt *fmt; 2152 int ret = 0; 2153 bool found = false; 2154 int i, j, k; 2155 2156 for (i = 0; i < ARRAY_SIZE(vpfe->cfg->asd); i++) { 2157 if (vpfe->cfg->asd[i]->match.fwnode == 2158 asd[i].match.fwnode) { 2159 sdinfo = &vpfe->cfg->sub_devs[i]; 2160 vpfe->sd[i] = subdev; 2161 vpfe->sd[i]->grp_id = sdinfo->grp_id; 2162 found = true; 2163 break; 2164 } 2165 } 2166 2167 if (!found) { 2168 vpfe_info(vpfe, "sub device (%s) not matched\n", subdev->name); 2169 return -EINVAL; 2170 } 2171 2172 vpfe->video_dev.tvnorms |= sdinfo->inputs[0].std; 2173 2174 vpfe->num_active_fmt = 0; 2175 for (j = 0, i = 0; (ret != -EINVAL); ++j) { 2176 memset(&mbus_code, 0, sizeof(mbus_code)); 2177 mbus_code.index = j; 2178 mbus_code.which = V4L2_SUBDEV_FORMAT_ACTIVE; 2179 ret = v4l2_subdev_call(subdev, pad, enum_mbus_code, 2180 NULL, &mbus_code); 2181 if (ret) 2182 continue; 2183 2184 vpfe_dbg(3, vpfe, 2185 "subdev %s: code: %04x idx: %d\n", 2186 subdev->name, mbus_code.code, j); 2187 2188 for (k = 0; k < ARRAY_SIZE(formats); k++) { 2189 fmt = &formats[k]; 2190 if (mbus_code.code != fmt->code) 2191 continue; 2192 vpfe->active_fmt[i] = fmt; 2193 vpfe_dbg(3, vpfe, 2194 "matched fourcc: %s code: %04x idx: %d\n", 2195 print_fourcc(fmt->fourcc), mbus_code.code, i); 2196 vpfe->num_active_fmt = ++i; 2197 } 2198 } 2199 2200 if (!i) { 2201 vpfe_err(vpfe, "No suitable format reported by subdev %s\n", 2202 subdev->name); 2203 return -EINVAL; 2204 } 2205 return 0; 2206 } 2207 2208 static int vpfe_probe_complete(struct vpfe_device *vpfe) 2209 { 2210 struct video_device *vdev; 2211 struct vb2_queue *q; 2212 int err; 2213 2214 spin_lock_init(&vpfe->dma_queue_lock); 2215 mutex_init(&vpfe->lock); 2216 2217 vpfe->fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 2218 2219 /* set first sub device as current one */ 2220 vpfe->current_subdev = &vpfe->cfg->sub_devs[0]; 2221 vpfe->v4l2_dev.ctrl_handler = vpfe->sd[0]->ctrl_handler; 2222 2223 err = vpfe_set_input(vpfe, 0); 2224 if (err) 2225 goto probe_out; 2226 2227 /* Initialize videobuf2 queue as per the buffer type */ 2228 q = &vpfe->buffer_queue; 2229 q->type = V4L2_BUF_TYPE_VIDEO_CAPTURE; 2230 q->io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ; 2231 q->drv_priv = vpfe; 2232 q->ops = &vpfe_video_qops; 2233 q->mem_ops = &vb2_dma_contig_memops; 2234 q->buf_struct_size = sizeof(struct vpfe_cap_buffer); 2235 q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; 2236 q->lock = &vpfe->lock; 2237 q->min_buffers_needed = 1; 2238 q->dev = vpfe->pdev; 2239 2240 err = vb2_queue_init(q); 2241 if (err) { 2242 vpfe_err(vpfe, "vb2_queue_init() failed\n"); 2243 goto probe_out; 2244 } 2245 2246 INIT_LIST_HEAD(&vpfe->dma_queue); 2247 2248 vdev = &vpfe->video_dev; 2249 strscpy(vdev->name, VPFE_MODULE_NAME, sizeof(vdev->name)); 2250 vdev->release = video_device_release_empty; 2251 vdev->fops = &vpfe_fops; 2252 vdev->ioctl_ops = &vpfe_ioctl_ops; 2253 vdev->v4l2_dev = &vpfe->v4l2_dev; 2254 vdev->vfl_dir = VFL_DIR_RX; 2255 vdev->queue = q; 2256 vdev->lock = &vpfe->lock; 2257 vdev->device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING | 2258 V4L2_CAP_READWRITE; 2259 video_set_drvdata(vdev, vpfe); 2260 err = video_register_device(&vpfe->video_dev, VFL_TYPE_VIDEO, -1); 2261 if (err) { 2262 vpfe_err(vpfe, 2263 "Unable to register video device.\n"); 2264 goto probe_out; 2265 } 2266 2267 return 0; 2268 2269 probe_out: 2270 v4l2_device_unregister(&vpfe->v4l2_dev); 2271 return err; 2272 } 2273 2274 static int vpfe_async_complete(struct v4l2_async_notifier *notifier) 2275 { 2276 struct vpfe_device *vpfe = container_of(notifier->v4l2_dev, 2277 struct vpfe_device, v4l2_dev); 2278 2279 return vpfe_probe_complete(vpfe); 2280 } 2281 2282 static const struct v4l2_async_notifier_operations vpfe_async_ops = { 2283 .bound = vpfe_async_bound, 2284 .complete = vpfe_async_complete, 2285 }; 2286 2287 static struct vpfe_config * 2288 vpfe_get_pdata(struct vpfe_device *vpfe) 2289 { 2290 struct device_node *endpoint = NULL; 2291 struct device *dev = vpfe->pdev; 2292 struct vpfe_subdev_info *sdinfo; 2293 struct vpfe_config *pdata; 2294 unsigned int flags; 2295 unsigned int i; 2296 int err; 2297 2298 dev_dbg(dev, "vpfe_get_pdata\n"); 2299 2300 v4l2_async_nf_init(&vpfe->notifier); 2301 2302 if (!IS_ENABLED(CONFIG_OF) || !dev->of_node) 2303 return dev->platform_data; 2304 2305 pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL); 2306 if (!pdata) 2307 return NULL; 2308 2309 for (i = 0; ; i++) { 2310 struct v4l2_fwnode_endpoint bus_cfg = { .bus_type = 0 }; 2311 struct device_node *rem; 2312 2313 endpoint = of_graph_get_next_endpoint(dev->of_node, endpoint); 2314 if (!endpoint) 2315 break; 2316 2317 sdinfo = &pdata->sub_devs[i]; 2318 sdinfo->grp_id = 0; 2319 2320 /* we only support camera */ 2321 sdinfo->inputs[0].index = i; 2322 strscpy(sdinfo->inputs[0].name, "Camera", 2323 sizeof(sdinfo->inputs[0].name)); 2324 sdinfo->inputs[0].type = V4L2_INPUT_TYPE_CAMERA; 2325 sdinfo->inputs[0].std = V4L2_STD_ALL; 2326 sdinfo->inputs[0].capabilities = V4L2_IN_CAP_STD; 2327 2328 sdinfo->can_route = 0; 2329 sdinfo->routes = NULL; 2330 2331 of_property_read_u32(endpoint, "ti,am437x-vpfe-interface", 2332 &sdinfo->vpfe_param.if_type); 2333 if (sdinfo->vpfe_param.if_type < 0 || 2334 sdinfo->vpfe_param.if_type > 4) { 2335 sdinfo->vpfe_param.if_type = VPFE_RAW_BAYER; 2336 } 2337 2338 err = v4l2_fwnode_endpoint_parse(of_fwnode_handle(endpoint), 2339 &bus_cfg); 2340 if (err) { 2341 dev_err(dev, "Could not parse the endpoint\n"); 2342 goto cleanup; 2343 } 2344 2345 sdinfo->vpfe_param.bus_width = bus_cfg.bus.parallel.bus_width; 2346 2347 if (sdinfo->vpfe_param.bus_width < 8 || 2348 sdinfo->vpfe_param.bus_width > 16) { 2349 dev_err(dev, "Invalid bus width.\n"); 2350 goto cleanup; 2351 } 2352 2353 flags = bus_cfg.bus.parallel.flags; 2354 2355 if (flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH) 2356 sdinfo->vpfe_param.hdpol = 1; 2357 2358 if (flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH) 2359 sdinfo->vpfe_param.vdpol = 1; 2360 2361 rem = of_graph_get_remote_port_parent(endpoint); 2362 if (!rem) { 2363 dev_err(dev, "Remote device at %pOF not found\n", 2364 endpoint); 2365 goto cleanup; 2366 } 2367 2368 pdata->asd[i] = v4l2_async_nf_add_fwnode(&vpfe->notifier, 2369 of_fwnode_handle(rem), 2370 struct 2371 v4l2_async_subdev); 2372 of_node_put(rem); 2373 if (IS_ERR(pdata->asd[i])) 2374 goto cleanup; 2375 } 2376 2377 of_node_put(endpoint); 2378 return pdata; 2379 2380 cleanup: 2381 v4l2_async_nf_cleanup(&vpfe->notifier); 2382 of_node_put(endpoint); 2383 return NULL; 2384 } 2385 2386 /* 2387 * vpfe_probe : This function creates device entries by register 2388 * itself to the V4L2 driver and initializes fields of each 2389 * device objects 2390 */ 2391 static int vpfe_probe(struct platform_device *pdev) 2392 { 2393 struct vpfe_config *vpfe_cfg; 2394 struct vpfe_device *vpfe; 2395 struct vpfe_ccdc *ccdc; 2396 int ret; 2397 2398 vpfe = devm_kzalloc(&pdev->dev, sizeof(*vpfe), GFP_KERNEL); 2399 if (!vpfe) 2400 return -ENOMEM; 2401 2402 vpfe->pdev = &pdev->dev; 2403 2404 vpfe_cfg = vpfe_get_pdata(vpfe); 2405 if (!vpfe_cfg) { 2406 dev_err(&pdev->dev, "No platform data\n"); 2407 return -EINVAL; 2408 } 2409 2410 vpfe->cfg = vpfe_cfg; 2411 ccdc = &vpfe->ccdc; 2412 2413 ccdc->ccdc_cfg.base_addr = devm_platform_ioremap_resource(pdev, 0); 2414 if (IS_ERR(ccdc->ccdc_cfg.base_addr)) { 2415 ret = PTR_ERR(ccdc->ccdc_cfg.base_addr); 2416 goto probe_out_cleanup; 2417 } 2418 2419 ret = platform_get_irq(pdev, 0); 2420 if (ret <= 0) { 2421 ret = -ENODEV; 2422 goto probe_out_cleanup; 2423 } 2424 vpfe->irq = ret; 2425 2426 ret = devm_request_irq(vpfe->pdev, vpfe->irq, vpfe_isr, 0, 2427 "vpfe_capture0", vpfe); 2428 if (ret) { 2429 dev_err(&pdev->dev, "Unable to request interrupt\n"); 2430 ret = -EINVAL; 2431 goto probe_out_cleanup; 2432 } 2433 2434 ret = v4l2_device_register(&pdev->dev, &vpfe->v4l2_dev); 2435 if (ret) { 2436 vpfe_err(vpfe, 2437 "Unable to register v4l2 device.\n"); 2438 goto probe_out_cleanup; 2439 } 2440 2441 /* set the driver data in platform device */ 2442 platform_set_drvdata(pdev, vpfe); 2443 /* Enabling module functional clock */ 2444 pm_runtime_enable(&pdev->dev); 2445 2446 /* for now just enable it here instead of waiting for the open */ 2447 ret = pm_runtime_resume_and_get(&pdev->dev); 2448 if (ret < 0) { 2449 vpfe_err(vpfe, "Unable to resume device.\n"); 2450 goto probe_out_v4l2_unregister; 2451 } 2452 2453 vpfe_ccdc_config_defaults(ccdc); 2454 2455 pm_runtime_put_sync(&pdev->dev); 2456 2457 vpfe->sd = devm_kcalloc(&pdev->dev, 2458 ARRAY_SIZE(vpfe->cfg->asd), 2459 sizeof(struct v4l2_subdev *), 2460 GFP_KERNEL); 2461 if (!vpfe->sd) { 2462 ret = -ENOMEM; 2463 goto probe_out_v4l2_unregister; 2464 } 2465 2466 vpfe->notifier.ops = &vpfe_async_ops; 2467 ret = v4l2_async_nf_register(&vpfe->v4l2_dev, &vpfe->notifier); 2468 if (ret) { 2469 vpfe_err(vpfe, "Error registering async notifier\n"); 2470 ret = -EINVAL; 2471 goto probe_out_v4l2_unregister; 2472 } 2473 2474 return 0; 2475 2476 probe_out_v4l2_unregister: 2477 v4l2_device_unregister(&vpfe->v4l2_dev); 2478 probe_out_cleanup: 2479 v4l2_async_nf_cleanup(&vpfe->notifier); 2480 return ret; 2481 } 2482 2483 /* 2484 * vpfe_remove : It un-register device from V4L2 driver 2485 */ 2486 static int vpfe_remove(struct platform_device *pdev) 2487 { 2488 struct vpfe_device *vpfe = platform_get_drvdata(pdev); 2489 2490 pm_runtime_disable(&pdev->dev); 2491 2492 v4l2_async_nf_unregister(&vpfe->notifier); 2493 v4l2_async_nf_cleanup(&vpfe->notifier); 2494 v4l2_device_unregister(&vpfe->v4l2_dev); 2495 video_unregister_device(&vpfe->video_dev); 2496 2497 return 0; 2498 } 2499 2500 #ifdef CONFIG_PM_SLEEP 2501 2502 static void vpfe_save_context(struct vpfe_ccdc *ccdc) 2503 { 2504 ccdc->ccdc_ctx[VPFE_PCR >> 2] = vpfe_reg_read(ccdc, VPFE_PCR); 2505 ccdc->ccdc_ctx[VPFE_SYNMODE >> 2] = vpfe_reg_read(ccdc, VPFE_SYNMODE); 2506 ccdc->ccdc_ctx[VPFE_SDOFST >> 2] = vpfe_reg_read(ccdc, VPFE_SDOFST); 2507 ccdc->ccdc_ctx[VPFE_SDR_ADDR >> 2] = vpfe_reg_read(ccdc, VPFE_SDR_ADDR); 2508 ccdc->ccdc_ctx[VPFE_CLAMP >> 2] = vpfe_reg_read(ccdc, VPFE_CLAMP); 2509 ccdc->ccdc_ctx[VPFE_DCSUB >> 2] = vpfe_reg_read(ccdc, VPFE_DCSUB); 2510 ccdc->ccdc_ctx[VPFE_COLPTN >> 2] = vpfe_reg_read(ccdc, VPFE_COLPTN); 2511 ccdc->ccdc_ctx[VPFE_BLKCMP >> 2] = vpfe_reg_read(ccdc, VPFE_BLKCMP); 2512 ccdc->ccdc_ctx[VPFE_VDINT >> 2] = vpfe_reg_read(ccdc, VPFE_VDINT); 2513 ccdc->ccdc_ctx[VPFE_ALAW >> 2] = vpfe_reg_read(ccdc, VPFE_ALAW); 2514 ccdc->ccdc_ctx[VPFE_REC656IF >> 2] = vpfe_reg_read(ccdc, VPFE_REC656IF); 2515 ccdc->ccdc_ctx[VPFE_CCDCFG >> 2] = vpfe_reg_read(ccdc, VPFE_CCDCFG); 2516 ccdc->ccdc_ctx[VPFE_CULLING >> 2] = vpfe_reg_read(ccdc, VPFE_CULLING); 2517 ccdc->ccdc_ctx[VPFE_HD_VD_WID >> 2] = vpfe_reg_read(ccdc, 2518 VPFE_HD_VD_WID); 2519 ccdc->ccdc_ctx[VPFE_PIX_LINES >> 2] = vpfe_reg_read(ccdc, 2520 VPFE_PIX_LINES); 2521 ccdc->ccdc_ctx[VPFE_HORZ_INFO >> 2] = vpfe_reg_read(ccdc, 2522 VPFE_HORZ_INFO); 2523 ccdc->ccdc_ctx[VPFE_VERT_START >> 2] = vpfe_reg_read(ccdc, 2524 VPFE_VERT_START); 2525 ccdc->ccdc_ctx[VPFE_VERT_LINES >> 2] = vpfe_reg_read(ccdc, 2526 VPFE_VERT_LINES); 2527 ccdc->ccdc_ctx[VPFE_HSIZE_OFF >> 2] = vpfe_reg_read(ccdc, 2528 VPFE_HSIZE_OFF); 2529 } 2530 2531 static int vpfe_suspend(struct device *dev) 2532 { 2533 struct vpfe_device *vpfe = dev_get_drvdata(dev); 2534 struct vpfe_ccdc *ccdc = &vpfe->ccdc; 2535 2536 /* only do full suspend if streaming has started */ 2537 if (vb2_start_streaming_called(&vpfe->buffer_queue)) { 2538 /* 2539 * ignore RPM resume errors here, as it is already too late. 2540 * A check like that should happen earlier, either at 2541 * open() or just before start streaming. 2542 */ 2543 pm_runtime_get_sync(dev); 2544 vpfe_config_enable(ccdc, 1); 2545 2546 /* Save VPFE context */ 2547 vpfe_save_context(ccdc); 2548 2549 /* Disable CCDC */ 2550 vpfe_pcr_enable(ccdc, 0); 2551 vpfe_config_enable(ccdc, 0); 2552 2553 /* Disable both master and slave clock */ 2554 pm_runtime_put_sync(dev); 2555 } 2556 2557 /* Select sleep pin state */ 2558 pinctrl_pm_select_sleep_state(dev); 2559 2560 return 0; 2561 } 2562 2563 static void vpfe_restore_context(struct vpfe_ccdc *ccdc) 2564 { 2565 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_SYNMODE >> 2], VPFE_SYNMODE); 2566 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_CULLING >> 2], VPFE_CULLING); 2567 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_SDOFST >> 2], VPFE_SDOFST); 2568 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_SDR_ADDR >> 2], VPFE_SDR_ADDR); 2569 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_CLAMP >> 2], VPFE_CLAMP); 2570 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_DCSUB >> 2], VPFE_DCSUB); 2571 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_COLPTN >> 2], VPFE_COLPTN); 2572 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_BLKCMP >> 2], VPFE_BLKCMP); 2573 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_VDINT >> 2], VPFE_VDINT); 2574 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_ALAW >> 2], VPFE_ALAW); 2575 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_REC656IF >> 2], VPFE_REC656IF); 2576 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_CCDCFG >> 2], VPFE_CCDCFG); 2577 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_PCR >> 2], VPFE_PCR); 2578 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_HD_VD_WID >> 2], 2579 VPFE_HD_VD_WID); 2580 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_PIX_LINES >> 2], 2581 VPFE_PIX_LINES); 2582 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_HORZ_INFO >> 2], 2583 VPFE_HORZ_INFO); 2584 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_VERT_START >> 2], 2585 VPFE_VERT_START); 2586 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_VERT_LINES >> 2], 2587 VPFE_VERT_LINES); 2588 vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_HSIZE_OFF >> 2], 2589 VPFE_HSIZE_OFF); 2590 } 2591 2592 static int vpfe_resume(struct device *dev) 2593 { 2594 struct vpfe_device *vpfe = dev_get_drvdata(dev); 2595 struct vpfe_ccdc *ccdc = &vpfe->ccdc; 2596 2597 /* only do full resume if streaming has started */ 2598 if (vb2_start_streaming_called(&vpfe->buffer_queue)) { 2599 /* Enable both master and slave clock */ 2600 pm_runtime_get_sync(dev); 2601 vpfe_config_enable(ccdc, 1); 2602 2603 /* Restore VPFE context */ 2604 vpfe_restore_context(ccdc); 2605 2606 vpfe_config_enable(ccdc, 0); 2607 pm_runtime_put_sync(dev); 2608 } 2609 2610 /* Select default pin state */ 2611 pinctrl_pm_select_default_state(dev); 2612 2613 return 0; 2614 } 2615 2616 #endif 2617 2618 static SIMPLE_DEV_PM_OPS(vpfe_pm_ops, vpfe_suspend, vpfe_resume); 2619 2620 static const struct of_device_id vpfe_of_match[] = { 2621 { .compatible = "ti,am437x-vpfe", }, 2622 { /* sentinel */ }, 2623 }; 2624 MODULE_DEVICE_TABLE(of, vpfe_of_match); 2625 2626 static struct platform_driver vpfe_driver = { 2627 .probe = vpfe_probe, 2628 .remove = vpfe_remove, 2629 .driver = { 2630 .name = VPFE_MODULE_NAME, 2631 .pm = &vpfe_pm_ops, 2632 .of_match_table = of_match_ptr(vpfe_of_match), 2633 }, 2634 }; 2635 2636 module_platform_driver(vpfe_driver); 2637 2638 MODULE_AUTHOR("Texas Instruments"); 2639 MODULE_DESCRIPTION("TI AM437x VPFE driver"); 2640 MODULE_LICENSE("GPL"); 2641 MODULE_VERSION(VPFE_VERSION); 2642