1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (c) 2022 MediaTek Inc. 4 * Author: Ping-Hsun Wu <ping-hsun.wu@mediatek.com> 5 */ 6 7 #include <linux/clk.h> 8 #include <linux/of_platform.h> 9 #include <linux/of_address.h> 10 #include <linux/pm_runtime.h> 11 #include "mtk-mdp3-cfg.h" 12 #include "mtk-mdp3-comp.h" 13 #include "mtk-mdp3-core.h" 14 #include "mtk-mdp3-regs.h" 15 16 #include "mdp_reg_rdma.h" 17 #include "mdp_reg_ccorr.h" 18 #include "mdp_reg_rsz.h" 19 #include "mdp_reg_wrot.h" 20 #include "mdp_reg_wdma.h" 21 22 static u32 mdp_comp_alias_id[MDP_COMP_TYPE_COUNT]; 23 static int p_id; 24 25 static inline const struct mdp_platform_config * 26 __get_plat_cfg(const struct mdp_comp_ctx *ctx) 27 { 28 if (!ctx) 29 return NULL; 30 31 return ctx->comp->mdp_dev->mdp_data->mdp_cfg; 32 } 33 34 static s64 get_comp_flag(const struct mdp_comp_ctx *ctx) 35 { 36 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 37 u32 rdma0, rsz1; 38 39 rdma0 = mdp_cfg_get_id_inner(ctx->comp->mdp_dev, MDP_COMP_RDMA0); 40 rsz1 = mdp_cfg_get_id_inner(ctx->comp->mdp_dev, MDP_COMP_RSZ1); 41 if (!rdma0 || !rsz1) 42 return MDP_COMP_NONE; 43 44 if (mdp_cfg && mdp_cfg->rdma_rsz1_sram_sharing) 45 if (ctx->comp->inner_id == rdma0) 46 return BIT(rdma0) | BIT(rsz1); 47 48 return BIT(ctx->comp->inner_id); 49 } 50 51 static int init_rdma(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 52 { 53 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 54 phys_addr_t base = ctx->comp->reg_base; 55 u8 subsys_id = ctx->comp->subsys_id; 56 s32 rdma0; 57 58 rdma0 = mdp_cfg_get_id_inner(ctx->comp->mdp_dev, MDP_COMP_RDMA0); 59 if (!rdma0) 60 return -EINVAL; 61 62 if (mdp_cfg && mdp_cfg->rdma_support_10bit) { 63 struct mdp_comp *prz1 = ctx->comp->mdp_dev->comp[MDP_COMP_RSZ1]; 64 65 /* Disable RSZ1 */ 66 if (ctx->comp->inner_id == rdma0 && prz1) 67 MM_REG_WRITE(cmd, subsys_id, prz1->reg_base, PRZ_ENABLE, 68 0x0, BIT(0)); 69 } 70 71 /* Reset RDMA */ 72 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_RESET, BIT(0), BIT(0)); 73 MM_REG_POLL(cmd, subsys_id, base, MDP_RDMA_MON_STA_1, BIT(8), BIT(8)); 74 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_RESET, 0x0, BIT(0)); 75 return 0; 76 } 77 78 static int config_rdma_frame(struct mdp_comp_ctx *ctx, 79 struct mdp_cmdq_cmd *cmd, 80 const struct v4l2_rect *compose) 81 { 82 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 83 u32 colorformat = ctx->input->buffer.format.colorformat; 84 bool block10bit = MDP_COLOR_IS_10BIT_PACKED(colorformat); 85 bool en_ufo = MDP_COLOR_IS_UFP(colorformat); 86 phys_addr_t base = ctx->comp->reg_base; 87 u8 subsys_id = ctx->comp->subsys_id; 88 u32 reg = 0; 89 90 if (mdp_cfg && mdp_cfg->rdma_support_10bit) { 91 if (block10bit) 92 MM_REG_WRITE(cmd, subsys_id, base, 93 MDP_RDMA_RESV_DUMMY_0, 0x7, 0x7); 94 else 95 MM_REG_WRITE(cmd, subsys_id, base, 96 MDP_RDMA_RESV_DUMMY_0, 0x0, 0x7); 97 } 98 99 /* Setup smi control */ 100 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_GMCIF_CON, 101 (7 << 4) + //burst type to 8 102 (1 << 16), //enable pre-ultra 103 0x00030071); 104 105 /* Setup source frame info */ 106 if (CFG_CHECK(MT8183, p_id)) 107 reg = CFG_COMP(MT8183, ctx->param, rdma.src_ctrl); 108 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_CON, reg, 109 0x03C8FE0F); 110 111 if (mdp_cfg) 112 if (mdp_cfg->rdma_support_10bit && en_ufo) { 113 /* Setup source buffer base */ 114 if (CFG_CHECK(MT8183, p_id)) 115 reg = CFG_COMP(MT8183, ctx->param, rdma.ufo_dec_y); 116 MM_REG_WRITE(cmd, subsys_id, 117 base, MDP_RDMA_UFO_DEC_LENGTH_BASE_Y, 118 reg, 0xFFFFFFFF); 119 if (CFG_CHECK(MT8183, p_id)) 120 reg = CFG_COMP(MT8183, ctx->param, rdma.ufo_dec_c); 121 MM_REG_WRITE(cmd, subsys_id, 122 base, MDP_RDMA_UFO_DEC_LENGTH_BASE_C, 123 reg, 0xFFFFFFFF); 124 /* Set 10bit source frame pitch */ 125 if (block10bit) { 126 if (CFG_CHECK(MT8183, p_id)) 127 reg = CFG_COMP(MT8183, ctx->param, rdma.mf_bkgd_in_pxl); 128 MM_REG_WRITE(cmd, subsys_id, 129 base, MDP_RDMA_MF_BKGD_SIZE_IN_PXL, 130 reg, 0x001FFFFF); 131 } 132 } 133 134 if (CFG_CHECK(MT8183, p_id)) 135 reg = CFG_COMP(MT8183, ctx->param, rdma.control); 136 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_CON, reg, 137 0x1110); 138 /* Setup source buffer base */ 139 if (CFG_CHECK(MT8183, p_id)) 140 reg = CFG_COMP(MT8183, ctx->param, rdma.iova[0]); 141 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_BASE_0, reg, 142 0xFFFFFFFF); 143 if (CFG_CHECK(MT8183, p_id)) 144 reg = CFG_COMP(MT8183, ctx->param, rdma.iova[1]); 145 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_BASE_1, reg, 146 0xFFFFFFFF); 147 if (CFG_CHECK(MT8183, p_id)) 148 reg = CFG_COMP(MT8183, ctx->param, rdma.iova[2]); 149 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_BASE_2, reg, 150 0xFFFFFFFF); 151 /* Setup source buffer end */ 152 if (CFG_CHECK(MT8183, p_id)) 153 reg = CFG_COMP(MT8183, ctx->param, rdma.iova_end[0]); 154 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_END_0, 155 reg, 0xFFFFFFFF); 156 if (CFG_CHECK(MT8183, p_id)) 157 reg = CFG_COMP(MT8183, ctx->param, rdma.iova_end[1]); 158 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_END_1, 159 reg, 0xFFFFFFFF); 160 if (CFG_CHECK(MT8183, p_id)) 161 reg = CFG_COMP(MT8183, ctx->param, rdma.iova_end[2]); 162 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_END_2, 163 reg, 0xFFFFFFFF); 164 /* Setup source frame pitch */ 165 if (CFG_CHECK(MT8183, p_id)) 166 reg = CFG_COMP(MT8183, ctx->param, rdma.mf_bkgd); 167 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_MF_BKGD_SIZE_IN_BYTE, 168 reg, 0x001FFFFF); 169 if (CFG_CHECK(MT8183, p_id)) 170 reg = CFG_COMP(MT8183, ctx->param, rdma.sf_bkgd); 171 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SF_BKGD_SIZE_IN_BYTE, 172 reg, 0x001FFFFF); 173 /* Setup color transform */ 174 if (CFG_CHECK(MT8183, p_id)) 175 reg = CFG_COMP(MT8183, ctx->param, rdma.transform); 176 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_TRANSFORM_0, 177 reg, 0x0F110000); 178 179 return 0; 180 } 181 182 static int config_rdma_subfrm(struct mdp_comp_ctx *ctx, 183 struct mdp_cmdq_cmd *cmd, u32 index) 184 { 185 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 186 u32 colorformat = ctx->input->buffer.format.colorformat; 187 bool block10bit = MDP_COLOR_IS_10BIT_PACKED(colorformat); 188 bool en_ufo = MDP_COLOR_IS_UFP(colorformat); 189 phys_addr_t base = ctx->comp->reg_base; 190 u8 subsys_id = ctx->comp->subsys_id; 191 u32 csf_l = 0, csf_r = 0; 192 u32 reg = 0; 193 194 /* Enable RDMA */ 195 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_EN, BIT(0), BIT(0)); 196 197 /* Set Y pixel offset */ 198 if (CFG_CHECK(MT8183, p_id)) 199 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].offset[0]); 200 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_OFFSET_0, 201 reg, 0xFFFFFFFF); 202 203 /* Set 10bit UFO mode */ 204 if (mdp_cfg) { 205 if (mdp_cfg->rdma_support_10bit && block10bit && en_ufo) { 206 if (CFG_CHECK(MT8183, p_id)) 207 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].offset_0_p); 208 MM_REG_WRITE(cmd, subsys_id, base, 209 MDP_RDMA_SRC_OFFSET_0_P, 210 reg, 0xFFFFFFFF); 211 } 212 } 213 214 /* Set U pixel offset */ 215 if (CFG_CHECK(MT8183, p_id)) 216 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].offset[1]); 217 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_OFFSET_1, 218 reg, 0xFFFFFFFF); 219 /* Set V pixel offset */ 220 if (CFG_CHECK(MT8183, p_id)) 221 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].offset[2]); 222 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_SRC_OFFSET_2, 223 reg, 0xFFFFFFFF); 224 /* Set source size */ 225 if (CFG_CHECK(MT8183, p_id)) 226 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].src); 227 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_MF_SRC_SIZE, reg, 228 0x1FFF1FFF); 229 /* Set target size */ 230 if (CFG_CHECK(MT8183, p_id)) 231 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].clip); 232 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_MF_CLIP_SIZE, 233 reg, 0x1FFF1FFF); 234 /* Set crop offset */ 235 if (CFG_CHECK(MT8183, p_id)) 236 reg = CFG_COMP(MT8183, ctx->param, rdma.subfrms[index].clip_ofst); 237 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_MF_OFFSET_1, 238 reg, 0x003F001F); 239 240 if (CFG_CHECK(MT8183, p_id)) { 241 csf_l = CFG_COMP(MT8183, ctx->param, subfrms[index].in.left); 242 csf_r = CFG_COMP(MT8183, ctx->param, subfrms[index].in.right); 243 } 244 if (mdp_cfg && mdp_cfg->rdma_upsample_repeat_only) 245 if ((csf_r - csf_l + 1) > 320) 246 MM_REG_WRITE(cmd, subsys_id, base, 247 MDP_RDMA_RESV_DUMMY_0, BIT(2), BIT(2)); 248 249 return 0; 250 } 251 252 static int wait_rdma_event(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 253 { 254 struct device *dev = &ctx->comp->mdp_dev->pdev->dev; 255 phys_addr_t base = ctx->comp->reg_base; 256 u8 subsys_id = ctx->comp->subsys_id; 257 258 if (ctx->comp->alias_id == 0) 259 MM_REG_WAIT(cmd, ctx->comp->gce_event[MDP_GCE_EVENT_EOF]); 260 else 261 dev_err(dev, "Do not support RDMA1_DONE event\n"); 262 263 /* Disable RDMA */ 264 MM_REG_WRITE(cmd, subsys_id, base, MDP_RDMA_EN, 0x0, BIT(0)); 265 return 0; 266 } 267 268 static const struct mdp_comp_ops rdma_ops = { 269 .get_comp_flag = get_comp_flag, 270 .init_comp = init_rdma, 271 .config_frame = config_rdma_frame, 272 .config_subfrm = config_rdma_subfrm, 273 .wait_comp_event = wait_rdma_event, 274 }; 275 276 static int init_rsz(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 277 { 278 phys_addr_t base = ctx->comp->reg_base; 279 u8 subsys_id = ctx->comp->subsys_id; 280 281 /* Reset RSZ */ 282 MM_REG_WRITE(cmd, subsys_id, base, PRZ_ENABLE, 0x10000, BIT(16)); 283 MM_REG_WRITE(cmd, subsys_id, base, PRZ_ENABLE, 0x0, BIT(16)); 284 /* Enable RSZ */ 285 MM_REG_WRITE(cmd, subsys_id, base, PRZ_ENABLE, BIT(0), BIT(0)); 286 return 0; 287 } 288 289 static int config_rsz_frame(struct mdp_comp_ctx *ctx, 290 struct mdp_cmdq_cmd *cmd, 291 const struct v4l2_rect *compose) 292 { 293 phys_addr_t base = ctx->comp->reg_base; 294 u8 subsys_id = ctx->comp->subsys_id; 295 bool bypass = FALSE; 296 u32 reg = 0; 297 298 if (CFG_CHECK(MT8183, p_id)) 299 bypass = CFG_COMP(MT8183, ctx->param, frame.bypass); 300 301 if (bypass) { 302 /* Disable RSZ */ 303 MM_REG_WRITE(cmd, subsys_id, base, PRZ_ENABLE, 0x0, BIT(0)); 304 return 0; 305 } 306 307 if (CFG_CHECK(MT8183, p_id)) 308 reg = CFG_COMP(MT8183, ctx->param, rsz.control1); 309 MM_REG_WRITE(cmd, subsys_id, base, PRZ_CONTROL_1, reg, 310 0x03FFFDF3); 311 if (CFG_CHECK(MT8183, p_id)) 312 reg = CFG_COMP(MT8183, ctx->param, rsz.control2); 313 MM_REG_WRITE(cmd, subsys_id, base, PRZ_CONTROL_2, reg, 314 0x0FFFC290); 315 if (CFG_CHECK(MT8183, p_id)) 316 reg = CFG_COMP(MT8183, ctx->param, rsz.coeff_step_x); 317 MM_REG_WRITE(cmd, subsys_id, base, PRZ_HORIZONTAL_COEFF_STEP, 318 reg, 0x007FFFFF); 319 if (CFG_CHECK(MT8183, p_id)) 320 reg = CFG_COMP(MT8183, ctx->param, rsz.coeff_step_y); 321 MM_REG_WRITE(cmd, subsys_id, base, PRZ_VERTICAL_COEFF_STEP, 322 reg, 0x007FFFFF); 323 return 0; 324 } 325 326 static int config_rsz_subfrm(struct mdp_comp_ctx *ctx, 327 struct mdp_cmdq_cmd *cmd, u32 index) 328 { 329 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 330 phys_addr_t base = ctx->comp->reg_base; 331 u8 subsys_id = ctx->comp->subsys_id; 332 u32 csf_l = 0, csf_r = 0; 333 u32 reg = 0; 334 335 if (CFG_CHECK(MT8183, p_id)) 336 reg = CFG_COMP(MT8183, ctx->param, rsz.subfrms[index].control2); 337 MM_REG_WRITE(cmd, subsys_id, base, PRZ_CONTROL_2, reg, 338 0x00003800); 339 if (CFG_CHECK(MT8183, p_id)) 340 reg = CFG_COMP(MT8183, ctx->param, rsz.subfrms[index].src); 341 MM_REG_WRITE(cmd, subsys_id, base, PRZ_INPUT_IMAGE, reg, 342 0xFFFFFFFF); 343 344 if (CFG_CHECK(MT8183, p_id)) { 345 csf_l = CFG_COMP(MT8183, ctx->param, subfrms[index].in.left); 346 csf_r = CFG_COMP(MT8183, ctx->param, subfrms[index].in.right); 347 } 348 if (mdp_cfg && mdp_cfg->rsz_disable_dcm_small_sample) 349 if ((csf_r - csf_l + 1) <= 16) 350 MM_REG_WRITE(cmd, subsys_id, base, PRZ_CONTROL_1, 351 BIT(27), BIT(27)); 352 353 if (CFG_CHECK(MT8183, p_id)) 354 reg = CFG_COMP(MT8183, ctx->param, subfrms[index].luma.left); 355 MM_REG_WRITE(cmd, subsys_id, base, PRZ_LUMA_HORIZONTAL_INTEGER_OFFSET, 356 reg, 0xFFFF); 357 if (CFG_CHECK(MT8183, p_id)) 358 reg = CFG_COMP(MT8183, ctx->param, subfrms[index].luma.left_subpix); 359 MM_REG_WRITE(cmd, subsys_id, 360 base, PRZ_LUMA_HORIZONTAL_SUBPIXEL_OFFSET, 361 reg, 0x1FFFFF); 362 if (CFG_CHECK(MT8183, p_id)) 363 reg = CFG_COMP(MT8183, ctx->param, subfrms[index].luma.top); 364 MM_REG_WRITE(cmd, subsys_id, base, PRZ_LUMA_VERTICAL_INTEGER_OFFSET, 365 reg, 0xFFFF); 366 if (CFG_CHECK(MT8183, p_id)) 367 reg = CFG_COMP(MT8183, ctx->param, subfrms[index].luma.top_subpix); 368 MM_REG_WRITE(cmd, subsys_id, base, PRZ_LUMA_VERTICAL_SUBPIXEL_OFFSET, 369 reg, 0x1FFFFF); 370 if (CFG_CHECK(MT8183, p_id)) 371 reg = CFG_COMP(MT8183, ctx->param, subfrms[index].chroma.left); 372 MM_REG_WRITE(cmd, subsys_id, 373 base, PRZ_CHROMA_HORIZONTAL_INTEGER_OFFSET, 374 reg, 0xFFFF); 375 if (CFG_CHECK(MT8183, p_id)) 376 reg = CFG_COMP(MT8183, ctx->param, subfrms[index].chroma.left_subpix); 377 MM_REG_WRITE(cmd, subsys_id, 378 base, PRZ_CHROMA_HORIZONTAL_SUBPIXEL_OFFSET, 379 reg, 0x1FFFFF); 380 381 if (CFG_CHECK(MT8183, p_id)) 382 reg = CFG_COMP(MT8183, ctx->param, rsz.subfrms[index].clip); 383 MM_REG_WRITE(cmd, subsys_id, base, PRZ_OUTPUT_IMAGE, reg, 384 0xFFFFFFFF); 385 386 return 0; 387 } 388 389 static int advance_rsz_subfrm(struct mdp_comp_ctx *ctx, 390 struct mdp_cmdq_cmd *cmd, u32 index) 391 { 392 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 393 394 if (mdp_cfg && mdp_cfg->rsz_disable_dcm_small_sample) { 395 phys_addr_t base = ctx->comp->reg_base; 396 u8 subsys_id = ctx->comp->subsys_id; 397 u32 csf_l = 0, csf_r = 0; 398 399 if (CFG_CHECK(MT8183, p_id)) { 400 csf_l = CFG_COMP(MT8183, ctx->param, subfrms[index].in.left); 401 csf_r = CFG_COMP(MT8183, ctx->param, subfrms[index].in.right); 402 } 403 404 if ((csf_r - csf_l + 1) <= 16) 405 MM_REG_WRITE(cmd, subsys_id, base, PRZ_CONTROL_1, 0x0, 406 BIT(27)); 407 } 408 409 return 0; 410 } 411 412 static const struct mdp_comp_ops rsz_ops = { 413 .get_comp_flag = get_comp_flag, 414 .init_comp = init_rsz, 415 .config_frame = config_rsz_frame, 416 .config_subfrm = config_rsz_subfrm, 417 .advance_subfrm = advance_rsz_subfrm, 418 }; 419 420 static int init_wrot(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 421 { 422 phys_addr_t base = ctx->comp->reg_base; 423 u8 subsys_id = ctx->comp->subsys_id; 424 425 /* Reset WROT */ 426 MM_REG_WRITE(cmd, subsys_id, base, VIDO_SOFT_RST, BIT(0), BIT(0)); 427 MM_REG_POLL(cmd, subsys_id, base, VIDO_SOFT_RST_STAT, BIT(0), BIT(0)); 428 MM_REG_WRITE(cmd, subsys_id, base, VIDO_SOFT_RST, 0x0, BIT(0)); 429 MM_REG_POLL(cmd, subsys_id, base, VIDO_SOFT_RST_STAT, 0x0, BIT(0)); 430 return 0; 431 } 432 433 static int config_wrot_frame(struct mdp_comp_ctx *ctx, 434 struct mdp_cmdq_cmd *cmd, 435 const struct v4l2_rect *compose) 436 { 437 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 438 phys_addr_t base = ctx->comp->reg_base; 439 u8 subsys_id = ctx->comp->subsys_id; 440 u32 reg = 0; 441 442 /* Write frame base address */ 443 if (CFG_CHECK(MT8183, p_id)) 444 reg = CFG_COMP(MT8183, ctx->param, wrot.iova[0]); 445 MM_REG_WRITE(cmd, subsys_id, base, VIDO_BASE_ADDR, reg, 446 0xFFFFFFFF); 447 if (CFG_CHECK(MT8183, p_id)) 448 reg = CFG_COMP(MT8183, ctx->param, wrot.iova[1]); 449 MM_REG_WRITE(cmd, subsys_id, base, VIDO_BASE_ADDR_C, reg, 450 0xFFFFFFFF); 451 if (CFG_CHECK(MT8183, p_id)) 452 reg = CFG_COMP(MT8183, ctx->param, wrot.iova[2]); 453 MM_REG_WRITE(cmd, subsys_id, base, VIDO_BASE_ADDR_V, reg, 454 0xFFFFFFFF); 455 /* Write frame related registers */ 456 if (CFG_CHECK(MT8183, p_id)) 457 reg = CFG_COMP(MT8183, ctx->param, wrot.control); 458 MM_REG_WRITE(cmd, subsys_id, base, VIDO_CTRL, reg, 459 0xF131510F); 460 /* Write frame Y pitch */ 461 if (CFG_CHECK(MT8183, p_id)) 462 reg = CFG_COMP(MT8183, ctx->param, wrot.stride[0]); 463 MM_REG_WRITE(cmd, subsys_id, base, VIDO_STRIDE, reg, 464 0x0000FFFF); 465 /* Write frame UV pitch */ 466 if (CFG_CHECK(MT8183, p_id)) 467 reg = CFG_COMP(MT8183, ctx->param, wrot.stride[1]); 468 MM_REG_WRITE(cmd, subsys_id, base, VIDO_STRIDE_C, reg, 469 0xFFFF); 470 if (CFG_CHECK(MT8183, p_id)) 471 reg = CFG_COMP(MT8183, ctx->param, wrot.stride[2]); 472 MM_REG_WRITE(cmd, subsys_id, base, VIDO_STRIDE_V, reg, 473 0xFFFF); 474 /* Write matrix control */ 475 if (CFG_CHECK(MT8183, p_id)) 476 reg = CFG_COMP(MT8183, ctx->param, wrot.mat_ctrl); 477 MM_REG_WRITE(cmd, subsys_id, base, VIDO_MAT_CTRL, reg, 0xF3); 478 479 /* Set the fixed ALPHA as 0xFF */ 480 MM_REG_WRITE(cmd, subsys_id, base, VIDO_DITHER, 0xFF000000, 481 0xFF000000); 482 /* Set VIDO_EOL_SEL */ 483 MM_REG_WRITE(cmd, subsys_id, base, VIDO_RSV_1, BIT(31), BIT(31)); 484 /* Set VIDO_FIFO_TEST */ 485 if (CFG_CHECK(MT8183, p_id)) 486 reg = CFG_COMP(MT8183, ctx->param, wrot.fifo_test); 487 if (reg != 0) 488 MM_REG_WRITE(cmd, subsys_id, base, VIDO_FIFO_TEST, 489 reg, 0xFFF); 490 /* Filter enable */ 491 if (mdp_cfg && mdp_cfg->wrot_filter_constraint) { 492 if (CFG_CHECK(MT8183, p_id)) 493 reg = CFG_COMP(MT8183, ctx->param, wrot.filter); 494 MM_REG_WRITE(cmd, subsys_id, base, VIDO_MAIN_BUF_SIZE, 495 reg, 0x77); 496 } 497 498 return 0; 499 } 500 501 static int config_wrot_subfrm(struct mdp_comp_ctx *ctx, 502 struct mdp_cmdq_cmd *cmd, u32 index) 503 { 504 phys_addr_t base = ctx->comp->reg_base; 505 u8 subsys_id = ctx->comp->subsys_id; 506 u32 reg = 0; 507 508 /* Write Y pixel offset */ 509 if (CFG_CHECK(MT8183, p_id)) 510 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].offset[0]); 511 MM_REG_WRITE(cmd, subsys_id, base, VIDO_OFST_ADDR, 512 reg, 0x0FFFFFFF); 513 /* Write U pixel offset */ 514 if (CFG_CHECK(MT8183, p_id)) 515 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].offset[1]); 516 MM_REG_WRITE(cmd, subsys_id, base, VIDO_OFST_ADDR_C, 517 reg, 0x0FFFFFFF); 518 /* Write V pixel offset */ 519 if (CFG_CHECK(MT8183, p_id)) 520 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].offset[2]); 521 MM_REG_WRITE(cmd, subsys_id, base, VIDO_OFST_ADDR_V, 522 reg, 0x0FFFFFFF); 523 /* Write source size */ 524 if (CFG_CHECK(MT8183, p_id)) 525 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].src); 526 MM_REG_WRITE(cmd, subsys_id, base, VIDO_IN_SIZE, reg, 527 0x1FFF1FFF); 528 /* Write target size */ 529 if (CFG_CHECK(MT8183, p_id)) 530 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].clip); 531 MM_REG_WRITE(cmd, subsys_id, base, VIDO_TAR_SIZE, reg, 532 0x1FFF1FFF); 533 if (CFG_CHECK(MT8183, p_id)) 534 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].clip_ofst); 535 MM_REG_WRITE(cmd, subsys_id, base, VIDO_CROP_OFST, reg, 536 0x1FFF1FFF); 537 538 if (CFG_CHECK(MT8183, p_id)) 539 reg = CFG_COMP(MT8183, ctx->param, wrot.subfrms[index].main_buf); 540 MM_REG_WRITE(cmd, subsys_id, base, VIDO_MAIN_BUF_SIZE, 541 reg, 0x1FFF7F00); 542 543 /* Enable WROT */ 544 MM_REG_WRITE(cmd, subsys_id, base, VIDO_ROT_EN, BIT(0), BIT(0)); 545 546 return 0; 547 } 548 549 static int wait_wrot_event(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 550 { 551 const struct mdp_platform_config *mdp_cfg = __get_plat_cfg(ctx); 552 struct device *dev = &ctx->comp->mdp_dev->pdev->dev; 553 phys_addr_t base = ctx->comp->reg_base; 554 u8 subsys_id = ctx->comp->subsys_id; 555 556 if (ctx->comp->alias_id == 0) 557 MM_REG_WAIT(cmd, ctx->comp->gce_event[MDP_GCE_EVENT_EOF]); 558 else 559 dev_err(dev, "Do not support WROT1_DONE event\n"); 560 561 if (mdp_cfg && mdp_cfg->wrot_filter_constraint) 562 MM_REG_WRITE(cmd, subsys_id, base, VIDO_MAIN_BUF_SIZE, 0x0, 563 0x77); 564 565 /* Disable WROT */ 566 MM_REG_WRITE(cmd, subsys_id, base, VIDO_ROT_EN, 0x0, BIT(0)); 567 568 return 0; 569 } 570 571 static const struct mdp_comp_ops wrot_ops = { 572 .get_comp_flag = get_comp_flag, 573 .init_comp = init_wrot, 574 .config_frame = config_wrot_frame, 575 .config_subfrm = config_wrot_subfrm, 576 .wait_comp_event = wait_wrot_event, 577 }; 578 579 static int init_wdma(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 580 { 581 phys_addr_t base = ctx->comp->reg_base; 582 u8 subsys_id = ctx->comp->subsys_id; 583 584 /* Reset WDMA */ 585 MM_REG_WRITE(cmd, subsys_id, base, WDMA_RST, BIT(0), BIT(0)); 586 MM_REG_POLL(cmd, subsys_id, base, WDMA_FLOW_CTRL_DBG, BIT(0), BIT(0)); 587 MM_REG_WRITE(cmd, subsys_id, base, WDMA_RST, 0x0, BIT(0)); 588 return 0; 589 } 590 591 static int config_wdma_frame(struct mdp_comp_ctx *ctx, 592 struct mdp_cmdq_cmd *cmd, 593 const struct v4l2_rect *compose) 594 { 595 phys_addr_t base = ctx->comp->reg_base; 596 u8 subsys_id = ctx->comp->subsys_id; 597 u32 reg = 0; 598 599 MM_REG_WRITE(cmd, subsys_id, base, WDMA_BUF_CON2, 0x10101050, 600 0xFFFFFFFF); 601 602 /* Setup frame information */ 603 if (CFG_CHECK(MT8183, p_id)) 604 reg = CFG_COMP(MT8183, ctx->param, wdma.wdma_cfg); 605 MM_REG_WRITE(cmd, subsys_id, base, WDMA_CFG, reg, 606 0x0F01B8F0); 607 /* Setup frame base address */ 608 if (CFG_CHECK(MT8183, p_id)) 609 reg = CFG_COMP(MT8183, ctx->param, wdma.iova[0]); 610 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_ADDR, reg, 611 0xFFFFFFFF); 612 if (CFG_CHECK(MT8183, p_id)) 613 reg = CFG_COMP(MT8183, ctx->param, wdma.iova[1]); 614 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_U_ADDR, reg, 615 0xFFFFFFFF); 616 if (CFG_CHECK(MT8183, p_id)) 617 reg = CFG_COMP(MT8183, ctx->param, wdma.iova[2]); 618 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_V_ADDR, reg, 619 0xFFFFFFFF); 620 /* Setup Y pitch */ 621 if (CFG_CHECK(MT8183, p_id)) 622 reg = CFG_COMP(MT8183, ctx->param, wdma.w_in_byte); 623 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_W_IN_BYTE, 624 reg, 0x0000FFFF); 625 /* Setup UV pitch */ 626 if (CFG_CHECK(MT8183, p_id)) 627 reg = CFG_COMP(MT8183, ctx->param, wdma.uv_stride); 628 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_UV_PITCH, 629 reg, 0x0000FFFF); 630 /* Set the fixed ALPHA as 0xFF */ 631 MM_REG_WRITE(cmd, subsys_id, base, WDMA_ALPHA, 0x800000FF, 632 0x800000FF); 633 634 return 0; 635 } 636 637 static int config_wdma_subfrm(struct mdp_comp_ctx *ctx, 638 struct mdp_cmdq_cmd *cmd, u32 index) 639 { 640 phys_addr_t base = ctx->comp->reg_base; 641 u8 subsys_id = ctx->comp->subsys_id; 642 u32 reg = 0; 643 644 /* Write Y pixel offset */ 645 if (CFG_CHECK(MT8183, p_id)) 646 reg = CFG_COMP(MT8183, ctx->param, wdma.subfrms[index].offset[0]); 647 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_ADDR_OFFSET, 648 reg, 0x0FFFFFFF); 649 /* Write U pixel offset */ 650 if (CFG_CHECK(MT8183, p_id)) 651 reg = CFG_COMP(MT8183, ctx->param, wdma.subfrms[index].offset[1]); 652 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_U_ADDR_OFFSET, 653 reg, 0x0FFFFFFF); 654 /* Write V pixel offset */ 655 if (CFG_CHECK(MT8183, p_id)) 656 reg = CFG_COMP(MT8183, ctx->param, wdma.subfrms[index].offset[2]); 657 MM_REG_WRITE(cmd, subsys_id, base, WDMA_DST_V_ADDR_OFFSET, 658 reg, 0x0FFFFFFF); 659 /* Write source size */ 660 if (CFG_CHECK(MT8183, p_id)) 661 reg = CFG_COMP(MT8183, ctx->param, wdma.subfrms[index].src); 662 MM_REG_WRITE(cmd, subsys_id, base, WDMA_SRC_SIZE, reg, 663 0x3FFF3FFF); 664 /* Write target size */ 665 if (CFG_CHECK(MT8183, p_id)) 666 reg = CFG_COMP(MT8183, ctx->param, wdma.subfrms[index].clip); 667 MM_REG_WRITE(cmd, subsys_id, base, WDMA_CLIP_SIZE, reg, 668 0x3FFF3FFF); 669 /* Write clip offset */ 670 if (CFG_CHECK(MT8183, p_id)) 671 reg = CFG_COMP(MT8183, ctx->param, wdma.subfrms[index].clip_ofst); 672 MM_REG_WRITE(cmd, subsys_id, base, WDMA_CLIP_COORD, reg, 673 0x3FFF3FFF); 674 675 /* Enable WDMA */ 676 MM_REG_WRITE(cmd, subsys_id, base, WDMA_EN, BIT(0), BIT(0)); 677 678 return 0; 679 } 680 681 static int wait_wdma_event(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 682 { 683 phys_addr_t base = ctx->comp->reg_base; 684 u8 subsys_id = ctx->comp->subsys_id; 685 686 MM_REG_WAIT(cmd, ctx->comp->gce_event[MDP_GCE_EVENT_EOF]); 687 /* Disable WDMA */ 688 MM_REG_WRITE(cmd, subsys_id, base, WDMA_EN, 0x0, BIT(0)); 689 return 0; 690 } 691 692 static const struct mdp_comp_ops wdma_ops = { 693 .get_comp_flag = get_comp_flag, 694 .init_comp = init_wdma, 695 .config_frame = config_wdma_frame, 696 .config_subfrm = config_wdma_subfrm, 697 .wait_comp_event = wait_wdma_event, 698 }; 699 700 static int init_ccorr(struct mdp_comp_ctx *ctx, struct mdp_cmdq_cmd *cmd) 701 { 702 phys_addr_t base = ctx->comp->reg_base; 703 u8 subsys_id = ctx->comp->subsys_id; 704 705 /* CCORR enable */ 706 MM_REG_WRITE(cmd, subsys_id, base, MDP_CCORR_EN, BIT(0), BIT(0)); 707 /* Relay mode */ 708 MM_REG_WRITE(cmd, subsys_id, base, MDP_CCORR_CFG, BIT(0), BIT(0)); 709 return 0; 710 } 711 712 static int config_ccorr_subfrm(struct mdp_comp_ctx *ctx, 713 struct mdp_cmdq_cmd *cmd, u32 index) 714 { 715 phys_addr_t base = ctx->comp->reg_base; 716 u8 subsys_id = ctx->comp->subsys_id; 717 u32 csf_l = 0, csf_r = 0; 718 u32 csf_t = 0, csf_b = 0; 719 u32 hsize, vsize; 720 721 if (CFG_CHECK(MT8183, p_id)) { 722 csf_l = CFG_COMP(MT8183, ctx->param, subfrms[index].in.left); 723 csf_r = CFG_COMP(MT8183, ctx->param, subfrms[index].in.right); 724 csf_t = CFG_COMP(MT8183, ctx->param, subfrms[index].in.top); 725 csf_b = CFG_COMP(MT8183, ctx->param, subfrms[index].in.bottom); 726 } 727 728 hsize = csf_r - csf_l + 1; 729 vsize = csf_b - csf_t + 1; 730 MM_REG_WRITE(cmd, subsys_id, base, MDP_CCORR_SIZE, 731 (hsize << 16) + (vsize << 0), 0x1FFF1FFF); 732 return 0; 733 } 734 735 static const struct mdp_comp_ops ccorr_ops = { 736 .get_comp_flag = get_comp_flag, 737 .init_comp = init_ccorr, 738 .config_subfrm = config_ccorr_subfrm, 739 }; 740 741 static const struct mdp_comp_ops *mdp_comp_ops[MDP_COMP_TYPE_COUNT] = { 742 [MDP_COMP_TYPE_RDMA] = &rdma_ops, 743 [MDP_COMP_TYPE_RSZ] = &rsz_ops, 744 [MDP_COMP_TYPE_WROT] = &wrot_ops, 745 [MDP_COMP_TYPE_WDMA] = &wdma_ops, 746 [MDP_COMP_TYPE_CCORR] = &ccorr_ops, 747 }; 748 749 static const struct of_device_id mdp_comp_dt_ids[] __maybe_unused = { 750 { 751 .compatible = "mediatek,mt8183-mdp3-rdma", 752 .data = (void *)MDP_COMP_TYPE_RDMA, 753 }, { 754 .compatible = "mediatek,mt8183-mdp3-ccorr", 755 .data = (void *)MDP_COMP_TYPE_CCORR, 756 }, { 757 .compatible = "mediatek,mt8183-mdp3-rsz", 758 .data = (void *)MDP_COMP_TYPE_RSZ, 759 }, { 760 .compatible = "mediatek,mt8183-mdp3-wrot", 761 .data = (void *)MDP_COMP_TYPE_WROT, 762 }, { 763 .compatible = "mediatek,mt8183-mdp3-wdma", 764 .data = (void *)MDP_COMP_TYPE_WDMA, 765 }, 766 {} 767 }; 768 769 static inline bool is_dma_capable(const enum mdp_comp_type type) 770 { 771 return (type == MDP_COMP_TYPE_RDMA || 772 type == MDP_COMP_TYPE_WROT || 773 type == MDP_COMP_TYPE_WDMA); 774 } 775 776 static inline bool is_bypass_gce_event(const enum mdp_comp_type type) 777 { 778 /* 779 * Subcomponent PATH is only used for the direction of data flow and 780 * dose not need to wait for GCE event. 781 */ 782 return (type == MDP_COMP_TYPE_PATH); 783 } 784 785 static int mdp_comp_get_id(struct mdp_dev *mdp, enum mdp_comp_type type, u32 alias_id) 786 { 787 int i; 788 789 for (i = 0; i < mdp->mdp_data->comp_data_len; i++) 790 if (mdp->mdp_data->comp_data[i].match.type == type && 791 mdp->mdp_data->comp_data[i].match.alias_id == alias_id) 792 return i; 793 return -ENODEV; 794 } 795 796 int mdp_comp_clock_on(struct device *dev, struct mdp_comp *comp) 797 { 798 int i, ret; 799 800 /* Only DMA capable components need the pm control */ 801 if (comp->comp_dev && is_dma_capable(comp->type)) { 802 ret = pm_runtime_resume_and_get(comp->comp_dev); 803 if (ret < 0) { 804 dev_err(dev, 805 "Failed to get power, err %d. type:%d id:%d\n", 806 ret, comp->type, comp->inner_id); 807 return ret; 808 } 809 } 810 811 for (i = 0; i < comp->clk_num; i++) { 812 if (IS_ERR_OR_NULL(comp->clks[i])) 813 continue; 814 ret = clk_prepare_enable(comp->clks[i]); 815 if (ret) { 816 dev_err(dev, 817 "Failed to enable clk %d. type:%d id:%d\n", 818 i, comp->type, comp->inner_id); 819 goto err_revert; 820 } 821 } 822 823 return 0; 824 825 err_revert: 826 while (--i >= 0) { 827 if (IS_ERR_OR_NULL(comp->clks[i])) 828 continue; 829 clk_disable_unprepare(comp->clks[i]); 830 } 831 if (comp->comp_dev && is_dma_capable(comp->type)) 832 pm_runtime_put_sync(comp->comp_dev); 833 834 return ret; 835 } 836 837 void mdp_comp_clock_off(struct device *dev, struct mdp_comp *comp) 838 { 839 int i; 840 841 for (i = 0; i < comp->clk_num; i++) { 842 if (IS_ERR_OR_NULL(comp->clks[i])) 843 continue; 844 clk_disable_unprepare(comp->clks[i]); 845 } 846 847 if (comp->comp_dev && is_dma_capable(comp->type)) 848 pm_runtime_put(comp->comp_dev); 849 } 850 851 int mdp_comp_clocks_on(struct device *dev, struct mdp_comp *comps, int num) 852 { 853 int i, ret; 854 855 for (i = 0; i < num; i++) { 856 ret = mdp_comp_clock_on(dev, &comps[i]); 857 if (ret) 858 return ret; 859 } 860 861 return 0; 862 } 863 864 void mdp_comp_clocks_off(struct device *dev, struct mdp_comp *comps, int num) 865 { 866 int i; 867 868 for (i = 0; i < num; i++) 869 mdp_comp_clock_off(dev, &comps[i]); 870 } 871 872 static int mdp_get_subsys_id(struct mdp_dev *mdp, struct device *dev, 873 struct device_node *node, struct mdp_comp *comp) 874 { 875 struct platform_device *comp_pdev; 876 struct cmdq_client_reg cmdq_reg; 877 int ret = 0; 878 int index = 0; 879 880 if (!dev || !node || !comp) 881 return -EINVAL; 882 883 comp_pdev = of_find_device_by_node(node); 884 885 if (!comp_pdev) { 886 dev_err(dev, "get comp_pdev fail! comp public id=%d, inner id=%d, type=%d\n", 887 comp->public_id, comp->inner_id, comp->type); 888 return -ENODEV; 889 } 890 891 index = mdp->mdp_data->comp_data[comp->public_id].info.dts_reg_ofst; 892 ret = cmdq_dev_get_client_reg(&comp_pdev->dev, &cmdq_reg, index); 893 if (ret != 0) { 894 dev_err(&comp_pdev->dev, "cmdq_dev_get_subsys fail!\n"); 895 put_device(&comp_pdev->dev); 896 return -EINVAL; 897 } 898 899 comp->subsys_id = cmdq_reg.subsys; 900 dev_dbg(&comp_pdev->dev, "subsys id=%d\n", cmdq_reg.subsys); 901 put_device(&comp_pdev->dev); 902 903 return 0; 904 } 905 906 static void __mdp_comp_init(struct mdp_dev *mdp, struct device_node *node, 907 struct mdp_comp *comp) 908 { 909 struct resource res; 910 phys_addr_t base; 911 int index; 912 913 index = mdp->mdp_data->comp_data[comp->public_id].info.dts_reg_ofst; 914 if (of_address_to_resource(node, index, &res) < 0) 915 base = 0L; 916 else 917 base = res.start; 918 919 comp->mdp_dev = mdp; 920 comp->regs = of_iomap(node, 0); 921 comp->reg_base = base; 922 } 923 924 static int mdp_comp_init(struct mdp_dev *mdp, struct device_node *node, 925 struct mdp_comp *comp, enum mtk_mdp_comp_id id) 926 { 927 struct device *dev = &mdp->pdev->dev; 928 struct platform_device *pdev_c; 929 int clk_ofst; 930 int i; 931 s32 event; 932 933 if (id < 0 || id >= MDP_MAX_COMP_COUNT) { 934 dev_err(dev, "Invalid component id %d\n", id); 935 return -EINVAL; 936 } 937 938 pdev_c = of_find_device_by_node(node); 939 if (!pdev_c) { 940 dev_warn(dev, "can't find platform device of node:%s\n", 941 node->name); 942 return -ENODEV; 943 } 944 945 comp->comp_dev = &pdev_c->dev; 946 comp->public_id = id; 947 comp->type = mdp->mdp_data->comp_data[id].match.type; 948 comp->inner_id = mdp->mdp_data->comp_data[id].match.inner_id; 949 comp->alias_id = mdp->mdp_data->comp_data[id].match.alias_id; 950 comp->ops = mdp_comp_ops[comp->type]; 951 __mdp_comp_init(mdp, node, comp); 952 953 comp->clk_num = mdp->mdp_data->comp_data[id].info.clk_num; 954 comp->clks = devm_kzalloc(dev, sizeof(struct clk *) * comp->clk_num, 955 GFP_KERNEL); 956 if (!comp->clks) 957 return -ENOMEM; 958 959 clk_ofst = mdp->mdp_data->comp_data[id].info.clk_ofst; 960 961 for (i = 0; i < comp->clk_num; i++) { 962 comp->clks[i] = of_clk_get(node, i + clk_ofst); 963 if (IS_ERR(comp->clks[i])) 964 break; 965 } 966 967 mdp_get_subsys_id(mdp, dev, node, comp); 968 969 /* Set GCE SOF event */ 970 if (is_bypass_gce_event(comp->type) || 971 of_property_read_u32_index(node, "mediatek,gce-events", 972 MDP_GCE_EVENT_SOF, &event)) 973 event = MDP_GCE_NO_EVENT; 974 975 comp->gce_event[MDP_GCE_EVENT_SOF] = event; 976 977 /* Set GCE EOF event */ 978 if (is_dma_capable(comp->type)) { 979 if (of_property_read_u32_index(node, "mediatek,gce-events", 980 MDP_GCE_EVENT_EOF, &event)) { 981 dev_err(dev, "Component id %d has no EOF\n", id); 982 return -EINVAL; 983 } 984 } else { 985 event = MDP_GCE_NO_EVENT; 986 } 987 988 comp->gce_event[MDP_GCE_EVENT_EOF] = event; 989 990 return 0; 991 } 992 993 static void mdp_comp_deinit(struct mdp_comp *comp) 994 { 995 if (!comp) 996 return; 997 998 if (comp->comp_dev && comp->clks) { 999 devm_kfree(&comp->mdp_dev->pdev->dev, comp->clks); 1000 comp->clks = NULL; 1001 } 1002 1003 if (comp->regs) 1004 iounmap(comp->regs); 1005 } 1006 1007 static struct mdp_comp *mdp_comp_create(struct mdp_dev *mdp, 1008 struct device_node *node, 1009 enum mtk_mdp_comp_id id) 1010 { 1011 struct device *dev = &mdp->pdev->dev; 1012 struct mdp_comp *comp; 1013 int ret; 1014 1015 if (mdp->comp[id]) 1016 return ERR_PTR(-EEXIST); 1017 1018 comp = devm_kzalloc(dev, sizeof(*comp), GFP_KERNEL); 1019 if (!comp) 1020 return ERR_PTR(-ENOMEM); 1021 1022 ret = mdp_comp_init(mdp, node, comp, id); 1023 if (ret) { 1024 devm_kfree(dev, comp); 1025 return ERR_PTR(ret); 1026 } 1027 mdp->comp[id] = comp; 1028 mdp->comp[id]->mdp_dev = mdp; 1029 1030 dev_dbg(dev, "%s type:%d alias:%d public id:%d inner id:%d base:%#x regs:%p\n", 1031 dev->of_node->name, comp->type, comp->alias_id, id, comp->inner_id, 1032 (u32)comp->reg_base, comp->regs); 1033 return comp; 1034 } 1035 1036 static int mdp_comp_sub_create(struct mdp_dev *mdp) 1037 { 1038 struct device *dev = &mdp->pdev->dev; 1039 struct device_node *node, *parent; 1040 int ret = 0; 1041 1042 parent = dev->of_node->parent; 1043 1044 for_each_child_of_node(parent, node) { 1045 const struct of_device_id *of_id; 1046 enum mdp_comp_type type; 1047 int id, alias_id; 1048 struct mdp_comp *comp; 1049 1050 of_id = of_match_node(mdp->mdp_data->mdp_sub_comp_dt_ids, node); 1051 if (!of_id) 1052 continue; 1053 if (!of_device_is_available(node)) { 1054 dev_dbg(dev, "Skipping disabled sub comp. %pOF\n", 1055 node); 1056 continue; 1057 } 1058 1059 type = (enum mdp_comp_type)(uintptr_t)of_id->data; 1060 alias_id = mdp_comp_alias_id[type]; 1061 id = mdp_comp_get_id(mdp, type, alias_id); 1062 if (id < 0) { 1063 dev_err(dev, 1064 "Fail to get sub comp. id: type %d alias %d\n", 1065 type, alias_id); 1066 ret = -EINVAL; 1067 goto err_free_node; 1068 } 1069 mdp_comp_alias_id[type]++; 1070 1071 comp = mdp_comp_create(mdp, node, id); 1072 if (IS_ERR(comp)) { 1073 ret = PTR_ERR(comp); 1074 goto err_free_node; 1075 } 1076 } 1077 return ret; 1078 1079 err_free_node: 1080 of_node_put(node); 1081 return ret; 1082 } 1083 1084 void mdp_comp_destroy(struct mdp_dev *mdp) 1085 { 1086 int i; 1087 1088 for (i = 0; i < ARRAY_SIZE(mdp->comp); i++) { 1089 if (mdp->comp[i]) { 1090 if (is_dma_capable(mdp->comp[i]->type)) 1091 pm_runtime_disable(mdp->comp[i]->comp_dev); 1092 mdp_comp_deinit(mdp->comp[i]); 1093 devm_kfree(mdp->comp[i]->comp_dev, mdp->comp[i]); 1094 mdp->comp[i] = NULL; 1095 } 1096 } 1097 } 1098 1099 int mdp_comp_config(struct mdp_dev *mdp) 1100 { 1101 struct device *dev = &mdp->pdev->dev; 1102 struct device_node *node, *parent; 1103 int ret; 1104 1105 memset(mdp_comp_alias_id, 0, sizeof(mdp_comp_alias_id)); 1106 p_id = mdp->mdp_data->mdp_plat_id; 1107 1108 parent = dev->of_node->parent; 1109 /* Iterate over sibling MDP function blocks */ 1110 for_each_child_of_node(parent, node) { 1111 const struct of_device_id *of_id; 1112 enum mdp_comp_type type; 1113 int id, alias_id; 1114 struct mdp_comp *comp; 1115 1116 of_id = of_match_node(mdp_comp_dt_ids, node); 1117 if (!of_id) 1118 continue; 1119 1120 if (!of_device_is_available(node)) { 1121 dev_dbg(dev, "Skipping disabled component %pOF\n", 1122 node); 1123 continue; 1124 } 1125 1126 type = (enum mdp_comp_type)(uintptr_t)of_id->data; 1127 alias_id = mdp_comp_alias_id[type]; 1128 id = mdp_comp_get_id(mdp, type, alias_id); 1129 if (id < 0) { 1130 dev_err(dev, 1131 "Fail to get component id: type %d alias %d\n", 1132 type, alias_id); 1133 continue; 1134 } 1135 mdp_comp_alias_id[type]++; 1136 1137 comp = mdp_comp_create(mdp, node, id); 1138 if (IS_ERR(comp)) { 1139 ret = PTR_ERR(comp); 1140 goto err_init_comps; 1141 } 1142 1143 /* Only DMA capable components need the pm control */ 1144 if (!is_dma_capable(comp->type)) 1145 continue; 1146 pm_runtime_enable(comp->comp_dev); 1147 } 1148 1149 ret = mdp_comp_sub_create(mdp); 1150 if (ret) 1151 goto err_init_comps; 1152 1153 return 0; 1154 1155 err_init_comps: 1156 mdp_comp_destroy(mdp); 1157 return ret; 1158 } 1159 1160 int mdp_comp_ctx_config(struct mdp_dev *mdp, struct mdp_comp_ctx *ctx, 1161 const struct img_compparam *param, 1162 const struct img_ipi_frameparam *frame) 1163 { 1164 struct device *dev = &mdp->pdev->dev; 1165 enum mtk_mdp_comp_id public_id = MDP_COMP_NONE; 1166 u32 arg; 1167 int i, idx; 1168 1169 if (!param) { 1170 dev_err(dev, "Invalid component param"); 1171 return -EINVAL; 1172 } 1173 1174 if (CFG_CHECK(MT8183, p_id)) 1175 arg = CFG_COMP(MT8183, param, type); 1176 else 1177 return -EINVAL; 1178 public_id = mdp_cfg_get_id_public(mdp, arg); 1179 if (public_id < 0) { 1180 dev_err(dev, "Invalid component id %d", public_id); 1181 return -EINVAL; 1182 } 1183 1184 ctx->comp = mdp->comp[public_id]; 1185 if (!ctx->comp) { 1186 dev_err(dev, "Uninit component inner id %d", arg); 1187 return -EINVAL; 1188 } 1189 1190 ctx->param = param; 1191 if (CFG_CHECK(MT8183, p_id)) 1192 arg = CFG_COMP(MT8183, param, input); 1193 else 1194 return -EINVAL; 1195 ctx->input = &frame->inputs[arg]; 1196 if (CFG_CHECK(MT8183, p_id)) 1197 idx = CFG_COMP(MT8183, param, num_outputs); 1198 else 1199 return -EINVAL; 1200 for (i = 0; i < idx; i++) { 1201 if (CFG_CHECK(MT8183, p_id)) 1202 arg = CFG_COMP(MT8183, param, outputs[i]); 1203 else 1204 return -EINVAL; 1205 ctx->outputs[i] = &frame->outputs[arg]; 1206 } 1207 return 0; 1208 } 1209