xref: /openbmc/linux/drivers/gpu/drm/mediatek/mtk_dp.c (revision 39f555fb)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (c) 2019-2022 MediaTek Inc.
4  * Copyright (c) 2022 BayLibre
5  */
6 
7 #include <drm/display/drm_dp_aux_bus.h>
8 #include <drm/display/drm_dp.h>
9 #include <drm/display/drm_dp_helper.h>
10 #include <drm/drm_atomic_helper.h>
11 #include <drm/drm_bridge.h>
12 #include <drm/drm_crtc.h>
13 #include <drm/drm_edid.h>
14 #include <drm/drm_of.h>
15 #include <drm/drm_panel.h>
16 #include <drm/drm_print.h>
17 #include <drm/drm_probe_helper.h>
18 #include <linux/arm-smccc.h>
19 #include <linux/clk.h>
20 #include <linux/delay.h>
21 #include <linux/errno.h>
22 #include <linux/kernel.h>
23 #include <linux/media-bus-format.h>
24 #include <linux/nvmem-consumer.h>
25 #include <linux/of.h>
26 #include <linux/of_irq.h>
27 #include <linux/of_platform.h>
28 #include <linux/phy/phy.h>
29 #include <linux/platform_device.h>
30 #include <linux/pm_runtime.h>
31 #include <linux/regmap.h>
32 #include <linux/soc/mediatek/mtk_sip_svc.h>
33 #include <sound/hdmi-codec.h>
34 #include <video/videomode.h>
35 
36 #include "mtk_dp_reg.h"
37 
38 #define MTK_DP_SIP_CONTROL_AARCH32	MTK_SIP_SMC_CMD(0x523)
39 #define MTK_DP_SIP_ATF_EDP_VIDEO_UNMUTE	(BIT(0) | BIT(5))
40 #define MTK_DP_SIP_ATF_VIDEO_UNMUTE	BIT(5)
41 
42 #define MTK_DP_THREAD_CABLE_STATE_CHG	BIT(0)
43 #define MTK_DP_THREAD_HPD_EVENT		BIT(1)
44 
45 #define MTK_DP_4P1T 4
46 #define MTK_DP_HDE 2
47 #define MTK_DP_PIX_PER_ADDR 2
48 #define MTK_DP_AUX_WAIT_REPLY_COUNT 20
49 #define MTK_DP_TBC_BUF_READ_START_ADDR 0x8
50 #define MTK_DP_TRAIN_VOLTAGE_LEVEL_RETRY 5
51 #define MTK_DP_TRAIN_DOWNSCALE_RETRY 10
52 #define MTK_DP_VERSION 0x11
53 #define MTK_DP_SDP_AUI 0x4
54 
55 enum {
56 	MTK_DP_CAL_GLB_BIAS_TRIM = 0,
57 	MTK_DP_CAL_CLKTX_IMPSE,
58 	MTK_DP_CAL_LN_TX_IMPSEL_PMOS_0,
59 	MTK_DP_CAL_LN_TX_IMPSEL_PMOS_1,
60 	MTK_DP_CAL_LN_TX_IMPSEL_PMOS_2,
61 	MTK_DP_CAL_LN_TX_IMPSEL_PMOS_3,
62 	MTK_DP_CAL_LN_TX_IMPSEL_NMOS_0,
63 	MTK_DP_CAL_LN_TX_IMPSEL_NMOS_1,
64 	MTK_DP_CAL_LN_TX_IMPSEL_NMOS_2,
65 	MTK_DP_CAL_LN_TX_IMPSEL_NMOS_3,
66 	MTK_DP_CAL_MAX,
67 };
68 
69 struct mtk_dp_train_info {
70 	bool sink_ssc;
71 	bool cable_plugged_in;
72 	/* link_rate is in multiple of 0.27Gbps */
73 	int link_rate;
74 	int lane_count;
75 	unsigned int channel_eq_pattern;
76 };
77 
78 struct mtk_dp_audio_cfg {
79 	bool detect_monitor;
80 	int sad_count;
81 	int sample_rate;
82 	int word_length_bits;
83 	int channels;
84 };
85 
86 struct mtk_dp_info {
87 	enum dp_pixelformat format;
88 	struct videomode vm;
89 	struct mtk_dp_audio_cfg audio_cur_cfg;
90 };
91 
92 struct mtk_dp_efuse_fmt {
93 	unsigned short idx;
94 	unsigned short shift;
95 	unsigned short mask;
96 	unsigned short min_val;
97 	unsigned short max_val;
98 	unsigned short default_val;
99 };
100 
101 struct mtk_dp {
102 	bool enabled;
103 	bool need_debounce;
104 	int irq;
105 	u8 max_lanes;
106 	u8 max_linkrate;
107 	u8 rx_cap[DP_RECEIVER_CAP_SIZE];
108 	u32 cal_data[MTK_DP_CAL_MAX];
109 	u32 irq_thread_handle;
110 	/* irq_thread_lock is used to protect irq_thread_handle */
111 	spinlock_t irq_thread_lock;
112 
113 	struct device *dev;
114 	struct drm_bridge bridge;
115 	struct drm_bridge *next_bridge;
116 	struct drm_connector *conn;
117 	struct drm_device *drm_dev;
118 	struct drm_dp_aux aux;
119 
120 	const struct mtk_dp_data *data;
121 	struct mtk_dp_info info;
122 	struct mtk_dp_train_info train_info;
123 
124 	struct platform_device *phy_dev;
125 	struct phy *phy;
126 	struct regmap *regs;
127 	struct timer_list debounce_timer;
128 
129 	/* For audio */
130 	bool audio_enable;
131 	hdmi_codec_plugged_cb plugged_cb;
132 	struct platform_device *audio_pdev;
133 
134 	struct device *codec_dev;
135 	/* protect the plugged_cb as it's used in both bridge ops and audio */
136 	struct mutex update_plugged_status_lock;
137 };
138 
139 struct mtk_dp_data {
140 	int bridge_type;
141 	unsigned int smc_cmd;
142 	const struct mtk_dp_efuse_fmt *efuse_fmt;
143 	bool audio_supported;
144 };
145 
146 static const struct mtk_dp_efuse_fmt mt8195_edp_efuse_fmt[MTK_DP_CAL_MAX] = {
147 	[MTK_DP_CAL_GLB_BIAS_TRIM] = {
148 		.idx = 3,
149 		.shift = 27,
150 		.mask = 0x1f,
151 		.min_val = 1,
152 		.max_val = 0x1e,
153 		.default_val = 0xf,
154 	},
155 	[MTK_DP_CAL_CLKTX_IMPSE] = {
156 		.idx = 0,
157 		.shift = 9,
158 		.mask = 0xf,
159 		.min_val = 1,
160 		.max_val = 0xe,
161 		.default_val = 0x8,
162 	},
163 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_0] = {
164 		.idx = 2,
165 		.shift = 28,
166 		.mask = 0xf,
167 		.min_val = 1,
168 		.max_val = 0xe,
169 		.default_val = 0x8,
170 	},
171 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_1] = {
172 		.idx = 2,
173 		.shift = 20,
174 		.mask = 0xf,
175 		.min_val = 1,
176 		.max_val = 0xe,
177 		.default_val = 0x8,
178 	},
179 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_2] = {
180 		.idx = 2,
181 		.shift = 12,
182 		.mask = 0xf,
183 		.min_val = 1,
184 		.max_val = 0xe,
185 		.default_val = 0x8,
186 	},
187 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_3] = {
188 		.idx = 2,
189 		.shift = 4,
190 		.mask = 0xf,
191 		.min_val = 1,
192 		.max_val = 0xe,
193 		.default_val = 0x8,
194 	},
195 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_0] = {
196 		.idx = 2,
197 		.shift = 24,
198 		.mask = 0xf,
199 		.min_val = 1,
200 		.max_val = 0xe,
201 		.default_val = 0x8,
202 	},
203 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_1] = {
204 		.idx = 2,
205 		.shift = 16,
206 		.mask = 0xf,
207 		.min_val = 1,
208 		.max_val = 0xe,
209 		.default_val = 0x8,
210 	},
211 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_2] = {
212 		.idx = 2,
213 		.shift = 8,
214 		.mask = 0xf,
215 		.min_val = 1,
216 		.max_val = 0xe,
217 		.default_val = 0x8,
218 	},
219 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_3] = {
220 		.idx = 2,
221 		.shift = 0,
222 		.mask = 0xf,
223 		.min_val = 1,
224 		.max_val = 0xe,
225 		.default_val = 0x8,
226 	},
227 };
228 
229 static const struct mtk_dp_efuse_fmt mt8195_dp_efuse_fmt[MTK_DP_CAL_MAX] = {
230 	[MTK_DP_CAL_GLB_BIAS_TRIM] = {
231 		.idx = 0,
232 		.shift = 27,
233 		.mask = 0x1f,
234 		.min_val = 1,
235 		.max_val = 0x1e,
236 		.default_val = 0xf,
237 	},
238 	[MTK_DP_CAL_CLKTX_IMPSE] = {
239 		.idx = 0,
240 		.shift = 13,
241 		.mask = 0xf,
242 		.min_val = 1,
243 		.max_val = 0xe,
244 		.default_val = 0x8,
245 	},
246 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_0] = {
247 		.idx = 1,
248 		.shift = 28,
249 		.mask = 0xf,
250 		.min_val = 1,
251 		.max_val = 0xe,
252 		.default_val = 0x8,
253 	},
254 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_1] = {
255 		.idx = 1,
256 		.shift = 20,
257 		.mask = 0xf,
258 		.min_val = 1,
259 		.max_val = 0xe,
260 		.default_val = 0x8,
261 	},
262 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_2] = {
263 		.idx = 1,
264 		.shift = 12,
265 		.mask = 0xf,
266 		.min_val = 1,
267 		.max_val = 0xe,
268 		.default_val = 0x8,
269 	},
270 	[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_3] = {
271 		.idx = 1,
272 		.shift = 4,
273 		.mask = 0xf,
274 		.min_val = 1,
275 		.max_val = 0xe,
276 		.default_val = 0x8,
277 	},
278 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_0] = {
279 		.idx = 1,
280 		.shift = 24,
281 		.mask = 0xf,
282 		.min_val = 1,
283 		.max_val = 0xe,
284 		.default_val = 0x8,
285 	},
286 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_1] = {
287 		.idx = 1,
288 		.shift = 16,
289 		.mask = 0xf,
290 		.min_val = 1,
291 		.max_val = 0xe,
292 		.default_val = 0x8,
293 	},
294 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_2] = {
295 		.idx = 1,
296 		.shift = 8,
297 		.mask = 0xf,
298 		.min_val = 1,
299 		.max_val = 0xe,
300 		.default_val = 0x8,
301 	},
302 	[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_3] = {
303 		.idx = 1,
304 		.shift = 0,
305 		.mask = 0xf,
306 		.min_val = 1,
307 		.max_val = 0xe,
308 		.default_val = 0x8,
309 	},
310 };
311 
312 static struct regmap_config mtk_dp_regmap_config = {
313 	.reg_bits = 32,
314 	.val_bits = 32,
315 	.reg_stride = 4,
316 	.max_register = SEC_OFFSET + 0x90,
317 	.name = "mtk-dp-registers",
318 };
319 
320 static struct mtk_dp *mtk_dp_from_bridge(struct drm_bridge *b)
321 {
322 	return container_of(b, struct mtk_dp, bridge);
323 }
324 
325 static u32 mtk_dp_read(struct mtk_dp *mtk_dp, u32 offset)
326 {
327 	u32 read_val;
328 	int ret;
329 
330 	ret = regmap_read(mtk_dp->regs, offset, &read_val);
331 	if (ret) {
332 		dev_err(mtk_dp->dev, "Failed to read register 0x%x: %d\n",
333 			offset, ret);
334 		return 0;
335 	}
336 
337 	return read_val;
338 }
339 
340 static int mtk_dp_write(struct mtk_dp *mtk_dp, u32 offset, u32 val)
341 {
342 	int ret = regmap_write(mtk_dp->regs, offset, val);
343 
344 	if (ret)
345 		dev_err(mtk_dp->dev,
346 			"Failed to write register 0x%x with value 0x%x\n",
347 			offset, val);
348 	return ret;
349 }
350 
351 static int mtk_dp_update_bits(struct mtk_dp *mtk_dp, u32 offset,
352 			      u32 val, u32 mask)
353 {
354 	int ret = regmap_update_bits(mtk_dp->regs, offset, mask, val);
355 
356 	if (ret)
357 		dev_err(mtk_dp->dev,
358 			"Failed to update register 0x%x with value 0x%x, mask 0x%x\n",
359 			offset, val, mask);
360 	return ret;
361 }
362 
363 static void mtk_dp_bulk_16bit_write(struct mtk_dp *mtk_dp, u32 offset, u8 *buf,
364 				    size_t length)
365 {
366 	int i;
367 
368 	/* 2 bytes per register */
369 	for (i = 0; i < length; i += 2) {
370 		u32 val = buf[i] | (i + 1 < length ? buf[i + 1] << 8 : 0);
371 
372 		if (mtk_dp_write(mtk_dp, offset + i * 2, val))
373 			return;
374 	}
375 }
376 
377 static void mtk_dp_msa_bypass_enable(struct mtk_dp *mtk_dp, bool enable)
378 {
379 	u32 mask = HTOTAL_SEL_DP_ENC0_P0 | VTOTAL_SEL_DP_ENC0_P0 |
380 		   HSTART_SEL_DP_ENC0_P0 | VSTART_SEL_DP_ENC0_P0 |
381 		   HWIDTH_SEL_DP_ENC0_P0 | VHEIGHT_SEL_DP_ENC0_P0 |
382 		   HSP_SEL_DP_ENC0_P0 | HSW_SEL_DP_ENC0_P0 |
383 		   VSP_SEL_DP_ENC0_P0 | VSW_SEL_DP_ENC0_P0;
384 
385 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3030, enable ? 0 : mask, mask);
386 }
387 
388 static void mtk_dp_set_msa(struct mtk_dp *mtk_dp)
389 {
390 	struct drm_display_mode mode;
391 	struct videomode *vm = &mtk_dp->info.vm;
392 
393 	drm_display_mode_from_videomode(vm, &mode);
394 
395 	/* horizontal */
396 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3010,
397 			   mode.htotal, HTOTAL_SW_DP_ENC0_P0_MASK);
398 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3018,
399 			   vm->hsync_len + vm->hback_porch,
400 			   HSTART_SW_DP_ENC0_P0_MASK);
401 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3028,
402 			   vm->hsync_len, HSW_SW_DP_ENC0_P0_MASK);
403 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3028,
404 			   0, HSP_SW_DP_ENC0_P0_MASK);
405 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3020,
406 			   vm->hactive, HWIDTH_SW_DP_ENC0_P0_MASK);
407 
408 	/* vertical */
409 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3014,
410 			   mode.vtotal, VTOTAL_SW_DP_ENC0_P0_MASK);
411 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_301C,
412 			   vm->vsync_len + vm->vback_porch,
413 			   VSTART_SW_DP_ENC0_P0_MASK);
414 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_302C,
415 			   vm->vsync_len, VSW_SW_DP_ENC0_P0_MASK);
416 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_302C,
417 			   0, VSP_SW_DP_ENC0_P0_MASK);
418 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3024,
419 			   vm->vactive, VHEIGHT_SW_DP_ENC0_P0_MASK);
420 
421 	/* horizontal */
422 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3064,
423 			   vm->hactive, HDE_NUM_LAST_DP_ENC0_P0_MASK);
424 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3154,
425 			   mode.htotal, PGEN_HTOTAL_DP_ENC0_P0_MASK);
426 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3158,
427 			   vm->hfront_porch,
428 			   PGEN_HSYNC_RISING_DP_ENC0_P0_MASK);
429 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_315C,
430 			   vm->hsync_len,
431 			   PGEN_HSYNC_PULSE_WIDTH_DP_ENC0_P0_MASK);
432 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3160,
433 			   vm->hback_porch + vm->hsync_len,
434 			   PGEN_HFDE_START_DP_ENC0_P0_MASK);
435 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3164,
436 			   vm->hactive,
437 			   PGEN_HFDE_ACTIVE_WIDTH_DP_ENC0_P0_MASK);
438 
439 	/* vertical */
440 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3168,
441 			   mode.vtotal,
442 			   PGEN_VTOTAL_DP_ENC0_P0_MASK);
443 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_316C,
444 			   vm->vfront_porch,
445 			   PGEN_VSYNC_RISING_DP_ENC0_P0_MASK);
446 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3170,
447 			   vm->vsync_len,
448 			   PGEN_VSYNC_PULSE_WIDTH_DP_ENC0_P0_MASK);
449 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3174,
450 			   vm->vback_porch + vm->vsync_len,
451 			   PGEN_VFDE_START_DP_ENC0_P0_MASK);
452 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3178,
453 			   vm->vactive,
454 			   PGEN_VFDE_ACTIVE_WIDTH_DP_ENC0_P0_MASK);
455 }
456 
457 static int mtk_dp_set_color_format(struct mtk_dp *mtk_dp,
458 				   enum dp_pixelformat color_format)
459 {
460 	u32 val;
461 
462 	/* update MISC0 */
463 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3034,
464 			   color_format << DP_TEST_COLOR_FORMAT_SHIFT,
465 			   DP_TEST_COLOR_FORMAT_MASK);
466 
467 	switch (color_format) {
468 	case DP_PIXELFORMAT_YUV422:
469 		val = PIXEL_ENCODE_FORMAT_DP_ENC0_P0_YCBCR422;
470 		break;
471 	case DP_PIXELFORMAT_RGB:
472 		val = PIXEL_ENCODE_FORMAT_DP_ENC0_P0_RGB;
473 		break;
474 	default:
475 		drm_warn(mtk_dp->drm_dev, "Unsupported color format: %d\n",
476 			 color_format);
477 		return -EINVAL;
478 	}
479 
480 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_303C,
481 			   val, PIXEL_ENCODE_FORMAT_DP_ENC0_P0_MASK);
482 	return 0;
483 }
484 
485 static void mtk_dp_set_color_depth(struct mtk_dp *mtk_dp)
486 {
487 	/* Only support 8 bits currently */
488 	/* Update MISC0 */
489 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3034,
490 			   DP_MSA_MISC_8_BPC, DP_TEST_BIT_DEPTH_MASK);
491 
492 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_303C,
493 			   VIDEO_COLOR_DEPTH_DP_ENC0_P0_8BIT,
494 			   VIDEO_COLOR_DEPTH_DP_ENC0_P0_MASK);
495 }
496 
497 static void mtk_dp_config_mn_mode(struct mtk_dp *mtk_dp)
498 {
499 	/* 0: hw mode, 1: sw mode */
500 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3004,
501 			   0, VIDEO_M_CODE_SEL_DP_ENC0_P0_MASK);
502 }
503 
504 static void mtk_dp_set_sram_read_start(struct mtk_dp *mtk_dp, u32 val)
505 {
506 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_303C,
507 			   val, SRAM_START_READ_THRD_DP_ENC0_P0_MASK);
508 }
509 
510 static void mtk_dp_setup_encoder(struct mtk_dp *mtk_dp)
511 {
512 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_303C,
513 			   VIDEO_MN_GEN_EN_DP_ENC0_P0,
514 			   VIDEO_MN_GEN_EN_DP_ENC0_P0);
515 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3040,
516 			   SDP_DOWN_CNT_DP_ENC0_P0_VAL,
517 			   SDP_DOWN_CNT_INIT_DP_ENC0_P0_MASK);
518 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3364,
519 			   SDP_DOWN_CNT_IN_HBLANK_DP_ENC1_P0_VAL,
520 			   SDP_DOWN_CNT_INIT_IN_HBLANK_DP_ENC1_P0_MASK);
521 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3300,
522 			   VIDEO_AFIFO_RDY_SEL_DP_ENC1_P0_VAL << 8,
523 			   VIDEO_AFIFO_RDY_SEL_DP_ENC1_P0_MASK);
524 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3364,
525 			   FIFO_READ_START_POINT_DP_ENC1_P0_VAL << 12,
526 			   FIFO_READ_START_POINT_DP_ENC1_P0_MASK);
527 	mtk_dp_write(mtk_dp, MTK_DP_ENC1_P0_3368, DP_ENC1_P0_3368_VAL);
528 }
529 
530 static void mtk_dp_pg_enable(struct mtk_dp *mtk_dp, bool enable)
531 {
532 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3038,
533 			   enable ? VIDEO_SOURCE_SEL_DP_ENC0_P0_MASK : 0,
534 			   VIDEO_SOURCE_SEL_DP_ENC0_P0_MASK);
535 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_31B0,
536 			   PGEN_PATTERN_SEL_VAL << 4, PGEN_PATTERN_SEL_MASK);
537 }
538 
539 static void mtk_dp_audio_setup_channels(struct mtk_dp *mtk_dp,
540 					struct mtk_dp_audio_cfg *cfg)
541 {
542 	u32 channel_enable_bits;
543 
544 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3324,
545 			   AUDIO_SOURCE_MUX_DP_ENC1_P0_DPRX,
546 			   AUDIO_SOURCE_MUX_DP_ENC1_P0_MASK);
547 
548 	/* audio channel count change reset */
549 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_33F4,
550 			   DP_ENC_DUMMY_RW_1, DP_ENC_DUMMY_RW_1);
551 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3304,
552 			   AU_PRTY_REGEN_DP_ENC1_P0_MASK |
553 			   AU_CH_STS_REGEN_DP_ENC1_P0_MASK |
554 			   AUDIO_SAMPLE_PRSENT_REGEN_DP_ENC1_P0_MASK,
555 			   AU_PRTY_REGEN_DP_ENC1_P0_MASK |
556 			   AU_CH_STS_REGEN_DP_ENC1_P0_MASK |
557 			   AUDIO_SAMPLE_PRSENT_REGEN_DP_ENC1_P0_MASK);
558 
559 	switch (cfg->channels) {
560 	case 2:
561 		channel_enable_bits = AUDIO_2CH_SEL_DP_ENC0_P0_MASK |
562 				      AUDIO_2CH_EN_DP_ENC0_P0_MASK;
563 		break;
564 	case 8:
565 	default:
566 		channel_enable_bits = AUDIO_8CH_SEL_DP_ENC0_P0_MASK |
567 				      AUDIO_8CH_EN_DP_ENC0_P0_MASK;
568 		break;
569 	}
570 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3088,
571 			   channel_enable_bits | AU_EN_DP_ENC0_P0,
572 			   AUDIO_2CH_SEL_DP_ENC0_P0_MASK |
573 			   AUDIO_2CH_EN_DP_ENC0_P0_MASK |
574 			   AUDIO_8CH_SEL_DP_ENC0_P0_MASK |
575 			   AUDIO_8CH_EN_DP_ENC0_P0_MASK |
576 			   AU_EN_DP_ENC0_P0);
577 
578 	/* audio channel count change reset */
579 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_33F4, 0, DP_ENC_DUMMY_RW_1);
580 
581 	/* enable audio reset */
582 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_33F4,
583 			   DP_ENC_DUMMY_RW_1_AUDIO_RST_EN,
584 			   DP_ENC_DUMMY_RW_1_AUDIO_RST_EN);
585 }
586 
587 static void mtk_dp_audio_channel_status_set(struct mtk_dp *mtk_dp,
588 					    struct mtk_dp_audio_cfg *cfg)
589 {
590 	struct snd_aes_iec958 iec = { 0 };
591 
592 	switch (cfg->sample_rate) {
593 	case 32000:
594 		iec.status[3] = IEC958_AES3_CON_FS_32000;
595 		break;
596 	case 44100:
597 		iec.status[3] = IEC958_AES3_CON_FS_44100;
598 		break;
599 	case 48000:
600 		iec.status[3] = IEC958_AES3_CON_FS_48000;
601 		break;
602 	case 88200:
603 		iec.status[3] = IEC958_AES3_CON_FS_88200;
604 		break;
605 	case 96000:
606 		iec.status[3] = IEC958_AES3_CON_FS_96000;
607 		break;
608 	case 192000:
609 		iec.status[3] = IEC958_AES3_CON_FS_192000;
610 		break;
611 	default:
612 		iec.status[3] = IEC958_AES3_CON_FS_NOTID;
613 		break;
614 	}
615 
616 	switch (cfg->word_length_bits) {
617 	case 16:
618 		iec.status[4] = IEC958_AES4_CON_WORDLEN_20_16;
619 		break;
620 	case 20:
621 		iec.status[4] = IEC958_AES4_CON_WORDLEN_20_16 |
622 				IEC958_AES4_CON_MAX_WORDLEN_24;
623 		break;
624 	case 24:
625 		iec.status[4] = IEC958_AES4_CON_WORDLEN_24_20 |
626 				IEC958_AES4_CON_MAX_WORDLEN_24;
627 		break;
628 	default:
629 		iec.status[4] = IEC958_AES4_CON_WORDLEN_NOTID;
630 	}
631 
632 	/* IEC 60958 consumer channel status bits */
633 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_308C,
634 			   0, CH_STATUS_0_DP_ENC0_P0_MASK);
635 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3090,
636 			   iec.status[3] << 8, CH_STATUS_1_DP_ENC0_P0_MASK);
637 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3094,
638 			   iec.status[4], CH_STATUS_2_DP_ENC0_P0_MASK);
639 }
640 
641 static void mtk_dp_audio_sdp_asp_set_channels(struct mtk_dp *mtk_dp,
642 					      int channels)
643 {
644 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_312C,
645 			   (min(8, channels) - 1) << 8,
646 			   ASP_HB2_DP_ENC0_P0_MASK | ASP_HB3_DP_ENC0_P0_MASK);
647 }
648 
649 static void mtk_dp_audio_set_divider(struct mtk_dp *mtk_dp)
650 {
651 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_30BC,
652 			   AUDIO_M_CODE_MULT_DIV_SEL_DP_ENC0_P0_DIV_2,
653 			   AUDIO_M_CODE_MULT_DIV_SEL_DP_ENC0_P0_MASK);
654 }
655 
656 static void mtk_dp_sdp_trigger_aui(struct mtk_dp *mtk_dp)
657 {
658 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3280,
659 			   MTK_DP_SDP_AUI, SDP_PACKET_TYPE_DP_ENC1_P0_MASK);
660 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3280,
661 			   SDP_PACKET_W_DP_ENC1_P0, SDP_PACKET_W_DP_ENC1_P0);
662 }
663 
664 static void mtk_dp_sdp_set_data(struct mtk_dp *mtk_dp, u8 *data_bytes)
665 {
666 	mtk_dp_bulk_16bit_write(mtk_dp, MTK_DP_ENC1_P0_3200,
667 				data_bytes, 0x10);
668 }
669 
670 static void mtk_dp_sdp_set_header_aui(struct mtk_dp *mtk_dp,
671 				      struct dp_sdp_header *header)
672 {
673 	u32 db_addr = MTK_DP_ENC0_P0_30D8 + (MTK_DP_SDP_AUI - 1) * 8;
674 
675 	mtk_dp_bulk_16bit_write(mtk_dp, db_addr, (u8 *)header, 4);
676 }
677 
678 static void mtk_dp_disable_sdp_aui(struct mtk_dp *mtk_dp)
679 {
680 	/* Disable periodic send */
681 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_30A8 & 0xfffc, 0,
682 			   0xff << ((MTK_DP_ENC0_P0_30A8 & 3) * 8));
683 }
684 
685 static void mtk_dp_setup_sdp_aui(struct mtk_dp *mtk_dp,
686 				 struct dp_sdp *sdp)
687 {
688 	u32 shift;
689 
690 	mtk_dp_sdp_set_data(mtk_dp, sdp->db);
691 	mtk_dp_sdp_set_header_aui(mtk_dp, &sdp->sdp_header);
692 	mtk_dp_disable_sdp_aui(mtk_dp);
693 
694 	shift = (MTK_DP_ENC0_P0_30A8 & 3) * 8;
695 
696 	mtk_dp_sdp_trigger_aui(mtk_dp);
697 	/* Enable periodic sending */
698 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_30A8 & 0xfffc,
699 			   0x05 << shift, 0xff << shift);
700 }
701 
702 static void mtk_dp_aux_irq_clear(struct mtk_dp *mtk_dp)
703 {
704 	mtk_dp_write(mtk_dp, MTK_DP_AUX_P0_3640, DP_AUX_P0_3640_VAL);
705 }
706 
707 static void mtk_dp_aux_set_cmd(struct mtk_dp *mtk_dp, u8 cmd, u32 addr)
708 {
709 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3644,
710 			   cmd, MCU_REQUEST_COMMAND_AUX_TX_P0_MASK);
711 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3648,
712 			   addr, MCU_REQUEST_ADDRESS_LSB_AUX_TX_P0_MASK);
713 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_364C,
714 			   addr >> 16, MCU_REQUEST_ADDRESS_MSB_AUX_TX_P0_MASK);
715 }
716 
717 static void mtk_dp_aux_clear_fifo(struct mtk_dp *mtk_dp)
718 {
719 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3650,
720 			   MCU_ACK_TRAN_COMPLETE_AUX_TX_P0,
721 			   MCU_ACK_TRAN_COMPLETE_AUX_TX_P0 |
722 			   PHY_FIFO_RST_AUX_TX_P0_MASK |
723 			   MCU_REQ_DATA_NUM_AUX_TX_P0_MASK);
724 }
725 
726 static void mtk_dp_aux_request_ready(struct mtk_dp *mtk_dp)
727 {
728 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3630,
729 			   AUX_TX_REQUEST_READY_AUX_TX_P0,
730 			   AUX_TX_REQUEST_READY_AUX_TX_P0);
731 }
732 
733 static void mtk_dp_aux_fill_write_fifo(struct mtk_dp *mtk_dp, u8 *buf,
734 				       size_t length)
735 {
736 	mtk_dp_bulk_16bit_write(mtk_dp, MTK_DP_AUX_P0_3708, buf, length);
737 }
738 
739 static void mtk_dp_aux_read_rx_fifo(struct mtk_dp *mtk_dp, u8 *buf,
740 				    size_t length, int read_delay)
741 {
742 	int read_pos;
743 
744 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3620,
745 			   0, AUX_RD_MODE_AUX_TX_P0_MASK);
746 
747 	for (read_pos = 0; read_pos < length; read_pos++) {
748 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3620,
749 				   AUX_RX_FIFO_READ_PULSE_TX_P0,
750 				   AUX_RX_FIFO_READ_PULSE_TX_P0);
751 
752 		/* Hardware needs time to update the data */
753 		usleep_range(read_delay, read_delay * 2);
754 		buf[read_pos] = (u8)(mtk_dp_read(mtk_dp, MTK_DP_AUX_P0_3620) &
755 				     AUX_RX_FIFO_READ_DATA_AUX_TX_P0_MASK);
756 	}
757 }
758 
759 static void mtk_dp_aux_set_length(struct mtk_dp *mtk_dp, size_t length)
760 {
761 	if (length > 0) {
762 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3650,
763 				   (length - 1) << 12,
764 				   MCU_REQ_DATA_NUM_AUX_TX_P0_MASK);
765 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_362C,
766 				   0,
767 				   AUX_NO_LENGTH_AUX_TX_P0 |
768 				   AUX_TX_AUXTX_OV_EN_AUX_TX_P0_MASK |
769 				   AUX_RESERVED_RW_0_AUX_TX_P0_MASK);
770 	} else {
771 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_362C,
772 				   AUX_NO_LENGTH_AUX_TX_P0,
773 				   AUX_NO_LENGTH_AUX_TX_P0 |
774 				   AUX_TX_AUXTX_OV_EN_AUX_TX_P0_MASK |
775 				   AUX_RESERVED_RW_0_AUX_TX_P0_MASK);
776 	}
777 }
778 
779 static int mtk_dp_aux_wait_for_completion(struct mtk_dp *mtk_dp, bool is_read)
780 {
781 	int wait_reply = MTK_DP_AUX_WAIT_REPLY_COUNT;
782 
783 	while (--wait_reply) {
784 		u32 aux_irq_status;
785 
786 		if (is_read) {
787 			u32 fifo_status = mtk_dp_read(mtk_dp, MTK_DP_AUX_P0_3618);
788 
789 			if (fifo_status &
790 			    (AUX_RX_FIFO_WRITE_POINTER_AUX_TX_P0_MASK |
791 			     AUX_RX_FIFO_FULL_AUX_TX_P0_MASK)) {
792 				return 0;
793 			}
794 		}
795 
796 		aux_irq_status = mtk_dp_read(mtk_dp, MTK_DP_AUX_P0_3640);
797 		if (aux_irq_status & AUX_RX_AUX_RECV_COMPLETE_IRQ_AUX_TX_P0)
798 			return 0;
799 
800 		if (aux_irq_status & AUX_400US_TIMEOUT_IRQ_AUX_TX_P0)
801 			return -ETIMEDOUT;
802 
803 		/* Give the hardware a chance to reach completion before retrying */
804 		usleep_range(100, 500);
805 	}
806 
807 	return -ETIMEDOUT;
808 }
809 
810 static int mtk_dp_aux_do_transfer(struct mtk_dp *mtk_dp, bool is_read, u8 cmd,
811 				  u32 addr, u8 *buf, size_t length, u8 *reply_cmd)
812 {
813 	int ret;
814 
815 	if (is_read && (length > DP_AUX_MAX_PAYLOAD_BYTES ||
816 			(cmd == DP_AUX_NATIVE_READ && !length)))
817 		return -EINVAL;
818 
819 	if (!is_read)
820 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3704,
821 				   AUX_TX_FIFO_NEW_MODE_EN_AUX_TX_P0,
822 				   AUX_TX_FIFO_NEW_MODE_EN_AUX_TX_P0);
823 
824 	/* We need to clear fifo and irq before sending commands to the sink device. */
825 	mtk_dp_aux_clear_fifo(mtk_dp);
826 	mtk_dp_aux_irq_clear(mtk_dp);
827 
828 	mtk_dp_aux_set_cmd(mtk_dp, cmd, addr);
829 	mtk_dp_aux_set_length(mtk_dp, length);
830 
831 	if (!is_read) {
832 		if (length)
833 			mtk_dp_aux_fill_write_fifo(mtk_dp, buf, length);
834 
835 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3704,
836 				   AUX_TX_FIFO_WDATA_NEW_MODE_T_AUX_TX_P0_MASK,
837 				   AUX_TX_FIFO_WDATA_NEW_MODE_T_AUX_TX_P0_MASK);
838 	}
839 
840 	mtk_dp_aux_request_ready(mtk_dp);
841 
842 	/* Wait for feedback from sink device. */
843 	ret = mtk_dp_aux_wait_for_completion(mtk_dp, is_read);
844 
845 	*reply_cmd = mtk_dp_read(mtk_dp, MTK_DP_AUX_P0_3624) &
846 		     AUX_RX_REPLY_COMMAND_AUX_TX_P0_MASK;
847 
848 	if (ret) {
849 		u32 phy_status = mtk_dp_read(mtk_dp, MTK_DP_AUX_P0_3628) &
850 				 AUX_RX_PHY_STATE_AUX_TX_P0_MASK;
851 		if (phy_status != AUX_RX_PHY_STATE_AUX_TX_P0_RX_IDLE) {
852 			dev_err(mtk_dp->dev,
853 				"AUX Rx Aux hang, need SW reset\n");
854 			return -EIO;
855 		}
856 
857 		return -ETIMEDOUT;
858 	}
859 
860 	if (!length) {
861 		mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_362C,
862 				   0,
863 				   AUX_NO_LENGTH_AUX_TX_P0 |
864 				   AUX_TX_AUXTX_OV_EN_AUX_TX_P0_MASK |
865 				   AUX_RESERVED_RW_0_AUX_TX_P0_MASK);
866 	} else if (is_read) {
867 		int read_delay;
868 
869 		if (cmd == (DP_AUX_I2C_READ | DP_AUX_I2C_MOT) ||
870 		    cmd == DP_AUX_I2C_READ)
871 			read_delay = 500;
872 		else
873 			read_delay = 100;
874 
875 		mtk_dp_aux_read_rx_fifo(mtk_dp, buf, length, read_delay);
876 	}
877 
878 	return 0;
879 }
880 
881 static void mtk_dp_set_swing_pre_emphasis(struct mtk_dp *mtk_dp, int lane_num,
882 					  int swing_val, int preemphasis)
883 {
884 	u32 lane_shift = lane_num * DP_TX1_VOLT_SWING_SHIFT;
885 
886 	dev_dbg(mtk_dp->dev,
887 		"link training: swing_val = 0x%x, pre-emphasis = 0x%x\n",
888 		swing_val, preemphasis);
889 
890 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_SWING_EMP,
891 			   swing_val << (DP_TX0_VOLT_SWING_SHIFT + lane_shift),
892 			   DP_TX0_VOLT_SWING_MASK << lane_shift);
893 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_SWING_EMP,
894 			   preemphasis << (DP_TX0_PRE_EMPH_SHIFT + lane_shift),
895 			   DP_TX0_PRE_EMPH_MASK << lane_shift);
896 }
897 
898 static void mtk_dp_reset_swing_pre_emphasis(struct mtk_dp *mtk_dp)
899 {
900 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_SWING_EMP,
901 			   0,
902 			   DP_TX0_VOLT_SWING_MASK |
903 			   DP_TX1_VOLT_SWING_MASK |
904 			   DP_TX2_VOLT_SWING_MASK |
905 			   DP_TX3_VOLT_SWING_MASK |
906 			   DP_TX0_PRE_EMPH_MASK |
907 			   DP_TX1_PRE_EMPH_MASK |
908 			   DP_TX2_PRE_EMPH_MASK |
909 			   DP_TX3_PRE_EMPH_MASK);
910 }
911 
912 static u32 mtk_dp_swirq_get_clear(struct mtk_dp *mtk_dp)
913 {
914 	u32 irq_status = mtk_dp_read(mtk_dp, MTK_DP_TRANS_P0_35D0) &
915 			 SW_IRQ_FINAL_STATUS_DP_TRANS_P0_MASK;
916 
917 	if (irq_status) {
918 		mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_35C8,
919 				   irq_status, SW_IRQ_CLR_DP_TRANS_P0_MASK);
920 		mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_35C8,
921 				   0, SW_IRQ_CLR_DP_TRANS_P0_MASK);
922 	}
923 
924 	return irq_status;
925 }
926 
927 static u32 mtk_dp_hwirq_get_clear(struct mtk_dp *mtk_dp)
928 {
929 	u32 irq_status = (mtk_dp_read(mtk_dp, MTK_DP_TRANS_P0_3418) &
930 			  IRQ_STATUS_DP_TRANS_P0_MASK) >> 12;
931 
932 	if (irq_status) {
933 		mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3418,
934 				   irq_status, IRQ_CLR_DP_TRANS_P0_MASK);
935 		mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3418,
936 				   0, IRQ_CLR_DP_TRANS_P0_MASK);
937 	}
938 
939 	return irq_status;
940 }
941 
942 static void mtk_dp_hwirq_enable(struct mtk_dp *mtk_dp, bool enable)
943 {
944 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3418,
945 			   enable ? 0 :
946 			   IRQ_MASK_DP_TRANS_P0_DISC_IRQ |
947 			   IRQ_MASK_DP_TRANS_P0_CONN_IRQ |
948 			   IRQ_MASK_DP_TRANS_P0_INT_IRQ,
949 			   IRQ_MASK_DP_TRANS_P0_MASK);
950 }
951 
952 static void mtk_dp_initialize_settings(struct mtk_dp *mtk_dp)
953 {
954 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_342C,
955 			   XTAL_FREQ_DP_TRANS_P0_DEFAULT,
956 			   XTAL_FREQ_DP_TRANS_P0_MASK);
957 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3540,
958 			   FEC_CLOCK_EN_MODE_DP_TRANS_P0,
959 			   FEC_CLOCK_EN_MODE_DP_TRANS_P0);
960 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_31EC,
961 			   AUDIO_CH_SRC_SEL_DP_ENC0_P0,
962 			   AUDIO_CH_SRC_SEL_DP_ENC0_P0);
963 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_304C,
964 			   0, SDP_VSYNC_RISING_MASK_DP_ENC0_P0_MASK);
965 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_IRQ_MASK,
966 			   IRQ_MASK_AUX_TOP_IRQ, IRQ_MASK_AUX_TOP_IRQ);
967 }
968 
969 static void mtk_dp_initialize_hpd_detect_settings(struct mtk_dp *mtk_dp)
970 {
971 	u32 val;
972 	/* Debounce threshold */
973 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3410,
974 			   8, HPD_DEB_THD_DP_TRANS_P0_MASK);
975 
976 	val = (HPD_INT_THD_DP_TRANS_P0_LOWER_500US |
977 	       HPD_INT_THD_DP_TRANS_P0_UPPER_1100US) << 4;
978 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3410,
979 			   val, HPD_INT_THD_DP_TRANS_P0_MASK);
980 
981 	/*
982 	 * Connect threshold 1.5ms + 5 x 0.1ms = 2ms
983 	 * Disconnect threshold 1.5ms + 5 x 0.1ms = 2ms
984 	 */
985 	val = (5 << 8) | (5 << 12);
986 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3410,
987 			   val,
988 			   HPD_DISC_THD_DP_TRANS_P0_MASK |
989 			   HPD_CONN_THD_DP_TRANS_P0_MASK);
990 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3430,
991 			   HPD_INT_THD_ECO_DP_TRANS_P0_HIGH_BOUND_EXT,
992 			   HPD_INT_THD_ECO_DP_TRANS_P0_MASK);
993 }
994 
995 static void mtk_dp_initialize_aux_settings(struct mtk_dp *mtk_dp)
996 {
997 	/* modify timeout threshold = 0x1595 */
998 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_360C,
999 			   AUX_TIMEOUT_THR_AUX_TX_P0_VAL,
1000 			   AUX_TIMEOUT_THR_AUX_TX_P0_MASK);
1001 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3658,
1002 			   0, AUX_TX_OV_EN_AUX_TX_P0_MASK);
1003 	/* 25 for 26M */
1004 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3634,
1005 			   AUX_TX_OVER_SAMPLE_RATE_FOR_26M << 8,
1006 			   AUX_TX_OVER_SAMPLE_RATE_AUX_TX_P0_MASK);
1007 	/* 13 for 26M */
1008 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3614,
1009 			   AUX_RX_UI_CNT_THR_AUX_FOR_26M,
1010 			   AUX_RX_UI_CNT_THR_AUX_TX_P0_MASK);
1011 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_37C8,
1012 			   MTK_ATOP_EN_AUX_TX_P0,
1013 			   MTK_ATOP_EN_AUX_TX_P0);
1014 
1015 	/* Set complete reply mode for AUX */
1016 	mtk_dp_update_bits(mtk_dp, MTK_DP_AUX_P0_3690,
1017 			   RX_REPLY_COMPLETE_MODE_AUX_TX_P0,
1018 			   RX_REPLY_COMPLETE_MODE_AUX_TX_P0);
1019 }
1020 
1021 static void mtk_dp_initialize_digital_settings(struct mtk_dp *mtk_dp)
1022 {
1023 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_304C,
1024 			   0, VBID_VIDEO_MUTE_DP_ENC0_P0_MASK);
1025 
1026 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3368,
1027 			   BS2BS_MODE_DP_ENC1_P0_VAL << 12,
1028 			   BS2BS_MODE_DP_ENC1_P0_MASK);
1029 
1030 	/* dp tx encoder reset all sw */
1031 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3004,
1032 			   DP_TX_ENCODER_4P_RESET_SW_DP_ENC0_P0,
1033 			   DP_TX_ENCODER_4P_RESET_SW_DP_ENC0_P0);
1034 
1035 	/* Wait for sw reset to complete */
1036 	usleep_range(1000, 5000);
1037 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3004,
1038 			   0, DP_TX_ENCODER_4P_RESET_SW_DP_ENC0_P0);
1039 }
1040 
1041 static void mtk_dp_digital_sw_reset(struct mtk_dp *mtk_dp)
1042 {
1043 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_340C,
1044 			   DP_TX_TRANSMITTER_4P_RESET_SW_DP_TRANS_P0,
1045 			   DP_TX_TRANSMITTER_4P_RESET_SW_DP_TRANS_P0);
1046 
1047 	/* Wait for sw reset to complete */
1048 	usleep_range(1000, 5000);
1049 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_340C,
1050 			   0, DP_TX_TRANSMITTER_4P_RESET_SW_DP_TRANS_P0);
1051 }
1052 
1053 static void mtk_dp_set_lanes(struct mtk_dp *mtk_dp, int lanes)
1054 {
1055 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_35F0,
1056 			   lanes == 0 ? 0 : DP_TRANS_DUMMY_RW_0,
1057 			   DP_TRANS_DUMMY_RW_0_MASK);
1058 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3000,
1059 			   lanes, LANE_NUM_DP_ENC0_P0_MASK);
1060 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_34A4,
1061 			   lanes << 2, LANE_NUM_DP_TRANS_P0_MASK);
1062 }
1063 
1064 static void mtk_dp_get_calibration_data(struct mtk_dp *mtk_dp)
1065 {
1066 	const struct mtk_dp_efuse_fmt *fmt;
1067 	struct device *dev = mtk_dp->dev;
1068 	struct nvmem_cell *cell;
1069 	u32 *cal_data = mtk_dp->cal_data;
1070 	u32 *buf;
1071 	int i;
1072 	size_t len;
1073 
1074 	cell = nvmem_cell_get(dev, "dp_calibration_data");
1075 	if (IS_ERR(cell)) {
1076 		dev_warn(dev, "Failed to get nvmem cell dp_calibration_data\n");
1077 		goto use_default_val;
1078 	}
1079 
1080 	buf = (u32 *)nvmem_cell_read(cell, &len);
1081 	nvmem_cell_put(cell);
1082 
1083 	if (IS_ERR(buf) || ((len / sizeof(u32)) != 4)) {
1084 		dev_warn(dev, "Failed to read nvmem_cell_read\n");
1085 
1086 		if (!IS_ERR(buf))
1087 			kfree(buf);
1088 
1089 		goto use_default_val;
1090 	}
1091 
1092 	for (i = 0; i < MTK_DP_CAL_MAX; i++) {
1093 		fmt = &mtk_dp->data->efuse_fmt[i];
1094 		cal_data[i] = (buf[fmt->idx] >> fmt->shift) & fmt->mask;
1095 
1096 		if (cal_data[i] < fmt->min_val || cal_data[i] > fmt->max_val) {
1097 			dev_warn(mtk_dp->dev, "Invalid efuse data, idx = %d\n", i);
1098 			kfree(buf);
1099 			goto use_default_val;
1100 		}
1101 	}
1102 	kfree(buf);
1103 
1104 	return;
1105 
1106 use_default_val:
1107 	dev_warn(mtk_dp->dev, "Use default calibration data\n");
1108 	for (i = 0; i < MTK_DP_CAL_MAX; i++)
1109 		cal_data[i] = mtk_dp->data->efuse_fmt[i].default_val;
1110 }
1111 
1112 static void mtk_dp_set_calibration_data(struct mtk_dp *mtk_dp)
1113 {
1114 	u32 *cal_data = mtk_dp->cal_data;
1115 
1116 	mtk_dp_update_bits(mtk_dp, DP_PHY_GLB_DPAUX_TX,
1117 			   cal_data[MTK_DP_CAL_CLKTX_IMPSE] << 20,
1118 			   RG_CKM_PT0_CKTX_IMPSEL);
1119 	mtk_dp_update_bits(mtk_dp, DP_PHY_GLB_BIAS_GEN_00,
1120 			   cal_data[MTK_DP_CAL_GLB_BIAS_TRIM] << 16,
1121 			   RG_XTP_GLB_BIAS_INTR_CTRL);
1122 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_0,
1123 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_0] << 12,
1124 			   RG_XTP_LN0_TX_IMPSEL_PMOS);
1125 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_0,
1126 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_0] << 16,
1127 			   RG_XTP_LN0_TX_IMPSEL_NMOS);
1128 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_1,
1129 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_1] << 12,
1130 			   RG_XTP_LN1_TX_IMPSEL_PMOS);
1131 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_1,
1132 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_1] << 16,
1133 			   RG_XTP_LN1_TX_IMPSEL_NMOS);
1134 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_2,
1135 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_2] << 12,
1136 			   RG_XTP_LN2_TX_IMPSEL_PMOS);
1137 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_2,
1138 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_2] << 16,
1139 			   RG_XTP_LN2_TX_IMPSEL_NMOS);
1140 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_3,
1141 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_PMOS_3] << 12,
1142 			   RG_XTP_LN3_TX_IMPSEL_PMOS);
1143 	mtk_dp_update_bits(mtk_dp, DP_PHY_LANE_TX_3,
1144 			   cal_data[MTK_DP_CAL_LN_TX_IMPSEL_NMOS_3] << 16,
1145 			   RG_XTP_LN3_TX_IMPSEL_NMOS);
1146 }
1147 
1148 static int mtk_dp_phy_configure(struct mtk_dp *mtk_dp,
1149 				u32 link_rate, int lane_count)
1150 {
1151 	int ret;
1152 	union phy_configure_opts phy_opts = {
1153 		.dp = {
1154 			.link_rate = drm_dp_bw_code_to_link_rate(link_rate) / 100,
1155 			.set_rate = 1,
1156 			.lanes = lane_count,
1157 			.set_lanes = 1,
1158 			.ssc = mtk_dp->train_info.sink_ssc,
1159 		}
1160 	};
1161 
1162 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE, DP_PWR_STATE_BANDGAP,
1163 			   DP_PWR_STATE_MASK);
1164 
1165 	ret = phy_configure(mtk_dp->phy, &phy_opts);
1166 	if (ret)
1167 		return ret;
1168 
1169 	mtk_dp_set_calibration_data(mtk_dp);
1170 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
1171 			   DP_PWR_STATE_BANDGAP_TPLL_LANE, DP_PWR_STATE_MASK);
1172 
1173 	return 0;
1174 }
1175 
1176 static void mtk_dp_set_idle_pattern(struct mtk_dp *mtk_dp, bool enable)
1177 {
1178 	u32 val = POST_MISC_DATA_LANE0_OV_DP_TRANS_P0_MASK |
1179 		  POST_MISC_DATA_LANE1_OV_DP_TRANS_P0_MASK |
1180 		  POST_MISC_DATA_LANE2_OV_DP_TRANS_P0_MASK |
1181 		  POST_MISC_DATA_LANE3_OV_DP_TRANS_P0_MASK;
1182 
1183 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3580,
1184 			   enable ? val : 0, val);
1185 }
1186 
1187 static void mtk_dp_train_set_pattern(struct mtk_dp *mtk_dp, int pattern)
1188 {
1189 	/* TPS1 */
1190 	if (pattern == 1)
1191 		mtk_dp_set_idle_pattern(mtk_dp, false);
1192 
1193 	mtk_dp_update_bits(mtk_dp,
1194 			   MTK_DP_TRANS_P0_3400,
1195 			   pattern ? BIT(pattern - 1) << 12 : 0,
1196 			   PATTERN1_EN_DP_TRANS_P0_MASK |
1197 			   PATTERN2_EN_DP_TRANS_P0_MASK |
1198 			   PATTERN3_EN_DP_TRANS_P0_MASK |
1199 			   PATTERN4_EN_DP_TRANS_P0_MASK);
1200 }
1201 
1202 static void mtk_dp_set_enhanced_frame_mode(struct mtk_dp *mtk_dp)
1203 {
1204 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3000,
1205 			   ENHANCED_FRAME_EN_DP_ENC0_P0,
1206 			   ENHANCED_FRAME_EN_DP_ENC0_P0);
1207 }
1208 
1209 static void mtk_dp_training_set_scramble(struct mtk_dp *mtk_dp, bool enable)
1210 {
1211 	mtk_dp_update_bits(mtk_dp, MTK_DP_TRANS_P0_3404,
1212 			   enable ? DP_SCR_EN_DP_TRANS_P0_MASK : 0,
1213 			   DP_SCR_EN_DP_TRANS_P0_MASK);
1214 }
1215 
1216 static void mtk_dp_video_mute(struct mtk_dp *mtk_dp, bool enable)
1217 {
1218 	struct arm_smccc_res res;
1219 	u32 val = VIDEO_MUTE_SEL_DP_ENC0_P0 |
1220 		  (enable ? VIDEO_MUTE_SW_DP_ENC0_P0 : 0);
1221 
1222 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3000,
1223 			   val,
1224 			   VIDEO_MUTE_SEL_DP_ENC0_P0 |
1225 			   VIDEO_MUTE_SW_DP_ENC0_P0);
1226 
1227 	arm_smccc_smc(MTK_DP_SIP_CONTROL_AARCH32,
1228 		      mtk_dp->data->smc_cmd, enable,
1229 		      0, 0, 0, 0, 0, &res);
1230 
1231 	dev_dbg(mtk_dp->dev, "smc cmd: 0x%x, p1: %s, ret: 0x%lx-0x%lx\n",
1232 		mtk_dp->data->smc_cmd, enable ? "enable" : "disable", res.a0, res.a1);
1233 }
1234 
1235 static void mtk_dp_audio_mute(struct mtk_dp *mtk_dp, bool mute)
1236 {
1237 	u32 val[3];
1238 
1239 	if (mute) {
1240 		val[0] = VBID_AUDIO_MUTE_FLAG_SW_DP_ENC0_P0 |
1241 			 VBID_AUDIO_MUTE_FLAG_SEL_DP_ENC0_P0;
1242 		val[1] = 0;
1243 		val[2] = 0;
1244 	} else {
1245 		val[0] = 0;
1246 		val[1] = AU_EN_DP_ENC0_P0;
1247 		/* Send one every two frames */
1248 		val[2] = 0x0F;
1249 	}
1250 
1251 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3030,
1252 			   val[0],
1253 			   VBID_AUDIO_MUTE_FLAG_SW_DP_ENC0_P0 |
1254 			   VBID_AUDIO_MUTE_FLAG_SEL_DP_ENC0_P0);
1255 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3088,
1256 			   val[1], AU_EN_DP_ENC0_P0);
1257 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_30A4,
1258 			   val[2], AU_TS_CFG_DP_ENC0_P0_MASK);
1259 }
1260 
1261 static void mtk_dp_aux_panel_poweron(struct mtk_dp *mtk_dp, bool pwron)
1262 {
1263 	if (pwron) {
1264 		/* power on aux */
1265 		mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
1266 				   DP_PWR_STATE_BANDGAP_TPLL_LANE,
1267 				   DP_PWR_STATE_MASK);
1268 
1269 		/* power on panel */
1270 		drm_dp_dpcd_writeb(&mtk_dp->aux, DP_SET_POWER, DP_SET_POWER_D0);
1271 		usleep_range(2000, 5000);
1272 	} else {
1273 		/* power off panel */
1274 		drm_dp_dpcd_writeb(&mtk_dp->aux, DP_SET_POWER, DP_SET_POWER_D3);
1275 		usleep_range(2000, 3000);
1276 
1277 		/* power off aux */
1278 		mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
1279 				   DP_PWR_STATE_BANDGAP_TPLL,
1280 				   DP_PWR_STATE_MASK);
1281 	}
1282 }
1283 
1284 static void mtk_dp_power_enable(struct mtk_dp *mtk_dp)
1285 {
1286 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_RESET_AND_PROBE,
1287 			   0, SW_RST_B_PHYD);
1288 
1289 	/* Wait for power enable */
1290 	usleep_range(10, 200);
1291 
1292 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_RESET_AND_PROBE,
1293 			   SW_RST_B_PHYD, SW_RST_B_PHYD);
1294 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
1295 			   DP_PWR_STATE_BANDGAP_TPLL, DP_PWR_STATE_MASK);
1296 	mtk_dp_write(mtk_dp, MTK_DP_1040,
1297 		     RG_DPAUX_RX_VALID_DEGLITCH_EN | RG_XTP_GLB_CKDET_EN |
1298 		     RG_DPAUX_RX_EN);
1299 	mtk_dp_update_bits(mtk_dp, MTK_DP_0034, 0, DA_CKM_CKTX0_EN_FORCE_EN);
1300 }
1301 
1302 static void mtk_dp_power_disable(struct mtk_dp *mtk_dp)
1303 {
1304 	mtk_dp_write(mtk_dp, MTK_DP_TOP_PWR_STATE, 0);
1305 
1306 	mtk_dp_update_bits(mtk_dp, MTK_DP_0034,
1307 			   DA_CKM_CKTX0_EN_FORCE_EN, DA_CKM_CKTX0_EN_FORCE_EN);
1308 
1309 	/* Disable RX */
1310 	mtk_dp_write(mtk_dp, MTK_DP_1040, 0);
1311 	mtk_dp_write(mtk_dp, MTK_DP_TOP_MEM_PD,
1312 		     0x550 | FUSE_SEL | MEM_ISO_EN);
1313 }
1314 
1315 static void mtk_dp_initialize_priv_data(struct mtk_dp *mtk_dp)
1316 {
1317 	bool plugged_in = (mtk_dp->bridge.type == DRM_MODE_CONNECTOR_eDP);
1318 
1319 	mtk_dp->train_info.link_rate = DP_LINK_BW_5_4;
1320 	mtk_dp->train_info.lane_count = mtk_dp->max_lanes;
1321 	mtk_dp->train_info.cable_plugged_in = plugged_in;
1322 
1323 	mtk_dp->info.format = DP_PIXELFORMAT_RGB;
1324 	memset(&mtk_dp->info.vm, 0, sizeof(struct videomode));
1325 	mtk_dp->audio_enable = false;
1326 }
1327 
1328 static void mtk_dp_sdp_set_down_cnt_init(struct mtk_dp *mtk_dp,
1329 					 u32 sram_read_start)
1330 {
1331 	u32 sdp_down_cnt_init = 0;
1332 	struct drm_display_mode mode;
1333 	struct videomode *vm = &mtk_dp->info.vm;
1334 
1335 	drm_display_mode_from_videomode(vm, &mode);
1336 
1337 	if (mode.clock > 0)
1338 		sdp_down_cnt_init = sram_read_start *
1339 				    mtk_dp->train_info.link_rate * 2700 * 8 /
1340 				    (mode.clock * 4);
1341 
1342 	switch (mtk_dp->train_info.lane_count) {
1343 	case 1:
1344 		sdp_down_cnt_init = max_t(u32, sdp_down_cnt_init, 0x1A);
1345 		break;
1346 	case 2:
1347 		/* case for LowResolution && High Audio Sample Rate */
1348 		sdp_down_cnt_init = max_t(u32, sdp_down_cnt_init, 0x10);
1349 		sdp_down_cnt_init += mode.vtotal <= 525 ? 4 : 0;
1350 		break;
1351 	case 4:
1352 	default:
1353 		sdp_down_cnt_init = max_t(u32, sdp_down_cnt_init, 6);
1354 		break;
1355 	}
1356 
1357 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC0_P0_3040,
1358 			   sdp_down_cnt_init,
1359 			   SDP_DOWN_CNT_INIT_DP_ENC0_P0_MASK);
1360 }
1361 
1362 static void mtk_dp_sdp_set_down_cnt_init_in_hblank(struct mtk_dp *mtk_dp)
1363 {
1364 	int pix_clk_mhz;
1365 	u32 dc_offset;
1366 	u32 spd_down_cnt_init = 0;
1367 	struct drm_display_mode mode;
1368 	struct videomode *vm = &mtk_dp->info.vm;
1369 
1370 	drm_display_mode_from_videomode(vm, &mode);
1371 
1372 	pix_clk_mhz = mtk_dp->info.format == DP_PIXELFORMAT_YUV420 ?
1373 		      mode.clock / 2000 : mode.clock / 1000;
1374 
1375 	switch (mtk_dp->train_info.lane_count) {
1376 	case 1:
1377 		spd_down_cnt_init = 0x20;
1378 		break;
1379 	case 2:
1380 		dc_offset = (mode.vtotal <= 525) ? 0x14 : 0x00;
1381 		spd_down_cnt_init = 0x18 + dc_offset;
1382 		break;
1383 	case 4:
1384 	default:
1385 		dc_offset = (mode.vtotal <= 525) ? 0x08 : 0x00;
1386 		if (pix_clk_mhz > mtk_dp->train_info.link_rate * 27)
1387 			spd_down_cnt_init = 0x8;
1388 		else
1389 			spd_down_cnt_init = 0x10 + dc_offset;
1390 		break;
1391 	}
1392 
1393 	mtk_dp_update_bits(mtk_dp, MTK_DP_ENC1_P0_3364, spd_down_cnt_init,
1394 			   SDP_DOWN_CNT_INIT_IN_HBLANK_DP_ENC1_P0_MASK);
1395 }
1396 
1397 static void mtk_dp_setup_tu(struct mtk_dp *mtk_dp)
1398 {
1399 	u32 sram_read_start = min_t(u32, MTK_DP_TBC_BUF_READ_START_ADDR,
1400 				    mtk_dp->info.vm.hactive /
1401 				    mtk_dp->train_info.lane_count /
1402 				    MTK_DP_4P1T / MTK_DP_HDE /
1403 				    MTK_DP_PIX_PER_ADDR);
1404 	mtk_dp_set_sram_read_start(mtk_dp, sram_read_start);
1405 	mtk_dp_setup_encoder(mtk_dp);
1406 	mtk_dp_sdp_set_down_cnt_init_in_hblank(mtk_dp);
1407 	mtk_dp_sdp_set_down_cnt_init(mtk_dp, sram_read_start);
1408 }
1409 
1410 static void mtk_dp_set_tx_out(struct mtk_dp *mtk_dp)
1411 {
1412 	mtk_dp_setup_tu(mtk_dp);
1413 }
1414 
1415 static void mtk_dp_train_update_swing_pre(struct mtk_dp *mtk_dp, int lanes,
1416 					  u8 dpcd_adjust_req[2])
1417 {
1418 	int lane;
1419 
1420 	for (lane = 0; lane < lanes; ++lane) {
1421 		u8 val;
1422 		u8 swing;
1423 		u8 preemphasis;
1424 		int index = lane / 2;
1425 		int shift = lane % 2 ? DP_ADJUST_VOLTAGE_SWING_LANE1_SHIFT : 0;
1426 
1427 		swing = (dpcd_adjust_req[index] >> shift) &
1428 			DP_ADJUST_VOLTAGE_SWING_LANE0_MASK;
1429 		preemphasis = ((dpcd_adjust_req[index] >> shift) &
1430 			       DP_ADJUST_PRE_EMPHASIS_LANE0_MASK) >>
1431 			      DP_ADJUST_PRE_EMPHASIS_LANE0_SHIFT;
1432 		val = swing << DP_TRAIN_VOLTAGE_SWING_SHIFT |
1433 		      preemphasis << DP_TRAIN_PRE_EMPHASIS_SHIFT;
1434 
1435 		if (swing == DP_TRAIN_VOLTAGE_SWING_LEVEL_3)
1436 			val |= DP_TRAIN_MAX_SWING_REACHED;
1437 		if (preemphasis == 3)
1438 			val |= DP_TRAIN_MAX_PRE_EMPHASIS_REACHED;
1439 
1440 		mtk_dp_set_swing_pre_emphasis(mtk_dp, lane, swing, preemphasis);
1441 		drm_dp_dpcd_writeb(&mtk_dp->aux, DP_TRAINING_LANE0_SET + lane,
1442 				   val);
1443 	}
1444 }
1445 
1446 static void mtk_dp_pattern(struct mtk_dp *mtk_dp, bool is_tps1)
1447 {
1448 	int pattern;
1449 	unsigned int aux_offset;
1450 
1451 	if (is_tps1) {
1452 		pattern = 1;
1453 		aux_offset = DP_LINK_SCRAMBLING_DISABLE | DP_TRAINING_PATTERN_1;
1454 	} else {
1455 		aux_offset = mtk_dp->train_info.channel_eq_pattern;
1456 
1457 		switch (mtk_dp->train_info.channel_eq_pattern) {
1458 		case DP_TRAINING_PATTERN_4:
1459 			pattern = 4;
1460 			break;
1461 		case DP_TRAINING_PATTERN_3:
1462 			pattern = 3;
1463 			aux_offset |= DP_LINK_SCRAMBLING_DISABLE;
1464 			break;
1465 		case DP_TRAINING_PATTERN_2:
1466 		default:
1467 			pattern = 2;
1468 			aux_offset |= DP_LINK_SCRAMBLING_DISABLE;
1469 			break;
1470 		}
1471 	}
1472 
1473 	mtk_dp_train_set_pattern(mtk_dp, pattern);
1474 	drm_dp_dpcd_writeb(&mtk_dp->aux, DP_TRAINING_PATTERN_SET, aux_offset);
1475 }
1476 
1477 static int mtk_dp_train_setting(struct mtk_dp *mtk_dp, u8 target_link_rate,
1478 				u8 target_lane_count)
1479 {
1480 	int ret;
1481 
1482 	drm_dp_dpcd_writeb(&mtk_dp->aux, DP_LINK_BW_SET, target_link_rate);
1483 	drm_dp_dpcd_writeb(&mtk_dp->aux, DP_LANE_COUNT_SET,
1484 			   target_lane_count | DP_LANE_COUNT_ENHANCED_FRAME_EN);
1485 
1486 	if (mtk_dp->train_info.sink_ssc)
1487 		drm_dp_dpcd_writeb(&mtk_dp->aux, DP_DOWNSPREAD_CTRL,
1488 				   DP_SPREAD_AMP_0_5);
1489 
1490 	mtk_dp_set_lanes(mtk_dp, target_lane_count / 2);
1491 	ret = mtk_dp_phy_configure(mtk_dp, target_link_rate, target_lane_count);
1492 	if (ret)
1493 		return ret;
1494 
1495 	dev_dbg(mtk_dp->dev,
1496 		"Link train target_link_rate = 0x%x, target_lane_count = 0x%x\n",
1497 		target_link_rate, target_lane_count);
1498 
1499 	return 0;
1500 }
1501 
1502 static int mtk_dp_train_cr(struct mtk_dp *mtk_dp, u8 target_lane_count)
1503 {
1504 	u8 lane_adjust[2] = {};
1505 	u8 link_status[DP_LINK_STATUS_SIZE] = {};
1506 	u8 prev_lane_adjust = 0xff;
1507 	int train_retries = 0;
1508 	int voltage_retries = 0;
1509 
1510 	mtk_dp_pattern(mtk_dp, true);
1511 
1512 	/* In DP spec 1.4, the retry count of CR is defined as 10. */
1513 	do {
1514 		train_retries++;
1515 		if (!mtk_dp->train_info.cable_plugged_in) {
1516 			mtk_dp_train_set_pattern(mtk_dp, 0);
1517 			return -ENODEV;
1518 		}
1519 
1520 		drm_dp_dpcd_read(&mtk_dp->aux, DP_ADJUST_REQUEST_LANE0_1,
1521 				 lane_adjust, sizeof(lane_adjust));
1522 		mtk_dp_train_update_swing_pre(mtk_dp, target_lane_count,
1523 					      lane_adjust);
1524 
1525 		drm_dp_link_train_clock_recovery_delay(&mtk_dp->aux,
1526 						       mtk_dp->rx_cap);
1527 
1528 		/* check link status from sink device */
1529 		drm_dp_dpcd_read_link_status(&mtk_dp->aux, link_status);
1530 		if (drm_dp_clock_recovery_ok(link_status,
1531 					     target_lane_count)) {
1532 			dev_dbg(mtk_dp->dev, "Link train CR pass\n");
1533 			return 0;
1534 		}
1535 
1536 		/*
1537 		 * In DP spec 1.4, if current voltage level is the same
1538 		 * with previous voltage level, we need to retry 5 times.
1539 		 */
1540 		if (prev_lane_adjust == link_status[4]) {
1541 			voltage_retries++;
1542 			/*
1543 			 * Condition of CR fail:
1544 			 * 1. Failed to pass CR using the same voltage
1545 			 *    level over five times.
1546 			 * 2. Failed to pass CR when the current voltage
1547 			 *    level is the same with previous voltage
1548 			 *    level and reach max voltage level (3).
1549 			 */
1550 			if (voltage_retries > MTK_DP_TRAIN_VOLTAGE_LEVEL_RETRY ||
1551 			    (prev_lane_adjust & DP_ADJUST_VOLTAGE_SWING_LANE0_MASK) == 3) {
1552 				dev_dbg(mtk_dp->dev, "Link train CR fail\n");
1553 				break;
1554 			}
1555 		} else {
1556 			/*
1557 			 * If the voltage level is changed, we need to
1558 			 * re-calculate this retry count.
1559 			 */
1560 			voltage_retries = 0;
1561 		}
1562 		prev_lane_adjust = link_status[4];
1563 	} while (train_retries < MTK_DP_TRAIN_DOWNSCALE_RETRY);
1564 
1565 	/* Failed to train CR, and disable pattern. */
1566 	drm_dp_dpcd_writeb(&mtk_dp->aux, DP_TRAINING_PATTERN_SET,
1567 			   DP_TRAINING_PATTERN_DISABLE);
1568 	mtk_dp_train_set_pattern(mtk_dp, 0);
1569 
1570 	return -ETIMEDOUT;
1571 }
1572 
1573 static int mtk_dp_train_eq(struct mtk_dp *mtk_dp, u8 target_lane_count)
1574 {
1575 	u8 lane_adjust[2] = {};
1576 	u8 link_status[DP_LINK_STATUS_SIZE] = {};
1577 	int train_retries = 0;
1578 
1579 	mtk_dp_pattern(mtk_dp, false);
1580 
1581 	do {
1582 		train_retries++;
1583 		if (!mtk_dp->train_info.cable_plugged_in) {
1584 			mtk_dp_train_set_pattern(mtk_dp, 0);
1585 			return -ENODEV;
1586 		}
1587 
1588 		drm_dp_dpcd_read(&mtk_dp->aux, DP_ADJUST_REQUEST_LANE0_1,
1589 				 lane_adjust, sizeof(lane_adjust));
1590 		mtk_dp_train_update_swing_pre(mtk_dp, target_lane_count,
1591 					      lane_adjust);
1592 
1593 		drm_dp_link_train_channel_eq_delay(&mtk_dp->aux,
1594 						   mtk_dp->rx_cap);
1595 
1596 		/* check link status from sink device */
1597 		drm_dp_dpcd_read_link_status(&mtk_dp->aux, link_status);
1598 		if (drm_dp_channel_eq_ok(link_status, target_lane_count)) {
1599 			dev_dbg(mtk_dp->dev, "Link train EQ pass\n");
1600 
1601 			/* Training done, and disable pattern. */
1602 			drm_dp_dpcd_writeb(&mtk_dp->aux, DP_TRAINING_PATTERN_SET,
1603 					   DP_TRAINING_PATTERN_DISABLE);
1604 			mtk_dp_train_set_pattern(mtk_dp, 0);
1605 			return 0;
1606 		}
1607 		dev_dbg(mtk_dp->dev, "Link train EQ fail\n");
1608 	} while (train_retries < MTK_DP_TRAIN_DOWNSCALE_RETRY);
1609 
1610 	/* Failed to train EQ, and disable pattern. */
1611 	drm_dp_dpcd_writeb(&mtk_dp->aux, DP_TRAINING_PATTERN_SET,
1612 			   DP_TRAINING_PATTERN_DISABLE);
1613 	mtk_dp_train_set_pattern(mtk_dp, 0);
1614 
1615 	return -ETIMEDOUT;
1616 }
1617 
1618 static int mtk_dp_parse_capabilities(struct mtk_dp *mtk_dp)
1619 {
1620 	u8 val;
1621 	ssize_t ret;
1622 
1623 	/*
1624 	 * If we're eDP and capabilities were already parsed we can skip
1625 	 * reading again because eDP panels aren't hotpluggable hence the
1626 	 * caps and training information won't ever change in a boot life
1627 	 */
1628 	if (mtk_dp->bridge.type == DRM_MODE_CONNECTOR_eDP &&
1629 	    mtk_dp->rx_cap[DP_MAX_LINK_RATE] &&
1630 	    mtk_dp->train_info.sink_ssc)
1631 		return 0;
1632 
1633 	ret = drm_dp_read_dpcd_caps(&mtk_dp->aux, mtk_dp->rx_cap);
1634 	if (ret < 0)
1635 		return ret;
1636 
1637 	if (drm_dp_tps4_supported(mtk_dp->rx_cap))
1638 		mtk_dp->train_info.channel_eq_pattern = DP_TRAINING_PATTERN_4;
1639 	else if (drm_dp_tps3_supported(mtk_dp->rx_cap))
1640 		mtk_dp->train_info.channel_eq_pattern = DP_TRAINING_PATTERN_3;
1641 	else
1642 		mtk_dp->train_info.channel_eq_pattern = DP_TRAINING_PATTERN_2;
1643 
1644 	mtk_dp->train_info.sink_ssc = drm_dp_max_downspread(mtk_dp->rx_cap);
1645 
1646 	ret = drm_dp_dpcd_readb(&mtk_dp->aux, DP_MSTM_CAP, &val);
1647 	if (ret < 1) {
1648 		drm_err(mtk_dp->drm_dev, "Read mstm cap failed\n");
1649 		return ret == 0 ? -EIO : ret;
1650 	}
1651 
1652 	if (val & DP_MST_CAP) {
1653 		/* Clear DP_DEVICE_SERVICE_IRQ_VECTOR_ESI0 */
1654 		ret = drm_dp_dpcd_readb(&mtk_dp->aux,
1655 					DP_DEVICE_SERVICE_IRQ_VECTOR_ESI0,
1656 					&val);
1657 		if (ret < 1) {
1658 			drm_err(mtk_dp->drm_dev, "Read irq vector failed\n");
1659 			return ret == 0 ? -EIO : ret;
1660 		}
1661 
1662 		if (val) {
1663 			ret = drm_dp_dpcd_writeb(&mtk_dp->aux,
1664 						 DP_DEVICE_SERVICE_IRQ_VECTOR_ESI0,
1665 						 val);
1666 			if (ret < 0)
1667 				return ret;
1668 		}
1669 	}
1670 
1671 	return 0;
1672 }
1673 
1674 static bool mtk_dp_edid_parse_audio_capabilities(struct mtk_dp *mtk_dp,
1675 						 struct mtk_dp_audio_cfg *cfg)
1676 {
1677 	if (!mtk_dp->data->audio_supported)
1678 		return false;
1679 
1680 	if (mtk_dp->info.audio_cur_cfg.sad_count <= 0) {
1681 		drm_info(mtk_dp->drm_dev, "The SADs is NULL\n");
1682 		return false;
1683 	}
1684 
1685 	return true;
1686 }
1687 
1688 static void mtk_dp_train_change_mode(struct mtk_dp *mtk_dp)
1689 {
1690 	phy_reset(mtk_dp->phy);
1691 	mtk_dp_reset_swing_pre_emphasis(mtk_dp);
1692 }
1693 
1694 static int mtk_dp_training(struct mtk_dp *mtk_dp)
1695 {
1696 	int ret;
1697 	u8 lane_count, link_rate, train_limit, max_link_rate;
1698 
1699 	link_rate = min_t(u8, mtk_dp->max_linkrate,
1700 			  mtk_dp->rx_cap[DP_MAX_LINK_RATE]);
1701 	max_link_rate = link_rate;
1702 	lane_count = min_t(u8, mtk_dp->max_lanes,
1703 			   drm_dp_max_lane_count(mtk_dp->rx_cap));
1704 
1705 	/*
1706 	 * TPS are generated by the hardware pattern generator. From the
1707 	 * hardware setting we need to disable this scramble setting before
1708 	 * use the TPS pattern generator.
1709 	 */
1710 	mtk_dp_training_set_scramble(mtk_dp, false);
1711 
1712 	for (train_limit = 6; train_limit > 0; train_limit--) {
1713 		mtk_dp_train_change_mode(mtk_dp);
1714 
1715 		ret = mtk_dp_train_setting(mtk_dp, link_rate, lane_count);
1716 		if (ret)
1717 			return ret;
1718 
1719 		ret = mtk_dp_train_cr(mtk_dp, lane_count);
1720 		if (ret == -ENODEV) {
1721 			return ret;
1722 		} else if (ret) {
1723 			/* reduce link rate */
1724 			switch (link_rate) {
1725 			case DP_LINK_BW_1_62:
1726 				lane_count = lane_count / 2;
1727 				link_rate = max_link_rate;
1728 				if (lane_count == 0)
1729 					return -EIO;
1730 				break;
1731 			case DP_LINK_BW_2_7:
1732 				link_rate = DP_LINK_BW_1_62;
1733 				break;
1734 			case DP_LINK_BW_5_4:
1735 				link_rate = DP_LINK_BW_2_7;
1736 				break;
1737 			case DP_LINK_BW_8_1:
1738 				link_rate = DP_LINK_BW_5_4;
1739 				break;
1740 			default:
1741 				return -EINVAL;
1742 			}
1743 			continue;
1744 		}
1745 
1746 		ret = mtk_dp_train_eq(mtk_dp, lane_count);
1747 		if (ret == -ENODEV) {
1748 			return ret;
1749 		} else if (ret) {
1750 			/* reduce lane count */
1751 			if (lane_count == 0)
1752 				return -EIO;
1753 			lane_count /= 2;
1754 			continue;
1755 		}
1756 
1757 		/* if we can run to this, training is done. */
1758 		break;
1759 	}
1760 
1761 	if (train_limit == 0)
1762 		return -ETIMEDOUT;
1763 
1764 	mtk_dp->train_info.link_rate = link_rate;
1765 	mtk_dp->train_info.lane_count = lane_count;
1766 
1767 	/*
1768 	 * After training done, we need to output normal stream instead of TPS,
1769 	 * so we need to enable scramble.
1770 	 */
1771 	mtk_dp_training_set_scramble(mtk_dp, true);
1772 	mtk_dp_set_enhanced_frame_mode(mtk_dp);
1773 
1774 	return 0;
1775 }
1776 
1777 static void mtk_dp_video_enable(struct mtk_dp *mtk_dp, bool enable)
1778 {
1779 	/* the mute sequence is different between enable and disable */
1780 	if (enable) {
1781 		mtk_dp_msa_bypass_enable(mtk_dp, false);
1782 		mtk_dp_pg_enable(mtk_dp, false);
1783 		mtk_dp_set_tx_out(mtk_dp);
1784 		mtk_dp_video_mute(mtk_dp, false);
1785 	} else {
1786 		mtk_dp_video_mute(mtk_dp, true);
1787 		mtk_dp_pg_enable(mtk_dp, true);
1788 		mtk_dp_msa_bypass_enable(mtk_dp, true);
1789 	}
1790 }
1791 
1792 static void mtk_dp_audio_sdp_setup(struct mtk_dp *mtk_dp,
1793 				   struct mtk_dp_audio_cfg *cfg)
1794 {
1795 	struct dp_sdp sdp;
1796 	struct hdmi_audio_infoframe frame;
1797 
1798 	hdmi_audio_infoframe_init(&frame);
1799 	frame.coding_type = HDMI_AUDIO_CODING_TYPE_PCM;
1800 	frame.channels = cfg->channels;
1801 	frame.sample_frequency = cfg->sample_rate;
1802 
1803 	switch (cfg->word_length_bits) {
1804 	case 16:
1805 		frame.sample_size = HDMI_AUDIO_SAMPLE_SIZE_16;
1806 		break;
1807 	case 20:
1808 		frame.sample_size = HDMI_AUDIO_SAMPLE_SIZE_20;
1809 		break;
1810 	case 24:
1811 	default:
1812 		frame.sample_size = HDMI_AUDIO_SAMPLE_SIZE_24;
1813 		break;
1814 	}
1815 
1816 	hdmi_audio_infoframe_pack_for_dp(&frame, &sdp, MTK_DP_VERSION);
1817 
1818 	mtk_dp_audio_sdp_asp_set_channels(mtk_dp, cfg->channels);
1819 	mtk_dp_setup_sdp_aui(mtk_dp, &sdp);
1820 }
1821 
1822 static void mtk_dp_audio_setup(struct mtk_dp *mtk_dp,
1823 			       struct mtk_dp_audio_cfg *cfg)
1824 {
1825 	mtk_dp_audio_sdp_setup(mtk_dp, cfg);
1826 	mtk_dp_audio_channel_status_set(mtk_dp, cfg);
1827 
1828 	mtk_dp_audio_setup_channels(mtk_dp, cfg);
1829 	mtk_dp_audio_set_divider(mtk_dp);
1830 }
1831 
1832 static int mtk_dp_video_config(struct mtk_dp *mtk_dp)
1833 {
1834 	mtk_dp_config_mn_mode(mtk_dp);
1835 	mtk_dp_set_msa(mtk_dp);
1836 	mtk_dp_set_color_depth(mtk_dp);
1837 	return mtk_dp_set_color_format(mtk_dp, mtk_dp->info.format);
1838 }
1839 
1840 static void mtk_dp_init_port(struct mtk_dp *mtk_dp)
1841 {
1842 	mtk_dp_set_idle_pattern(mtk_dp, true);
1843 	mtk_dp_initialize_priv_data(mtk_dp);
1844 
1845 	mtk_dp_initialize_settings(mtk_dp);
1846 	mtk_dp_initialize_aux_settings(mtk_dp);
1847 	mtk_dp_initialize_digital_settings(mtk_dp);
1848 	mtk_dp_initialize_hpd_detect_settings(mtk_dp);
1849 
1850 	mtk_dp_digital_sw_reset(mtk_dp);
1851 }
1852 
1853 static irqreturn_t mtk_dp_hpd_event_thread(int hpd, void *dev)
1854 {
1855 	struct mtk_dp *mtk_dp = dev;
1856 	unsigned long flags;
1857 	u32 status;
1858 
1859 	if (mtk_dp->need_debounce && mtk_dp->train_info.cable_plugged_in)
1860 		msleep(100);
1861 
1862 	spin_lock_irqsave(&mtk_dp->irq_thread_lock, flags);
1863 	status = mtk_dp->irq_thread_handle;
1864 	mtk_dp->irq_thread_handle = 0;
1865 	spin_unlock_irqrestore(&mtk_dp->irq_thread_lock, flags);
1866 
1867 	if (status & MTK_DP_THREAD_CABLE_STATE_CHG) {
1868 		if (mtk_dp->bridge.dev)
1869 			drm_helper_hpd_irq_event(mtk_dp->bridge.dev);
1870 
1871 		if (!mtk_dp->train_info.cable_plugged_in) {
1872 			mtk_dp_disable_sdp_aui(mtk_dp);
1873 			memset(&mtk_dp->info.audio_cur_cfg, 0,
1874 			       sizeof(mtk_dp->info.audio_cur_cfg));
1875 
1876 			mtk_dp->need_debounce = false;
1877 			mod_timer(&mtk_dp->debounce_timer,
1878 				  jiffies + msecs_to_jiffies(100) - 1);
1879 		}
1880 	}
1881 
1882 	if (status & MTK_DP_THREAD_HPD_EVENT)
1883 		dev_dbg(mtk_dp->dev, "Receive IRQ from sink devices\n");
1884 
1885 	return IRQ_HANDLED;
1886 }
1887 
1888 static irqreturn_t mtk_dp_hpd_event(int hpd, void *dev)
1889 {
1890 	struct mtk_dp *mtk_dp = dev;
1891 	bool cable_sta_chg = false;
1892 	unsigned long flags;
1893 	u32 irq_status = mtk_dp_swirq_get_clear(mtk_dp) |
1894 			 mtk_dp_hwirq_get_clear(mtk_dp);
1895 
1896 	if (!irq_status)
1897 		return IRQ_HANDLED;
1898 
1899 	spin_lock_irqsave(&mtk_dp->irq_thread_lock, flags);
1900 
1901 	if (irq_status & MTK_DP_HPD_INTERRUPT)
1902 		mtk_dp->irq_thread_handle |= MTK_DP_THREAD_HPD_EVENT;
1903 
1904 	/* Cable state is changed. */
1905 	if (irq_status != MTK_DP_HPD_INTERRUPT) {
1906 		mtk_dp->irq_thread_handle |= MTK_DP_THREAD_CABLE_STATE_CHG;
1907 		cable_sta_chg = true;
1908 	}
1909 
1910 	spin_unlock_irqrestore(&mtk_dp->irq_thread_lock, flags);
1911 
1912 	if (cable_sta_chg) {
1913 		if (!!(mtk_dp_read(mtk_dp, MTK_DP_TRANS_P0_3414) &
1914 		       HPD_DB_DP_TRANS_P0_MASK))
1915 			mtk_dp->train_info.cable_plugged_in = true;
1916 		else
1917 			mtk_dp->train_info.cable_plugged_in = false;
1918 	}
1919 
1920 	return IRQ_WAKE_THREAD;
1921 }
1922 
1923 static int mtk_dp_wait_hpd_asserted(struct drm_dp_aux *mtk_aux, unsigned long wait_us)
1924 {
1925 	struct mtk_dp *mtk_dp = container_of(mtk_aux, struct mtk_dp, aux);
1926 	u32 val;
1927 	int ret;
1928 
1929 	ret = regmap_read_poll_timeout(mtk_dp->regs, MTK_DP_TRANS_P0_3414,
1930 				       val, !!(val & HPD_DB_DP_TRANS_P0_MASK),
1931 				       wait_us / 100, wait_us);
1932 	if (ret) {
1933 		mtk_dp->train_info.cable_plugged_in = false;
1934 		return ret;
1935 	}
1936 
1937 	mtk_dp->train_info.cable_plugged_in = true;
1938 
1939 	ret = mtk_dp_parse_capabilities(mtk_dp);
1940 	if (ret) {
1941 		drm_err(mtk_dp->drm_dev, "Can't parse capabilities\n");
1942 		return ret;
1943 	}
1944 
1945 	return 0;
1946 }
1947 
1948 static int mtk_dp_dt_parse(struct mtk_dp *mtk_dp,
1949 			   struct platform_device *pdev)
1950 {
1951 	struct device_node *endpoint;
1952 	struct device *dev = &pdev->dev;
1953 	int ret;
1954 	void __iomem *base;
1955 	u32 linkrate;
1956 	int len;
1957 
1958 	base = devm_platform_ioremap_resource(pdev, 0);
1959 	if (IS_ERR(base))
1960 		return PTR_ERR(base);
1961 
1962 	mtk_dp->regs = devm_regmap_init_mmio(dev, base, &mtk_dp_regmap_config);
1963 	if (IS_ERR(mtk_dp->regs))
1964 		return PTR_ERR(mtk_dp->regs);
1965 
1966 	endpoint = of_graph_get_endpoint_by_regs(pdev->dev.of_node, 1, -1);
1967 	len = of_property_count_elems_of_size(endpoint,
1968 					      "data-lanes", sizeof(u32));
1969 	if (len < 0 || len > 4 || len == 3) {
1970 		dev_err(dev, "invalid data lane size: %d\n", len);
1971 		return -EINVAL;
1972 	}
1973 
1974 	mtk_dp->max_lanes = len;
1975 
1976 	ret = device_property_read_u32(dev, "max-linkrate-mhz", &linkrate);
1977 	if (ret) {
1978 		dev_err(dev, "failed to read max linkrate: %d\n", ret);
1979 		return ret;
1980 	}
1981 
1982 	mtk_dp->max_linkrate = drm_dp_link_rate_to_bw_code(linkrate * 100);
1983 
1984 	return 0;
1985 }
1986 
1987 static void mtk_dp_update_plugged_status(struct mtk_dp *mtk_dp)
1988 {
1989 	if (!mtk_dp->data->audio_supported || !mtk_dp->audio_enable)
1990 		return;
1991 
1992 	mutex_lock(&mtk_dp->update_plugged_status_lock);
1993 	if (mtk_dp->plugged_cb && mtk_dp->codec_dev)
1994 		mtk_dp->plugged_cb(mtk_dp->codec_dev,
1995 				   mtk_dp->enabled &
1996 				   mtk_dp->info.audio_cur_cfg.detect_monitor);
1997 	mutex_unlock(&mtk_dp->update_plugged_status_lock);
1998 }
1999 
2000 static enum drm_connector_status mtk_dp_bdg_detect(struct drm_bridge *bridge)
2001 {
2002 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2003 	enum drm_connector_status ret = connector_status_disconnected;
2004 	bool enabled = mtk_dp->enabled;
2005 	u8 sink_count = 0;
2006 
2007 	if (!mtk_dp->train_info.cable_plugged_in)
2008 		return ret;
2009 
2010 	if (!enabled)
2011 		mtk_dp_aux_panel_poweron(mtk_dp, true);
2012 
2013 	/*
2014 	 * Some dongles still source HPD when they do not connect to any
2015 	 * sink device. To avoid this, we need to read the sink count
2016 	 * to make sure we do connect to sink devices. After this detect
2017 	 * function, we just need to check the HPD connection to check
2018 	 * whether we connect to a sink device.
2019 	 */
2020 	drm_dp_dpcd_readb(&mtk_dp->aux, DP_SINK_COUNT, &sink_count);
2021 	if (DP_GET_SINK_COUNT(sink_count))
2022 		ret = connector_status_connected;
2023 
2024 	if (!enabled)
2025 		mtk_dp_aux_panel_poweron(mtk_dp, false);
2026 
2027 	return ret;
2028 }
2029 
2030 static struct edid *mtk_dp_get_edid(struct drm_bridge *bridge,
2031 				    struct drm_connector *connector)
2032 {
2033 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2034 	bool enabled = mtk_dp->enabled;
2035 	struct edid *new_edid = NULL;
2036 	struct mtk_dp_audio_cfg *audio_caps = &mtk_dp->info.audio_cur_cfg;
2037 	struct cea_sad *sads;
2038 
2039 	if (!enabled) {
2040 		drm_atomic_bridge_chain_pre_enable(bridge, connector->state->state);
2041 		mtk_dp_aux_panel_poweron(mtk_dp, true);
2042 	}
2043 
2044 	new_edid = drm_get_edid(connector, &mtk_dp->aux.ddc);
2045 
2046 	/*
2047 	 * Parse capability here to let atomic_get_input_bus_fmts and
2048 	 * mode_valid use the capability to calculate sink bitrates.
2049 	 */
2050 	if (mtk_dp_parse_capabilities(mtk_dp)) {
2051 		drm_err(mtk_dp->drm_dev, "Can't parse capabilities\n");
2052 		new_edid = NULL;
2053 	}
2054 
2055 	if (new_edid) {
2056 		audio_caps->sad_count = drm_edid_to_sad(new_edid, &sads);
2057 		audio_caps->detect_monitor = drm_detect_monitor_audio(new_edid);
2058 	}
2059 
2060 	if (!enabled) {
2061 		mtk_dp_aux_panel_poweron(mtk_dp, false);
2062 		drm_atomic_bridge_chain_post_disable(bridge, connector->state->state);
2063 	}
2064 
2065 	return new_edid;
2066 }
2067 
2068 static ssize_t mtk_dp_aux_transfer(struct drm_dp_aux *mtk_aux,
2069 				   struct drm_dp_aux_msg *msg)
2070 {
2071 	struct mtk_dp *mtk_dp = container_of(mtk_aux, struct mtk_dp, aux);
2072 	bool is_read;
2073 	u8 request;
2074 	size_t accessed_bytes = 0;
2075 	int ret;
2076 
2077 	if (mtk_dp->bridge.type != DRM_MODE_CONNECTOR_eDP &&
2078 	    !mtk_dp->train_info.cable_plugged_in) {
2079 		ret = -EAGAIN;
2080 		goto err;
2081 	}
2082 
2083 	switch (msg->request) {
2084 	case DP_AUX_I2C_MOT:
2085 	case DP_AUX_I2C_WRITE:
2086 	case DP_AUX_NATIVE_WRITE:
2087 	case DP_AUX_I2C_WRITE_STATUS_UPDATE:
2088 	case DP_AUX_I2C_WRITE_STATUS_UPDATE | DP_AUX_I2C_MOT:
2089 		request = msg->request & ~DP_AUX_I2C_WRITE_STATUS_UPDATE;
2090 		is_read = false;
2091 		break;
2092 	case DP_AUX_I2C_READ:
2093 	case DP_AUX_NATIVE_READ:
2094 	case DP_AUX_I2C_READ | DP_AUX_I2C_MOT:
2095 		request = msg->request;
2096 		is_read = true;
2097 		break;
2098 	default:
2099 		dev_err(mtk_dp->dev, "invalid aux cmd = %d\n",
2100 			msg->request);
2101 		ret = -EINVAL;
2102 		goto err;
2103 	}
2104 
2105 	do {
2106 		size_t to_access = min_t(size_t, DP_AUX_MAX_PAYLOAD_BYTES,
2107 					 msg->size - accessed_bytes);
2108 
2109 		ret = mtk_dp_aux_do_transfer(mtk_dp, is_read, request,
2110 					     msg->address + accessed_bytes,
2111 					     msg->buffer + accessed_bytes,
2112 					     to_access, &msg->reply);
2113 
2114 		if (ret) {
2115 			dev_info(mtk_dp->dev,
2116 				 "Failed to do AUX transfer: %d\n", ret);
2117 			goto err;
2118 		}
2119 		accessed_bytes += to_access;
2120 	} while (accessed_bytes < msg->size);
2121 
2122 	return msg->size;
2123 err:
2124 	msg->reply = DP_AUX_NATIVE_REPLY_NACK | DP_AUX_I2C_REPLY_NACK;
2125 	return ret;
2126 }
2127 
2128 static int mtk_dp_poweron(struct mtk_dp *mtk_dp)
2129 {
2130 	int ret;
2131 
2132 	ret = phy_init(mtk_dp->phy);
2133 	if (ret) {
2134 		dev_err(mtk_dp->dev, "Failed to initialize phy: %d\n", ret);
2135 		return ret;
2136 	}
2137 
2138 	mtk_dp_init_port(mtk_dp);
2139 	mtk_dp_power_enable(mtk_dp);
2140 
2141 	return 0;
2142 }
2143 
2144 static void mtk_dp_poweroff(struct mtk_dp *mtk_dp)
2145 {
2146 	mtk_dp_power_disable(mtk_dp);
2147 	phy_exit(mtk_dp->phy);
2148 }
2149 
2150 static int mtk_dp_bridge_attach(struct drm_bridge *bridge,
2151 				enum drm_bridge_attach_flags flags)
2152 {
2153 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2154 	int ret;
2155 
2156 	if (!(flags & DRM_BRIDGE_ATTACH_NO_CONNECTOR)) {
2157 		dev_err(mtk_dp->dev, "Driver does not provide a connector!");
2158 		return -EINVAL;
2159 	}
2160 
2161 	mtk_dp->aux.drm_dev = bridge->dev;
2162 	ret = drm_dp_aux_register(&mtk_dp->aux);
2163 	if (ret) {
2164 		dev_err(mtk_dp->dev,
2165 			"failed to register DP AUX channel: %d\n", ret);
2166 		return ret;
2167 	}
2168 
2169 	ret = mtk_dp_poweron(mtk_dp);
2170 	if (ret)
2171 		goto err_aux_register;
2172 
2173 	if (mtk_dp->next_bridge) {
2174 		ret = drm_bridge_attach(bridge->encoder, mtk_dp->next_bridge,
2175 					&mtk_dp->bridge, flags);
2176 		if (ret) {
2177 			drm_warn(mtk_dp->drm_dev,
2178 				 "Failed to attach external bridge: %d\n", ret);
2179 			goto err_bridge_attach;
2180 		}
2181 	}
2182 
2183 	mtk_dp->drm_dev = bridge->dev;
2184 
2185 	if (mtk_dp->bridge.type != DRM_MODE_CONNECTOR_eDP) {
2186 		irq_clear_status_flags(mtk_dp->irq, IRQ_NOAUTOEN);
2187 		enable_irq(mtk_dp->irq);
2188 		mtk_dp_hwirq_enable(mtk_dp, true);
2189 	}
2190 
2191 	return 0;
2192 
2193 err_bridge_attach:
2194 	mtk_dp_poweroff(mtk_dp);
2195 err_aux_register:
2196 	drm_dp_aux_unregister(&mtk_dp->aux);
2197 	return ret;
2198 }
2199 
2200 static void mtk_dp_bridge_detach(struct drm_bridge *bridge)
2201 {
2202 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2203 
2204 	if (mtk_dp->bridge.type != DRM_MODE_CONNECTOR_eDP) {
2205 		mtk_dp_hwirq_enable(mtk_dp, false);
2206 		disable_irq(mtk_dp->irq);
2207 	}
2208 	mtk_dp->drm_dev = NULL;
2209 	mtk_dp_poweroff(mtk_dp);
2210 	drm_dp_aux_unregister(&mtk_dp->aux);
2211 }
2212 
2213 static void mtk_dp_bridge_atomic_enable(struct drm_bridge *bridge,
2214 					struct drm_bridge_state *old_state)
2215 {
2216 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2217 	int ret;
2218 
2219 	mtk_dp->conn = drm_atomic_get_new_connector_for_encoder(old_state->base.state,
2220 								bridge->encoder);
2221 	if (!mtk_dp->conn) {
2222 		drm_err(mtk_dp->drm_dev,
2223 			"Can't enable bridge as connector is missing\n");
2224 		return;
2225 	}
2226 
2227 	mtk_dp_aux_panel_poweron(mtk_dp, true);
2228 
2229 	/* Training */
2230 	ret = mtk_dp_training(mtk_dp);
2231 	if (ret) {
2232 		drm_err(mtk_dp->drm_dev, "Training failed, %d\n", ret);
2233 		goto power_off_aux;
2234 	}
2235 
2236 	ret = mtk_dp_video_config(mtk_dp);
2237 	if (ret)
2238 		goto power_off_aux;
2239 
2240 	mtk_dp_video_enable(mtk_dp, true);
2241 
2242 	mtk_dp->audio_enable =
2243 		mtk_dp_edid_parse_audio_capabilities(mtk_dp,
2244 						     &mtk_dp->info.audio_cur_cfg);
2245 	if (mtk_dp->audio_enable) {
2246 		mtk_dp_audio_setup(mtk_dp, &mtk_dp->info.audio_cur_cfg);
2247 		mtk_dp_audio_mute(mtk_dp, false);
2248 	} else {
2249 		memset(&mtk_dp->info.audio_cur_cfg, 0,
2250 		       sizeof(mtk_dp->info.audio_cur_cfg));
2251 	}
2252 
2253 	mtk_dp->enabled = true;
2254 	mtk_dp_update_plugged_status(mtk_dp);
2255 
2256 	return;
2257 power_off_aux:
2258 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
2259 			   DP_PWR_STATE_BANDGAP_TPLL,
2260 			   DP_PWR_STATE_MASK);
2261 }
2262 
2263 static void mtk_dp_bridge_atomic_disable(struct drm_bridge *bridge,
2264 					 struct drm_bridge_state *old_state)
2265 {
2266 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2267 
2268 	mtk_dp->enabled = false;
2269 	mtk_dp_update_plugged_status(mtk_dp);
2270 	mtk_dp_video_enable(mtk_dp, false);
2271 	mtk_dp_audio_mute(mtk_dp, true);
2272 
2273 	if (mtk_dp->train_info.cable_plugged_in) {
2274 		drm_dp_dpcd_writeb(&mtk_dp->aux, DP_SET_POWER, DP_SET_POWER_D3);
2275 		usleep_range(2000, 3000);
2276 	}
2277 
2278 	/* power off aux */
2279 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
2280 			   DP_PWR_STATE_BANDGAP_TPLL,
2281 			   DP_PWR_STATE_MASK);
2282 
2283 	/* Ensure the sink is muted */
2284 	msleep(20);
2285 }
2286 
2287 static enum drm_mode_status
2288 mtk_dp_bridge_mode_valid(struct drm_bridge *bridge,
2289 			 const struct drm_display_info *info,
2290 			 const struct drm_display_mode *mode)
2291 {
2292 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2293 	u32 bpp = info->color_formats & DRM_COLOR_FORMAT_YCBCR422 ? 16 : 24;
2294 	u32 rate = min_t(u32, drm_dp_max_link_rate(mtk_dp->rx_cap) *
2295 			      drm_dp_max_lane_count(mtk_dp->rx_cap),
2296 			 drm_dp_bw_code_to_link_rate(mtk_dp->max_linkrate) *
2297 			 mtk_dp->max_lanes);
2298 
2299 	if (rate < mode->clock * bpp / 8)
2300 		return MODE_CLOCK_HIGH;
2301 
2302 	return MODE_OK;
2303 }
2304 
2305 static u32 *mtk_dp_bridge_atomic_get_output_bus_fmts(struct drm_bridge *bridge,
2306 						     struct drm_bridge_state *bridge_state,
2307 						     struct drm_crtc_state *crtc_state,
2308 						     struct drm_connector_state *conn_state,
2309 						     unsigned int *num_output_fmts)
2310 {
2311 	u32 *output_fmts;
2312 
2313 	*num_output_fmts = 0;
2314 	output_fmts = kmalloc(sizeof(*output_fmts), GFP_KERNEL);
2315 	if (!output_fmts)
2316 		return NULL;
2317 	*num_output_fmts = 1;
2318 	output_fmts[0] = MEDIA_BUS_FMT_FIXED;
2319 	return output_fmts;
2320 }
2321 
2322 static const u32 mt8195_input_fmts[] = {
2323 	MEDIA_BUS_FMT_RGB888_1X24,
2324 	MEDIA_BUS_FMT_YUV8_1X24,
2325 	MEDIA_BUS_FMT_YUYV8_1X16,
2326 };
2327 
2328 static u32 *mtk_dp_bridge_atomic_get_input_bus_fmts(struct drm_bridge *bridge,
2329 						    struct drm_bridge_state *bridge_state,
2330 						    struct drm_crtc_state *crtc_state,
2331 						    struct drm_connector_state *conn_state,
2332 						    u32 output_fmt,
2333 						    unsigned int *num_input_fmts)
2334 {
2335 	u32 *input_fmts;
2336 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2337 	struct drm_display_mode *mode = &crtc_state->adjusted_mode;
2338 	struct drm_display_info *display_info =
2339 		&conn_state->connector->display_info;
2340 	u32 rate = min_t(u32, drm_dp_max_link_rate(mtk_dp->rx_cap) *
2341 			      drm_dp_max_lane_count(mtk_dp->rx_cap),
2342 			 drm_dp_bw_code_to_link_rate(mtk_dp->max_linkrate) *
2343 			 mtk_dp->max_lanes);
2344 
2345 	*num_input_fmts = 0;
2346 
2347 	/*
2348 	 * If the linkrate is smaller than datarate of RGB888, larger than
2349 	 * datarate of YUV422 and sink device supports YUV422, we output YUV422
2350 	 * format. Use this condition, we can support more resolution.
2351 	 */
2352 	if ((rate < (mode->clock * 24 / 8)) &&
2353 	    (rate > (mode->clock * 16 / 8)) &&
2354 	    (display_info->color_formats & DRM_COLOR_FORMAT_YCBCR422)) {
2355 		input_fmts = kcalloc(1, sizeof(*input_fmts), GFP_KERNEL);
2356 		if (!input_fmts)
2357 			return NULL;
2358 		*num_input_fmts = 1;
2359 		input_fmts[0] = MEDIA_BUS_FMT_YUYV8_1X16;
2360 	} else {
2361 		input_fmts = kcalloc(ARRAY_SIZE(mt8195_input_fmts),
2362 				     sizeof(*input_fmts),
2363 				     GFP_KERNEL);
2364 		if (!input_fmts)
2365 			return NULL;
2366 
2367 		*num_input_fmts = ARRAY_SIZE(mt8195_input_fmts);
2368 		memcpy(input_fmts, mt8195_input_fmts, sizeof(mt8195_input_fmts));
2369 	}
2370 
2371 	return input_fmts;
2372 }
2373 
2374 static int mtk_dp_bridge_atomic_check(struct drm_bridge *bridge,
2375 				      struct drm_bridge_state *bridge_state,
2376 				      struct drm_crtc_state *crtc_state,
2377 				      struct drm_connector_state *conn_state)
2378 {
2379 	struct mtk_dp *mtk_dp = mtk_dp_from_bridge(bridge);
2380 	struct drm_crtc *crtc = conn_state->crtc;
2381 	unsigned int input_bus_format;
2382 
2383 	input_bus_format = bridge_state->input_bus_cfg.format;
2384 
2385 	dev_dbg(mtk_dp->dev, "input format 0x%04x, output format 0x%04x\n",
2386 		bridge_state->input_bus_cfg.format,
2387 		 bridge_state->output_bus_cfg.format);
2388 
2389 	if (input_bus_format == MEDIA_BUS_FMT_YUYV8_1X16)
2390 		mtk_dp->info.format = DP_PIXELFORMAT_YUV422;
2391 	else
2392 		mtk_dp->info.format = DP_PIXELFORMAT_RGB;
2393 
2394 	if (!crtc) {
2395 		drm_err(mtk_dp->drm_dev,
2396 			"Can't enable bridge as connector state doesn't have a crtc\n");
2397 		return -EINVAL;
2398 	}
2399 
2400 	drm_display_mode_to_videomode(&crtc_state->adjusted_mode, &mtk_dp->info.vm);
2401 
2402 	return 0;
2403 }
2404 
2405 static const struct drm_bridge_funcs mtk_dp_bridge_funcs = {
2406 	.atomic_check = mtk_dp_bridge_atomic_check,
2407 	.atomic_duplicate_state = drm_atomic_helper_bridge_duplicate_state,
2408 	.atomic_destroy_state = drm_atomic_helper_bridge_destroy_state,
2409 	.atomic_get_output_bus_fmts = mtk_dp_bridge_atomic_get_output_bus_fmts,
2410 	.atomic_get_input_bus_fmts = mtk_dp_bridge_atomic_get_input_bus_fmts,
2411 	.atomic_reset = drm_atomic_helper_bridge_reset,
2412 	.attach = mtk_dp_bridge_attach,
2413 	.detach = mtk_dp_bridge_detach,
2414 	.atomic_enable = mtk_dp_bridge_atomic_enable,
2415 	.atomic_disable = mtk_dp_bridge_atomic_disable,
2416 	.mode_valid = mtk_dp_bridge_mode_valid,
2417 	.get_edid = mtk_dp_get_edid,
2418 	.detect = mtk_dp_bdg_detect,
2419 };
2420 
2421 static void mtk_dp_debounce_timer(struct timer_list *t)
2422 {
2423 	struct mtk_dp *mtk_dp = from_timer(mtk_dp, t, debounce_timer);
2424 
2425 	mtk_dp->need_debounce = true;
2426 }
2427 
2428 /*
2429  * HDMI audio codec callbacks
2430  */
2431 static int mtk_dp_audio_hw_params(struct device *dev, void *data,
2432 				  struct hdmi_codec_daifmt *daifmt,
2433 				  struct hdmi_codec_params *params)
2434 {
2435 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2436 
2437 	if (!mtk_dp->enabled) {
2438 		dev_err(mtk_dp->dev, "%s, DP is not ready!\n", __func__);
2439 		return -ENODEV;
2440 	}
2441 
2442 	mtk_dp->info.audio_cur_cfg.channels = params->cea.channels;
2443 	mtk_dp->info.audio_cur_cfg.sample_rate = params->sample_rate;
2444 
2445 	mtk_dp_audio_setup(mtk_dp, &mtk_dp->info.audio_cur_cfg);
2446 
2447 	return 0;
2448 }
2449 
2450 static int mtk_dp_audio_startup(struct device *dev, void *data)
2451 {
2452 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2453 
2454 	mtk_dp_audio_mute(mtk_dp, false);
2455 
2456 	return 0;
2457 }
2458 
2459 static void mtk_dp_audio_shutdown(struct device *dev, void *data)
2460 {
2461 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2462 
2463 	mtk_dp_audio_mute(mtk_dp, true);
2464 }
2465 
2466 static int mtk_dp_audio_get_eld(struct device *dev, void *data, uint8_t *buf,
2467 				size_t len)
2468 {
2469 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2470 
2471 	if (mtk_dp->enabled)
2472 		memcpy(buf, mtk_dp->conn->eld, len);
2473 	else
2474 		memset(buf, 0, len);
2475 
2476 	return 0;
2477 }
2478 
2479 static int mtk_dp_audio_hook_plugged_cb(struct device *dev, void *data,
2480 					hdmi_codec_plugged_cb fn,
2481 					struct device *codec_dev)
2482 {
2483 	struct mtk_dp *mtk_dp = data;
2484 
2485 	mutex_lock(&mtk_dp->update_plugged_status_lock);
2486 	mtk_dp->plugged_cb = fn;
2487 	mtk_dp->codec_dev = codec_dev;
2488 	mutex_unlock(&mtk_dp->update_plugged_status_lock);
2489 
2490 	mtk_dp_update_plugged_status(mtk_dp);
2491 
2492 	return 0;
2493 }
2494 
2495 static const struct hdmi_codec_ops mtk_dp_audio_codec_ops = {
2496 	.hw_params = mtk_dp_audio_hw_params,
2497 	.audio_startup = mtk_dp_audio_startup,
2498 	.audio_shutdown = mtk_dp_audio_shutdown,
2499 	.get_eld = mtk_dp_audio_get_eld,
2500 	.hook_plugged_cb = mtk_dp_audio_hook_plugged_cb,
2501 	.no_capture_mute = 1,
2502 };
2503 
2504 static int mtk_dp_register_audio_driver(struct device *dev)
2505 {
2506 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2507 	struct hdmi_codec_pdata codec_data = {
2508 		.ops = &mtk_dp_audio_codec_ops,
2509 		.max_i2s_channels = 8,
2510 		.i2s = 1,
2511 		.data = mtk_dp,
2512 	};
2513 
2514 	mtk_dp->audio_pdev = platform_device_register_data(dev,
2515 							   HDMI_CODEC_DRV_NAME,
2516 							   PLATFORM_DEVID_AUTO,
2517 							   &codec_data,
2518 							   sizeof(codec_data));
2519 	return PTR_ERR_OR_ZERO(mtk_dp->audio_pdev);
2520 }
2521 
2522 static int mtk_dp_register_phy(struct mtk_dp *mtk_dp)
2523 {
2524 	struct device *dev = mtk_dp->dev;
2525 
2526 	mtk_dp->phy_dev = platform_device_register_data(dev, "mediatek-dp-phy",
2527 							PLATFORM_DEVID_AUTO,
2528 							&mtk_dp->regs,
2529 							sizeof(struct regmap *));
2530 	if (IS_ERR(mtk_dp->phy_dev))
2531 		return dev_err_probe(dev, PTR_ERR(mtk_dp->phy_dev),
2532 				     "Failed to create device mediatek-dp-phy\n");
2533 
2534 	mtk_dp_get_calibration_data(mtk_dp);
2535 
2536 	mtk_dp->phy = devm_phy_get(&mtk_dp->phy_dev->dev, "dp");
2537 	if (IS_ERR(mtk_dp->phy)) {
2538 		platform_device_unregister(mtk_dp->phy_dev);
2539 		return dev_err_probe(dev, PTR_ERR(mtk_dp->phy), "Failed to get phy\n");
2540 	}
2541 
2542 	return 0;
2543 }
2544 
2545 static int mtk_dp_edp_link_panel(struct drm_dp_aux *mtk_aux)
2546 {
2547 	struct mtk_dp *mtk_dp = container_of(mtk_aux, struct mtk_dp, aux);
2548 	struct device *dev = mtk_aux->dev;
2549 	int ret;
2550 
2551 	mtk_dp->next_bridge = devm_drm_of_get_bridge(dev, dev->of_node, 1, 0);
2552 
2553 	/* Power off the DP and AUX: either detection is done, or no panel present */
2554 	mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
2555 			   DP_PWR_STATE_BANDGAP_TPLL,
2556 			   DP_PWR_STATE_MASK);
2557 	mtk_dp_power_disable(mtk_dp);
2558 
2559 	if (IS_ERR(mtk_dp->next_bridge)) {
2560 		ret = PTR_ERR(mtk_dp->next_bridge);
2561 		mtk_dp->next_bridge = NULL;
2562 		return ret;
2563 	}
2564 
2565 	/* For eDP, we add the bridge only if the panel was found */
2566 	ret = devm_drm_bridge_add(dev, &mtk_dp->bridge);
2567 	if (ret)
2568 		return ret;
2569 
2570 	return 0;
2571 }
2572 
2573 static int mtk_dp_probe(struct platform_device *pdev)
2574 {
2575 	struct mtk_dp *mtk_dp;
2576 	struct device *dev = &pdev->dev;
2577 	int ret;
2578 
2579 	mtk_dp = devm_kzalloc(dev, sizeof(*mtk_dp), GFP_KERNEL);
2580 	if (!mtk_dp)
2581 		return -ENOMEM;
2582 
2583 	mtk_dp->dev = dev;
2584 	mtk_dp->data = (struct mtk_dp_data *)of_device_get_match_data(dev);
2585 
2586 	ret = mtk_dp_dt_parse(mtk_dp, pdev);
2587 	if (ret)
2588 		return dev_err_probe(dev, ret, "Failed to parse dt\n");
2589 
2590 	/*
2591 	 * Request the interrupt and install service routine only if we are
2592 	 * on full DisplayPort.
2593 	 * For eDP, polling the HPD instead is more convenient because we
2594 	 * don't expect any (un)plug events during runtime, hence we can
2595 	 * avoid some locking.
2596 	 */
2597 	if (mtk_dp->data->bridge_type != DRM_MODE_CONNECTOR_eDP) {
2598 		mtk_dp->irq = platform_get_irq(pdev, 0);
2599 		if (mtk_dp->irq < 0)
2600 			return dev_err_probe(dev, mtk_dp->irq,
2601 					     "failed to request dp irq resource\n");
2602 
2603 		spin_lock_init(&mtk_dp->irq_thread_lock);
2604 
2605 		irq_set_status_flags(mtk_dp->irq, IRQ_NOAUTOEN);
2606 		ret = devm_request_threaded_irq(dev, mtk_dp->irq, mtk_dp_hpd_event,
2607 						mtk_dp_hpd_event_thread,
2608 						IRQ_TYPE_LEVEL_HIGH, dev_name(dev),
2609 						mtk_dp);
2610 		if (ret)
2611 			return dev_err_probe(dev, ret,
2612 					     "failed to request mediatek dptx irq\n");
2613 
2614 		mtk_dp->need_debounce = true;
2615 		timer_setup(&mtk_dp->debounce_timer, mtk_dp_debounce_timer, 0);
2616 	}
2617 
2618 	mtk_dp->aux.name = "aux_mtk_dp";
2619 	mtk_dp->aux.dev = dev;
2620 	mtk_dp->aux.transfer = mtk_dp_aux_transfer;
2621 	mtk_dp->aux.wait_hpd_asserted = mtk_dp_wait_hpd_asserted;
2622 	drm_dp_aux_init(&mtk_dp->aux);
2623 
2624 	platform_set_drvdata(pdev, mtk_dp);
2625 
2626 	if (mtk_dp->data->audio_supported) {
2627 		mutex_init(&mtk_dp->update_plugged_status_lock);
2628 
2629 		ret = mtk_dp_register_audio_driver(dev);
2630 		if (ret) {
2631 			dev_err(dev, "Failed to register audio driver: %d\n",
2632 				ret);
2633 			return ret;
2634 		}
2635 	}
2636 
2637 	ret = mtk_dp_register_phy(mtk_dp);
2638 	if (ret)
2639 		return ret;
2640 
2641 	mtk_dp->bridge.funcs = &mtk_dp_bridge_funcs;
2642 	mtk_dp->bridge.of_node = dev->of_node;
2643 	mtk_dp->bridge.type = mtk_dp->data->bridge_type;
2644 
2645 	if (mtk_dp->bridge.type == DRM_MODE_CONNECTOR_eDP) {
2646 		/*
2647 		 * Set the data lanes to idle in case the bootloader didn't
2648 		 * properly close the eDP port to avoid stalls and then
2649 		 * reinitialize, reset and power on the AUX block.
2650 		 */
2651 		mtk_dp_set_idle_pattern(mtk_dp, true);
2652 		mtk_dp_initialize_aux_settings(mtk_dp);
2653 		mtk_dp_power_enable(mtk_dp);
2654 
2655 		/* Disable HW interrupts: we don't need any for eDP */
2656 		mtk_dp_hwirq_enable(mtk_dp, false);
2657 
2658 		/*
2659 		 * Power on the AUX to allow reading the EDID from aux-bus:
2660 		 * please note that it is necessary to call power off in the
2661 		 * .done_probing() callback (mtk_dp_edp_link_panel), as only
2662 		 * there we can safely assume that we finished reading EDID.
2663 		 */
2664 		mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
2665 				   DP_PWR_STATE_BANDGAP_TPLL_LANE,
2666 				   DP_PWR_STATE_MASK);
2667 
2668 		ret = devm_of_dp_aux_populate_bus(&mtk_dp->aux, mtk_dp_edp_link_panel);
2669 		if (ret) {
2670 			/* -ENODEV this means that the panel is not on the aux-bus */
2671 			if (ret == -ENODEV) {
2672 				ret = mtk_dp_edp_link_panel(&mtk_dp->aux);
2673 				if (ret)
2674 					return ret;
2675 			} else {
2676 				mtk_dp_update_bits(mtk_dp, MTK_DP_TOP_PWR_STATE,
2677 						   DP_PWR_STATE_BANDGAP_TPLL,
2678 						   DP_PWR_STATE_MASK);
2679 				mtk_dp_power_disable(mtk_dp);
2680 				return ret;
2681 			}
2682 		}
2683 	} else {
2684 		mtk_dp->bridge.ops = DRM_BRIDGE_OP_DETECT |
2685 				     DRM_BRIDGE_OP_EDID | DRM_BRIDGE_OP_HPD;
2686 		ret = devm_drm_bridge_add(dev, &mtk_dp->bridge);
2687 		if (ret)
2688 			return dev_err_probe(dev, ret, "Failed to add bridge\n");
2689 	}
2690 
2691 	pm_runtime_enable(dev);
2692 	pm_runtime_get_sync(dev);
2693 
2694 	return 0;
2695 }
2696 
2697 static void mtk_dp_remove(struct platform_device *pdev)
2698 {
2699 	struct mtk_dp *mtk_dp = platform_get_drvdata(pdev);
2700 
2701 	pm_runtime_put(&pdev->dev);
2702 	pm_runtime_disable(&pdev->dev);
2703 	if (mtk_dp->data->bridge_type != DRM_MODE_CONNECTOR_eDP)
2704 		del_timer_sync(&mtk_dp->debounce_timer);
2705 	platform_device_unregister(mtk_dp->phy_dev);
2706 	if (mtk_dp->audio_pdev)
2707 		platform_device_unregister(mtk_dp->audio_pdev);
2708 }
2709 
2710 #ifdef CONFIG_PM_SLEEP
2711 static int mtk_dp_suspend(struct device *dev)
2712 {
2713 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2714 
2715 	mtk_dp_power_disable(mtk_dp);
2716 	if (mtk_dp->bridge.type != DRM_MODE_CONNECTOR_eDP)
2717 		mtk_dp_hwirq_enable(mtk_dp, false);
2718 	pm_runtime_put_sync(dev);
2719 
2720 	return 0;
2721 }
2722 
2723 static int mtk_dp_resume(struct device *dev)
2724 {
2725 	struct mtk_dp *mtk_dp = dev_get_drvdata(dev);
2726 
2727 	pm_runtime_get_sync(dev);
2728 	mtk_dp_init_port(mtk_dp);
2729 	if (mtk_dp->bridge.type != DRM_MODE_CONNECTOR_eDP)
2730 		mtk_dp_hwirq_enable(mtk_dp, true);
2731 	mtk_dp_power_enable(mtk_dp);
2732 
2733 	return 0;
2734 }
2735 #endif
2736 
2737 static SIMPLE_DEV_PM_OPS(mtk_dp_pm_ops, mtk_dp_suspend, mtk_dp_resume);
2738 
2739 static const struct mtk_dp_data mt8195_edp_data = {
2740 	.bridge_type = DRM_MODE_CONNECTOR_eDP,
2741 	.smc_cmd = MTK_DP_SIP_ATF_EDP_VIDEO_UNMUTE,
2742 	.efuse_fmt = mt8195_edp_efuse_fmt,
2743 	.audio_supported = false,
2744 };
2745 
2746 static const struct mtk_dp_data mt8195_dp_data = {
2747 	.bridge_type = DRM_MODE_CONNECTOR_DisplayPort,
2748 	.smc_cmd = MTK_DP_SIP_ATF_VIDEO_UNMUTE,
2749 	.efuse_fmt = mt8195_dp_efuse_fmt,
2750 	.audio_supported = true,
2751 };
2752 
2753 static const struct of_device_id mtk_dp_of_match[] = {
2754 	{
2755 		.compatible = "mediatek,mt8195-edp-tx",
2756 		.data = &mt8195_edp_data,
2757 	},
2758 	{
2759 		.compatible = "mediatek,mt8195-dp-tx",
2760 		.data = &mt8195_dp_data,
2761 	},
2762 	{},
2763 };
2764 MODULE_DEVICE_TABLE(of, mtk_dp_of_match);
2765 
2766 static struct platform_driver mtk_dp_driver = {
2767 	.probe = mtk_dp_probe,
2768 	.remove_new = mtk_dp_remove,
2769 	.driver = {
2770 		.name = "mediatek-drm-dp",
2771 		.of_match_table = mtk_dp_of_match,
2772 		.pm = &mtk_dp_pm_ops,
2773 	},
2774 };
2775 
2776 module_platform_driver(mtk_dp_driver);
2777 
2778 MODULE_AUTHOR("Jitao Shi <jitao.shi@mediatek.com>");
2779 MODULE_AUTHOR("Markus Schneider-Pargmann <msp@baylibre.com>");
2780 MODULE_AUTHOR("Bo-Chen Chen <rex-bc.chen@mediatek.com>");
2781 MODULE_DESCRIPTION("MediaTek DisplayPort Driver");
2782 MODULE_LICENSE("GPL");
2783