xref: /openbmc/linux/drivers/media/i2c/ds90ub960.c (revision 3180449e)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Driver for the Texas Instruments DS90UB960-Q1 video deserializer
4  *
5  * Copyright (c) 2019 Luca Ceresoli <luca@lucaceresoli.net>
6  * Copyright (c) 2023 Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>
7  */
8 
9 /*
10  * (Possible) TODOs:
11  *
12  * - PM for serializer and remote peripherals. We need to manage:
13  *   - VPOC
14  *     - Power domain? Regulator? Somehow any remote device should be able to
15  *       cause the VPOC to be turned on.
16  *   - Link between the deserializer and the serializer
17  *     - Related to VPOC management. We probably always want to turn on the VPOC
18  *       and then enable the link.
19  *   - Serializer's services: i2c, gpios, power
20  *     - The serializer needs to resume before the remote peripherals can
21  *       e.g. use the i2c.
22  *     - How to handle gpios? Reserving a gpio essentially keeps the provider
23  *       (serializer) always powered on.
24  * - Do we need a new bus for the FPD-Link? At the moment the serializers
25  *   are children of the same i2c-adapter where the deserializer resides.
26  * - i2c-atr could be made embeddable instead of allocatable.
27  */
28 
29 #include <linux/bitops.h>
30 #include <linux/clk.h>
31 #include <linux/delay.h>
32 #include <linux/fwnode.h>
33 #include <linux/gpio/consumer.h>
34 #include <linux/i2c-atr.h>
35 #include <linux/i2c.h>
36 #include <linux/init.h>
37 #include <linux/interrupt.h>
38 #include <linux/kernel.h>
39 #include <linux/kthread.h>
40 #include <linux/module.h>
41 #include <linux/mutex.h>
42 #include <linux/property.h>
43 #include <linux/regmap.h>
44 #include <linux/regulator/consumer.h>
45 #include <linux/slab.h>
46 #include <linux/workqueue.h>
47 
48 #include <media/i2c/ds90ub9xx.h>
49 #include <media/mipi-csi2.h>
50 #include <media/v4l2-ctrls.h>
51 #include <media/v4l2-event.h>
52 #include <media/v4l2-fwnode.h>
53 #include <media/v4l2-subdev.h>
54 
55 #define MHZ(v) ((u32)((v) * 1000000U))
56 
57 #define UB960_POLL_TIME_MS	500
58 
59 #define UB960_MAX_RX_NPORTS	4
60 #define UB960_MAX_TX_NPORTS	2
61 #define UB960_MAX_NPORTS	(UB960_MAX_RX_NPORTS + UB960_MAX_TX_NPORTS)
62 
63 #define UB960_MAX_PORT_ALIASES	8
64 
65 #define UB960_NUM_BC_GPIOS		4
66 
67 /*
68  * Register map
69  *
70  * 0x00-0x32   Shared (UB960_SR)
71  * 0x33-0x3a   CSI-2 TX (per-port paged on DS90UB960, shared on 954) (UB960_TR)
72  * 0x4c        Shared (UB960_SR)
73  * 0x4d-0x7f   FPD-Link RX, per-port paged (UB960_RR)
74  * 0xb0-0xbf   Shared (UB960_SR)
75  * 0xd0-0xdf   FPD-Link RX, per-port paged (UB960_RR)
76  * 0xf0-0xf5   Shared (UB960_SR)
77  * 0xf8-0xfb   Shared (UB960_SR)
78  * All others  Reserved
79  *
80  * Register prefixes:
81  * UB960_SR_* = Shared register
82  * UB960_RR_* = FPD-Link RX, per-port paged register
83  * UB960_TR_* = CSI-2 TX, per-port paged register
84  * UB960_XR_* = Reserved register
85  * UB960_IR_* = Indirect register
86  */
87 
88 #define UB960_SR_I2C_DEV_ID			0x00
89 #define UB960_SR_RESET				0x01
90 #define UB960_SR_RESET_DIGITAL_RESET1		BIT(1)
91 #define UB960_SR_RESET_DIGITAL_RESET0		BIT(0)
92 #define UB960_SR_RESET_GPIO_LOCK_RELEASE	BIT(5)
93 
94 #define UB960_SR_GEN_CONFIG			0x02
95 #define UB960_SR_REV_MASK			0x03
96 #define UB960_SR_DEVICE_STS			0x04
97 #define UB960_SR_PAR_ERR_THOLD_HI		0x05
98 #define UB960_SR_PAR_ERR_THOLD_LO		0x06
99 #define UB960_SR_BCC_WDOG_CTL			0x07
100 #define UB960_SR_I2C_CTL1			0x08
101 #define UB960_SR_I2C_CTL2			0x09
102 #define UB960_SR_SCL_HIGH_TIME			0x0a
103 #define UB960_SR_SCL_LOW_TIME			0x0b
104 #define UB960_SR_RX_PORT_CTL			0x0c
105 #define UB960_SR_IO_CTL				0x0d
106 #define UB960_SR_GPIO_PIN_STS			0x0e
107 #define UB960_SR_GPIO_INPUT_CTL			0x0f
108 #define UB960_SR_GPIO_PIN_CTL(n)		(0x10 + (n)) /* n < UB960_NUM_GPIOS */
109 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SEL		5
110 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_SRC_SHIFT	2
111 #define UB960_SR_GPIO_PIN_CTL_GPIO_OUT_EN		BIT(0)
112 
113 #define UB960_SR_FS_CTL				0x18
114 #define UB960_SR_FS_HIGH_TIME_1			0x19
115 #define UB960_SR_FS_HIGH_TIME_0			0x1a
116 #define UB960_SR_FS_LOW_TIME_1			0x1b
117 #define UB960_SR_FS_LOW_TIME_0			0x1c
118 #define UB960_SR_MAX_FRM_HI			0x1d
119 #define UB960_SR_MAX_FRM_LO			0x1e
120 #define UB960_SR_CSI_PLL_CTL			0x1f
121 
122 #define UB960_SR_FWD_CTL1			0x20
123 #define UB960_SR_FWD_CTL1_PORT_DIS(n)		BIT((n) + 4)
124 
125 #define UB960_SR_FWD_CTL2			0x21
126 #define UB960_SR_FWD_STS			0x22
127 
128 #define UB960_SR_INTERRUPT_CTL			0x23
129 #define UB960_SR_INTERRUPT_CTL_INT_EN		BIT(7)
130 #define UB960_SR_INTERRUPT_CTL_IE_CSI_TX0	BIT(4)
131 #define UB960_SR_INTERRUPT_CTL_IE_RX(n)		BIT((n)) /* rxport[n] IRQ */
132 
133 #define UB960_SR_INTERRUPT_STS			0x24
134 #define UB960_SR_INTERRUPT_STS_INT		BIT(7)
135 #define UB960_SR_INTERRUPT_STS_IS_CSI_TX(n)	BIT(4 + (n)) /* txport[n] IRQ */
136 #define UB960_SR_INTERRUPT_STS_IS_RX(n)		BIT((n)) /* rxport[n] IRQ */
137 
138 #define UB960_SR_TS_CONFIG			0x25
139 #define UB960_SR_TS_CONTROL			0x26
140 #define UB960_SR_TS_LINE_HI			0x27
141 #define UB960_SR_TS_LINE_LO			0x28
142 #define UB960_SR_TS_STATUS			0x29
143 #define UB960_SR_TIMESTAMP_P0_HI		0x2a
144 #define UB960_SR_TIMESTAMP_P0_LO		0x2b
145 #define UB960_SR_TIMESTAMP_P1_HI		0x2c
146 #define UB960_SR_TIMESTAMP_P1_LO		0x2d
147 
148 #define UB960_SR_CSI_PORT_SEL			0x32
149 
150 #define UB960_TR_CSI_CTL			0x33
151 #define UB960_TR_CSI_CTL_CSI_CAL_EN		BIT(6)
152 #define UB960_TR_CSI_CTL_CSI_ENABLE		BIT(0)
153 
154 #define UB960_TR_CSI_CTL2			0x34
155 #define UB960_TR_CSI_STS			0x35
156 #define UB960_TR_CSI_TX_ICR			0x36
157 
158 #define UB960_TR_CSI_TX_ISR			0x37
159 #define UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR	BIT(3)
160 #define UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR	BIT(1)
161 
162 #define UB960_TR_CSI_TEST_CTL			0x38
163 #define UB960_TR_CSI_TEST_PATT_HI		0x39
164 #define UB960_TR_CSI_TEST_PATT_LO		0x3a
165 
166 #define UB960_XR_SFILTER_CFG			0x41
167 #define UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT	4
168 #define UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT	0
169 
170 #define UB960_XR_AEQ_CTL1			0x42
171 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK	BIT(6)
172 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING	BIT(5)
173 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY	BIT(4)
174 #define UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK        \
175 	(UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_FPD_CLK |  \
176 	 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_ENCODING | \
177 	 UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_PARITY)
178 #define UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN	BIT(0)
179 
180 #define UB960_XR_AEQ_ERR_THOLD			0x43
181 
182 #define UB960_RR_BCC_ERR_CTL			0x46
183 #define UB960_RR_BCC_STATUS			0x47
184 #define UB960_RR_BCC_STATUS_SEQ_ERROR		BIT(5)
185 #define UB960_RR_BCC_STATUS_MASTER_ERR		BIT(4)
186 #define UB960_RR_BCC_STATUS_MASTER_TO		BIT(3)
187 #define UB960_RR_BCC_STATUS_SLAVE_ERR		BIT(2)
188 #define UB960_RR_BCC_STATUS_SLAVE_TO		BIT(1)
189 #define UB960_RR_BCC_STATUS_RESP_ERR		BIT(0)
190 #define UB960_RR_BCC_STATUS_ERROR_MASK                                    \
191 	(UB960_RR_BCC_STATUS_SEQ_ERROR | UB960_RR_BCC_STATUS_MASTER_ERR | \
192 	 UB960_RR_BCC_STATUS_MASTER_TO | UB960_RR_BCC_STATUS_SLAVE_ERR |  \
193 	 UB960_RR_BCC_STATUS_SLAVE_TO | UB960_RR_BCC_STATUS_RESP_ERR)
194 
195 #define UB960_RR_FPD3_CAP			0x4a
196 #define UB960_RR_RAW_EMBED_DTYPE		0x4b
197 #define UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT	6
198 
199 #define UB960_SR_FPD3_PORT_SEL			0x4c
200 
201 #define UB960_RR_RX_PORT_STS1			0x4d
202 #define UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR	BIT(5)
203 #define UB960_RR_RX_PORT_STS1_LOCK_STS_CHG	BIT(4)
204 #define UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR	BIT(3)
205 #define UB960_RR_RX_PORT_STS1_PARITY_ERROR	BIT(2)
206 #define UB960_RR_RX_PORT_STS1_PORT_PASS		BIT(1)
207 #define UB960_RR_RX_PORT_STS1_LOCK_STS		BIT(0)
208 #define UB960_RR_RX_PORT_STS1_ERROR_MASK       \
209 	(UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR | \
210 	 UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR | \
211 	 UB960_RR_RX_PORT_STS1_PARITY_ERROR)
212 
213 #define UB960_RR_RX_PORT_STS2			0x4e
214 #define UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE	BIT(7)
215 #define UB960_RR_RX_PORT_STS2_LINE_LEN_CHG	BIT(6)
216 #define UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR	BIT(5)
217 #define UB960_RR_RX_PORT_STS2_BUFFER_ERROR	BIT(4)
218 #define UB960_RR_RX_PORT_STS2_CSI_ERROR		BIT(3)
219 #define UB960_RR_RX_PORT_STS2_FREQ_STABLE	BIT(2)
220 #define UB960_RR_RX_PORT_STS2_CABLE_FAULT	BIT(1)
221 #define UB960_RR_RX_PORT_STS2_LINE_CNT_CHG	BIT(0)
222 #define UB960_RR_RX_PORT_STS2_ERROR_MASK       \
223 	UB960_RR_RX_PORT_STS2_BUFFER_ERROR
224 
225 #define UB960_RR_RX_FREQ_HIGH			0x4f
226 #define UB960_RR_RX_FREQ_LOW			0x50
227 #define UB960_RR_SENSOR_STS_0			0x51
228 #define UB960_RR_SENSOR_STS_1			0x52
229 #define UB960_RR_SENSOR_STS_2			0x53
230 #define UB960_RR_SENSOR_STS_3			0x54
231 #define UB960_RR_RX_PAR_ERR_HI			0x55
232 #define UB960_RR_RX_PAR_ERR_LO			0x56
233 #define UB960_RR_BIST_ERR_COUNT			0x57
234 
235 #define UB960_RR_BCC_CONFIG			0x58
236 #define UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH	BIT(6)
237 #define UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK	GENMASK(2, 0)
238 
239 #define UB960_RR_DATAPATH_CTL1			0x59
240 #define UB960_RR_DATAPATH_CTL2			0x5a
241 #define UB960_RR_SER_ID				0x5b
242 #define UB960_RR_SER_ALIAS_ID			0x5c
243 
244 /* For these two register sets: n < UB960_MAX_PORT_ALIASES */
245 #define UB960_RR_SLAVE_ID(n)			(0x5d + (n))
246 #define UB960_RR_SLAVE_ALIAS(n)			(0x65 + (n))
247 
248 #define UB960_RR_PORT_CONFIG			0x6d
249 #define UB960_RR_PORT_CONFIG_FPD3_MODE_MASK	GENMASK(1, 0)
250 
251 #define UB960_RR_BC_GPIO_CTL(n)			(0x6e + (n)) /* n < 2 */
252 #define UB960_RR_RAW10_ID			0x70
253 #define UB960_RR_RAW10_ID_VC_SHIFT		6
254 #define UB960_RR_RAW10_ID_DT_SHIFT		0
255 
256 #define UB960_RR_RAW12_ID			0x71
257 #define UB960_RR_CSI_VC_MAP			0x72
258 #define UB960_RR_CSI_VC_MAP_SHIFT(x)		((x) * 2)
259 
260 #define UB960_RR_LINE_COUNT_HI			0x73
261 #define UB960_RR_LINE_COUNT_LO			0x74
262 #define UB960_RR_LINE_LEN_1			0x75
263 #define UB960_RR_LINE_LEN_0			0x76
264 #define UB960_RR_FREQ_DET_CTL			0x77
265 #define UB960_RR_MAILBOX_1			0x78
266 #define UB960_RR_MAILBOX_2			0x79
267 
268 #define UB960_RR_CSI_RX_STS			0x7a
269 #define UB960_RR_CSI_RX_STS_LENGTH_ERR		BIT(3)
270 #define UB960_RR_CSI_RX_STS_CKSUM_ERR		BIT(2)
271 #define UB960_RR_CSI_RX_STS_ECC2_ERR		BIT(1)
272 #define UB960_RR_CSI_RX_STS_ECC1_ERR		BIT(0)
273 #define UB960_RR_CSI_RX_STS_ERROR_MASK                                    \
274 	(UB960_RR_CSI_RX_STS_LENGTH_ERR | UB960_RR_CSI_RX_STS_CKSUM_ERR | \
275 	 UB960_RR_CSI_RX_STS_ECC2_ERR | UB960_RR_CSI_RX_STS_ECC1_ERR)
276 
277 #define UB960_RR_CSI_ERR_COUNTER		0x7b
278 #define UB960_RR_PORT_CONFIG2			0x7c
279 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK GENMASK(7, 6)
280 #define UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT 6
281 
282 #define UB960_RR_PORT_CONFIG2_LV_POL_LOW	BIT(1)
283 #define UB960_RR_PORT_CONFIG2_FV_POL_LOW	BIT(0)
284 
285 #define UB960_RR_PORT_PASS_CTL			0x7d
286 #define UB960_RR_SEN_INT_RISE_CTL		0x7e
287 #define UB960_RR_SEN_INT_FALL_CTL		0x7f
288 
289 #define UB960_SR_CSI_FRAME_COUNT_HI(n)		(0x90 + 8 * (n))
290 #define UB960_SR_CSI_FRAME_COUNT_LO(n)		(0x91 + 8 * (n))
291 #define UB960_SR_CSI_FRAME_ERR_COUNT_HI(n)	(0x92 + 8 * (n))
292 #define UB960_SR_CSI_FRAME_ERR_COUNT_LO(n)	(0x93 + 8 * (n))
293 #define UB960_SR_CSI_LINE_COUNT_HI(n)		(0x94 + 8 * (n))
294 #define UB960_SR_CSI_LINE_COUNT_LO(n)		(0x95 + 8 * (n))
295 #define UB960_SR_CSI_LINE_ERR_COUNT_HI(n)	(0x96 + 8 * (n))
296 #define UB960_SR_CSI_LINE_ERR_COUNT_LO(n)	(0x97 + 8 * (n))
297 
298 #define UB960_XR_REFCLK_FREQ			0xa5	/* UB960 */
299 
300 #define UB960_RR_VC_ID_MAP(x)			(0xa0 + (x)) /* UB9702 */
301 
302 #define UB960_SR_IND_ACC_CTL			0xb0
303 #define UB960_SR_IND_ACC_CTL_IA_AUTO_INC	BIT(1)
304 
305 #define UB960_SR_IND_ACC_ADDR			0xb1
306 #define UB960_SR_IND_ACC_DATA			0xb2
307 #define UB960_SR_BIST_CONTROL			0xb3
308 #define UB960_SR_MODE_IDX_STS			0xb8
309 #define UB960_SR_LINK_ERROR_COUNT		0xb9
310 #define UB960_SR_FPD3_ENC_CTL			0xba
311 #define UB960_SR_FV_MIN_TIME			0xbc
312 #define UB960_SR_GPIO_PD_CTL			0xbe
313 
314 #define UB960_SR_FPD_RATE_CFG			0xc2	/* UB9702 */
315 #define UB960_SR_CSI_PLL_DIV			0xc9	/* UB9702 */
316 
317 #define UB960_RR_PORT_DEBUG			0xd0
318 #define UB960_RR_AEQ_CTL2			0xd2
319 #define UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR		BIT(2)
320 
321 #define UB960_RR_AEQ_STATUS			0xd3
322 #define UB960_RR_AEQ_STATUS_STATUS_2		GENMASK(5, 3)
323 #define UB960_RR_AEQ_STATUS_STATUS_1		GENMASK(2, 0)
324 
325 #define UB960_RR_AEQ_BYPASS			0xd4
326 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT	5
327 #define UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK	GENMASK(7, 5)
328 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT	1
329 #define UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK	GENMASK(3, 1)
330 #define UB960_RR_AEQ_BYPASS_ENABLE			BIT(0)
331 
332 #define UB960_RR_AEQ_MIN_MAX			0xd5
333 #define UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT	4
334 #define UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT	0
335 
336 #define UB960_RR_SFILTER_STS_0			0xd6
337 #define UB960_RR_SFILTER_STS_1			0xd7
338 #define UB960_RR_PORT_ICR_HI			0xd8
339 #define UB960_RR_PORT_ICR_LO			0xd9
340 #define UB960_RR_PORT_ISR_HI			0xda
341 #define UB960_RR_PORT_ISR_LO			0xdb
342 #define UB960_RR_FC_GPIO_STS			0xdc
343 #define UB960_RR_FC_GPIO_ICR			0xdd
344 #define UB960_RR_SEN_INT_RISE_STS		0xde
345 #define UB960_RR_SEN_INT_FALL_STS		0xdf
346 
347 #define UB960_RR_CHANNEL_MODE			0xe4	/* UB9702 */
348 
349 #define UB960_SR_FPD3_RX_ID(n)			(0xf0 + (n))
350 #define UB960_SR_FPD3_RX_ID_LEN			6
351 
352 #define UB960_SR_I2C_RX_ID(n)			(0xf8 + (n)) /* < UB960_FPD_RX_NPORTS */
353 
354 /* Indirect register blocks */
355 #define UB960_IND_TARGET_PAT_GEN		0x00
356 #define UB960_IND_TARGET_RX_ANA(n)		(0x01 + (n))
357 #define UB960_IND_TARGET_CSI_CSIPLL_REG_1	0x92	/* UB9702 */
358 #define UB960_IND_TARGET_CSI_ANA		0x07
359 
360 /* UB960_IR_PGEN_*: Indirect Registers for Test Pattern Generator */
361 
362 #define UB960_IR_PGEN_CTL			0x01
363 #define UB960_IR_PGEN_CTL_PGEN_ENABLE		BIT(0)
364 
365 #define UB960_IR_PGEN_CFG			0x02
366 #define UB960_IR_PGEN_CSI_DI			0x03
367 #define UB960_IR_PGEN_LINE_SIZE1		0x04
368 #define UB960_IR_PGEN_LINE_SIZE0		0x05
369 #define UB960_IR_PGEN_BAR_SIZE1			0x06
370 #define UB960_IR_PGEN_BAR_SIZE0			0x07
371 #define UB960_IR_PGEN_ACT_LPF1			0x08
372 #define UB960_IR_PGEN_ACT_LPF0			0x09
373 #define UB960_IR_PGEN_TOT_LPF1			0x0a
374 #define UB960_IR_PGEN_TOT_LPF0			0x0b
375 #define UB960_IR_PGEN_LINE_PD1			0x0c
376 #define UB960_IR_PGEN_LINE_PD0			0x0d
377 #define UB960_IR_PGEN_VBP			0x0e
378 #define UB960_IR_PGEN_VFP			0x0f
379 #define UB960_IR_PGEN_COLOR(n)			(0x10 + (n)) /* n < 15 */
380 
381 #define UB960_IR_RX_ANA_STROBE_SET_CLK		0x08
382 #define UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY	BIT(3)
383 #define UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK	GENMASK(2, 0)
384 
385 #define UB960_IR_RX_ANA_STROBE_SET_DATA		0x09
386 #define UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY	BIT(3)
387 #define UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK	GENMASK(2, 0)
388 
389 /* EQ related */
390 
391 #define UB960_MIN_AEQ_STROBE_POS -7
392 #define UB960_MAX_AEQ_STROBE_POS  7
393 
394 #define UB960_MANUAL_STROBE_EXTRA_DELAY 6
395 
396 #define UB960_MIN_MANUAL_STROBE_POS -(7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
397 #define UB960_MAX_MANUAL_STROBE_POS  (7 + UB960_MANUAL_STROBE_EXTRA_DELAY)
398 #define UB960_NUM_MANUAL_STROBE_POS  (UB960_MAX_MANUAL_STROBE_POS - UB960_MIN_MANUAL_STROBE_POS + 1)
399 
400 #define UB960_MIN_EQ_LEVEL  0
401 #define UB960_MAX_EQ_LEVEL  14
402 #define UB960_NUM_EQ_LEVELS (UB960_MAX_EQ_LEVEL - UB960_MIN_EQ_LEVEL + 1)
403 
404 struct ub960_hw_data {
405 	const char *model;
406 	u8 num_rxports;
407 	u8 num_txports;
408 	bool is_ub9702;
409 	bool is_fpdlink4;
410 };
411 
412 enum ub960_rxport_mode {
413 	RXPORT_MODE_RAW10 = 0,
414 	RXPORT_MODE_RAW12_HF = 1,
415 	RXPORT_MODE_RAW12_LF = 2,
416 	RXPORT_MODE_CSI2_SYNC = 3,
417 	RXPORT_MODE_CSI2_ASYNC = 4,
418 	RXPORT_MODE_LAST = RXPORT_MODE_CSI2_ASYNC,
419 };
420 
421 enum ub960_rxport_cdr {
422 	RXPORT_CDR_FPD3 = 0,
423 	RXPORT_CDR_FPD4 = 1,
424 	RXPORT_CDR_LAST = RXPORT_CDR_FPD4,
425 };
426 
427 struct ub960_rxport {
428 	struct ub960_data      *priv;
429 	u8                      nport;	/* RX port number, and index in priv->rxport[] */
430 
431 	struct {
432 		struct v4l2_subdev *sd;
433 		u16 pad;
434 		struct fwnode_handle *ep_fwnode;
435 	} source;
436 
437 	/* Serializer */
438 	struct {
439 		struct fwnode_handle *fwnode;
440 		struct i2c_client *client;
441 		unsigned short alias; /* I2C alias (lower 7 bits) */
442 		struct ds90ub9xx_platform_data pdata;
443 	} ser;
444 
445 	enum ub960_rxport_mode  rx_mode;
446 	enum ub960_rxport_cdr	cdr_mode;
447 
448 	u8			lv_fv_pol;	/* LV and FV polarities */
449 
450 	struct regulator	*vpoc;
451 
452 	/* EQ settings */
453 	struct {
454 		bool manual_eq;
455 
456 		s8 strobe_pos;
457 
458 		union {
459 			struct {
460 				u8 eq_level_min;
461 				u8 eq_level_max;
462 			} aeq;
463 
464 			struct {
465 				u8 eq_level;
466 			} manual;
467 		};
468 	} eq;
469 
470 	const struct i2c_client *aliased_clients[UB960_MAX_PORT_ALIASES];
471 };
472 
473 struct ub960_asd {
474 	struct v4l2_async_subdev base;
475 	struct ub960_rxport *rxport;
476 };
477 
478 static inline struct ub960_asd *to_ub960_asd(struct v4l2_async_subdev *asd)
479 {
480 	return container_of(asd, struct ub960_asd, base);
481 }
482 
483 struct ub960_txport {
484 	struct ub960_data      *priv;
485 	u8                      nport;	/* TX port number, and index in priv->txport[] */
486 
487 	u32 num_data_lanes;
488 };
489 
490 struct ub960_data {
491 	const struct ub960_hw_data	*hw_data;
492 	struct i2c_client	*client; /* for shared local registers */
493 	struct regmap		*regmap;
494 
495 	/* lock for register access */
496 	struct mutex		reg_lock;
497 
498 	struct clk		*refclk;
499 
500 	struct regulator	*vddio;
501 
502 	struct gpio_desc	*pd_gpio;
503 	struct delayed_work	poll_work;
504 	struct ub960_rxport	*rxports[UB960_MAX_RX_NPORTS];
505 	struct ub960_txport	*txports[UB960_MAX_TX_NPORTS];
506 
507 	struct v4l2_subdev	sd;
508 	struct media_pad	pads[UB960_MAX_NPORTS];
509 
510 	struct v4l2_ctrl_handler   ctrl_handler;
511 	struct v4l2_async_notifier notifier;
512 
513 	u32 tx_data_rate;		/* Nominal data rate (Gb/s) */
514 	s64 tx_link_freq[1];
515 
516 	struct i2c_atr *atr;
517 
518 	struct {
519 		u8 rxport;
520 		u8 txport;
521 		u8 indirect_target;
522 	} reg_current;
523 
524 	bool streaming;
525 
526 	u8 stored_fwd_ctl;
527 
528 	u64 stream_enable_mask[UB960_MAX_NPORTS];
529 
530 	/* These are common to all ports */
531 	struct {
532 		bool manual;
533 
534 		s8 min;
535 		s8 max;
536 	} strobe;
537 };
538 
539 static inline struct ub960_data *sd_to_ub960(struct v4l2_subdev *sd)
540 {
541 	return container_of(sd, struct ub960_data, sd);
542 }
543 
544 static inline bool ub960_pad_is_sink(struct ub960_data *priv, u32 pad)
545 {
546 	return pad < priv->hw_data->num_rxports;
547 }
548 
549 static inline bool ub960_pad_is_source(struct ub960_data *priv, u32 pad)
550 {
551 	return pad >= priv->hw_data->num_rxports;
552 }
553 
554 static inline unsigned int ub960_pad_to_port(struct ub960_data *priv, u32 pad)
555 {
556 	if (ub960_pad_is_sink(priv, pad))
557 		return pad;
558 	else
559 		return pad - priv->hw_data->num_rxports;
560 }
561 
562 struct ub960_format_info {
563 	u32 code;
564 	u32 bpp;
565 	u8 datatype;
566 	bool meta;
567 };
568 
569 static const struct ub960_format_info ub960_formats[] = {
570 	{ .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
571 	{ .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
572 	{ .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
573 	{ .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .datatype = MIPI_CSI2_DT_YUV422_8B, },
574 
575 	{ .code = MEDIA_BUS_FMT_SBGGR12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
576 	{ .code = MEDIA_BUS_FMT_SGBRG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
577 	{ .code = MEDIA_BUS_FMT_SGRBG12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
578 	{ .code = MEDIA_BUS_FMT_SRGGB12_1X12, .bpp = 12, .datatype = MIPI_CSI2_DT_RAW12, },
579 };
580 
581 static const struct ub960_format_info *ub960_find_format(u32 code)
582 {
583 	unsigned int i;
584 
585 	for (i = 0; i < ARRAY_SIZE(ub960_formats); i++) {
586 		if (ub960_formats[i].code == code)
587 			return &ub960_formats[i];
588 	}
589 
590 	return NULL;
591 }
592 
593 /* -----------------------------------------------------------------------------
594  * Basic device access
595  */
596 
597 static int ub960_read(struct ub960_data *priv, u8 reg, u8 *val)
598 {
599 	struct device *dev = &priv->client->dev;
600 	unsigned int v;
601 	int ret;
602 
603 	mutex_lock(&priv->reg_lock);
604 
605 	ret = regmap_read(priv->regmap, reg, &v);
606 	if (ret) {
607 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
608 			__func__, reg, ret);
609 		goto out_unlock;
610 	}
611 
612 	*val = v;
613 
614 out_unlock:
615 	mutex_unlock(&priv->reg_lock);
616 
617 	return ret;
618 }
619 
620 static int ub960_write(struct ub960_data *priv, u8 reg, u8 val)
621 {
622 	struct device *dev = &priv->client->dev;
623 	int ret;
624 
625 	mutex_lock(&priv->reg_lock);
626 
627 	ret = regmap_write(priv->regmap, reg, val);
628 	if (ret)
629 		dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
630 			__func__, reg, ret);
631 
632 	mutex_unlock(&priv->reg_lock);
633 
634 	return ret;
635 }
636 
637 static int ub960_update_bits(struct ub960_data *priv, u8 reg, u8 mask, u8 val)
638 {
639 	struct device *dev = &priv->client->dev;
640 	int ret;
641 
642 	mutex_lock(&priv->reg_lock);
643 
644 	ret = regmap_update_bits(priv->regmap, reg, mask, val);
645 	if (ret)
646 		dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
647 			__func__, reg, ret);
648 
649 	mutex_unlock(&priv->reg_lock);
650 
651 	return ret;
652 }
653 
654 static int ub960_read16(struct ub960_data *priv, u8 reg, u16 *val)
655 {
656 	struct device *dev = &priv->client->dev;
657 	__be16 __v;
658 	int ret;
659 
660 	mutex_lock(&priv->reg_lock);
661 
662 	ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
663 	if (ret) {
664 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
665 			__func__, reg, ret);
666 		goto out_unlock;
667 	}
668 
669 	*val = be16_to_cpu(__v);
670 
671 out_unlock:
672 	mutex_unlock(&priv->reg_lock);
673 
674 	return ret;
675 }
676 
677 static int ub960_rxport_select(struct ub960_data *priv, u8 nport)
678 {
679 	struct device *dev = &priv->client->dev;
680 	int ret;
681 
682 	lockdep_assert_held(&priv->reg_lock);
683 
684 	if (priv->reg_current.rxport == nport)
685 		return 0;
686 
687 	ret = regmap_write(priv->regmap, UB960_SR_FPD3_PORT_SEL,
688 			   (nport << 4) | BIT(nport));
689 	if (ret) {
690 		dev_err(dev, "%s: cannot select rxport %d (%d)!\n", __func__,
691 			nport, ret);
692 		return ret;
693 	}
694 
695 	priv->reg_current.rxport = nport;
696 
697 	return 0;
698 }
699 
700 static int ub960_rxport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
701 {
702 	struct device *dev = &priv->client->dev;
703 	unsigned int v;
704 	int ret;
705 
706 	mutex_lock(&priv->reg_lock);
707 
708 	ret = ub960_rxport_select(priv, nport);
709 	if (ret)
710 		goto out_unlock;
711 
712 	ret = regmap_read(priv->regmap, reg, &v);
713 	if (ret) {
714 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
715 			__func__, reg, ret);
716 		goto out_unlock;
717 	}
718 
719 	*val = v;
720 
721 out_unlock:
722 	mutex_unlock(&priv->reg_lock);
723 
724 	return ret;
725 }
726 
727 static int ub960_rxport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
728 {
729 	struct device *dev = &priv->client->dev;
730 	int ret;
731 
732 	mutex_lock(&priv->reg_lock);
733 
734 	ret = ub960_rxport_select(priv, nport);
735 	if (ret)
736 		goto out_unlock;
737 
738 	ret = regmap_write(priv->regmap, reg, val);
739 	if (ret)
740 		dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
741 			__func__, reg, ret);
742 
743 out_unlock:
744 	mutex_unlock(&priv->reg_lock);
745 
746 	return ret;
747 }
748 
749 static int ub960_rxport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
750 				    u8 mask, u8 val)
751 {
752 	struct device *dev = &priv->client->dev;
753 	int ret;
754 
755 	mutex_lock(&priv->reg_lock);
756 
757 	ret = ub960_rxport_select(priv, nport);
758 	if (ret)
759 		goto out_unlock;
760 
761 	ret = regmap_update_bits(priv->regmap, reg, mask, val);
762 	if (ret)
763 		dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
764 			__func__, reg, ret);
765 
766 out_unlock:
767 	mutex_unlock(&priv->reg_lock);
768 
769 	return ret;
770 }
771 
772 static int ub960_rxport_read16(struct ub960_data *priv, u8 nport, u8 reg,
773 			       u16 *val)
774 {
775 	struct device *dev = &priv->client->dev;
776 	__be16 __v;
777 	int ret;
778 
779 	mutex_lock(&priv->reg_lock);
780 
781 	ret = ub960_rxport_select(priv, nport);
782 	if (ret)
783 		goto out_unlock;
784 
785 	ret = regmap_bulk_read(priv->regmap, reg, &__v, sizeof(__v));
786 	if (ret) {
787 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
788 			__func__, reg, ret);
789 		goto out_unlock;
790 	}
791 
792 	*val = be16_to_cpu(__v);
793 
794 out_unlock:
795 	mutex_unlock(&priv->reg_lock);
796 
797 	return ret;
798 }
799 
800 static int ub960_txport_select(struct ub960_data *priv, u8 nport)
801 {
802 	struct device *dev = &priv->client->dev;
803 	int ret;
804 
805 	lockdep_assert_held(&priv->reg_lock);
806 
807 	if (priv->reg_current.txport == nport)
808 		return 0;
809 
810 	ret = regmap_write(priv->regmap, UB960_SR_CSI_PORT_SEL,
811 			   (nport << 4) | BIT(nport));
812 	if (ret) {
813 		dev_err(dev, "%s: cannot select tx port %d (%d)!\n", __func__,
814 			nport, ret);
815 		return ret;
816 	}
817 
818 	priv->reg_current.txport = nport;
819 
820 	return 0;
821 }
822 
823 static int ub960_txport_read(struct ub960_data *priv, u8 nport, u8 reg, u8 *val)
824 {
825 	struct device *dev = &priv->client->dev;
826 	unsigned int v;
827 	int ret;
828 
829 	mutex_lock(&priv->reg_lock);
830 
831 	ret = ub960_txport_select(priv, nport);
832 	if (ret)
833 		goto out_unlock;
834 
835 	ret = regmap_read(priv->regmap, reg, &v);
836 	if (ret) {
837 		dev_err(dev, "%s: cannot read register 0x%02x (%d)!\n",
838 			__func__, reg, ret);
839 		goto out_unlock;
840 	}
841 
842 	*val = v;
843 
844 out_unlock:
845 	mutex_unlock(&priv->reg_lock);
846 
847 	return ret;
848 }
849 
850 static int ub960_txport_write(struct ub960_data *priv, u8 nport, u8 reg, u8 val)
851 {
852 	struct device *dev = &priv->client->dev;
853 	int ret;
854 
855 	mutex_lock(&priv->reg_lock);
856 
857 	ret = ub960_txport_select(priv, nport);
858 	if (ret)
859 		goto out_unlock;
860 
861 	ret = regmap_write(priv->regmap, reg, val);
862 	if (ret)
863 		dev_err(dev, "%s: cannot write register 0x%02x (%d)!\n",
864 			__func__, reg, ret);
865 
866 out_unlock:
867 	mutex_unlock(&priv->reg_lock);
868 
869 	return ret;
870 }
871 
872 static int ub960_txport_update_bits(struct ub960_data *priv, u8 nport, u8 reg,
873 				    u8 mask, u8 val)
874 {
875 	struct device *dev = &priv->client->dev;
876 	int ret;
877 
878 	mutex_lock(&priv->reg_lock);
879 
880 	ret = ub960_txport_select(priv, nport);
881 	if (ret)
882 		goto out_unlock;
883 
884 	ret = regmap_update_bits(priv->regmap, reg, mask, val);
885 	if (ret)
886 		dev_err(dev, "%s: cannot update register 0x%02x (%d)!\n",
887 			__func__, reg, ret);
888 
889 out_unlock:
890 	mutex_unlock(&priv->reg_lock);
891 
892 	return ret;
893 }
894 
895 static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
896 {
897 	struct device *dev = &priv->client->dev;
898 	int ret;
899 
900 	lockdep_assert_held(&priv->reg_lock);
901 
902 	if (priv->reg_current.indirect_target == block)
903 		return 0;
904 
905 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
906 	if (ret) {
907 		dev_err(dev, "%s: cannot select indirect target %u (%d)!\n",
908 			__func__, block, ret);
909 		return ret;
910 	}
911 
912 	priv->reg_current.indirect_target = block;
913 
914 	return 0;
915 }
916 
917 static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val)
918 {
919 	struct device *dev = &priv->client->dev;
920 	unsigned int v;
921 	int ret;
922 
923 	mutex_lock(&priv->reg_lock);
924 
925 	ret = ub960_select_ind_reg_block(priv, block);
926 	if (ret)
927 		goto out_unlock;
928 
929 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
930 	if (ret) {
931 		dev_err(dev,
932 			"Write to IND_ACC_ADDR failed when reading %u:%x02x: %d\n",
933 			block, reg, ret);
934 		goto out_unlock;
935 	}
936 
937 	ret = regmap_read(priv->regmap, UB960_SR_IND_ACC_DATA, &v);
938 	if (ret) {
939 		dev_err(dev,
940 			"Write to IND_ACC_DATA failed when reading %u:%x02x: %d\n",
941 			block, reg, ret);
942 		goto out_unlock;
943 	}
944 
945 	*val = v;
946 
947 out_unlock:
948 	mutex_unlock(&priv->reg_lock);
949 
950 	return ret;
951 }
952 
953 static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val)
954 {
955 	struct device *dev = &priv->client->dev;
956 	int ret;
957 
958 	mutex_lock(&priv->reg_lock);
959 
960 	ret = ub960_select_ind_reg_block(priv, block);
961 	if (ret)
962 		goto out_unlock;
963 
964 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
965 	if (ret) {
966 		dev_err(dev,
967 			"Write to IND_ACC_ADDR failed when writing %u:%x02x: %d\n",
968 			block, reg, ret);
969 		goto out_unlock;
970 	}
971 
972 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_DATA, val);
973 	if (ret) {
974 		dev_err(dev,
975 			"Write to IND_ACC_DATA failed when writing %u:%x02x: %d\n",
976 			block, reg, ret);
977 		goto out_unlock;
978 	}
979 
980 out_unlock:
981 	mutex_unlock(&priv->reg_lock);
982 
983 	return ret;
984 }
985 
986 static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
987 				 u8 mask, u8 val)
988 {
989 	struct device *dev = &priv->client->dev;
990 	int ret;
991 
992 	mutex_lock(&priv->reg_lock);
993 
994 	ret = ub960_select_ind_reg_block(priv, block);
995 	if (ret)
996 		goto out_unlock;
997 
998 	ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_ADDR, reg);
999 	if (ret) {
1000 		dev_err(dev,
1001 			"Write to IND_ACC_ADDR failed when updating %u:%x02x: %d\n",
1002 			block, reg, ret);
1003 		goto out_unlock;
1004 	}
1005 
1006 	ret = regmap_update_bits(priv->regmap, UB960_SR_IND_ACC_DATA, mask,
1007 				 val);
1008 	if (ret) {
1009 		dev_err(dev,
1010 			"Write to IND_ACC_DATA failed when updating %u:%x02x: %d\n",
1011 			block, reg, ret);
1012 		goto out_unlock;
1013 	}
1014 
1015 out_unlock:
1016 	mutex_unlock(&priv->reg_lock);
1017 
1018 	return ret;
1019 }
1020 
1021 /* -----------------------------------------------------------------------------
1022  * I2C-ATR (address translator)
1023  */
1024 
1025 static int ub960_atr_attach_client(struct i2c_atr *atr, u32 chan_id,
1026 				   const struct i2c_client *client, u16 alias)
1027 {
1028 	struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1029 	struct ub960_rxport *rxport = priv->rxports[chan_id];
1030 	struct device *dev = &priv->client->dev;
1031 	unsigned int reg_idx;
1032 
1033 	for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1034 		if (!rxport->aliased_clients[reg_idx])
1035 			break;
1036 	}
1037 
1038 	if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1039 		dev_err(dev, "rx%u: alias pool exhausted\n", rxport->nport);
1040 		return -EADDRNOTAVAIL;
1041 	}
1042 
1043 	rxport->aliased_clients[reg_idx] = client;
1044 
1045 	ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ID(reg_idx),
1046 			   client->addr << 1);
1047 	ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx),
1048 			   alias << 1);
1049 
1050 	dev_dbg(dev, "rx%u: client 0x%02x assigned alias 0x%02x at slot %u\n",
1051 		rxport->nport, client->addr, alias, reg_idx);
1052 
1053 	return 0;
1054 }
1055 
1056 static void ub960_atr_detach_client(struct i2c_atr *atr, u32 chan_id,
1057 				    const struct i2c_client *client)
1058 {
1059 	struct ub960_data *priv = i2c_atr_get_driver_data(atr);
1060 	struct ub960_rxport *rxport = priv->rxports[chan_id];
1061 	struct device *dev = &priv->client->dev;
1062 	unsigned int reg_idx;
1063 
1064 	for (reg_idx = 0; reg_idx < ARRAY_SIZE(rxport->aliased_clients); reg_idx++) {
1065 		if (rxport->aliased_clients[reg_idx] == client)
1066 			break;
1067 	}
1068 
1069 	if (reg_idx == ARRAY_SIZE(rxport->aliased_clients)) {
1070 		dev_err(dev, "rx%u: client 0x%02x is not mapped!\n",
1071 			rxport->nport, client->addr);
1072 		return;
1073 	}
1074 
1075 	rxport->aliased_clients[reg_idx] = NULL;
1076 
1077 	ub960_rxport_write(priv, chan_id, UB960_RR_SLAVE_ALIAS(reg_idx), 0);
1078 
1079 	dev_dbg(dev, "rx%u: client 0x%02x released at slot %u\n", rxport->nport,
1080 		client->addr, reg_idx);
1081 }
1082 
1083 static const struct i2c_atr_ops ub960_atr_ops = {
1084 	.attach_client = ub960_atr_attach_client,
1085 	.detach_client = ub960_atr_detach_client,
1086 };
1087 
1088 static int ub960_init_atr(struct ub960_data *priv)
1089 {
1090 	struct device *dev = &priv->client->dev;
1091 	struct i2c_adapter *parent_adap = priv->client->adapter;
1092 
1093 	priv->atr = i2c_atr_new(parent_adap, dev, &ub960_atr_ops,
1094 				priv->hw_data->num_rxports);
1095 	if (IS_ERR(priv->atr))
1096 		return PTR_ERR(priv->atr);
1097 
1098 	i2c_atr_set_driver_data(priv->atr, priv);
1099 
1100 	return 0;
1101 }
1102 
1103 static void ub960_uninit_atr(struct ub960_data *priv)
1104 {
1105 	i2c_atr_delete(priv->atr);
1106 	priv->atr = NULL;
1107 }
1108 
1109 /* -----------------------------------------------------------------------------
1110  * TX ports
1111  */
1112 
1113 static int ub960_parse_dt_txport(struct ub960_data *priv,
1114 				 struct fwnode_handle *ep_fwnode,
1115 				 u8 nport)
1116 {
1117 	struct device *dev = &priv->client->dev;
1118 	struct v4l2_fwnode_endpoint vep = {};
1119 	struct ub960_txport *txport;
1120 	int ret;
1121 
1122 	txport = kzalloc(sizeof(*txport), GFP_KERNEL);
1123 	if (!txport)
1124 		return -ENOMEM;
1125 
1126 	txport->priv = priv;
1127 	txport->nport = nport;
1128 
1129 	vep.bus_type = V4L2_MBUS_CSI2_DPHY;
1130 	ret = v4l2_fwnode_endpoint_alloc_parse(ep_fwnode, &vep);
1131 	if (ret) {
1132 		dev_err(dev, "tx%u: failed to parse endpoint data\n", nport);
1133 		goto err_free_txport;
1134 	}
1135 
1136 	txport->num_data_lanes = vep.bus.mipi_csi2.num_data_lanes;
1137 
1138 	if (vep.nr_of_link_frequencies != 1) {
1139 		ret = -EINVAL;
1140 		goto err_free_vep;
1141 	}
1142 
1143 	priv->tx_link_freq[0] = vep.link_frequencies[0];
1144 	priv->tx_data_rate = priv->tx_link_freq[0] * 2;
1145 
1146 	if (priv->tx_data_rate != MHZ(1600) &&
1147 	    priv->tx_data_rate != MHZ(1200) &&
1148 	    priv->tx_data_rate != MHZ(800) &&
1149 	    priv->tx_data_rate != MHZ(400)) {
1150 		dev_err(dev, "tx%u: invalid 'link-frequencies' value\n", nport);
1151 		ret = -EINVAL;
1152 		goto err_free_vep;
1153 	}
1154 
1155 	v4l2_fwnode_endpoint_free(&vep);
1156 
1157 	priv->txports[nport] = txport;
1158 
1159 	return 0;
1160 
1161 err_free_vep:
1162 	v4l2_fwnode_endpoint_free(&vep);
1163 err_free_txport:
1164 	kfree(txport);
1165 
1166 	return ret;
1167 }
1168 
1169 static void ub960_csi_handle_events(struct ub960_data *priv, u8 nport)
1170 {
1171 	struct device *dev = &priv->client->dev;
1172 	u8 csi_tx_isr;
1173 	int ret;
1174 
1175 	ret = ub960_txport_read(priv, nport, UB960_TR_CSI_TX_ISR, &csi_tx_isr);
1176 	if (ret)
1177 		return;
1178 
1179 	if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_SYNC_ERROR)
1180 		dev_warn(dev, "TX%u: CSI_SYNC_ERROR\n", nport);
1181 
1182 	if (csi_tx_isr & UB960_TR_CSI_TX_ISR_IS_CSI_PASS_ERROR)
1183 		dev_warn(dev, "TX%u: CSI_PASS_ERROR\n", nport);
1184 }
1185 
1186 /* -----------------------------------------------------------------------------
1187  * RX ports
1188  */
1189 
1190 static int ub960_rxport_enable_vpocs(struct ub960_data *priv)
1191 {
1192 	unsigned int nport;
1193 	int ret;
1194 
1195 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1196 		struct ub960_rxport *rxport = priv->rxports[nport];
1197 
1198 		if (!rxport || !rxport->vpoc)
1199 			continue;
1200 
1201 		ret = regulator_enable(rxport->vpoc);
1202 		if (ret)
1203 			goto err_disable_vpocs;
1204 	}
1205 
1206 	return 0;
1207 
1208 err_disable_vpocs:
1209 	while (nport--) {
1210 		struct ub960_rxport *rxport = priv->rxports[nport];
1211 
1212 		if (!rxport || !rxport->vpoc)
1213 			continue;
1214 
1215 		regulator_disable(rxport->vpoc);
1216 	}
1217 
1218 	return ret;
1219 }
1220 
1221 static void ub960_rxport_disable_vpocs(struct ub960_data *priv)
1222 {
1223 	unsigned int nport;
1224 
1225 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1226 		struct ub960_rxport *rxport = priv->rxports[nport];
1227 
1228 		if (!rxport || !rxport->vpoc)
1229 			continue;
1230 
1231 		regulator_disable(rxport->vpoc);
1232 	}
1233 }
1234 
1235 static void ub960_rxport_clear_errors(struct ub960_data *priv,
1236 				      unsigned int nport)
1237 {
1238 	u8 v;
1239 
1240 	ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
1241 	ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
1242 	ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &v);
1243 	ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &v);
1244 
1245 	ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v);
1246 	ub960_rxport_read(priv, nport, UB960_RR_RX_PAR_ERR_LO, &v);
1247 
1248 	ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
1249 }
1250 
1251 static void ub960_clear_rx_errors(struct ub960_data *priv)
1252 {
1253 	unsigned int nport;
1254 
1255 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++)
1256 		ub960_rxport_clear_errors(priv, nport);
1257 }
1258 
1259 static int ub960_rxport_get_strobe_pos(struct ub960_data *priv,
1260 				       unsigned int nport, s8 *strobe_pos)
1261 {
1262 	u8 v;
1263 	u8 clk_delay, data_delay;
1264 	int ret;
1265 
1266 	ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1267 		       UB960_IR_RX_ANA_STROBE_SET_CLK, &v);
1268 
1269 	clk_delay = (v & UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY) ?
1270 			    0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1271 
1272 	ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1273 		       UB960_IR_RX_ANA_STROBE_SET_DATA, &v);
1274 
1275 	data_delay = (v & UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY) ?
1276 			     0 : UB960_MANUAL_STROBE_EXTRA_DELAY;
1277 
1278 	ret = ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_0, &v);
1279 	if (ret)
1280 		return ret;
1281 
1282 	clk_delay += v & UB960_IR_RX_ANA_STROBE_SET_CLK_DELAY_MASK;
1283 
1284 	ub960_rxport_read(priv, nport, UB960_RR_SFILTER_STS_1, &v);
1285 	if (ret)
1286 		return ret;
1287 
1288 	data_delay += v & UB960_IR_RX_ANA_STROBE_SET_DATA_DELAY_MASK;
1289 
1290 	*strobe_pos = data_delay - clk_delay;
1291 
1292 	return 0;
1293 }
1294 
1295 static void ub960_rxport_set_strobe_pos(struct ub960_data *priv,
1296 					unsigned int nport, s8 strobe_pos)
1297 {
1298 	u8 clk_delay, data_delay;
1299 
1300 	clk_delay = UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1301 	data_delay = UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1302 
1303 	if (strobe_pos < UB960_MIN_AEQ_STROBE_POS)
1304 		clk_delay = abs(strobe_pos) - UB960_MANUAL_STROBE_EXTRA_DELAY;
1305 	else if (strobe_pos > UB960_MAX_AEQ_STROBE_POS)
1306 		data_delay = strobe_pos - UB960_MANUAL_STROBE_EXTRA_DELAY;
1307 	else if (strobe_pos < 0)
1308 		clk_delay = abs(strobe_pos) | UB960_IR_RX_ANA_STROBE_SET_CLK_NO_EXTRA_DELAY;
1309 	else if (strobe_pos > 0)
1310 		data_delay = strobe_pos | UB960_IR_RX_ANA_STROBE_SET_DATA_NO_EXTRA_DELAY;
1311 
1312 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1313 			UB960_IR_RX_ANA_STROBE_SET_CLK, clk_delay);
1314 
1315 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport),
1316 			UB960_IR_RX_ANA_STROBE_SET_DATA, data_delay);
1317 }
1318 
1319 static void ub960_rxport_set_strobe_range(struct ub960_data *priv,
1320 					  s8 strobe_min, s8 strobe_max)
1321 {
1322 	/* Convert the signed strobe pos to positive zero based value */
1323 	strobe_min -= UB960_MIN_AEQ_STROBE_POS;
1324 	strobe_max -= UB960_MIN_AEQ_STROBE_POS;
1325 
1326 	ub960_write(priv, UB960_XR_SFILTER_CFG,
1327 		    ((u8)strobe_min << UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) |
1328 		    ((u8)strobe_max << UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT));
1329 }
1330 
1331 static int ub960_rxport_get_eq_level(struct ub960_data *priv,
1332 				     unsigned int nport, u8 *eq_level)
1333 {
1334 	int ret;
1335 	u8 v;
1336 
1337 	ret = ub960_rxport_read(priv, nport, UB960_RR_AEQ_STATUS, &v);
1338 	if (ret)
1339 		return ret;
1340 
1341 	*eq_level = (v & UB960_RR_AEQ_STATUS_STATUS_1) +
1342 		    (v & UB960_RR_AEQ_STATUS_STATUS_2);
1343 
1344 	return 0;
1345 }
1346 
1347 static void ub960_rxport_set_eq_level(struct ub960_data *priv,
1348 				      unsigned int nport, u8 eq_level)
1349 {
1350 	u8 eq_stage_1_select_value, eq_stage_2_select_value;
1351 	const unsigned int eq_stage_max = 7;
1352 	u8 v;
1353 
1354 	if (eq_level <= eq_stage_max) {
1355 		eq_stage_1_select_value = eq_level;
1356 		eq_stage_2_select_value = 0;
1357 	} else {
1358 		eq_stage_1_select_value = eq_stage_max;
1359 		eq_stage_2_select_value = eq_level - eq_stage_max;
1360 	}
1361 
1362 	ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
1363 
1364 	v &= ~(UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_MASK |
1365 	       UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_MASK);
1366 	v |= eq_stage_1_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE1_VALUE_SHIFT;
1367 	v |= eq_stage_2_select_value << UB960_RR_AEQ_BYPASS_EQ_STAGE2_VALUE_SHIFT;
1368 	v |= UB960_RR_AEQ_BYPASS_ENABLE;
1369 
1370 	ub960_rxport_write(priv, nport, UB960_RR_AEQ_BYPASS, v);
1371 }
1372 
1373 static void ub960_rxport_set_eq_range(struct ub960_data *priv,
1374 				      unsigned int nport, u8 eq_min, u8 eq_max)
1375 {
1376 	ub960_rxport_write(priv, nport, UB960_RR_AEQ_MIN_MAX,
1377 			   (eq_min << UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) |
1378 			   (eq_max << UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT));
1379 
1380 	/* Enable AEQ min setting */
1381 	ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_CTL2,
1382 				 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR,
1383 				 UB960_RR_AEQ_CTL2_SET_AEQ_FLOOR);
1384 }
1385 
1386 static void ub960_rxport_config_eq(struct ub960_data *priv, unsigned int nport)
1387 {
1388 	struct ub960_rxport *rxport = priv->rxports[nport];
1389 
1390 	/* We also set common settings here. Should be moved elsewhere. */
1391 
1392 	if (priv->strobe.manual) {
1393 		/* Disable AEQ_SFILTER_EN */
1394 		ub960_update_bits(priv, UB960_XR_AEQ_CTL1,
1395 				  UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN, 0);
1396 	} else {
1397 		/* Enable SFILTER and error control */
1398 		ub960_write(priv, UB960_XR_AEQ_CTL1,
1399 			    UB960_XR_AEQ_CTL1_AEQ_ERR_CTL_MASK |
1400 				    UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN);
1401 
1402 		/* Set AEQ strobe range */
1403 		ub960_rxport_set_strobe_range(priv, priv->strobe.min,
1404 					      priv->strobe.max);
1405 	}
1406 
1407 	/* The rest are port specific */
1408 
1409 	if (priv->strobe.manual)
1410 		ub960_rxport_set_strobe_pos(priv, nport, rxport->eq.strobe_pos);
1411 	else
1412 		ub960_rxport_set_strobe_pos(priv, nport, 0);
1413 
1414 	if (rxport->eq.manual_eq) {
1415 		ub960_rxport_set_eq_level(priv, nport,
1416 					  rxport->eq.manual.eq_level);
1417 
1418 		/* Enable AEQ Bypass */
1419 		ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1420 					 UB960_RR_AEQ_BYPASS_ENABLE,
1421 					 UB960_RR_AEQ_BYPASS_ENABLE);
1422 	} else {
1423 		ub960_rxport_set_eq_range(priv, nport,
1424 					  rxport->eq.aeq.eq_level_min,
1425 					  rxport->eq.aeq.eq_level_max);
1426 
1427 		/* Disable AEQ Bypass */
1428 		ub960_rxport_update_bits(priv, nport, UB960_RR_AEQ_BYPASS,
1429 					 UB960_RR_AEQ_BYPASS_ENABLE, 0);
1430 	}
1431 }
1432 
1433 static int ub960_rxport_link_ok(struct ub960_data *priv, unsigned int nport,
1434 				bool *ok)
1435 {
1436 	u8 rx_port_sts1, rx_port_sts2;
1437 	u16 parity_errors;
1438 	u8 csi_rx_sts;
1439 	u8 csi_err_cnt;
1440 	u8 bcc_sts;
1441 	int ret;
1442 	bool errors;
1443 
1444 	ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
1445 				&rx_port_sts1);
1446 	if (ret)
1447 		return ret;
1448 
1449 	if (!(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS)) {
1450 		*ok = false;
1451 		return 0;
1452 	}
1453 
1454 	ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
1455 				&rx_port_sts2);
1456 	if (ret)
1457 		return ret;
1458 
1459 	ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS, &csi_rx_sts);
1460 	if (ret)
1461 		return ret;
1462 
1463 	ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER,
1464 				&csi_err_cnt);
1465 	if (ret)
1466 		return ret;
1467 
1468 	ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS, &bcc_sts);
1469 	if (ret)
1470 		return ret;
1471 
1472 	ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
1473 				  &parity_errors);
1474 	if (ret)
1475 		return ret;
1476 
1477 	errors = (rx_port_sts1 & UB960_RR_RX_PORT_STS1_ERROR_MASK) ||
1478 		 (rx_port_sts2 & UB960_RR_RX_PORT_STS2_ERROR_MASK) ||
1479 		 (bcc_sts & UB960_RR_BCC_STATUS_ERROR_MASK) ||
1480 		 (csi_rx_sts & UB960_RR_CSI_RX_STS_ERROR_MASK) || csi_err_cnt ||
1481 		 parity_errors;
1482 
1483 	*ok = !errors;
1484 
1485 	return 0;
1486 }
1487 
1488 /*
1489  * Wait for the RX ports to lock, have no errors and have stable strobe position
1490  * and EQ level.
1491  */
1492 static int ub960_rxport_wait_locks(struct ub960_data *priv,
1493 				   unsigned long port_mask,
1494 				   unsigned int *lock_mask)
1495 {
1496 	struct device *dev = &priv->client->dev;
1497 	unsigned long timeout;
1498 	unsigned int link_ok_mask;
1499 	unsigned int missing;
1500 	unsigned int loops;
1501 	u8 nport;
1502 	int ret;
1503 
1504 	if (port_mask == 0) {
1505 		if (lock_mask)
1506 			*lock_mask = 0;
1507 		return 0;
1508 	}
1509 
1510 	if (port_mask >= BIT(priv->hw_data->num_rxports))
1511 		return -EINVAL;
1512 
1513 	timeout = jiffies + msecs_to_jiffies(1000);
1514 	loops = 0;
1515 	link_ok_mask = 0;
1516 
1517 	while (time_before(jiffies, timeout)) {
1518 		missing = 0;
1519 
1520 		for_each_set_bit(nport, &port_mask,
1521 				 priv->hw_data->num_rxports) {
1522 			struct ub960_rxport *rxport = priv->rxports[nport];
1523 			bool ok;
1524 
1525 			if (!rxport)
1526 				continue;
1527 
1528 			ret = ub960_rxport_link_ok(priv, nport, &ok);
1529 			if (ret)
1530 				return ret;
1531 
1532 			/*
1533 			 * We want the link to be ok for two consecutive loops,
1534 			 * as a link could get established just before our test
1535 			 * and drop soon after.
1536 			 */
1537 			if (!ok || !(link_ok_mask & BIT(nport)))
1538 				missing++;
1539 
1540 			if (ok)
1541 				link_ok_mask |= BIT(nport);
1542 			else
1543 				link_ok_mask &= ~BIT(nport);
1544 		}
1545 
1546 		loops++;
1547 
1548 		if (missing == 0)
1549 			break;
1550 
1551 		msleep(50);
1552 	}
1553 
1554 	if (lock_mask)
1555 		*lock_mask = link_ok_mask;
1556 
1557 	dev_dbg(dev, "Wait locks done in %u loops\n", loops);
1558 	for_each_set_bit(nport, &port_mask, priv->hw_data->num_rxports) {
1559 		struct ub960_rxport *rxport = priv->rxports[nport];
1560 		s8 strobe_pos, eq_level;
1561 		u16 v;
1562 
1563 		if (!rxport)
1564 			continue;
1565 
1566 		if (!(link_ok_mask & BIT(nport))) {
1567 			dev_dbg(dev, "\trx%u: not locked\n", nport);
1568 			continue;
1569 		}
1570 
1571 		ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v);
1572 
1573 		ret = ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
1574 		if (ret)
1575 			return ret;
1576 
1577 		ret = ub960_rxport_get_eq_level(priv, nport, &eq_level);
1578 		if (ret)
1579 			return ret;
1580 
1581 		dev_dbg(dev, "\trx%u: locked, SP: %d, EQ: %u, freq %llu Hz\n",
1582 			nport, strobe_pos, eq_level, (v * 1000000ULL) >> 8);
1583 	}
1584 
1585 	return 0;
1586 }
1587 
1588 static unsigned long ub960_calc_bc_clk_rate_ub960(struct ub960_data *priv,
1589 						  struct ub960_rxport *rxport)
1590 {
1591 	unsigned int mult;
1592 	unsigned int div;
1593 
1594 	switch (rxport->rx_mode) {
1595 	case RXPORT_MODE_RAW10:
1596 	case RXPORT_MODE_RAW12_HF:
1597 	case RXPORT_MODE_RAW12_LF:
1598 		mult = 1;
1599 		div = 10;
1600 		break;
1601 
1602 	case RXPORT_MODE_CSI2_SYNC:
1603 		mult = 2;
1604 		div = 1;
1605 		break;
1606 
1607 	case RXPORT_MODE_CSI2_ASYNC:
1608 		mult = 2;
1609 		div = 5;
1610 		break;
1611 
1612 	default:
1613 		return 0;
1614 	}
1615 
1616 	return clk_get_rate(priv->refclk) * mult / div;
1617 }
1618 
1619 static unsigned long ub960_calc_bc_clk_rate_ub9702(struct ub960_data *priv,
1620 						   struct ub960_rxport *rxport)
1621 {
1622 	switch (rxport->rx_mode) {
1623 	case RXPORT_MODE_RAW10:
1624 	case RXPORT_MODE_RAW12_HF:
1625 	case RXPORT_MODE_RAW12_LF:
1626 		return 2359400;
1627 
1628 	case RXPORT_MODE_CSI2_SYNC:
1629 		return 47187500;
1630 
1631 	case RXPORT_MODE_CSI2_ASYNC:
1632 		return 9437500;
1633 
1634 	default:
1635 		return 0;
1636 	}
1637 }
1638 
1639 static int ub960_rxport_add_serializer(struct ub960_data *priv, u8 nport)
1640 {
1641 	struct ub960_rxport *rxport = priv->rxports[nport];
1642 	struct device *dev = &priv->client->dev;
1643 	struct ds90ub9xx_platform_data *ser_pdata = &rxport->ser.pdata;
1644 	struct i2c_board_info ser_info = {
1645 		.of_node = to_of_node(rxport->ser.fwnode),
1646 		.fwnode = rxport->ser.fwnode,
1647 		.platform_data = ser_pdata,
1648 	};
1649 
1650 	ser_pdata->port = nport;
1651 	ser_pdata->atr = priv->atr;
1652 	if (priv->hw_data->is_ub9702)
1653 		ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub9702(priv, rxport);
1654 	else
1655 		ser_pdata->bc_rate = ub960_calc_bc_clk_rate_ub960(priv, rxport);
1656 
1657 	/*
1658 	 * The serializer is added under the same i2c adapter as the
1659 	 * deserializer. This is not quite right, as the serializer is behind
1660 	 * the FPD-Link.
1661 	 */
1662 	ser_info.addr = rxport->ser.alias;
1663 	rxport->ser.client =
1664 		i2c_new_client_device(priv->client->adapter, &ser_info);
1665 	if (IS_ERR(rxport->ser.client)) {
1666 		dev_err(dev, "rx%u: cannot add %s i2c device", nport,
1667 			ser_info.type);
1668 		return PTR_ERR(rxport->ser.client);
1669 	}
1670 
1671 	dev_dbg(dev, "rx%u: remote serializer at alias 0x%02x (%u-%04x)\n",
1672 		nport, rxport->ser.client->addr,
1673 		rxport->ser.client->adapter->nr, rxport->ser.client->addr);
1674 
1675 	return 0;
1676 }
1677 
1678 static void ub960_rxport_remove_serializer(struct ub960_data *priv, u8 nport)
1679 {
1680 	struct ub960_rxport *rxport = priv->rxports[nport];
1681 
1682 	i2c_unregister_device(rxport->ser.client);
1683 	rxport->ser.client = NULL;
1684 }
1685 
1686 /* Add serializer i2c devices for all initialized ports */
1687 static int ub960_rxport_add_serializers(struct ub960_data *priv)
1688 {
1689 	unsigned int nport;
1690 	int ret;
1691 
1692 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1693 		struct ub960_rxport *rxport = priv->rxports[nport];
1694 
1695 		if (!rxport)
1696 			continue;
1697 
1698 		ret = ub960_rxport_add_serializer(priv, nport);
1699 		if (ret)
1700 			goto err_remove_sers;
1701 	}
1702 
1703 	return 0;
1704 
1705 err_remove_sers:
1706 	while (nport--) {
1707 		struct ub960_rxport *rxport = priv->rxports[nport];
1708 
1709 		if (!rxport)
1710 			continue;
1711 
1712 		ub960_rxport_remove_serializer(priv, nport);
1713 	}
1714 
1715 	return ret;
1716 }
1717 
1718 static void ub960_rxport_remove_serializers(struct ub960_data *priv)
1719 {
1720 	unsigned int nport;
1721 
1722 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
1723 		struct ub960_rxport *rxport = priv->rxports[nport];
1724 
1725 		if (!rxport)
1726 			continue;
1727 
1728 		ub960_rxport_remove_serializer(priv, nport);
1729 	}
1730 }
1731 
1732 static void ub960_init_tx_port(struct ub960_data *priv,
1733 			       struct ub960_txport *txport)
1734 {
1735 	unsigned int nport = txport->nport;
1736 	u8 csi_ctl = 0;
1737 
1738 	/*
1739 	 * From the datasheet: "initial CSI Skew-Calibration
1740 	 * sequence [...] should be set when operating at 1.6 Gbps"
1741 	 */
1742 	if (priv->tx_data_rate == MHZ(1600))
1743 		csi_ctl |= UB960_TR_CSI_CTL_CSI_CAL_EN;
1744 
1745 	csi_ctl |= (4 - txport->num_data_lanes) << 4;
1746 
1747 	ub960_txport_write(priv, nport, UB960_TR_CSI_CTL, csi_ctl);
1748 }
1749 
1750 static int ub960_init_tx_ports(struct ub960_data *priv)
1751 {
1752 	unsigned int nport;
1753 	u8 speed_select;
1754 	u8 pll_div;
1755 
1756 	/* TX ports */
1757 
1758 	switch (priv->tx_data_rate) {
1759 	case MHZ(1600):
1760 	default:
1761 		speed_select = 0;
1762 		pll_div = 0x10;
1763 		break;
1764 	case MHZ(1200):
1765 		speed_select = 1;
1766 		break;
1767 	case MHZ(800):
1768 		speed_select = 2;
1769 		pll_div = 0x10;
1770 		break;
1771 	case MHZ(400):
1772 		speed_select = 3;
1773 		pll_div = 0x10;
1774 		break;
1775 	}
1776 
1777 	ub960_write(priv, UB960_SR_CSI_PLL_CTL, speed_select);
1778 
1779 	if (priv->hw_data->is_ub9702) {
1780 		ub960_write(priv, UB960_SR_CSI_PLL_DIV, pll_div);
1781 
1782 		switch (priv->tx_data_rate) {
1783 		case MHZ(1600):
1784 		default:
1785 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x80);
1786 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1787 			break;
1788 		case MHZ(800):
1789 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0x90);
1790 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4f, 0x2a);
1791 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x4b, 0x2a);
1792 			break;
1793 		case MHZ(400):
1794 			ub960_write_ind(priv, UB960_IND_TARGET_CSI_ANA, 0x92, 0xa0);
1795 			break;
1796 		}
1797 	}
1798 
1799 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
1800 		struct ub960_txport *txport = priv->txports[nport];
1801 
1802 		if (!txport)
1803 			continue;
1804 
1805 		ub960_init_tx_port(priv, txport);
1806 	}
1807 
1808 	return 0;
1809 }
1810 
1811 static void ub960_init_rx_port_ub960(struct ub960_data *priv,
1812 				     struct ub960_rxport *rxport)
1813 {
1814 	unsigned int nport = rxport->nport;
1815 	u32 bc_freq_val;
1816 
1817 	/*
1818 	 * Back channel frequency select.
1819 	 * Override FREQ_SELECT from the strap.
1820 	 * 0 - 2.5 Mbps (DS90UB913A-Q1 / DS90UB933-Q1)
1821 	 * 2 - 10 Mbps
1822 	 * 6 - 50 Mbps (DS90UB953-Q1)
1823 	 *
1824 	 * Note that changing this setting will result in some errors on the back
1825 	 * channel for a short period of time.
1826 	 */
1827 
1828 	switch (rxport->rx_mode) {
1829 	case RXPORT_MODE_RAW10:
1830 	case RXPORT_MODE_RAW12_HF:
1831 	case RXPORT_MODE_RAW12_LF:
1832 		bc_freq_val = 0;
1833 		break;
1834 
1835 	case RXPORT_MODE_CSI2_ASYNC:
1836 		bc_freq_val = 2;
1837 		break;
1838 
1839 	case RXPORT_MODE_CSI2_SYNC:
1840 		bc_freq_val = 6;
1841 		break;
1842 
1843 	default:
1844 		return;
1845 	}
1846 
1847 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1848 				 UB960_RR_BCC_CONFIG_BC_FREQ_SEL_MASK,
1849 				 bc_freq_val);
1850 
1851 	switch (rxport->rx_mode) {
1852 	case RXPORT_MODE_RAW10:
1853 		/* FPD3_MODE = RAW10 Mode (DS90UB913A-Q1 / DS90UB933-Q1 compatible) */
1854 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG,
1855 					 UB960_RR_PORT_CONFIG_FPD3_MODE_MASK,
1856 					 0x3);
1857 
1858 		/*
1859 		 * RAW10_8BIT_CTL = 0b10 : 8-bit processing using upper 8 bits
1860 		 */
1861 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
1862 			UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_MASK,
1863 			0x2 << UB960_RR_PORT_CONFIG2_RAW10_8BIT_CTL_SHIFT);
1864 
1865 		break;
1866 
1867 	case RXPORT_MODE_RAW12_HF:
1868 	case RXPORT_MODE_RAW12_LF:
1869 		/* Not implemented */
1870 		return;
1871 
1872 	case RXPORT_MODE_CSI2_SYNC:
1873 	case RXPORT_MODE_CSI2_ASYNC:
1874 		/* CSI-2 Mode (DS90UB953-Q1 compatible) */
1875 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG, 0x3,
1876 					 0x0);
1877 
1878 		break;
1879 	}
1880 
1881 	/* LV_POLARITY & FV_POLARITY */
1882 	ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
1883 				 rxport->lv_fv_pol);
1884 
1885 	/* Enable all interrupt sources from this port */
1886 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
1887 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
1888 
1889 	/* Enable I2C_PASS_THROUGH */
1890 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
1891 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
1892 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
1893 
1894 	/* Enable I2C communication to the serializer via the alias addr */
1895 	ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
1896 			   rxport->ser.alias << 1);
1897 
1898 	/* Configure EQ related settings */
1899 	ub960_rxport_config_eq(priv, nport);
1900 
1901 	/* Enable RX port */
1902 	ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
1903 }
1904 
1905 static void ub960_init_rx_port_ub9702_fpd3(struct ub960_data *priv,
1906 					   struct ub960_rxport *rxport)
1907 {
1908 	unsigned int nport = rxport->nport;
1909 	u8 bc_freq_val;
1910 	u8 fpd_func_mode;
1911 
1912 	switch (rxport->rx_mode) {
1913 	case RXPORT_MODE_RAW10:
1914 		bc_freq_val = 0;
1915 		fpd_func_mode = 5;
1916 		break;
1917 
1918 	case RXPORT_MODE_RAW12_HF:
1919 		bc_freq_val = 0;
1920 		fpd_func_mode = 4;
1921 		break;
1922 
1923 	case RXPORT_MODE_RAW12_LF:
1924 		bc_freq_val = 0;
1925 		fpd_func_mode = 6;
1926 		break;
1927 
1928 	case RXPORT_MODE_CSI2_SYNC:
1929 		bc_freq_val = 6;
1930 		fpd_func_mode = 2;
1931 		break;
1932 
1933 	case RXPORT_MODE_CSI2_ASYNC:
1934 		bc_freq_val = 2;
1935 		fpd_func_mode = 2;
1936 		break;
1937 
1938 	default:
1939 		return;
1940 	}
1941 
1942 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
1943 				 bc_freq_val);
1944 	ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, fpd_func_mode);
1945 
1946 	/* set serdes_eq_mode = 1 */
1947 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa8, 0x80);
1948 
1949 	/* enable serdes driver */
1950 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x0d, 0x7f);
1951 
1952 	/* set serdes_eq_offset=4 */
1953 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
1954 
1955 	/* init default serdes_eq_max in 0xa9 */
1956 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xa9, 0x23);
1957 
1958 	/* init serdes_eq_min in 0xaa */
1959 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xaa, 0);
1960 
1961 	/* serdes_driver_ctl2 control: DS90UB953-Q1/DS90UB933-Q1/DS90UB913A-Q1 */
1962 	ub960_ind_update_bits(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b,
1963 			      BIT(3), BIT(3));
1964 
1965 	/* RX port to half-rate */
1966 	ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
1967 			  BIT(nport * 2));
1968 }
1969 
1970 static void ub960_init_rx_port_ub9702_fpd4_aeq(struct ub960_data *priv,
1971 					       struct ub960_rxport *rxport)
1972 {
1973 	unsigned int nport = rxport->nport;
1974 	bool first_time_power_up = true;
1975 
1976 	if (first_time_power_up) {
1977 		u8 v;
1978 
1979 		/* AEQ init */
1980 		ub960_read_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2c, &v);
1981 
1982 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, v);
1983 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, v + 1);
1984 
1985 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x00);
1986 	}
1987 
1988 	/* enable serdes_eq_ctl2 */
1989 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x9e, 0x00);
1990 
1991 	/* enable serdes_eq_ctl1 */
1992 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x90, 0x40);
1993 
1994 	/* enable serdes_eq_en */
1995 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2e, 0x40);
1996 
1997 	/* disable serdes_eq_override */
1998 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0xf0, 0x00);
1999 
2000 	/* disable serdes_gain_override */
2001 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x71, 0x00);
2002 }
2003 
2004 static void ub960_init_rx_port_ub9702_fpd4(struct ub960_data *priv,
2005 					   struct ub960_rxport *rxport)
2006 {
2007 	unsigned int nport = rxport->nport;
2008 	u8 bc_freq_val;
2009 
2010 	switch (rxport->rx_mode) {
2011 	case RXPORT_MODE_RAW10:
2012 		bc_freq_val = 0;
2013 		break;
2014 
2015 	case RXPORT_MODE_RAW12_HF:
2016 		bc_freq_val = 0;
2017 		break;
2018 
2019 	case RXPORT_MODE_RAW12_LF:
2020 		bc_freq_val = 0;
2021 		break;
2022 
2023 	case RXPORT_MODE_CSI2_SYNC:
2024 		bc_freq_val = 6;
2025 		break;
2026 
2027 	case RXPORT_MODE_CSI2_ASYNC:
2028 		bc_freq_val = 2;
2029 		break;
2030 
2031 	default:
2032 		return;
2033 	}
2034 
2035 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG, 0x7,
2036 				 bc_freq_val);
2037 
2038 	/* FPD4 Sync Mode */
2039 	ub960_rxport_write(priv, nport, UB960_RR_CHANNEL_MODE, 0);
2040 
2041 	/* add serdes_eq_offset of 4 */
2042 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x2b, 0x04);
2043 
2044 	/* FPD4 serdes_start_eq in 0x27: assign default */
2045 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x27, 0x0);
2046 	/* FPD4 serdes_end_eq in 0x28: assign default */
2047 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x28, 0x23);
2048 
2049 	/* set serdes_driver_mode into FPD IV mode */
2050 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x04, 0x00);
2051 	/* set FPD PBC drv into FPD IV mode */
2052 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x1b, 0x00);
2053 
2054 	/* set serdes_system_init to 0x2f */
2055 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x21, 0x2f);
2056 	/* set serdes_system_rst in reset mode */
2057 	ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0xc1);
2058 
2059 	/* RX port to 7.55G mode */
2060 	ub960_update_bits(priv, UB960_SR_FPD_RATE_CFG, 0x3 << (nport * 2),
2061 			  0 << (nport * 2));
2062 
2063 	ub960_init_rx_port_ub9702_fpd4_aeq(priv, rxport);
2064 }
2065 
2066 static void ub960_init_rx_port_ub9702(struct ub960_data *priv,
2067 				      struct ub960_rxport *rxport)
2068 {
2069 	unsigned int nport = rxport->nport;
2070 
2071 	if (rxport->cdr_mode == RXPORT_CDR_FPD3)
2072 		ub960_init_rx_port_ub9702_fpd3(priv, rxport);
2073 	else /* RXPORT_CDR_FPD4 */
2074 		ub960_init_rx_port_ub9702_fpd4(priv, rxport);
2075 
2076 	switch (rxport->rx_mode) {
2077 	case RXPORT_MODE_RAW10:
2078 		/*
2079 		 * RAW10_8BIT_CTL = 0b11 : 8-bit processing using lower 8 bits
2080 		 * 0b10 : 8-bit processing using upper 8 bits
2081 		 */
2082 		ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2,
2083 					 0x3 << 6, 0x2 << 6);
2084 
2085 		break;
2086 
2087 	case RXPORT_MODE_RAW12_HF:
2088 	case RXPORT_MODE_RAW12_LF:
2089 		/* Not implemented */
2090 		return;
2091 
2092 	case RXPORT_MODE_CSI2_SYNC:
2093 	case RXPORT_MODE_CSI2_ASYNC:
2094 
2095 		break;
2096 	}
2097 
2098 	/* LV_POLARITY & FV_POLARITY */
2099 	ub960_rxport_update_bits(priv, nport, UB960_RR_PORT_CONFIG2, 0x3,
2100 				 rxport->lv_fv_pol);
2101 
2102 	/* Enable all interrupt sources from this port */
2103 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_HI, 0x07);
2104 	ub960_rxport_write(priv, nport, UB960_RR_PORT_ICR_LO, 0x7f);
2105 
2106 	/* Enable I2C_PASS_THROUGH */
2107 	ub960_rxport_update_bits(priv, nport, UB960_RR_BCC_CONFIG,
2108 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH,
2109 				 UB960_RR_BCC_CONFIG_I2C_PASS_THROUGH);
2110 
2111 	/* Enable I2C communication to the serializer via the alias addr */
2112 	ub960_rxport_write(priv, nport, UB960_RR_SER_ALIAS_ID,
2113 			   rxport->ser.alias << 1);
2114 
2115 	/* Enable RX port */
2116 	ub960_update_bits(priv, UB960_SR_RX_PORT_CTL, BIT(nport), BIT(nport));
2117 
2118 	if (rxport->cdr_mode == RXPORT_CDR_FPD4) {
2119 		/* unreset 960 AEQ */
2120 		ub960_write_ind(priv, UB960_IND_TARGET_RX_ANA(nport), 0x25, 0x41);
2121 	}
2122 }
2123 
2124 static int ub960_init_rx_ports(struct ub960_data *priv)
2125 {
2126 	unsigned int nport;
2127 
2128 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2129 		struct ub960_rxport *rxport = priv->rxports[nport];
2130 
2131 		if (!rxport)
2132 			continue;
2133 
2134 		if (priv->hw_data->is_ub9702)
2135 			ub960_init_rx_port_ub9702(priv, rxport);
2136 		else
2137 			ub960_init_rx_port_ub960(priv, rxport);
2138 	}
2139 
2140 	return 0;
2141 }
2142 
2143 static void ub960_rxport_handle_events(struct ub960_data *priv, u8 nport)
2144 {
2145 	struct device *dev = &priv->client->dev;
2146 	u8 rx_port_sts1;
2147 	u8 rx_port_sts2;
2148 	u8 csi_rx_sts;
2149 	u8 bcc_sts;
2150 	int ret = 0;
2151 
2152 	/* Read interrupts (also clears most of them) */
2153 	if (!ret)
2154 		ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1,
2155 					&rx_port_sts1);
2156 	if (!ret)
2157 		ret = ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2,
2158 					&rx_port_sts2);
2159 	if (!ret)
2160 		ret = ub960_rxport_read(priv, nport, UB960_RR_CSI_RX_STS,
2161 					&csi_rx_sts);
2162 	if (!ret)
2163 		ret = ub960_rxport_read(priv, nport, UB960_RR_BCC_STATUS,
2164 					&bcc_sts);
2165 
2166 	if (ret)
2167 		return;
2168 
2169 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_PARITY_ERROR) {
2170 		u16 v;
2171 
2172 		ret = ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI,
2173 					  &v);
2174 		if (!ret)
2175 			dev_err(dev, "rx%u parity errors: %u\n", nport, v);
2176 	}
2177 
2178 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_CRC_ERROR)
2179 		dev_err(dev, "rx%u BCC CRC error\n", nport);
2180 
2181 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_BCC_SEQ_ERROR)
2182 		dev_err(dev, "rx%u BCC SEQ error\n", nport);
2183 
2184 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_UNSTABLE)
2185 		dev_err(dev, "rx%u line length unstable\n", nport);
2186 
2187 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_FPD3_ENCODE_ERROR)
2188 		dev_err(dev, "rx%u FPD3 encode error\n", nport);
2189 
2190 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_BUFFER_ERROR)
2191 		dev_err(dev, "rx%u buffer error\n", nport);
2192 
2193 	if (csi_rx_sts)
2194 		dev_err(dev, "rx%u CSI error: %#02x\n", nport, csi_rx_sts);
2195 
2196 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC1_ERR)
2197 		dev_err(dev, "rx%u CSI ECC1 error\n", nport);
2198 
2199 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_ECC2_ERR)
2200 		dev_err(dev, "rx%u CSI ECC2 error\n", nport);
2201 
2202 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_CKSUM_ERR)
2203 		dev_err(dev, "rx%u CSI checksum error\n", nport);
2204 
2205 	if (csi_rx_sts & UB960_RR_CSI_RX_STS_LENGTH_ERR)
2206 		dev_err(dev, "rx%u CSI length error\n", nport);
2207 
2208 	if (bcc_sts)
2209 		dev_err(dev, "rx%u BCC error: %#02x\n", nport, bcc_sts);
2210 
2211 	if (bcc_sts & UB960_RR_BCC_STATUS_RESP_ERR)
2212 		dev_err(dev, "rx%u BCC response error", nport);
2213 
2214 	if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_TO)
2215 		dev_err(dev, "rx%u BCC slave timeout", nport);
2216 
2217 	if (bcc_sts & UB960_RR_BCC_STATUS_SLAVE_ERR)
2218 		dev_err(dev, "rx%u BCC slave error", nport);
2219 
2220 	if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_TO)
2221 		dev_err(dev, "rx%u BCC master timeout", nport);
2222 
2223 	if (bcc_sts & UB960_RR_BCC_STATUS_MASTER_ERR)
2224 		dev_err(dev, "rx%u BCC master error", nport);
2225 
2226 	if (bcc_sts & UB960_RR_BCC_STATUS_SEQ_ERROR)
2227 		dev_err(dev, "rx%u BCC sequence error", nport);
2228 
2229 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_LEN_CHG) {
2230 		u16 v;
2231 
2232 		ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v);
2233 		if (!ret)
2234 			dev_dbg(dev, "rx%u line len changed: %u\n", nport, v);
2235 	}
2236 
2237 	if (rx_port_sts2 & UB960_RR_RX_PORT_STS2_LINE_CNT_CHG) {
2238 		u16 v;
2239 
2240 		ret = ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI,
2241 					  &v);
2242 		if (!ret)
2243 			dev_dbg(dev, "rx%u line count changed: %u\n", nport, v);
2244 	}
2245 
2246 	if (rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS_CHG) {
2247 		dev_dbg(dev, "rx%u: %s, %s, %s, %s\n", nport,
2248 			(rx_port_sts1 & UB960_RR_RX_PORT_STS1_LOCK_STS) ?
2249 				"locked" :
2250 				"unlocked",
2251 			(rx_port_sts1 & UB960_RR_RX_PORT_STS1_PORT_PASS) ?
2252 				"passed" :
2253 				"not passed",
2254 			(rx_port_sts2 & UB960_RR_RX_PORT_STS2_CABLE_FAULT) ?
2255 				"no clock" :
2256 				"clock ok",
2257 			(rx_port_sts2 & UB960_RR_RX_PORT_STS2_FREQ_STABLE) ?
2258 				"stable freq" :
2259 				"unstable freq");
2260 	}
2261 }
2262 
2263 /* -----------------------------------------------------------------------------
2264  * V4L2
2265  */
2266 
2267 /*
2268  * The current implementation only supports a simple VC mapping, where all VCs
2269  * from a one RX port will be mapped to the same VC. Also, the hardware
2270  * dictates that all streams from an RX port must go to a single TX port.
2271  *
2272  * This function decides the target VC numbers for each RX port with a simple
2273  * algorithm, so that for each TX port, we get VC numbers starting from 0,
2274  * and counting up.
2275  *
2276  * E.g. if all four RX ports are in use, of which the first two go to the
2277  * first TX port and the secont two go to the second TX port, we would get
2278  * the following VCs for the four RX ports: 0, 1, 0, 1.
2279  *
2280  * TODO: implement a more sophisticated VC mapping. As the driver cannot know
2281  * what VCs the sinks expect (say, an FPGA with hardcoded VC routing), this
2282  * probably needs to be somehow configurable. Device tree?
2283  */
2284 static void ub960_get_vc_maps(struct ub960_data *priv,
2285 			      struct v4l2_subdev_state *state, u8 *vc)
2286 {
2287 	u8 cur_vc[UB960_MAX_TX_NPORTS] = {};
2288 	struct v4l2_subdev_route *route;
2289 	u8 handled_mask = 0;
2290 
2291 	for_each_active_route(&state->routing, route) {
2292 		unsigned int rx, tx;
2293 
2294 		rx = ub960_pad_to_port(priv, route->sink_pad);
2295 		if (BIT(rx) & handled_mask)
2296 			continue;
2297 
2298 		tx = ub960_pad_to_port(priv, route->source_pad);
2299 
2300 		vc[rx] = cur_vc[tx]++;
2301 		handled_mask |= BIT(rx);
2302 	}
2303 }
2304 
2305 static int ub960_enable_tx_port(struct ub960_data *priv, unsigned int nport)
2306 {
2307 	struct device *dev = &priv->client->dev;
2308 
2309 	dev_dbg(dev, "enable TX port %u\n", nport);
2310 
2311 	return ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2312 					UB960_TR_CSI_CTL_CSI_ENABLE,
2313 					UB960_TR_CSI_CTL_CSI_ENABLE);
2314 }
2315 
2316 static void ub960_disable_tx_port(struct ub960_data *priv, unsigned int nport)
2317 {
2318 	struct device *dev = &priv->client->dev;
2319 
2320 	dev_dbg(dev, "disable TX port %u\n", nport);
2321 
2322 	ub960_txport_update_bits(priv, nport, UB960_TR_CSI_CTL,
2323 				 UB960_TR_CSI_CTL_CSI_ENABLE, 0);
2324 }
2325 
2326 static int ub960_enable_rx_port(struct ub960_data *priv, unsigned int nport)
2327 {
2328 	struct device *dev = &priv->client->dev;
2329 
2330 	dev_dbg(dev, "enable RX port %u\n", nport);
2331 
2332 	/* Enable forwarding */
2333 	return ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2334 				 UB960_SR_FWD_CTL1_PORT_DIS(nport), 0);
2335 }
2336 
2337 static void ub960_disable_rx_port(struct ub960_data *priv, unsigned int nport)
2338 {
2339 	struct device *dev = &priv->client->dev;
2340 
2341 	dev_dbg(dev, "disable RX port %u\n", nport);
2342 
2343 	/* Disable forwarding */
2344 	ub960_update_bits(priv, UB960_SR_FWD_CTL1,
2345 			  UB960_SR_FWD_CTL1_PORT_DIS(nport),
2346 			  UB960_SR_FWD_CTL1_PORT_DIS(nport));
2347 }
2348 
2349 /*
2350  * The driver only supports using a single VC for each source. This function
2351  * checks that each source only provides streams using a single VC.
2352  */
2353 static int ub960_validate_stream_vcs(struct ub960_data *priv)
2354 {
2355 	unsigned int nport;
2356 	unsigned int i;
2357 
2358 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2359 		struct ub960_rxport *rxport = priv->rxports[nport];
2360 		struct v4l2_mbus_frame_desc desc;
2361 		int ret;
2362 		u8 vc;
2363 
2364 		if (!rxport)
2365 			continue;
2366 
2367 		ret = v4l2_subdev_call(rxport->source.sd, pad, get_frame_desc,
2368 				       rxport->source.pad, &desc);
2369 		if (ret)
2370 			return ret;
2371 
2372 		if (desc.type != V4L2_MBUS_FRAME_DESC_TYPE_CSI2)
2373 			continue;
2374 
2375 		if (desc.num_entries == 0)
2376 			continue;
2377 
2378 		vc = desc.entry[0].bus.csi2.vc;
2379 
2380 		for (i = 1; i < desc.num_entries; i++) {
2381 			if (vc == desc.entry[i].bus.csi2.vc)
2382 				continue;
2383 
2384 			dev_err(&priv->client->dev,
2385 				"rx%u: source with multiple virtual-channels is not supported\n",
2386 				nport);
2387 			return -ENODEV;
2388 		}
2389 	}
2390 
2391 	return 0;
2392 }
2393 
2394 static int ub960_configure_ports_for_streaming(struct ub960_data *priv,
2395 					       struct v4l2_subdev_state *state)
2396 {
2397 	u8 fwd_ctl;
2398 	struct {
2399 		u32 num_streams;
2400 		u8 pixel_dt;
2401 		u8 meta_dt;
2402 		u32 meta_lines;
2403 		u32 tx_port;
2404 	} rx_data[UB960_MAX_RX_NPORTS] = {};
2405 	u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2406 	struct v4l2_subdev_route *route;
2407 	unsigned int nport;
2408 	int ret;
2409 
2410 	ret = ub960_validate_stream_vcs(priv);
2411 	if (ret)
2412 		return ret;
2413 
2414 	ub960_get_vc_maps(priv, state, vc_map);
2415 
2416 	for_each_active_route(&state->routing, route) {
2417 		struct ub960_rxport *rxport;
2418 		struct ub960_txport *txport;
2419 		struct v4l2_mbus_framefmt *fmt;
2420 		const struct ub960_format_info *ub960_fmt;
2421 		unsigned int nport;
2422 
2423 		nport = ub960_pad_to_port(priv, route->sink_pad);
2424 
2425 		rxport = priv->rxports[nport];
2426 		if (!rxport)
2427 			return -EINVAL;
2428 
2429 		txport = priv->txports[ub960_pad_to_port(priv, route->source_pad)];
2430 		if (!txport)
2431 			return -EINVAL;
2432 
2433 		rx_data[nport].tx_port = ub960_pad_to_port(priv, route->source_pad);
2434 
2435 		rx_data[nport].num_streams++;
2436 
2437 		/* For the rest, we are only interested in parallel busses */
2438 		if (rxport->rx_mode == RXPORT_MODE_CSI2_SYNC ||
2439 		    rxport->rx_mode == RXPORT_MODE_CSI2_ASYNC)
2440 			continue;
2441 
2442 		if (rx_data[nport].num_streams > 2)
2443 			return -EPIPE;
2444 
2445 		fmt = v4l2_subdev_state_get_stream_format(state,
2446 							  route->sink_pad,
2447 							  route->sink_stream);
2448 		if (!fmt)
2449 			return -EPIPE;
2450 
2451 		ub960_fmt = ub960_find_format(fmt->code);
2452 		if (!ub960_fmt)
2453 			return -EPIPE;
2454 
2455 		if (ub960_fmt->meta) {
2456 			if (fmt->height > 3) {
2457 				dev_err(&priv->client->dev,
2458 					"rx%u: unsupported metadata height %u\n",
2459 					nport, fmt->height);
2460 				return -EPIPE;
2461 			}
2462 
2463 			rx_data[nport].meta_dt = ub960_fmt->datatype;
2464 			rx_data[nport].meta_lines = fmt->height;
2465 		} else {
2466 			rx_data[nport].pixel_dt = ub960_fmt->datatype;
2467 		}
2468 	}
2469 
2470 	/* Configure RX ports */
2471 
2472 	/*
2473 	 * Keep all port forwardings disabled by default. Forwarding will be
2474 	 * enabled in ub960_enable_rx_port.
2475 	 */
2476 	fwd_ctl = GENMASK(7, 4);
2477 
2478 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2479 		struct ub960_rxport *rxport = priv->rxports[nport];
2480 		u8 vc = vc_map[nport];
2481 
2482 		if (rx_data[nport].num_streams == 0)
2483 			continue;
2484 
2485 		switch (rxport->rx_mode) {
2486 		case RXPORT_MODE_RAW10:
2487 			ub960_rxport_write(priv, nport, UB960_RR_RAW10_ID,
2488 				rx_data[nport].pixel_dt | (vc << UB960_RR_RAW10_ID_VC_SHIFT));
2489 
2490 			ub960_rxport_write(priv, rxport->nport,
2491 				UB960_RR_RAW_EMBED_DTYPE,
2492 				(rx_data[nport].meta_lines << UB960_RR_RAW_EMBED_DTYPE_LINES_SHIFT) |
2493 					rx_data[nport].meta_dt);
2494 
2495 			break;
2496 
2497 		case RXPORT_MODE_RAW12_HF:
2498 		case RXPORT_MODE_RAW12_LF:
2499 			/* Not implemented */
2500 			break;
2501 
2502 		case RXPORT_MODE_CSI2_SYNC:
2503 		case RXPORT_MODE_CSI2_ASYNC:
2504 			if (!priv->hw_data->is_ub9702) {
2505 				/* Map all VCs from this port to the same VC */
2506 				ub960_rxport_write(priv, nport, UB960_RR_CSI_VC_MAP,
2507 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(3)) |
2508 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(2)) |
2509 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(1)) |
2510 						   (vc << UB960_RR_CSI_VC_MAP_SHIFT(0)));
2511 			} else {
2512 				unsigned int i;
2513 
2514 				/* Map all VCs from this port to VC(nport) */
2515 				for (i = 0; i < 8; i++)
2516 					ub960_rxport_write(priv, nport,
2517 							   UB960_RR_VC_ID_MAP(i),
2518 							   nport);
2519 			}
2520 
2521 			break;
2522 		}
2523 
2524 		if (rx_data[nport].tx_port == 1)
2525 			fwd_ctl |= BIT(nport); /* forward to TX1 */
2526 		else
2527 			fwd_ctl &= ~BIT(nport); /* forward to TX0 */
2528 	}
2529 
2530 	ub960_write(priv, UB960_SR_FWD_CTL1, fwd_ctl);
2531 
2532 	return 0;
2533 }
2534 
2535 static void ub960_update_streaming_status(struct ub960_data *priv)
2536 {
2537 	unsigned int i;
2538 
2539 	for (i = 0; i < UB960_MAX_NPORTS; i++) {
2540 		if (priv->stream_enable_mask[i])
2541 			break;
2542 	}
2543 
2544 	priv->streaming = i < UB960_MAX_NPORTS;
2545 }
2546 
2547 static int ub960_enable_streams(struct v4l2_subdev *sd,
2548 				struct v4l2_subdev_state *state, u32 source_pad,
2549 				u64 source_streams_mask)
2550 {
2551 	struct ub960_data *priv = sd_to_ub960(sd);
2552 	struct device *dev = &priv->client->dev;
2553 	u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2554 	struct v4l2_subdev_route *route;
2555 	unsigned int failed_port;
2556 	unsigned int nport;
2557 	int ret;
2558 
2559 	if (!priv->streaming) {
2560 		dev_dbg(dev, "Prepare for streaming\n");
2561 		ret = ub960_configure_ports_for_streaming(priv, state);
2562 		if (ret)
2563 			return ret;
2564 	}
2565 
2566 	/* Enable TX port if not yet enabled */
2567 	if (!priv->stream_enable_mask[source_pad]) {
2568 		ret = ub960_enable_tx_port(priv,
2569 					   ub960_pad_to_port(priv, source_pad));
2570 		if (ret)
2571 			return ret;
2572 	}
2573 
2574 	priv->stream_enable_mask[source_pad] |= source_streams_mask;
2575 
2576 	/* Collect sink streams per pad which we need to enable */
2577 	for_each_active_route(&state->routing, route) {
2578 		if (route->source_pad != source_pad)
2579 			continue;
2580 
2581 		if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2582 			continue;
2583 
2584 		nport = ub960_pad_to_port(priv, route->sink_pad);
2585 
2586 		sink_streams[nport] |= BIT_ULL(route->sink_stream);
2587 	}
2588 
2589 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2590 		if (!sink_streams[nport])
2591 			continue;
2592 
2593 		/* Enable the RX port if not yet enabled */
2594 		if (!priv->stream_enable_mask[nport]) {
2595 			ret = ub960_enable_rx_port(priv, nport);
2596 			if (ret) {
2597 				failed_port = nport;
2598 				goto err;
2599 			}
2600 		}
2601 
2602 		priv->stream_enable_mask[nport] |= sink_streams[nport];
2603 
2604 		dev_dbg(dev, "enable RX port %u streams %#llx\n", nport,
2605 			sink_streams[nport]);
2606 
2607 		ret = v4l2_subdev_enable_streams(
2608 			priv->rxports[nport]->source.sd,
2609 			priv->rxports[nport]->source.pad,
2610 			sink_streams[nport]);
2611 		if (ret) {
2612 			priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2613 
2614 			if (!priv->stream_enable_mask[nport])
2615 				ub960_disable_rx_port(priv, nport);
2616 
2617 			failed_port = nport;
2618 			goto err;
2619 		}
2620 	}
2621 
2622 	priv->streaming = true;
2623 
2624 	return 0;
2625 
2626 err:
2627 	for (nport = 0; nport < failed_port; nport++) {
2628 		if (!sink_streams[nport])
2629 			continue;
2630 
2631 		dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2632 			sink_streams[nport]);
2633 
2634 		ret = v4l2_subdev_disable_streams(
2635 			priv->rxports[nport]->source.sd,
2636 			priv->rxports[nport]->source.pad,
2637 			sink_streams[nport]);
2638 		if (ret)
2639 			dev_err(dev, "Failed to disable streams: %d\n", ret);
2640 
2641 		priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2642 
2643 		/* Disable RX port if no active streams */
2644 		if (!priv->stream_enable_mask[nport])
2645 			ub960_disable_rx_port(priv, nport);
2646 	}
2647 
2648 	priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2649 
2650 	if (!priv->stream_enable_mask[source_pad])
2651 		ub960_disable_tx_port(priv,
2652 				      ub960_pad_to_port(priv, source_pad));
2653 
2654 	ub960_update_streaming_status(priv);
2655 
2656 	return ret;
2657 }
2658 
2659 static int ub960_disable_streams(struct v4l2_subdev *sd,
2660 				 struct v4l2_subdev_state *state,
2661 				 u32 source_pad, u64 source_streams_mask)
2662 {
2663 	struct ub960_data *priv = sd_to_ub960(sd);
2664 	struct device *dev = &priv->client->dev;
2665 	u64 sink_streams[UB960_MAX_RX_NPORTS] = {};
2666 	struct v4l2_subdev_route *route;
2667 	unsigned int nport;
2668 	int ret;
2669 
2670 	/* Collect sink streams per pad which we need to disable */
2671 	for_each_active_route(&state->routing, route) {
2672 		if (route->source_pad != source_pad)
2673 			continue;
2674 
2675 		if (!(source_streams_mask & BIT_ULL(route->source_stream)))
2676 			continue;
2677 
2678 		nport = ub960_pad_to_port(priv, route->sink_pad);
2679 
2680 		sink_streams[nport] |= BIT_ULL(route->sink_stream);
2681 	}
2682 
2683 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2684 		if (!sink_streams[nport])
2685 			continue;
2686 
2687 		dev_dbg(dev, "disable RX port %u streams %#llx\n", nport,
2688 			sink_streams[nport]);
2689 
2690 		ret = v4l2_subdev_disable_streams(
2691 			priv->rxports[nport]->source.sd,
2692 			priv->rxports[nport]->source.pad,
2693 			sink_streams[nport]);
2694 		if (ret)
2695 			dev_err(dev, "Failed to disable streams: %d\n", ret);
2696 
2697 		priv->stream_enable_mask[nport] &= ~sink_streams[nport];
2698 
2699 		/* Disable RX port if no active streams */
2700 		if (!priv->stream_enable_mask[nport])
2701 			ub960_disable_rx_port(priv, nport);
2702 	}
2703 
2704 	/* Disable TX port if no active streams */
2705 
2706 	priv->stream_enable_mask[source_pad] &= ~source_streams_mask;
2707 
2708 	if (!priv->stream_enable_mask[source_pad])
2709 		ub960_disable_tx_port(priv,
2710 				      ub960_pad_to_port(priv, source_pad));
2711 
2712 	ub960_update_streaming_status(priv);
2713 
2714 	return 0;
2715 }
2716 
2717 static int _ub960_set_routing(struct v4l2_subdev *sd,
2718 			      struct v4l2_subdev_state *state,
2719 			      struct v4l2_subdev_krouting *routing)
2720 {
2721 	static const struct v4l2_mbus_framefmt format = {
2722 		.width = 640,
2723 		.height = 480,
2724 		.code = MEDIA_BUS_FMT_UYVY8_1X16,
2725 		.field = V4L2_FIELD_NONE,
2726 		.colorspace = V4L2_COLORSPACE_SRGB,
2727 		.ycbcr_enc = V4L2_YCBCR_ENC_601,
2728 		.quantization = V4L2_QUANTIZATION_LIM_RANGE,
2729 		.xfer_func = V4L2_XFER_FUNC_SRGB,
2730 	};
2731 	int ret;
2732 
2733 	/*
2734 	 * Note: we can only support up to V4L2_FRAME_DESC_ENTRY_MAX, until
2735 	 * frame desc is made dynamically allocated.
2736 	 */
2737 
2738 	if (routing->num_routes > V4L2_FRAME_DESC_ENTRY_MAX)
2739 		return -E2BIG;
2740 
2741 	ret = v4l2_subdev_routing_validate(sd, routing,
2742 					   V4L2_SUBDEV_ROUTING_ONLY_1_TO_1 |
2743 					   V4L2_SUBDEV_ROUTING_NO_SINK_STREAM_MIX);
2744 	if (ret)
2745 		return ret;
2746 
2747 	ret = v4l2_subdev_set_routing_with_fmt(sd, state, routing, &format);
2748 	if (ret)
2749 		return ret;
2750 
2751 	return 0;
2752 }
2753 
2754 static int ub960_set_routing(struct v4l2_subdev *sd,
2755 			     struct v4l2_subdev_state *state,
2756 			     enum v4l2_subdev_format_whence which,
2757 			     struct v4l2_subdev_krouting *routing)
2758 {
2759 	struct ub960_data *priv = sd_to_ub960(sd);
2760 
2761 	if (which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2762 		return -EBUSY;
2763 
2764 	return _ub960_set_routing(sd, state, routing);
2765 }
2766 
2767 static int ub960_get_frame_desc(struct v4l2_subdev *sd, unsigned int pad,
2768 				struct v4l2_mbus_frame_desc *fd)
2769 {
2770 	struct ub960_data *priv = sd_to_ub960(sd);
2771 	struct v4l2_subdev_route *route;
2772 	struct v4l2_subdev_state *state;
2773 	int ret = 0;
2774 	struct device *dev = &priv->client->dev;
2775 	u8 vc_map[UB960_MAX_RX_NPORTS] = {};
2776 
2777 	if (!ub960_pad_is_source(priv, pad))
2778 		return -EINVAL;
2779 
2780 	memset(fd, 0, sizeof(*fd));
2781 
2782 	fd->type = V4L2_MBUS_FRAME_DESC_TYPE_CSI2;
2783 
2784 	state = v4l2_subdev_lock_and_get_active_state(&priv->sd);
2785 
2786 	ub960_get_vc_maps(priv, state, vc_map);
2787 
2788 	for_each_active_route(&state->routing, route) {
2789 		struct v4l2_mbus_frame_desc_entry *source_entry = NULL;
2790 		struct v4l2_mbus_frame_desc source_fd;
2791 		unsigned int nport;
2792 		unsigned int i;
2793 
2794 		if (route->source_pad != pad)
2795 			continue;
2796 
2797 		nport = ub960_pad_to_port(priv, route->sink_pad);
2798 
2799 		ret = v4l2_subdev_call(priv->rxports[nport]->source.sd, pad,
2800 				       get_frame_desc,
2801 				       priv->rxports[nport]->source.pad,
2802 				       &source_fd);
2803 		if (ret) {
2804 			dev_err(dev,
2805 				"Failed to get source frame desc for pad %u\n",
2806 				route->sink_pad);
2807 			goto out_unlock;
2808 		}
2809 
2810 		for (i = 0; i < source_fd.num_entries; i++) {
2811 			if (source_fd.entry[i].stream == route->sink_stream) {
2812 				source_entry = &source_fd.entry[i];
2813 				break;
2814 			}
2815 		}
2816 
2817 		if (!source_entry) {
2818 			dev_err(dev,
2819 				"Failed to find stream from source frame desc\n");
2820 			ret = -EPIPE;
2821 			goto out_unlock;
2822 		}
2823 
2824 		fd->entry[fd->num_entries].stream = route->source_stream;
2825 		fd->entry[fd->num_entries].flags = source_entry->flags;
2826 		fd->entry[fd->num_entries].length = source_entry->length;
2827 		fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
2828 
2829 		fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
2830 
2831 		if (source_fd.type == V4L2_MBUS_FRAME_DESC_TYPE_CSI2) {
2832 			fd->entry[fd->num_entries].bus.csi2.dt =
2833 				source_entry->bus.csi2.dt;
2834 		} else {
2835 			const struct ub960_format_info *ub960_fmt;
2836 			struct v4l2_mbus_framefmt *fmt;
2837 
2838 			fmt = v4l2_subdev_state_get_stream_format(state, pad,
2839 								  route->source_stream);
2840 
2841 			if (!fmt) {
2842 				ret = -EINVAL;
2843 				goto out_unlock;
2844 			}
2845 
2846 			ub960_fmt = ub960_find_format(fmt->code);
2847 			if (!ub960_fmt) {
2848 				dev_err(dev, "Unable to find format\n");
2849 				ret = -EINVAL;
2850 				goto out_unlock;
2851 			}
2852 
2853 			fd->entry[fd->num_entries].bus.csi2.dt =
2854 				ub960_fmt->datatype;
2855 		}
2856 
2857 		fd->num_entries++;
2858 	}
2859 
2860 out_unlock:
2861 	v4l2_subdev_unlock_state(state);
2862 
2863 	return ret;
2864 }
2865 
2866 static int ub960_set_fmt(struct v4l2_subdev *sd,
2867 			 struct v4l2_subdev_state *state,
2868 			 struct v4l2_subdev_format *format)
2869 {
2870 	struct ub960_data *priv = sd_to_ub960(sd);
2871 	struct v4l2_mbus_framefmt *fmt;
2872 
2873 	if (format->which == V4L2_SUBDEV_FORMAT_ACTIVE && priv->streaming)
2874 		return -EBUSY;
2875 
2876 	/* No transcoding, source and sink formats must match. */
2877 	if (ub960_pad_is_source(priv, format->pad))
2878 		return v4l2_subdev_get_fmt(sd, state, format);
2879 
2880 	/*
2881 	 * Default to the first format if the requested media bus code isn't
2882 	 * supported.
2883 	 */
2884 	if (!ub960_find_format(format->format.code))
2885 		format->format.code = ub960_formats[0].code;
2886 
2887 	fmt = v4l2_subdev_state_get_stream_format(state, format->pad,
2888 						  format->stream);
2889 	if (!fmt)
2890 		return -EINVAL;
2891 
2892 	*fmt = format->format;
2893 
2894 	fmt = v4l2_subdev_state_get_opposite_stream_format(state, format->pad,
2895 							   format->stream);
2896 	if (!fmt)
2897 		return -EINVAL;
2898 
2899 	*fmt = format->format;
2900 
2901 	return 0;
2902 }
2903 
2904 static int ub960_init_cfg(struct v4l2_subdev *sd,
2905 			  struct v4l2_subdev_state *state)
2906 {
2907 	struct ub960_data *priv = sd_to_ub960(sd);
2908 
2909 	struct v4l2_subdev_route routes[] = {
2910 		{
2911 			.sink_pad = 0,
2912 			.sink_stream = 0,
2913 			.source_pad = priv->hw_data->num_rxports,
2914 			.source_stream = 0,
2915 			.flags = V4L2_SUBDEV_ROUTE_FL_ACTIVE,
2916 		},
2917 	};
2918 
2919 	struct v4l2_subdev_krouting routing = {
2920 		.num_routes = ARRAY_SIZE(routes),
2921 		.routes = routes,
2922 	};
2923 
2924 	return _ub960_set_routing(sd, state, &routing);
2925 }
2926 
2927 static const struct v4l2_subdev_pad_ops ub960_pad_ops = {
2928 	.enable_streams = ub960_enable_streams,
2929 	.disable_streams = ub960_disable_streams,
2930 
2931 	.set_routing = ub960_set_routing,
2932 	.get_frame_desc = ub960_get_frame_desc,
2933 
2934 	.get_fmt = v4l2_subdev_get_fmt,
2935 	.set_fmt = ub960_set_fmt,
2936 
2937 	.init_cfg = ub960_init_cfg,
2938 };
2939 
2940 static int ub960_log_status(struct v4l2_subdev *sd)
2941 {
2942 	struct ub960_data *priv = sd_to_ub960(sd);
2943 	struct device *dev = &priv->client->dev;
2944 	struct v4l2_subdev_state *state;
2945 	unsigned int nport;
2946 	unsigned int i;
2947 	u16 v16 = 0;
2948 	u8 v = 0;
2949 	u8 id[UB960_SR_FPD3_RX_ID_LEN];
2950 
2951 	state = v4l2_subdev_lock_and_get_active_state(sd);
2952 
2953 	for (i = 0; i < sizeof(id); i++)
2954 		ub960_read(priv, UB960_SR_FPD3_RX_ID(i), &id[i]);
2955 
2956 	dev_info(dev, "ID '%.*s'\n", (int)sizeof(id), id);
2957 
2958 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
2959 		struct ub960_txport *txport = priv->txports[nport];
2960 
2961 		dev_info(dev, "TX %u\n", nport);
2962 
2963 		if (!txport) {
2964 			dev_info(dev, "\tNot initialized\n");
2965 			continue;
2966 		}
2967 
2968 		ub960_txport_read(priv, nport, UB960_TR_CSI_STS, &v);
2969 		dev_info(dev, "\tsync %u, pass %u\n", v & (u8)BIT(1),
2970 			 v & (u8)BIT(0));
2971 
2972 		ub960_read16(priv, UB960_SR_CSI_FRAME_COUNT_HI(nport), &v16);
2973 		dev_info(dev, "\tframe counter %u\n", v16);
2974 
2975 		ub960_read16(priv, UB960_SR_CSI_FRAME_ERR_COUNT_HI(nport), &v16);
2976 		dev_info(dev, "\tframe error counter %u\n", v16);
2977 
2978 		ub960_read16(priv, UB960_SR_CSI_LINE_COUNT_HI(nport), &v16);
2979 		dev_info(dev, "\tline counter %u\n", v16);
2980 
2981 		ub960_read16(priv, UB960_SR_CSI_LINE_ERR_COUNT_HI(nport), &v16);
2982 		dev_info(dev, "\tline error counter %u\n", v16);
2983 	}
2984 
2985 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
2986 		struct ub960_rxport *rxport = priv->rxports[nport];
2987 		u8 eq_level;
2988 		s8 strobe_pos;
2989 		unsigned int i;
2990 
2991 		dev_info(dev, "RX %u\n", nport);
2992 
2993 		if (!rxport) {
2994 			dev_info(dev, "\tNot initialized\n");
2995 			continue;
2996 		}
2997 
2998 		ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS1, &v);
2999 
3000 		if (v & UB960_RR_RX_PORT_STS1_LOCK_STS)
3001 			dev_info(dev, "\tLocked\n");
3002 		else
3003 			dev_info(dev, "\tNot locked\n");
3004 
3005 		dev_info(dev, "\trx_port_sts1 %#02x\n", v);
3006 		ub960_rxport_read(priv, nport, UB960_RR_RX_PORT_STS2, &v);
3007 		dev_info(dev, "\trx_port_sts2 %#02x\n", v);
3008 
3009 		ub960_rxport_read16(priv, nport, UB960_RR_RX_FREQ_HIGH, &v16);
3010 		dev_info(dev, "\tlink freq %llu Hz\n", (v16 * 1000000ULL) >> 8);
3011 
3012 		ub960_rxport_read16(priv, nport, UB960_RR_RX_PAR_ERR_HI, &v16);
3013 		dev_info(dev, "\tparity errors %u\n", v16);
3014 
3015 		ub960_rxport_read16(priv, nport, UB960_RR_LINE_COUNT_HI, &v16);
3016 		dev_info(dev, "\tlines per frame %u\n", v16);
3017 
3018 		ub960_rxport_read16(priv, nport, UB960_RR_LINE_LEN_1, &v16);
3019 		dev_info(dev, "\tbytes per line %u\n", v16);
3020 
3021 		ub960_rxport_read(priv, nport, UB960_RR_CSI_ERR_COUNTER, &v);
3022 		dev_info(dev, "\tcsi_err_counter %u\n", v);
3023 
3024 		/* Strobe */
3025 
3026 		ub960_read(priv, UB960_XR_AEQ_CTL1, &v);
3027 
3028 		dev_info(dev, "\t%s strobe\n",
3029 			 (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) ? "Adaptive" :
3030 								  "Manual");
3031 
3032 		if (v & UB960_XR_AEQ_CTL1_AEQ_SFILTER_EN) {
3033 			ub960_read(priv, UB960_XR_SFILTER_CFG, &v);
3034 
3035 			dev_info(dev, "\tStrobe range [%d, %d]\n",
3036 				 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MIN_SHIFT) & 0xf) - 7,
3037 				 ((v >> UB960_XR_SFILTER_CFG_SFILTER_MAX_SHIFT) & 0xf) - 7);
3038 		}
3039 
3040 		ub960_rxport_get_strobe_pos(priv, nport, &strobe_pos);
3041 
3042 		dev_info(dev, "\tStrobe pos %d\n", strobe_pos);
3043 
3044 		/* EQ */
3045 
3046 		ub960_rxport_read(priv, nport, UB960_RR_AEQ_BYPASS, &v);
3047 
3048 		dev_info(dev, "\t%s EQ\n",
3049 			 (v & UB960_RR_AEQ_BYPASS_ENABLE) ? "Manual" :
3050 							    "Adaptive");
3051 
3052 		if (!(v & UB960_RR_AEQ_BYPASS_ENABLE)) {
3053 			ub960_rxport_read(priv, nport, UB960_RR_AEQ_MIN_MAX, &v);
3054 
3055 			dev_info(dev, "\tEQ range [%u, %u]\n",
3056 				 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_FLOOR_SHIFT) & 0xf,
3057 				 (v >> UB960_RR_AEQ_MIN_MAX_AEQ_MAX_SHIFT) & 0xf);
3058 		}
3059 
3060 		if (ub960_rxport_get_eq_level(priv, nport, &eq_level) == 0)
3061 			dev_info(dev, "\tEQ level %u\n", eq_level);
3062 
3063 		/* GPIOs */
3064 		for (i = 0; i < UB960_NUM_BC_GPIOS; i++) {
3065 			u8 ctl_reg;
3066 			u8 ctl_shift;
3067 
3068 			ctl_reg = UB960_RR_BC_GPIO_CTL(i / 2);
3069 			ctl_shift = (i % 2) * 4;
3070 
3071 			ub960_rxport_read(priv, nport, ctl_reg, &v);
3072 
3073 			dev_info(dev, "\tGPIO%u: mode %u\n", i,
3074 				 (v >> ctl_shift) & 0xf);
3075 		}
3076 	}
3077 
3078 	v4l2_subdev_unlock_state(state);
3079 
3080 	return 0;
3081 }
3082 
3083 static const struct v4l2_subdev_core_ops ub960_subdev_core_ops = {
3084 	.log_status = ub960_log_status,
3085 	.subscribe_event = v4l2_ctrl_subdev_subscribe_event,
3086 	.unsubscribe_event = v4l2_event_subdev_unsubscribe,
3087 };
3088 
3089 static const struct v4l2_subdev_ops ub960_subdev_ops = {
3090 	.core = &ub960_subdev_core_ops,
3091 	.pad = &ub960_pad_ops,
3092 };
3093 
3094 static const struct media_entity_operations ub960_entity_ops = {
3095 	.get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
3096 	.link_validate = v4l2_subdev_link_validate,
3097 	.has_pad_interdep = v4l2_subdev_has_pad_interdep,
3098 };
3099 
3100 /* -----------------------------------------------------------------------------
3101  * Core
3102  */
3103 
3104 static irqreturn_t ub960_handle_events(int irq, void *arg)
3105 {
3106 	struct ub960_data *priv = arg;
3107 	unsigned int i;
3108 	u8 int_sts;
3109 	u8 fwd_sts;
3110 	int ret;
3111 
3112 	ret = ub960_read(priv, UB960_SR_INTERRUPT_STS, &int_sts);
3113 	if (ret || !int_sts)
3114 		return IRQ_NONE;
3115 
3116 	dev_dbg(&priv->client->dev, "INTERRUPT_STS %x\n", int_sts);
3117 
3118 	ret = ub960_read(priv, UB960_SR_FWD_STS, &fwd_sts);
3119 	if (ret)
3120 		return IRQ_NONE;
3121 
3122 	dev_dbg(&priv->client->dev, "FWD_STS %#02x\n", fwd_sts);
3123 
3124 	for (i = 0; i < priv->hw_data->num_txports; i++) {
3125 		if (int_sts & UB960_SR_INTERRUPT_STS_IS_CSI_TX(i))
3126 			ub960_csi_handle_events(priv, i);
3127 	}
3128 
3129 	for (i = 0; i < priv->hw_data->num_rxports; i++) {
3130 		if (!priv->rxports[i])
3131 			continue;
3132 
3133 		if (int_sts & UB960_SR_INTERRUPT_STS_IS_RX(i))
3134 			ub960_rxport_handle_events(priv, i);
3135 	}
3136 
3137 	return IRQ_HANDLED;
3138 }
3139 
3140 static void ub960_handler_work(struct work_struct *work)
3141 {
3142 	struct delayed_work *dwork = to_delayed_work(work);
3143 	struct ub960_data *priv =
3144 		container_of(dwork, struct ub960_data, poll_work);
3145 
3146 	ub960_handle_events(0, priv);
3147 
3148 	schedule_delayed_work(&priv->poll_work,
3149 			      msecs_to_jiffies(UB960_POLL_TIME_MS));
3150 }
3151 
3152 static void ub960_txport_free_ports(struct ub960_data *priv)
3153 {
3154 	unsigned int nport;
3155 
3156 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3157 		struct ub960_txport *txport = priv->txports[nport];
3158 
3159 		if (!txport)
3160 			continue;
3161 
3162 		kfree(txport);
3163 		priv->txports[nport] = NULL;
3164 	}
3165 }
3166 
3167 static void ub960_rxport_free_ports(struct ub960_data *priv)
3168 {
3169 	unsigned int nport;
3170 
3171 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3172 		struct ub960_rxport *rxport = priv->rxports[nport];
3173 
3174 		if (!rxport)
3175 			continue;
3176 
3177 		fwnode_handle_put(rxport->source.ep_fwnode);
3178 		fwnode_handle_put(rxport->ser.fwnode);
3179 
3180 		kfree(rxport);
3181 		priv->rxports[nport] = NULL;
3182 	}
3183 }
3184 
3185 static int
3186 ub960_parse_dt_rxport_link_properties(struct ub960_data *priv,
3187 				      struct fwnode_handle *link_fwnode,
3188 				      struct ub960_rxport *rxport)
3189 {
3190 	struct device *dev = &priv->client->dev;
3191 	unsigned int nport = rxport->nport;
3192 	u32 rx_mode;
3193 	u32 cdr_mode;
3194 	s32 strobe_pos;
3195 	u32 eq_level;
3196 	u32 ser_i2c_alias;
3197 	int ret;
3198 
3199 	cdr_mode = RXPORT_CDR_FPD3;
3200 
3201 	ret = fwnode_property_read_u32(link_fwnode, "ti,cdr-mode", &cdr_mode);
3202 	if (ret < 0 && ret != -EINVAL) {
3203 		dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3204 			"ti,cdr-mode", ret);
3205 		return ret;
3206 	}
3207 
3208 	if (cdr_mode > RXPORT_CDR_LAST) {
3209 		dev_err(dev, "rx%u: bad 'ti,cdr-mode' %u\n", nport, cdr_mode);
3210 		return -EINVAL;
3211 	}
3212 
3213 	if (!priv->hw_data->is_fpdlink4 && cdr_mode == RXPORT_CDR_FPD4) {
3214 		dev_err(dev, "rx%u: FPD-Link 4 CDR not supported\n", nport);
3215 		return -EINVAL;
3216 	}
3217 
3218 	rxport->cdr_mode = cdr_mode;
3219 
3220 	ret = fwnode_property_read_u32(link_fwnode, "ti,rx-mode", &rx_mode);
3221 	if (ret < 0) {
3222 		dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3223 			"ti,rx-mode", ret);
3224 		return ret;
3225 	}
3226 
3227 	if (rx_mode > RXPORT_MODE_LAST) {
3228 		dev_err(dev, "rx%u: bad 'ti,rx-mode' %u\n", nport, rx_mode);
3229 		return -EINVAL;
3230 	}
3231 
3232 	switch (rx_mode) {
3233 	case RXPORT_MODE_RAW12_HF:
3234 	case RXPORT_MODE_RAW12_LF:
3235 	case RXPORT_MODE_CSI2_ASYNC:
3236 		dev_err(dev, "rx%u: unsupported 'ti,rx-mode' %u\n", nport,
3237 			rx_mode);
3238 		return -EINVAL;
3239 	default:
3240 		break;
3241 	}
3242 
3243 	rxport->rx_mode = rx_mode;
3244 
3245 	/* EQ & Strobe related */
3246 
3247 	/* Defaults */
3248 	rxport->eq.manual_eq = false;
3249 	rxport->eq.aeq.eq_level_min = UB960_MIN_EQ_LEVEL;
3250 	rxport->eq.aeq.eq_level_max = UB960_MAX_EQ_LEVEL;
3251 
3252 	ret = fwnode_property_read_u32(link_fwnode, "ti,strobe-pos",
3253 				       &strobe_pos);
3254 	if (ret) {
3255 		if (ret != -EINVAL) {
3256 			dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3257 				"ti,strobe-pos", ret);
3258 			return ret;
3259 		}
3260 	} else {
3261 		if (strobe_pos < UB960_MIN_MANUAL_STROBE_POS ||
3262 		    strobe_pos > UB960_MAX_MANUAL_STROBE_POS) {
3263 			dev_err(dev, "rx%u: illegal 'strobe-pos' value: %d\n",
3264 				nport, strobe_pos);
3265 			return -EINVAL;
3266 		}
3267 
3268 		/* NOTE: ignored unless global manual strobe pos is also set */
3269 		rxport->eq.strobe_pos = strobe_pos;
3270 		if (!priv->strobe.manual)
3271 			dev_warn(dev,
3272 				 "rx%u: 'ti,strobe-pos' ignored as 'ti,manual-strobe' not set\n",
3273 				 nport);
3274 	}
3275 
3276 	ret = fwnode_property_read_u32(link_fwnode, "ti,eq-level", &eq_level);
3277 	if (ret) {
3278 		if (ret != -EINVAL) {
3279 			dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3280 				"ti,eq-level", ret);
3281 			return ret;
3282 		}
3283 	} else {
3284 		if (eq_level > UB960_MAX_EQ_LEVEL) {
3285 			dev_err(dev, "rx%u: illegal 'ti,eq-level' value: %d\n",
3286 				nport, eq_level);
3287 			return -EINVAL;
3288 		}
3289 
3290 		rxport->eq.manual_eq = true;
3291 		rxport->eq.manual.eq_level = eq_level;
3292 	}
3293 
3294 	ret = fwnode_property_read_u32(link_fwnode, "i2c-alias",
3295 				       &ser_i2c_alias);
3296 	if (ret) {
3297 		dev_err(dev, "rx%u: failed to read '%s': %d\n", nport,
3298 			"i2c-alias", ret);
3299 		return ret;
3300 	}
3301 	rxport->ser.alias = ser_i2c_alias;
3302 
3303 	rxport->ser.fwnode = fwnode_get_named_child_node(link_fwnode, "serializer");
3304 	if (!rxport->ser.fwnode) {
3305 		dev_err(dev, "rx%u: missing 'serializer' node\n", nport);
3306 		return -EINVAL;
3307 	}
3308 
3309 	return 0;
3310 }
3311 
3312 static int ub960_parse_dt_rxport_ep_properties(struct ub960_data *priv,
3313 					       struct fwnode_handle *ep_fwnode,
3314 					       struct ub960_rxport *rxport)
3315 {
3316 	struct device *dev = &priv->client->dev;
3317 	struct v4l2_fwnode_endpoint vep = {};
3318 	unsigned int nport = rxport->nport;
3319 	bool hsync_hi;
3320 	bool vsync_hi;
3321 	int ret;
3322 
3323 	rxport->source.ep_fwnode = fwnode_graph_get_remote_endpoint(ep_fwnode);
3324 	if (!rxport->source.ep_fwnode) {
3325 		dev_err(dev, "rx%u: no remote endpoint\n", nport);
3326 		return -ENODEV;
3327 	}
3328 
3329 	/* We currently have properties only for RAW modes */
3330 
3331 	switch (rxport->rx_mode) {
3332 	case RXPORT_MODE_RAW10:
3333 	case RXPORT_MODE_RAW12_HF:
3334 	case RXPORT_MODE_RAW12_LF:
3335 		break;
3336 	default:
3337 		return 0;
3338 	}
3339 
3340 	vep.bus_type = V4L2_MBUS_PARALLEL;
3341 	ret = v4l2_fwnode_endpoint_parse(ep_fwnode, &vep);
3342 	if (ret) {
3343 		dev_err(dev, "rx%u: failed to parse endpoint data\n", nport);
3344 		goto err_put_source_ep_fwnode;
3345 	}
3346 
3347 	hsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH);
3348 	vsync_hi = !!(vep.bus.parallel.flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH);
3349 
3350 	/* LineValid and FrameValid are inverse to the h/vsync active */
3351 	rxport->lv_fv_pol = (hsync_hi ? UB960_RR_PORT_CONFIG2_LV_POL_LOW : 0) |
3352 			    (vsync_hi ? UB960_RR_PORT_CONFIG2_FV_POL_LOW : 0);
3353 
3354 	return 0;
3355 
3356 err_put_source_ep_fwnode:
3357 	fwnode_handle_put(rxport->source.ep_fwnode);
3358 	return ret;
3359 }
3360 
3361 static int ub960_parse_dt_rxport(struct ub960_data *priv, unsigned int nport,
3362 				 struct fwnode_handle *link_fwnode,
3363 				 struct fwnode_handle *ep_fwnode)
3364 {
3365 	static const char *vpoc_names[UB960_MAX_RX_NPORTS] = {
3366 		"vpoc0", "vpoc1", "vpoc2", "vpoc3"
3367 	};
3368 	struct device *dev = &priv->client->dev;
3369 	struct ub960_rxport *rxport;
3370 	int ret;
3371 
3372 	rxport = kzalloc(sizeof(*rxport), GFP_KERNEL);
3373 	if (!rxport)
3374 		return -ENOMEM;
3375 
3376 	priv->rxports[nport] = rxport;
3377 
3378 	rxport->nport = nport;
3379 	rxport->priv = priv;
3380 
3381 	ret = ub960_parse_dt_rxport_link_properties(priv, link_fwnode, rxport);
3382 	if (ret)
3383 		goto err_free_rxport;
3384 
3385 	rxport->vpoc = devm_regulator_get_optional(dev, vpoc_names[nport]);
3386 	if (IS_ERR(rxport->vpoc)) {
3387 		ret = PTR_ERR(rxport->vpoc);
3388 		if (ret == -ENODEV) {
3389 			rxport->vpoc = NULL;
3390 		} else {
3391 			dev_err(dev, "rx%u: failed to get VPOC supply: %d\n",
3392 				nport, ret);
3393 			goto err_put_remote_fwnode;
3394 		}
3395 	}
3396 
3397 	ret = ub960_parse_dt_rxport_ep_properties(priv, ep_fwnode, rxport);
3398 	if (ret)
3399 		goto err_put_remote_fwnode;
3400 
3401 	return 0;
3402 
3403 err_put_remote_fwnode:
3404 	fwnode_handle_put(rxport->ser.fwnode);
3405 err_free_rxport:
3406 	priv->rxports[nport] = NULL;
3407 	kfree(rxport);
3408 	return ret;
3409 }
3410 
3411 static struct fwnode_handle *
3412 ub960_fwnode_get_link_by_regs(struct fwnode_handle *links_fwnode,
3413 			      unsigned int nport)
3414 {
3415 	struct fwnode_handle *link_fwnode;
3416 	int ret;
3417 
3418 	fwnode_for_each_child_node(links_fwnode, link_fwnode) {
3419 		u32 link_num;
3420 
3421 		if (!str_has_prefix(fwnode_get_name(link_fwnode), "link@"))
3422 			continue;
3423 
3424 		ret = fwnode_property_read_u32(link_fwnode, "reg", &link_num);
3425 		if (ret) {
3426 			fwnode_handle_put(link_fwnode);
3427 			return NULL;
3428 		}
3429 
3430 		if (nport == link_num)
3431 			return link_fwnode;
3432 	}
3433 
3434 	return NULL;
3435 }
3436 
3437 static int ub960_parse_dt_rxports(struct ub960_data *priv)
3438 {
3439 	struct device *dev = &priv->client->dev;
3440 	struct fwnode_handle *links_fwnode;
3441 	unsigned int nport;
3442 	int ret;
3443 
3444 	links_fwnode = fwnode_get_named_child_node(dev_fwnode(dev), "links");
3445 	if (!links_fwnode) {
3446 		dev_err(dev, "'links' node missing\n");
3447 		return -ENODEV;
3448 	}
3449 
3450 	/* Defaults, recommended by TI */
3451 	priv->strobe.min = 2;
3452 	priv->strobe.max = 3;
3453 
3454 	priv->strobe.manual = fwnode_property_read_bool(links_fwnode, "ti,manual-strobe");
3455 
3456 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3457 		struct fwnode_handle *link_fwnode;
3458 		struct fwnode_handle *ep_fwnode;
3459 
3460 		link_fwnode = ub960_fwnode_get_link_by_regs(links_fwnode, nport);
3461 		if (!link_fwnode)
3462 			continue;
3463 
3464 		ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3465 							    nport, 0, 0);
3466 		if (!ep_fwnode) {
3467 			fwnode_handle_put(link_fwnode);
3468 			continue;
3469 		}
3470 
3471 		ret = ub960_parse_dt_rxport(priv, nport, link_fwnode,
3472 					    ep_fwnode);
3473 
3474 		fwnode_handle_put(link_fwnode);
3475 		fwnode_handle_put(ep_fwnode);
3476 
3477 		if (ret) {
3478 			dev_err(dev, "rx%u: failed to parse RX port\n", nport);
3479 			goto err_put_links;
3480 		}
3481 	}
3482 
3483 	fwnode_handle_put(links_fwnode);
3484 
3485 	return 0;
3486 
3487 err_put_links:
3488 	fwnode_handle_put(links_fwnode);
3489 
3490 	return ret;
3491 }
3492 
3493 static int ub960_parse_dt_txports(struct ub960_data *priv)
3494 {
3495 	struct device *dev = &priv->client->dev;
3496 	u32 nport;
3497 	int ret;
3498 
3499 	for (nport = 0; nport < priv->hw_data->num_txports; nport++) {
3500 		unsigned int port = nport + priv->hw_data->num_rxports;
3501 		struct fwnode_handle *ep_fwnode;
3502 
3503 		ep_fwnode = fwnode_graph_get_endpoint_by_id(dev_fwnode(dev),
3504 							    port, 0, 0);
3505 		if (!ep_fwnode)
3506 			continue;
3507 
3508 		ret = ub960_parse_dt_txport(priv, ep_fwnode, nport);
3509 
3510 		fwnode_handle_put(ep_fwnode);
3511 
3512 		if (ret)
3513 			break;
3514 	}
3515 
3516 	return 0;
3517 }
3518 
3519 static int ub960_parse_dt(struct ub960_data *priv)
3520 {
3521 	int ret;
3522 
3523 	ret = ub960_parse_dt_rxports(priv);
3524 	if (ret)
3525 		return ret;
3526 
3527 	ret = ub960_parse_dt_txports(priv);
3528 	if (ret)
3529 		goto err_free_rxports;
3530 
3531 	return 0;
3532 
3533 err_free_rxports:
3534 	ub960_rxport_free_ports(priv);
3535 
3536 	return ret;
3537 }
3538 
3539 static int ub960_notify_bound(struct v4l2_async_notifier *notifier,
3540 			      struct v4l2_subdev *subdev,
3541 			      struct v4l2_async_subdev *asd)
3542 {
3543 	struct ub960_data *priv = sd_to_ub960(notifier->sd);
3544 	struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3545 	struct device *dev = &priv->client->dev;
3546 	u8 nport = rxport->nport;
3547 	unsigned int i;
3548 	int ret;
3549 
3550 	ret = media_entity_get_fwnode_pad(&subdev->entity,
3551 					  rxport->source.ep_fwnode,
3552 					  MEDIA_PAD_FL_SOURCE);
3553 	if (ret < 0) {
3554 		dev_err(dev, "Failed to find pad for %s\n", subdev->name);
3555 		return ret;
3556 	}
3557 
3558 	rxport->source.sd = subdev;
3559 	rxport->source.pad = ret;
3560 
3561 	ret = media_create_pad_link(&rxport->source.sd->entity,
3562 				    rxport->source.pad, &priv->sd.entity, nport,
3563 				    MEDIA_LNK_FL_ENABLED |
3564 					    MEDIA_LNK_FL_IMMUTABLE);
3565 	if (ret) {
3566 		dev_err(dev, "Unable to link %s:%u -> %s:%u\n",
3567 			rxport->source.sd->name, rxport->source.pad,
3568 			priv->sd.name, nport);
3569 		return ret;
3570 	}
3571 
3572 	for (i = 0; i < priv->hw_data->num_rxports; i++) {
3573 		if (priv->rxports[i] && !priv->rxports[i]->source.sd) {
3574 			dev_dbg(dev, "Waiting for more subdevs to be bound\n");
3575 			return 0;
3576 		}
3577 	}
3578 
3579 	return 0;
3580 }
3581 
3582 static void ub960_notify_unbind(struct v4l2_async_notifier *notifier,
3583 				struct v4l2_subdev *subdev,
3584 				struct v4l2_async_subdev *asd)
3585 {
3586 	struct ub960_rxport *rxport = to_ub960_asd(asd)->rxport;
3587 
3588 	rxport->source.sd = NULL;
3589 }
3590 
3591 static const struct v4l2_async_notifier_operations ub960_notify_ops = {
3592 	.bound = ub960_notify_bound,
3593 	.unbind = ub960_notify_unbind,
3594 };
3595 
3596 static int ub960_v4l2_notifier_register(struct ub960_data *priv)
3597 {
3598 	struct device *dev = &priv->client->dev;
3599 	unsigned int i;
3600 	int ret;
3601 
3602 	v4l2_async_nf_init(&priv->notifier);
3603 
3604 	for (i = 0; i < priv->hw_data->num_rxports; i++) {
3605 		struct ub960_rxport *rxport = priv->rxports[i];
3606 		struct ub960_asd *asd;
3607 
3608 		if (!rxport)
3609 			continue;
3610 
3611 		asd = v4l2_async_nf_add_fwnode(&priv->notifier,
3612 					       rxport->source.ep_fwnode,
3613 					       struct ub960_asd);
3614 		if (IS_ERR(asd)) {
3615 			dev_err(dev, "Failed to add subdev for source %u: %pe",
3616 				i, asd);
3617 			v4l2_async_nf_cleanup(&priv->notifier);
3618 			return PTR_ERR(asd);
3619 		}
3620 
3621 		asd->rxport = rxport;
3622 	}
3623 
3624 	priv->notifier.ops = &ub960_notify_ops;
3625 
3626 	ret = v4l2_async_subdev_nf_register(&priv->sd, &priv->notifier);
3627 	if (ret) {
3628 		dev_err(dev, "Failed to register subdev_notifier");
3629 		v4l2_async_nf_cleanup(&priv->notifier);
3630 		return ret;
3631 	}
3632 
3633 	return 0;
3634 }
3635 
3636 static void ub960_v4l2_notifier_unregister(struct ub960_data *priv)
3637 {
3638 	v4l2_async_nf_unregister(&priv->notifier);
3639 	v4l2_async_nf_cleanup(&priv->notifier);
3640 }
3641 
3642 static int ub960_create_subdev(struct ub960_data *priv)
3643 {
3644 	struct device *dev = &priv->client->dev;
3645 	unsigned int i;
3646 	int ret;
3647 
3648 	v4l2_i2c_subdev_init(&priv->sd, priv->client, &ub960_subdev_ops);
3649 
3650 	v4l2_ctrl_handler_init(&priv->ctrl_handler, 1);
3651 	priv->sd.ctrl_handler = &priv->ctrl_handler;
3652 
3653 	v4l2_ctrl_new_int_menu(&priv->ctrl_handler, NULL, V4L2_CID_LINK_FREQ,
3654 			       ARRAY_SIZE(priv->tx_link_freq) - 1, 0,
3655 			       priv->tx_link_freq);
3656 
3657 	if (priv->ctrl_handler.error) {
3658 		ret = priv->ctrl_handler.error;
3659 		goto err_free_ctrl;
3660 	}
3661 
3662 	priv->sd.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE |
3663 			  V4L2_SUBDEV_FL_HAS_EVENTS | V4L2_SUBDEV_FL_STREAMS;
3664 	priv->sd.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
3665 	priv->sd.entity.ops = &ub960_entity_ops;
3666 
3667 	for (i = 0; i < priv->hw_data->num_rxports + priv->hw_data->num_txports; i++) {
3668 		priv->pads[i].flags = ub960_pad_is_sink(priv, i) ?
3669 					      MEDIA_PAD_FL_SINK :
3670 					      MEDIA_PAD_FL_SOURCE;
3671 	}
3672 
3673 	ret = media_entity_pads_init(&priv->sd.entity,
3674 				     priv->hw_data->num_rxports +
3675 					     priv->hw_data->num_txports,
3676 				     priv->pads);
3677 	if (ret)
3678 		goto err_free_ctrl;
3679 
3680 	priv->sd.state_lock = priv->sd.ctrl_handler->lock;
3681 
3682 	ret = v4l2_subdev_init_finalize(&priv->sd);
3683 	if (ret)
3684 		goto err_entity_cleanup;
3685 
3686 	ret = ub960_v4l2_notifier_register(priv);
3687 	if (ret) {
3688 		dev_err(dev, "v4l2 subdev notifier register failed: %d\n", ret);
3689 		goto err_subdev_cleanup;
3690 	}
3691 
3692 	ret = v4l2_async_register_subdev(&priv->sd);
3693 	if (ret) {
3694 		dev_err(dev, "v4l2_async_register_subdev error: %d\n", ret);
3695 		goto err_unreg_notif;
3696 	}
3697 
3698 	return 0;
3699 
3700 err_unreg_notif:
3701 	ub960_v4l2_notifier_unregister(priv);
3702 err_subdev_cleanup:
3703 	v4l2_subdev_cleanup(&priv->sd);
3704 err_entity_cleanup:
3705 	media_entity_cleanup(&priv->sd.entity);
3706 err_free_ctrl:
3707 	v4l2_ctrl_handler_free(&priv->ctrl_handler);
3708 
3709 	return ret;
3710 }
3711 
3712 static void ub960_destroy_subdev(struct ub960_data *priv)
3713 {
3714 	ub960_v4l2_notifier_unregister(priv);
3715 	v4l2_async_unregister_subdev(&priv->sd);
3716 
3717 	v4l2_subdev_cleanup(&priv->sd);
3718 
3719 	media_entity_cleanup(&priv->sd.entity);
3720 	v4l2_ctrl_handler_free(&priv->ctrl_handler);
3721 }
3722 
3723 static const struct regmap_config ub960_regmap_config = {
3724 	.name = "ds90ub960",
3725 
3726 	.reg_bits = 8,
3727 	.val_bits = 8,
3728 
3729 	.max_register = 0xff,
3730 
3731 	/*
3732 	 * We do locking in the driver to cover the TX/RX port selection and the
3733 	 * indirect register access.
3734 	 */
3735 	.disable_locking = true,
3736 };
3737 
3738 static void ub960_reset(struct ub960_data *priv, bool reset_regs)
3739 {
3740 	struct device *dev = &priv->client->dev;
3741 	unsigned int v;
3742 	int ret;
3743 	u8 bit;
3744 
3745 	bit = reset_regs ? UB960_SR_RESET_DIGITAL_RESET1 :
3746 			   UB960_SR_RESET_DIGITAL_RESET0;
3747 
3748 	ub960_write(priv, UB960_SR_RESET, bit);
3749 
3750 	mutex_lock(&priv->reg_lock);
3751 
3752 	ret = regmap_read_poll_timeout(priv->regmap, UB960_SR_RESET, v,
3753 				       (v & bit) == 0, 2000, 100000);
3754 
3755 	mutex_unlock(&priv->reg_lock);
3756 
3757 	if (ret)
3758 		dev_err(dev, "reset failed: %d\n", ret);
3759 }
3760 
3761 static int ub960_get_hw_resources(struct ub960_data *priv)
3762 {
3763 	struct device *dev = &priv->client->dev;
3764 
3765 	priv->regmap = devm_regmap_init_i2c(priv->client, &ub960_regmap_config);
3766 	if (IS_ERR(priv->regmap))
3767 		return PTR_ERR(priv->regmap);
3768 
3769 	priv->vddio = devm_regulator_get(dev, "vddio");
3770 	if (IS_ERR(priv->vddio))
3771 		return dev_err_probe(dev, PTR_ERR(priv->vddio),
3772 				     "cannot get VDDIO regulator\n");
3773 
3774 	/* get power-down pin from DT */
3775 	priv->pd_gpio =
3776 		devm_gpiod_get_optional(dev, "powerdown", GPIOD_OUT_HIGH);
3777 	if (IS_ERR(priv->pd_gpio))
3778 		return dev_err_probe(dev, PTR_ERR(priv->pd_gpio),
3779 				     "Cannot get powerdown GPIO\n");
3780 
3781 	priv->refclk = devm_clk_get(dev, "refclk");
3782 	if (IS_ERR(priv->refclk))
3783 		return dev_err_probe(dev, PTR_ERR(priv->refclk),
3784 				     "Cannot get REFCLK\n");
3785 
3786 	return 0;
3787 }
3788 
3789 static int ub960_enable_core_hw(struct ub960_data *priv)
3790 {
3791 	struct device *dev = &priv->client->dev;
3792 	u8 rev_mask;
3793 	int ret;
3794 	u8 dev_sts;
3795 	u8 refclk_freq;
3796 
3797 	ret = regulator_enable(priv->vddio);
3798 	if (ret)
3799 		return dev_err_probe(dev, ret,
3800 				     "failed to enable VDDIO regulator\n");
3801 
3802 	ret = clk_prepare_enable(priv->refclk);
3803 	if (ret) {
3804 		dev_err_probe(dev, ret, "Failed to enable refclk\n");
3805 		goto err_disable_vddio;
3806 	}
3807 
3808 	if (priv->pd_gpio) {
3809 		gpiod_set_value_cansleep(priv->pd_gpio, 1);
3810 		/* wait min 2 ms for reset to complete */
3811 		fsleep(2000);
3812 		gpiod_set_value_cansleep(priv->pd_gpio, 0);
3813 		/* wait min 2 ms for power up to finish */
3814 		fsleep(2000);
3815 	}
3816 
3817 	ub960_reset(priv, true);
3818 
3819 	/* Runtime check register accessibility */
3820 	ret = ub960_read(priv, UB960_SR_REV_MASK, &rev_mask);
3821 	if (ret) {
3822 		dev_err_probe(dev, ret, "Cannot read first register, abort\n");
3823 		goto err_pd_gpio;
3824 	}
3825 
3826 	dev_dbg(dev, "Found %s (rev/mask %#04x)\n", priv->hw_data->model,
3827 		rev_mask);
3828 
3829 	ret = ub960_read(priv, UB960_SR_DEVICE_STS, &dev_sts);
3830 	if (ret)
3831 		goto err_pd_gpio;
3832 
3833 	ret = ub960_read(priv, UB960_XR_REFCLK_FREQ, &refclk_freq);
3834 	if (ret)
3835 		goto err_pd_gpio;
3836 
3837 	dev_dbg(dev, "refclk valid %u freq %u MHz (clk fw freq %lu MHz)\n",
3838 		!!(dev_sts & BIT(4)), refclk_freq,
3839 		clk_get_rate(priv->refclk) / 1000000);
3840 
3841 	/* Disable all RX ports by default */
3842 	ret = ub960_write(priv, UB960_SR_RX_PORT_CTL, 0);
3843 	if (ret)
3844 		goto err_pd_gpio;
3845 
3846 	/* release GPIO lock */
3847 	if (priv->hw_data->is_ub9702) {
3848 		ret = ub960_update_bits(priv, UB960_SR_RESET,
3849 					UB960_SR_RESET_GPIO_LOCK_RELEASE,
3850 					UB960_SR_RESET_GPIO_LOCK_RELEASE);
3851 		if (ret)
3852 			goto err_pd_gpio;
3853 	}
3854 
3855 	return 0;
3856 
3857 err_pd_gpio:
3858 	gpiod_set_value_cansleep(priv->pd_gpio, 1);
3859 	clk_disable_unprepare(priv->refclk);
3860 err_disable_vddio:
3861 	regulator_disable(priv->vddio);
3862 
3863 	return ret;
3864 }
3865 
3866 static void ub960_disable_core_hw(struct ub960_data *priv)
3867 {
3868 	gpiod_set_value_cansleep(priv->pd_gpio, 1);
3869 	clk_disable_unprepare(priv->refclk);
3870 	regulator_disable(priv->vddio);
3871 }
3872 
3873 static int ub960_probe(struct i2c_client *client)
3874 {
3875 	struct device *dev = &client->dev;
3876 	struct ub960_data *priv;
3877 	unsigned int port_lock_mask;
3878 	unsigned int port_mask;
3879 	unsigned int nport;
3880 	int ret;
3881 
3882 	priv = devm_kzalloc(dev, sizeof(*priv), GFP_KERNEL);
3883 	if (!priv)
3884 		return -ENOMEM;
3885 
3886 	priv->client = client;
3887 
3888 	priv->hw_data = device_get_match_data(dev);
3889 
3890 	mutex_init(&priv->reg_lock);
3891 
3892 	INIT_DELAYED_WORK(&priv->poll_work, ub960_handler_work);
3893 
3894 	/*
3895 	 * Initialize these to invalid values so that the first reg writes will
3896 	 * configure the target.
3897 	 */
3898 	priv->reg_current.indirect_target = 0xff;
3899 	priv->reg_current.rxport = 0xff;
3900 	priv->reg_current.txport = 0xff;
3901 
3902 	ret = ub960_get_hw_resources(priv);
3903 	if (ret)
3904 		goto err_mutex_destroy;
3905 
3906 	ret = ub960_enable_core_hw(priv);
3907 	if (ret)
3908 		goto err_mutex_destroy;
3909 
3910 	ret = ub960_parse_dt(priv);
3911 	if (ret)
3912 		goto err_disable_core_hw;
3913 
3914 	ret = ub960_init_tx_ports(priv);
3915 	if (ret)
3916 		goto err_free_ports;
3917 
3918 	ret = ub960_rxport_enable_vpocs(priv);
3919 	if (ret)
3920 		goto err_free_ports;
3921 
3922 	ret = ub960_init_rx_ports(priv);
3923 	if (ret)
3924 		goto err_disable_vpocs;
3925 
3926 	ub960_reset(priv, false);
3927 
3928 	port_mask = 0;
3929 
3930 	for (nport = 0; nport < priv->hw_data->num_rxports; nport++) {
3931 		struct ub960_rxport *rxport = priv->rxports[nport];
3932 
3933 		if (!rxport)
3934 			continue;
3935 
3936 		port_mask |= BIT(nport);
3937 	}
3938 
3939 	ret = ub960_rxport_wait_locks(priv, port_mask, &port_lock_mask);
3940 	if (ret)
3941 		goto err_disable_vpocs;
3942 
3943 	if (port_mask != port_lock_mask) {
3944 		ret = -EIO;
3945 		dev_err_probe(dev, ret, "Failed to lock all RX ports\n");
3946 		goto err_disable_vpocs;
3947 	}
3948 
3949 	/*
3950 	 * Clear any errors caused by switching the RX port settings while
3951 	 * probing.
3952 	 */
3953 	ub960_clear_rx_errors(priv);
3954 
3955 	ret = ub960_init_atr(priv);
3956 	if (ret)
3957 		goto err_disable_vpocs;
3958 
3959 	ret = ub960_rxport_add_serializers(priv);
3960 	if (ret)
3961 		goto err_uninit_atr;
3962 
3963 	ret = ub960_create_subdev(priv);
3964 	if (ret)
3965 		goto err_free_sers;
3966 
3967 	if (client->irq)
3968 		dev_warn(dev, "irq support not implemented, using polling\n");
3969 
3970 	schedule_delayed_work(&priv->poll_work,
3971 			      msecs_to_jiffies(UB960_POLL_TIME_MS));
3972 
3973 	return 0;
3974 
3975 err_free_sers:
3976 	ub960_rxport_remove_serializers(priv);
3977 err_uninit_atr:
3978 	ub960_uninit_atr(priv);
3979 err_disable_vpocs:
3980 	ub960_rxport_disable_vpocs(priv);
3981 err_free_ports:
3982 	ub960_rxport_free_ports(priv);
3983 	ub960_txport_free_ports(priv);
3984 err_disable_core_hw:
3985 	ub960_disable_core_hw(priv);
3986 err_mutex_destroy:
3987 	mutex_destroy(&priv->reg_lock);
3988 	return ret;
3989 }
3990 
3991 static void ub960_remove(struct i2c_client *client)
3992 {
3993 	struct v4l2_subdev *sd = i2c_get_clientdata(client);
3994 	struct ub960_data *priv = sd_to_ub960(sd);
3995 
3996 	cancel_delayed_work_sync(&priv->poll_work);
3997 
3998 	ub960_destroy_subdev(priv);
3999 	ub960_rxport_remove_serializers(priv);
4000 	ub960_uninit_atr(priv);
4001 	ub960_rxport_disable_vpocs(priv);
4002 	ub960_rxport_free_ports(priv);
4003 	ub960_txport_free_ports(priv);
4004 	ub960_disable_core_hw(priv);
4005 	mutex_destroy(&priv->reg_lock);
4006 }
4007 
4008 static const struct ub960_hw_data ds90ub960_hw = {
4009 	.model = "ub960",
4010 	.num_rxports = 4,
4011 	.num_txports = 2,
4012 };
4013 
4014 static const struct ub960_hw_data ds90ub9702_hw = {
4015 	.model = "ub9702",
4016 	.num_rxports = 4,
4017 	.num_txports = 2,
4018 	.is_ub9702 = true,
4019 	.is_fpdlink4 = true,
4020 };
4021 
4022 static const struct i2c_device_id ub960_id[] = {
4023 	{ "ds90ub960-q1", (kernel_ulong_t)&ds90ub960_hw },
4024 	{ "ds90ub9702-q1", (kernel_ulong_t)&ds90ub9702_hw },
4025 	{}
4026 };
4027 MODULE_DEVICE_TABLE(i2c, ub960_id);
4028 
4029 static const struct of_device_id ub960_dt_ids[] = {
4030 	{ .compatible = "ti,ds90ub960-q1", .data = &ds90ub960_hw },
4031 	{ .compatible = "ti,ds90ub9702-q1", .data = &ds90ub9702_hw },
4032 	{}
4033 };
4034 MODULE_DEVICE_TABLE(of, ub960_dt_ids);
4035 
4036 static struct i2c_driver ds90ub960_driver = {
4037 	.probe		= ub960_probe,
4038 	.remove		= ub960_remove,
4039 	.id_table	= ub960_id,
4040 	.driver = {
4041 		.name	= "ds90ub960",
4042 		.of_match_table = ub960_dt_ids,
4043 	},
4044 };
4045 module_i2c_driver(ds90ub960_driver);
4046 
4047 MODULE_LICENSE("GPL");
4048 MODULE_DESCRIPTION("Texas Instruments FPD-Link III/IV Deserializers Driver");
4049 MODULE_AUTHOR("Luca Ceresoli <luca@lucaceresoli.net>");
4050 MODULE_AUTHOR("Tomi Valkeinen <tomi.valkeinen@ideasonboard.com>");
4051 MODULE_IMPORT_NS(I2C_ATR);
4052