1 // SPDX-License-Identifier: GPL-2.0
2 #include <media/drv-intf/saa7146_vv.h>
3 
4 static int vbi_pixel_to_capture = 720 * 2;
5 
6 static int vbi_workaround(struct saa7146_dev *dev)
7 {
8 	struct saa7146_vv *vv = dev->vv_data;
9 
10 	u32          *cpu;
11 	dma_addr_t   dma_addr;
12 
13 	int count = 0;
14 	int i;
15 
16 	DECLARE_WAITQUEUE(wait, current);
17 
18 	DEB_VBI("dev:%p\n", dev);
19 
20 	/* once again, a bug in the saa7146: the brs acquisition
21 	   is buggy and especially the BXO-counter does not work
22 	   as specified. there is this workaround, but please
23 	   don't let me explain it. ;-) */
24 
25 	cpu = dma_alloc_coherent(&dev->pci->dev, 4096, &dma_addr, GFP_KERNEL);
26 	if (NULL == cpu)
27 		return -ENOMEM;
28 
29 	/* setup some basic programming, just for the workaround */
30 	saa7146_write(dev, BASE_EVEN3,	dma_addr);
31 	saa7146_write(dev, BASE_ODD3,	dma_addr+vbi_pixel_to_capture);
32 	saa7146_write(dev, PROT_ADDR3,	dma_addr+4096);
33 	saa7146_write(dev, PITCH3,	vbi_pixel_to_capture);
34 	saa7146_write(dev, BASE_PAGE3,	0x0);
35 	saa7146_write(dev, NUM_LINE_BYTE3, (2<<16)|((vbi_pixel_to_capture)<<0));
36 	saa7146_write(dev, MC2, MASK_04|MASK_20);
37 
38 	/* load brs-control register */
39 	WRITE_RPS1(CMD_WR_REG | (1 << 8) | (BRS_CTRL/4));
40 	/* BXO = 1h, BRS to outbound */
41 	WRITE_RPS1(0xc000008c);
42 	/* wait for vbi_a or vbi_b*/
43 	if ( 0 != (SAA7146_USE_PORT_B_FOR_VBI & dev->ext_vv_data->flags)) {
44 		DEB_D("...using port b\n");
45 		WRITE_RPS1(CMD_PAUSE | CMD_OAN | CMD_SIG1 | CMD_E_FID_B);
46 		WRITE_RPS1(CMD_PAUSE | CMD_OAN | CMD_SIG1 | CMD_O_FID_B);
47 /*
48 		WRITE_RPS1(CMD_PAUSE | MASK_09);
49 */
50 	} else {
51 		DEB_D("...using port a\n");
52 		WRITE_RPS1(CMD_PAUSE | MASK_10);
53 	}
54 	/* upload brs */
55 	WRITE_RPS1(CMD_UPLOAD | MASK_08);
56 	/* load brs-control register */
57 	WRITE_RPS1(CMD_WR_REG | (1 << 8) | (BRS_CTRL/4));
58 	/* BYO = 1, BXO = NQBIL (=1728 for PAL, for NTSC this is 858*2) - NumByte3 (=1440) = 288 */
59 	WRITE_RPS1(((1728-(vbi_pixel_to_capture)) << 7) | MASK_19);
60 	/* wait for brs_done */
61 	WRITE_RPS1(CMD_PAUSE | MASK_08);
62 	/* upload brs */
63 	WRITE_RPS1(CMD_UPLOAD | MASK_08);
64 	/* load video-dma3 NumLines3 and NumBytes3 */
65 	WRITE_RPS1(CMD_WR_REG | (1 << 8) | (NUM_LINE_BYTE3/4));
66 	/* dev->vbi_count*2 lines, 720 pixel (= 1440 Bytes) */
67 	WRITE_RPS1((2 << 16) | (vbi_pixel_to_capture));
68 	/* load brs-control register */
69 	WRITE_RPS1(CMD_WR_REG | (1 << 8) | (BRS_CTRL/4));
70 	/* Set BRS right: note: this is an experimental value for BXO (=> PAL!) */
71 	WRITE_RPS1((540 << 7) | (5 << 19));  // 5 == vbi_start
72 	/* wait for brs_done */
73 	WRITE_RPS1(CMD_PAUSE | MASK_08);
74 	/* upload brs and video-dma3*/
75 	WRITE_RPS1(CMD_UPLOAD | MASK_08 | MASK_04);
76 	/* load mc2 register: enable dma3 */
77 	WRITE_RPS1(CMD_WR_REG | (1 << 8) | (MC1/4));
78 	WRITE_RPS1(MASK_20 | MASK_04);
79 	/* generate interrupt */
80 	WRITE_RPS1(CMD_INTERRUPT);
81 	/* stop rps1 */
82 	WRITE_RPS1(CMD_STOP);
83 
84 	/* we have to do the workaround twice to be sure that
85 	   everything is ok */
86 	for(i = 0; i < 2; i++) {
87 
88 		/* indicate to the irq handler that we do the workaround */
89 		saa7146_write(dev, MC2, MASK_31|MASK_15);
90 
91 		saa7146_write(dev, NUM_LINE_BYTE3, (1<<16)|(2<<0));
92 		saa7146_write(dev, MC2, MASK_04|MASK_20);
93 
94 		/* enable rps1 irqs */
95 		SAA7146_IER_ENABLE(dev,MASK_28);
96 
97 		/* prepare to wait to be woken up by the irq-handler */
98 		add_wait_queue(&vv->vbi_wq, &wait);
99 		set_current_state(TASK_INTERRUPTIBLE);
100 
101 		/* start rps1 to enable workaround */
102 		saa7146_write(dev, RPS_ADDR1, dev->d_rps1.dma_handle);
103 		saa7146_write(dev, MC1, (MASK_13 | MASK_29));
104 
105 		schedule();
106 
107 		DEB_VBI("brs bug workaround %d/1\n", i);
108 
109 		remove_wait_queue(&vv->vbi_wq, &wait);
110 		__set_current_state(TASK_RUNNING);
111 
112 		/* disable rps1 irqs */
113 		SAA7146_IER_DISABLE(dev,MASK_28);
114 
115 		/* stop video-dma3 */
116 		saa7146_write(dev, MC1, MASK_20);
117 
118 		if(signal_pending(current)) {
119 
120 			DEB_VBI("aborted (rps:0x%08x)\n",
121 				saa7146_read(dev, RPS_ADDR1));
122 
123 			/* stop rps1 for sure */
124 			saa7146_write(dev, MC1, MASK_29);
125 
126 			dma_free_coherent(&dev->pci->dev, 4096, cpu, dma_addr);
127 			return -EINTR;
128 		}
129 	}
130 
131 	dma_free_coherent(&dev->pci->dev, 4096, cpu, dma_addr);
132 	return 0;
133 }
134 
135 static void saa7146_set_vbi_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next)
136 {
137 	struct saa7146_vv *vv = dev->vv_data;
138 
139 	struct saa7146_video_dma vdma3;
140 
141 	int count = 0;
142 	unsigned long e_wait = vv->current_hps_sync == SAA7146_HPS_SYNC_PORT_A ? CMD_E_FID_A : CMD_E_FID_B;
143 	unsigned long o_wait = vv->current_hps_sync == SAA7146_HPS_SYNC_PORT_A ? CMD_O_FID_A : CMD_O_FID_B;
144 
145 /*
146 	vdma3.base_even	= 0xc8000000+2560*70;
147 	vdma3.base_odd	= 0xc8000000;
148 	vdma3.prot_addr	= 0xc8000000+2560*164;
149 	vdma3.pitch	= 2560;
150 	vdma3.base_page	= 0;
151 	vdma3.num_line_byte = (64<<16)|((vbi_pixel_to_capture)<<0); // set above!
152 */
153 	vdma3.base_even	= buf->pt[2].offset;
154 	vdma3.base_odd	= buf->pt[2].offset + 16 * vbi_pixel_to_capture;
155 	vdma3.prot_addr	= buf->pt[2].offset + 16 * 2 * vbi_pixel_to_capture;
156 	vdma3.pitch	= vbi_pixel_to_capture;
157 	vdma3.base_page	= buf->pt[2].dma | ME1;
158 	vdma3.num_line_byte = (16 << 16) | vbi_pixel_to_capture;
159 
160 	saa7146_write_out_dma(dev, 3, &vdma3);
161 
162 	/* write beginning of rps-program */
163 	count = 0;
164 
165 	/* wait for o_fid_a/b / e_fid_a/b toggle only if bit 1 is not set */
166 
167 	/* we don't wait here for the first field anymore. this is different from the video
168 	   capture and might cause that the first buffer is only half filled (with only
169 	   one field). but since this is some sort of streaming data, this is not that negative.
170 	   but by doing this, we can use the whole engine from videobuf-dma-sg.c... */
171 
172 /*
173 	WRITE_RPS1(CMD_PAUSE | CMD_OAN | CMD_SIG1 | e_wait);
174 	WRITE_RPS1(CMD_PAUSE | CMD_OAN | CMD_SIG1 | o_wait);
175 */
176 	/* set bit 1 */
177 	WRITE_RPS1(CMD_WR_REG | (1 << 8) | (MC2/4));
178 	WRITE_RPS1(MASK_28 | MASK_12);
179 
180 	/* turn on video-dma3 */
181 	WRITE_RPS1(CMD_WR_REG_MASK | (MC1/4));
182 	WRITE_RPS1(MASK_04 | MASK_20);			/* => mask */
183 	WRITE_RPS1(MASK_04 | MASK_20);			/* => values */
184 
185 	/* wait for o_fid_a/b / e_fid_a/b toggle */
186 	WRITE_RPS1(CMD_PAUSE | o_wait);
187 	WRITE_RPS1(CMD_PAUSE | e_wait);
188 
189 	/* generate interrupt */
190 	WRITE_RPS1(CMD_INTERRUPT);
191 
192 	/* stop */
193 	WRITE_RPS1(CMD_STOP);
194 
195 	/* enable rps1 irqs */
196 	SAA7146_IER_ENABLE(dev, MASK_28);
197 
198 	/* write the address of the rps-program */
199 	saa7146_write(dev, RPS_ADDR1, dev->d_rps1.dma_handle);
200 
201 	/* turn on rps */
202 	saa7146_write(dev, MC1, (MASK_13 | MASK_29));
203 }
204 
205 static int buffer_activate(struct saa7146_dev *dev,
206 			   struct saa7146_buf *buf,
207 			   struct saa7146_buf *next)
208 {
209 	struct saa7146_vv *vv = dev->vv_data;
210 
211 	DEB_VBI("dev:%p, buf:%p, next:%p\n", dev, buf, next);
212 	saa7146_set_vbi_capture(dev,buf,next);
213 
214 	mod_timer(&vv->vbi_dmaq.timeout, jiffies+BUFFER_TIMEOUT);
215 	return 0;
216 }
217 
218 /* ------------------------------------------------------------------ */
219 
220 static int queue_setup(struct vb2_queue *q,
221 		       unsigned int *num_buffers, unsigned int *num_planes,
222 		       unsigned int sizes[], struct device *alloc_devs[])
223 {
224 	unsigned int size = 16 * 2 * vbi_pixel_to_capture;
225 
226 	if (*num_planes)
227 		return sizes[0] < size ? -EINVAL : 0;
228 	*num_planes = 1;
229 	sizes[0] = size;
230 
231 	return 0;
232 }
233 
234 static void buf_queue(struct vb2_buffer *vb)
235 {
236 	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
237 	struct vb2_queue *vq = vb->vb2_queue;
238 	struct saa7146_dev *dev = vb2_get_drv_priv(vq);
239 	struct saa7146_buf *buf = container_of(vbuf, struct saa7146_buf, vb);
240 	unsigned long flags;
241 
242 	spin_lock_irqsave(&dev->slock, flags);
243 
244 	saa7146_buffer_queue(dev, &dev->vv_data->vbi_dmaq, buf);
245 	spin_unlock_irqrestore(&dev->slock, flags);
246 }
247 
248 static int buf_init(struct vb2_buffer *vb)
249 {
250 	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
251 	struct saa7146_buf *buf = container_of(vbuf, struct saa7146_buf, vb);
252 	struct sg_table *sgt = vb2_dma_sg_plane_desc(&buf->vb.vb2_buf, 0);
253 	struct scatterlist *list = sgt->sgl;
254 	int length = sgt->nents;
255 	struct vb2_queue *vq = vb->vb2_queue;
256 	struct saa7146_dev *dev = vb2_get_drv_priv(vq);
257 	int ret;
258 
259 	buf->activate = buffer_activate;
260 
261 	saa7146_pgtable_alloc(dev->pci, &buf->pt[2]);
262 
263 	ret = saa7146_pgtable_build_single(dev->pci, &buf->pt[2],
264 					   list, length);
265 	if (ret)
266 		saa7146_pgtable_free(dev->pci, &buf->pt[2]);
267 	return ret;
268 }
269 
270 static int buf_prepare(struct vb2_buffer *vb)
271 {
272 	unsigned int size = 16 * 2 * vbi_pixel_to_capture;
273 
274 	if (vb2_plane_size(vb, 0) < size)
275 		return -EINVAL;
276 	vb2_set_plane_payload(vb, 0, size);
277 	return 0;
278 }
279 
280 static void buf_cleanup(struct vb2_buffer *vb)
281 {
282 	struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
283 	struct saa7146_buf *buf = container_of(vbuf, struct saa7146_buf, vb);
284 	struct vb2_queue *vq = vb->vb2_queue;
285 	struct saa7146_dev *dev = vb2_get_drv_priv(vq);
286 
287 	saa7146_pgtable_free(dev->pci, &buf->pt[2]);
288 }
289 
290 static void return_buffers(struct vb2_queue *q, int state)
291 {
292 	struct saa7146_dev *dev = vb2_get_drv_priv(q);
293 	struct saa7146_dmaqueue *dq = &dev->vv_data->vbi_dmaq;
294 	struct saa7146_buf *buf;
295 
296 	if (dq->curr) {
297 		buf = dq->curr;
298 		dq->curr = NULL;
299 		vb2_buffer_done(&buf->vb.vb2_buf, state);
300 	}
301 	while (!list_empty(&dq->queue)) {
302 		buf = list_entry(dq->queue.next, struct saa7146_buf, list);
303 		list_del(&buf->list);
304 		vb2_buffer_done(&buf->vb.vb2_buf, state);
305 	}
306 }
307 
308 static void vbi_stop(struct saa7146_dev *dev)
309 {
310 	struct saa7146_vv *vv = dev->vv_data;
311 	unsigned long flags;
312 	DEB_VBI("dev:%p\n", dev);
313 
314 	spin_lock_irqsave(&dev->slock,flags);
315 
316 	/* disable rps1  */
317 	saa7146_write(dev, MC1, MASK_29);
318 
319 	/* disable rps1 irqs */
320 	SAA7146_IER_DISABLE(dev, MASK_28);
321 
322 	/* shut down dma 3 transfers */
323 	saa7146_write(dev, MC1, MASK_20);
324 
325 	del_timer(&vv->vbi_dmaq.timeout);
326 	del_timer(&vv->vbi_read_timeout);
327 
328 	spin_unlock_irqrestore(&dev->slock, flags);
329 }
330 
331 static void vbi_read_timeout(struct timer_list *t)
332 {
333 	struct saa7146_vv *vv = from_timer(vv, t, vbi_read_timeout);
334 	struct saa7146_dev *dev = vv->vbi_dmaq.dev;
335 
336 	DEB_VBI("dev:%p\n", dev);
337 
338 	vbi_stop(dev);
339 }
340 
341 static int vbi_begin(struct saa7146_dev *dev)
342 {
343 	struct saa7146_vv *vv = dev->vv_data;
344 	u32 arbtr_ctrl	= saa7146_read(dev, PCI_BT_V1);
345 	int ret = 0;
346 
347 	DEB_VBI("dev:%p\n", dev);
348 
349 	ret = saa7146_res_get(dev, RESOURCE_DMA3_BRS);
350 	if (0 == ret) {
351 		DEB_S("cannot get vbi RESOURCE_DMA3_BRS resource\n");
352 		return -EBUSY;
353 	}
354 
355 	/* adjust arbitrition control for video dma 3 */
356 	arbtr_ctrl &= ~0x1f0000;
357 	arbtr_ctrl |=  0x1d0000;
358 	saa7146_write(dev, PCI_BT_V1, arbtr_ctrl);
359 	saa7146_write(dev, MC2, (MASK_04|MASK_20));
360 
361 	vv->vbi_read_timeout.function = vbi_read_timeout;
362 
363 	/* initialize the brs */
364 	if ( 0 != (SAA7146_USE_PORT_B_FOR_VBI & dev->ext_vv_data->flags)) {
365 		saa7146_write(dev, BRS_CTRL, MASK_30|MASK_29 | (7 << 19));
366 	} else {
367 		saa7146_write(dev, BRS_CTRL, 0x00000001);
368 
369 		if (0 != (ret = vbi_workaround(dev))) {
370 			DEB_VBI("vbi workaround failed!\n");
371 			/* return ret;*/
372 		}
373 	}
374 
375 	/* upload brs register */
376 	saa7146_write(dev, MC2, (MASK_08|MASK_24));
377 	return 0;
378 }
379 
380 static int start_streaming(struct vb2_queue *q, unsigned int count)
381 {
382 	struct saa7146_dev *dev = vb2_get_drv_priv(q);
383 	int ret;
384 
385 	if (!vb2_is_streaming(&dev->vv_data->vbi_dmaq.q))
386 		dev->vv_data->seqnr = 0;
387 	ret = vbi_begin(dev);
388 	if (ret)
389 		return_buffers(q, VB2_BUF_STATE_QUEUED);
390 	return ret;
391 }
392 
393 static void stop_streaming(struct vb2_queue *q)
394 {
395 	struct saa7146_dev *dev = vb2_get_drv_priv(q);
396 
397 	vbi_stop(dev);
398 	return_buffers(q, VB2_BUF_STATE_ERROR);
399 	saa7146_res_free(dev, RESOURCE_DMA3_BRS);
400 }
401 
402 const struct vb2_ops vbi_qops = {
403 	.queue_setup	= queue_setup,
404 	.buf_queue	= buf_queue,
405 	.buf_init	= buf_init,
406 	.buf_prepare	= buf_prepare,
407 	.buf_cleanup	= buf_cleanup,
408 	.start_streaming = start_streaming,
409 	.stop_streaming = stop_streaming,
410 	.wait_prepare	= vb2_ops_wait_prepare,
411 	.wait_finish	= vb2_ops_wait_finish,
412 };
413 
414 /* ------------------------------------------------------------------ */
415 
416 static void vbi_init(struct saa7146_dev *dev, struct saa7146_vv *vv)
417 {
418 	DEB_VBI("dev:%p\n", dev);
419 
420 	INIT_LIST_HEAD(&vv->vbi_dmaq.queue);
421 
422 	timer_setup(&vv->vbi_dmaq.timeout, saa7146_buffer_timeout, 0);
423 	vv->vbi_dmaq.dev              = dev;
424 
425 	init_waitqueue_head(&vv->vbi_wq);
426 }
427 
428 static void vbi_irq_done(struct saa7146_dev *dev, unsigned long status)
429 {
430 	struct saa7146_vv *vv = dev->vv_data;
431 	spin_lock(&dev->slock);
432 
433 	if (vv->vbi_dmaq.curr) {
434 		DEB_VBI("dev:%p, curr:%p\n", dev, vv->vbi_dmaq.curr);
435 		saa7146_buffer_finish(dev, &vv->vbi_dmaq, VB2_BUF_STATE_DONE);
436 	} else {
437 		DEB_VBI("dev:%p\n", dev);
438 	}
439 	saa7146_buffer_next(dev, &vv->vbi_dmaq, 1);
440 
441 	spin_unlock(&dev->slock);
442 }
443 
444 const struct saa7146_use_ops saa7146_vbi_uops = {
445 	.init		= vbi_init,
446 	.irq_done	= vbi_irq_done,
447 };
448