xref: /openbmc/linux/drivers/media/pci/ivtv/ivtv-irq.c (revision e3b9f1e8)
1 /* interrupt handling
2     Copyright (C) 2003-2004  Kevin Thayer <nufan_wfk at yahoo.com>
3     Copyright (C) 2004  Chris Kennedy <c@groovy.org>
4     Copyright (C) 2005-2007  Hans Verkuil <hverkuil@xs4all.nl>
5 
6     This program is free software; you can redistribute it and/or modify
7     it under the terms of the GNU General Public License as published by
8     the Free Software Foundation; either version 2 of the License, or
9     (at your option) any later version.
10 
11     This program is distributed in the hope that it will be useful,
12     but WITHOUT ANY WARRANTY; without even the implied warranty of
13     MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
14     GNU General Public License for more details.
15 
16     You should have received a copy of the GNU General Public License
17     along with this program; if not, write to the Free Software
18     Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
19  */
20 
21 #include "ivtv-driver.h"
22 #include "ivtv-queue.h"
23 #include "ivtv-udma.h"
24 #include "ivtv-irq.h"
25 #include "ivtv-mailbox.h"
26 #include "ivtv-vbi.h"
27 #include "ivtv-yuv.h"
28 #include <media/v4l2-event.h>
29 
30 #define DMA_MAGIC_COOKIE 0x000001fe
31 
32 static void ivtv_dma_dec_start(struct ivtv_stream *s);
33 
34 static const int ivtv_stream_map[] = {
35 	IVTV_ENC_STREAM_TYPE_MPG,
36 	IVTV_ENC_STREAM_TYPE_YUV,
37 	IVTV_ENC_STREAM_TYPE_PCM,
38 	IVTV_ENC_STREAM_TYPE_VBI,
39 };
40 
41 static void ivtv_pcm_work_handler(struct ivtv *itv)
42 {
43 	struct ivtv_stream *s = &itv->streams[IVTV_ENC_STREAM_TYPE_PCM];
44 	struct ivtv_buffer *buf;
45 
46 	/* Pass the PCM data to ivtv-alsa */
47 
48 	while (1) {
49 		/*
50 		 * Users should not be using both the ALSA and V4L2 PCM audio
51 		 * capture interfaces at the same time.  If the user is doing
52 		 * this, there maybe a buffer in q_io to grab, use, and put
53 		 * back in rotation.
54 		 */
55 		buf = ivtv_dequeue(s, &s->q_io);
56 		if (buf == NULL)
57 			buf = ivtv_dequeue(s, &s->q_full);
58 		if (buf == NULL)
59 			break;
60 
61 		if (buf->readpos < buf->bytesused)
62 			itv->pcm_announce_callback(itv->alsa,
63 				(u8 *)(buf->buf + buf->readpos),
64 				(size_t)(buf->bytesused - buf->readpos));
65 
66 		ivtv_enqueue(s, buf, &s->q_free);
67 	}
68 }
69 
70 static void ivtv_pio_work_handler(struct ivtv *itv)
71 {
72 	struct ivtv_stream *s = &itv->streams[itv->cur_pio_stream];
73 	struct ivtv_buffer *buf;
74 	int i = 0;
75 
76 	IVTV_DEBUG_HI_DMA("ivtv_pio_work_handler\n");
77 	if (itv->cur_pio_stream < 0 || itv->cur_pio_stream >= IVTV_MAX_STREAMS ||
78 			s->vdev.v4l2_dev == NULL || !ivtv_use_pio(s)) {
79 		itv->cur_pio_stream = -1;
80 		/* trigger PIO complete user interrupt */
81 		write_reg(IVTV_IRQ_ENC_PIO_COMPLETE, 0x44);
82 		return;
83 	}
84 	IVTV_DEBUG_HI_DMA("Process PIO %s\n", s->name);
85 	list_for_each_entry(buf, &s->q_dma.list, list) {
86 		u32 size = s->sg_processing[i].size & 0x3ffff;
87 
88 		/* Copy the data from the card to the buffer */
89 		if (s->type == IVTV_DEC_STREAM_TYPE_VBI) {
90 			memcpy_fromio(buf->buf, itv->dec_mem + s->sg_processing[i].src - IVTV_DECODER_OFFSET, size);
91 		}
92 		else {
93 			memcpy_fromio(buf->buf, itv->enc_mem + s->sg_processing[i].src, size);
94 		}
95 		i++;
96 		if (i == s->sg_processing_size)
97 			break;
98 	}
99 	write_reg(IVTV_IRQ_ENC_PIO_COMPLETE, 0x44);
100 }
101 
102 void ivtv_irq_work_handler(struct kthread_work *work)
103 {
104 	struct ivtv *itv = container_of(work, struct ivtv, irq_work);
105 
106 	if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_PIO, &itv->i_flags))
107 		ivtv_pio_work_handler(itv);
108 
109 	if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_VBI, &itv->i_flags))
110 		ivtv_vbi_work_handler(itv);
111 
112 	if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_YUV, &itv->i_flags))
113 		ivtv_yuv_work_handler(itv);
114 
115 	if (test_and_clear_bit(IVTV_F_I_WORK_HANDLER_PCM, &itv->i_flags))
116 		ivtv_pcm_work_handler(itv);
117 }
118 
119 /* Determine the required DMA size, setup enough buffers in the predma queue and
120    actually copy the data from the card to the buffers in case a PIO transfer is
121    required for this stream.
122  */
123 static int stream_enc_dma_append(struct ivtv_stream *s, u32 data[CX2341X_MBOX_MAX_DATA])
124 {
125 	struct ivtv *itv = s->itv;
126 	struct ivtv_buffer *buf;
127 	u32 bytes_needed = 0;
128 	u32 offset, size;
129 	u32 UVoffset = 0, UVsize = 0;
130 	int skip_bufs = s->q_predma.buffers;
131 	int idx = s->sg_pending_size;
132 	int rc;
133 
134 	/* sanity checks */
135 	if (s->vdev.v4l2_dev == NULL) {
136 		IVTV_DEBUG_WARN("Stream %s not started\n", s->name);
137 		return -1;
138 	}
139 	if (!test_bit(IVTV_F_S_CLAIMED, &s->s_flags)) {
140 		IVTV_DEBUG_WARN("Stream %s not open\n", s->name);
141 		return -1;
142 	}
143 
144 	/* determine offset, size and PTS for the various streams */
145 	switch (s->type) {
146 		case IVTV_ENC_STREAM_TYPE_MPG:
147 			offset = data[1];
148 			size = data[2];
149 			s->pending_pts = 0;
150 			break;
151 
152 		case IVTV_ENC_STREAM_TYPE_YUV:
153 			offset = data[1];
154 			size = data[2];
155 			UVoffset = data[3];
156 			UVsize = data[4];
157 			s->pending_pts = ((u64) data[5] << 32) | data[6];
158 			break;
159 
160 		case IVTV_ENC_STREAM_TYPE_PCM:
161 			offset = data[1] + 12;
162 			size = data[2] - 12;
163 			s->pending_pts = read_dec(offset - 8) |
164 				((u64)(read_dec(offset - 12)) << 32);
165 			if (itv->has_cx23415)
166 				offset += IVTV_DECODER_OFFSET;
167 			break;
168 
169 		case IVTV_ENC_STREAM_TYPE_VBI:
170 			size = itv->vbi.enc_size * itv->vbi.fpi;
171 			offset = read_enc(itv->vbi.enc_start - 4) + 12;
172 			if (offset == 12) {
173 				IVTV_DEBUG_INFO("VBI offset == 0\n");
174 				return -1;
175 			}
176 			s->pending_pts = read_enc(offset - 4) | ((u64)read_enc(offset - 8) << 32);
177 			break;
178 
179 		case IVTV_DEC_STREAM_TYPE_VBI:
180 			size = read_dec(itv->vbi.dec_start + 4) + 8;
181 			offset = read_dec(itv->vbi.dec_start) + itv->vbi.dec_start;
182 			s->pending_pts = 0;
183 			offset += IVTV_DECODER_OFFSET;
184 			break;
185 		default:
186 			/* shouldn't happen */
187 			return -1;
188 	}
189 
190 	/* if this is the start of the DMA then fill in the magic cookie */
191 	if (s->sg_pending_size == 0 && ivtv_use_dma(s)) {
192 		if (itv->has_cx23415 && (s->type == IVTV_ENC_STREAM_TYPE_PCM ||
193 		    s->type == IVTV_DEC_STREAM_TYPE_VBI)) {
194 			s->pending_backup = read_dec(offset - IVTV_DECODER_OFFSET);
195 			write_dec_sync(DMA_MAGIC_COOKIE, offset - IVTV_DECODER_OFFSET);
196 		}
197 		else {
198 			s->pending_backup = read_enc(offset);
199 			write_enc_sync(DMA_MAGIC_COOKIE, offset);
200 		}
201 		s->pending_offset = offset;
202 	}
203 
204 	bytes_needed = size;
205 	if (s->type == IVTV_ENC_STREAM_TYPE_YUV) {
206 		/* The size for the Y samples needs to be rounded upwards to a
207 		   multiple of the buf_size. The UV samples then start in the
208 		   next buffer. */
209 		bytes_needed = s->buf_size * ((bytes_needed + s->buf_size - 1) / s->buf_size);
210 		bytes_needed += UVsize;
211 	}
212 
213 	IVTV_DEBUG_HI_DMA("%s %s: 0x%08x bytes at 0x%08x\n",
214 		ivtv_use_pio(s) ? "PIO" : "DMA", s->name, bytes_needed, offset);
215 
216 	rc = ivtv_queue_move(s, &s->q_free, &s->q_full, &s->q_predma, bytes_needed);
217 	if (rc < 0) { /* Insufficient buffers */
218 		IVTV_DEBUG_WARN("Cannot obtain %d bytes for %s data transfer\n",
219 				bytes_needed, s->name);
220 		return -1;
221 	}
222 	if (rc && !s->buffers_stolen && test_bit(IVTV_F_S_APPL_IO, &s->s_flags)) {
223 		IVTV_WARN("All %s stream buffers are full. Dropping data.\n", s->name);
224 		IVTV_WARN("Cause: the application is not reading fast enough.\n");
225 	}
226 	s->buffers_stolen = rc;
227 
228 	/* got the buffers, now fill in sg_pending */
229 	buf = list_entry(s->q_predma.list.next, struct ivtv_buffer, list);
230 	memset(buf->buf, 0, 128);
231 	list_for_each_entry(buf, &s->q_predma.list, list) {
232 		if (skip_bufs-- > 0)
233 			continue;
234 		s->sg_pending[idx].dst = buf->dma_handle;
235 		s->sg_pending[idx].src = offset;
236 		s->sg_pending[idx].size = s->buf_size;
237 		buf->bytesused = min(size, s->buf_size);
238 		buf->dma_xfer_cnt = s->dma_xfer_cnt;
239 
240 		s->q_predma.bytesused += buf->bytesused;
241 		size -= buf->bytesused;
242 		offset += s->buf_size;
243 
244 		/* Sync SG buffers */
245 		ivtv_buf_sync_for_device(s, buf);
246 
247 		if (size == 0) {	/* YUV */
248 			/* process the UV section */
249 			offset = UVoffset;
250 			size = UVsize;
251 		}
252 		idx++;
253 	}
254 	s->sg_pending_size = idx;
255 	return 0;
256 }
257 
258 static void dma_post(struct ivtv_stream *s)
259 {
260 	struct ivtv *itv = s->itv;
261 	struct ivtv_buffer *buf = NULL;
262 	struct list_head *p;
263 	u32 offset;
264 	__le32 *u32buf;
265 	int x = 0;
266 
267 	IVTV_DEBUG_HI_DMA("%s %s completed (%x)\n", ivtv_use_pio(s) ? "PIO" : "DMA",
268 			s->name, s->dma_offset);
269 	list_for_each(p, &s->q_dma.list) {
270 		buf = list_entry(p, struct ivtv_buffer, list);
271 		u32buf = (__le32 *)buf->buf;
272 
273 		/* Sync Buffer */
274 		ivtv_buf_sync_for_cpu(s, buf);
275 
276 		if (x == 0 && ivtv_use_dma(s)) {
277 			offset = s->dma_last_offset;
278 			if (le32_to_cpu(u32buf[offset / 4]) != DMA_MAGIC_COOKIE)
279 			{
280 				for (offset = 0; offset < 64; offset++)
281 					if (le32_to_cpu(u32buf[offset]) == DMA_MAGIC_COOKIE)
282 						break;
283 				offset *= 4;
284 				if (offset == 256) {
285 					IVTV_DEBUG_WARN("%s: Couldn't find start of buffer within the first 256 bytes\n", s->name);
286 					offset = s->dma_last_offset;
287 				}
288 				if (s->dma_last_offset != offset)
289 					IVTV_DEBUG_WARN("%s: offset %d -> %d\n", s->name, s->dma_last_offset, offset);
290 				s->dma_last_offset = offset;
291 			}
292 			if (itv->has_cx23415 && (s->type == IVTV_ENC_STREAM_TYPE_PCM ||
293 						s->type == IVTV_DEC_STREAM_TYPE_VBI)) {
294 				write_dec_sync(0, s->dma_offset - IVTV_DECODER_OFFSET);
295 			}
296 			else {
297 				write_enc_sync(0, s->dma_offset);
298 			}
299 			if (offset) {
300 				buf->bytesused -= offset;
301 				memcpy(buf->buf, buf->buf + offset, buf->bytesused + offset);
302 			}
303 			*u32buf = cpu_to_le32(s->dma_backup);
304 		}
305 		x++;
306 		/* flag byteswap ABCD -> DCBA for MPG & VBI data outside irq */
307 		if (s->type == IVTV_ENC_STREAM_TYPE_MPG ||
308 		    s->type == IVTV_ENC_STREAM_TYPE_VBI)
309 			buf->b_flags |= IVTV_F_B_NEED_BUF_SWAP;
310 	}
311 	if (buf)
312 		buf->bytesused += s->dma_last_offset;
313 	if (buf && s->type == IVTV_DEC_STREAM_TYPE_VBI) {
314 		list_for_each_entry(buf, &s->q_dma.list, list) {
315 			/* Parse and Groom VBI Data */
316 			s->q_dma.bytesused -= buf->bytesused;
317 			ivtv_process_vbi_data(itv, buf, 0, s->type);
318 			s->q_dma.bytesused += buf->bytesused;
319 		}
320 		if (s->fh == NULL) {
321 			ivtv_queue_move(s, &s->q_dma, NULL, &s->q_free, 0);
322 			return;
323 		}
324 	}
325 
326 	ivtv_queue_move(s, &s->q_dma, NULL, &s->q_full, s->q_dma.bytesused);
327 
328 	if (s->type == IVTV_ENC_STREAM_TYPE_PCM &&
329 	    itv->pcm_announce_callback != NULL) {
330 		/*
331 		 * Set up the work handler to pass the data to ivtv-alsa.
332 		 *
333 		 * We just use q_full and let the work handler race with users
334 		 * making ivtv-fileops.c calls on the PCM device node.
335 		 *
336 		 * Users should not be using both the ALSA and V4L2 PCM audio
337 		 * capture interfaces at the same time.  If the user does this,
338 		 * fragments of data will just go out each interface as they
339 		 * race for PCM data.
340 		 */
341 		set_bit(IVTV_F_I_WORK_HANDLER_PCM, &itv->i_flags);
342 		set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
343 	}
344 
345 	if (s->fh)
346 		wake_up(&s->waitq);
347 }
348 
349 void ivtv_dma_stream_dec_prepare(struct ivtv_stream *s, u32 offset, int lock)
350 {
351 	struct ivtv *itv = s->itv;
352 	struct yuv_playback_info *yi = &itv->yuv_info;
353 	u8 frame = yi->draw_frame;
354 	struct yuv_frame_info *f = &yi->new_frame_info[frame];
355 	struct ivtv_buffer *buf;
356 	u32 y_size = 720 * ((f->src_h + 31) & ~31);
357 	u32 uv_offset = offset + IVTV_YUV_BUFFER_UV_OFFSET;
358 	int y_done = 0;
359 	int bytes_written = 0;
360 	int idx = 0;
361 
362 	IVTV_DEBUG_HI_DMA("DEC PREPARE DMA %s: %08x %08x\n", s->name, s->q_predma.bytesused, offset);
363 
364 	/* Insert buffer block for YUV if needed */
365 	if (s->type == IVTV_DEC_STREAM_TYPE_YUV && f->offset_y) {
366 		if (yi->blanking_dmaptr) {
367 			s->sg_pending[idx].src = yi->blanking_dmaptr;
368 			s->sg_pending[idx].dst = offset;
369 			s->sg_pending[idx].size = 720 * 16;
370 		}
371 		offset += 720 * 16;
372 		idx++;
373 	}
374 
375 	list_for_each_entry(buf, &s->q_predma.list, list) {
376 		/* YUV UV Offset from Y Buffer */
377 		if (s->type == IVTV_DEC_STREAM_TYPE_YUV && !y_done &&
378 				(bytes_written + buf->bytesused) >= y_size) {
379 			s->sg_pending[idx].src = buf->dma_handle;
380 			s->sg_pending[idx].dst = offset;
381 			s->sg_pending[idx].size = y_size - bytes_written;
382 			offset = uv_offset;
383 			if (s->sg_pending[idx].size != buf->bytesused) {
384 				idx++;
385 				s->sg_pending[idx].src =
386 				  buf->dma_handle + s->sg_pending[idx - 1].size;
387 				s->sg_pending[idx].dst = offset;
388 				s->sg_pending[idx].size =
389 				   buf->bytesused - s->sg_pending[idx - 1].size;
390 				offset += s->sg_pending[idx].size;
391 			}
392 			y_done = 1;
393 		} else {
394 			s->sg_pending[idx].src = buf->dma_handle;
395 			s->sg_pending[idx].dst = offset;
396 			s->sg_pending[idx].size = buf->bytesused;
397 			offset += buf->bytesused;
398 		}
399 		bytes_written += buf->bytesused;
400 
401 		/* Sync SG buffers */
402 		ivtv_buf_sync_for_device(s, buf);
403 		idx++;
404 	}
405 	s->sg_pending_size = idx;
406 
407 	/* Sync Hardware SG List of buffers */
408 	ivtv_stream_sync_for_device(s);
409 	if (lock) {
410 		unsigned long flags = 0;
411 
412 		spin_lock_irqsave(&itv->dma_reg_lock, flags);
413 		if (!test_bit(IVTV_F_I_DMA, &itv->i_flags))
414 			ivtv_dma_dec_start(s);
415 		else
416 			set_bit(IVTV_F_S_DMA_PENDING, &s->s_flags);
417 		spin_unlock_irqrestore(&itv->dma_reg_lock, flags);
418 	} else {
419 		if (!test_bit(IVTV_F_I_DMA, &itv->i_flags))
420 			ivtv_dma_dec_start(s);
421 		else
422 			set_bit(IVTV_F_S_DMA_PENDING, &s->s_flags);
423 	}
424 }
425 
426 static void ivtv_dma_enc_start_xfer(struct ivtv_stream *s)
427 {
428 	struct ivtv *itv = s->itv;
429 
430 	s->sg_dma->src = cpu_to_le32(s->sg_processing[s->sg_processed].src);
431 	s->sg_dma->dst = cpu_to_le32(s->sg_processing[s->sg_processed].dst);
432 	s->sg_dma->size = cpu_to_le32(s->sg_processing[s->sg_processed].size | 0x80000000);
433 	s->sg_processed++;
434 	/* Sync Hardware SG List of buffers */
435 	ivtv_stream_sync_for_device(s);
436 	write_reg(s->sg_handle, IVTV_REG_ENCDMAADDR);
437 	write_reg_sync(read_reg(IVTV_REG_DMAXFER) | 0x02, IVTV_REG_DMAXFER);
438 	itv->dma_timer.expires = jiffies + msecs_to_jiffies(300);
439 	add_timer(&itv->dma_timer);
440 }
441 
442 static void ivtv_dma_dec_start_xfer(struct ivtv_stream *s)
443 {
444 	struct ivtv *itv = s->itv;
445 
446 	s->sg_dma->src = cpu_to_le32(s->sg_processing[s->sg_processed].src);
447 	s->sg_dma->dst = cpu_to_le32(s->sg_processing[s->sg_processed].dst);
448 	s->sg_dma->size = cpu_to_le32(s->sg_processing[s->sg_processed].size | 0x80000000);
449 	s->sg_processed++;
450 	/* Sync Hardware SG List of buffers */
451 	ivtv_stream_sync_for_device(s);
452 	write_reg(s->sg_handle, IVTV_REG_DECDMAADDR);
453 	write_reg_sync(read_reg(IVTV_REG_DMAXFER) | 0x01, IVTV_REG_DMAXFER);
454 	itv->dma_timer.expires = jiffies + msecs_to_jiffies(300);
455 	add_timer(&itv->dma_timer);
456 }
457 
458 /* start the encoder DMA */
459 static void ivtv_dma_enc_start(struct ivtv_stream *s)
460 {
461 	struct ivtv *itv = s->itv;
462 	struct ivtv_stream *s_vbi = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
463 	int i;
464 
465 	IVTV_DEBUG_HI_DMA("start %s for %s\n", ivtv_use_dma(s) ? "DMA" : "PIO", s->name);
466 
467 	if (s->q_predma.bytesused)
468 		ivtv_queue_move(s, &s->q_predma, NULL, &s->q_dma, s->q_predma.bytesused);
469 
470 	if (ivtv_use_dma(s))
471 		s->sg_pending[s->sg_pending_size - 1].size += 256;
472 
473 	/* If this is an MPEG stream, and VBI data is also pending, then append the
474 	   VBI DMA to the MPEG DMA and transfer both sets of data at once.
475 
476 	   VBI DMA is a second class citizen compared to MPEG and mixing them together
477 	   will confuse the firmware (the end of a VBI DMA is seen as the end of a
478 	   MPEG DMA, thus effectively dropping an MPEG frame). So instead we make
479 	   sure we only use the MPEG DMA to transfer the VBI DMA if both are in
480 	   use. This way no conflicts occur. */
481 	clear_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags);
482 	if (s->type == IVTV_ENC_STREAM_TYPE_MPG && s_vbi->sg_pending_size &&
483 			s->sg_pending_size + s_vbi->sg_pending_size <= s->buffers) {
484 		ivtv_queue_move(s_vbi, &s_vbi->q_predma, NULL, &s_vbi->q_dma, s_vbi->q_predma.bytesused);
485 		if (ivtv_use_dma(s_vbi))
486 			s_vbi->sg_pending[s_vbi->sg_pending_size - 1].size += 256;
487 		for (i = 0; i < s_vbi->sg_pending_size; i++) {
488 			s->sg_pending[s->sg_pending_size++] = s_vbi->sg_pending[i];
489 		}
490 		s_vbi->dma_offset = s_vbi->pending_offset;
491 		s_vbi->sg_pending_size = 0;
492 		s_vbi->dma_xfer_cnt++;
493 		set_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags);
494 		IVTV_DEBUG_HI_DMA("include DMA for %s\n", s_vbi->name);
495 	}
496 
497 	s->dma_xfer_cnt++;
498 	memcpy(s->sg_processing, s->sg_pending, sizeof(struct ivtv_sg_host_element) * s->sg_pending_size);
499 	s->sg_processing_size = s->sg_pending_size;
500 	s->sg_pending_size = 0;
501 	s->sg_processed = 0;
502 	s->dma_offset = s->pending_offset;
503 	s->dma_backup = s->pending_backup;
504 	s->dma_pts = s->pending_pts;
505 
506 	if (ivtv_use_pio(s)) {
507 		set_bit(IVTV_F_I_WORK_HANDLER_PIO, &itv->i_flags);
508 		set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
509 		set_bit(IVTV_F_I_PIO, &itv->i_flags);
510 		itv->cur_pio_stream = s->type;
511 	}
512 	else {
513 		itv->dma_retries = 0;
514 		ivtv_dma_enc_start_xfer(s);
515 		set_bit(IVTV_F_I_DMA, &itv->i_flags);
516 		itv->cur_dma_stream = s->type;
517 	}
518 }
519 
520 static void ivtv_dma_dec_start(struct ivtv_stream *s)
521 {
522 	struct ivtv *itv = s->itv;
523 
524 	if (s->q_predma.bytesused)
525 		ivtv_queue_move(s, &s->q_predma, NULL, &s->q_dma, s->q_predma.bytesused);
526 	s->dma_xfer_cnt++;
527 	memcpy(s->sg_processing, s->sg_pending, sizeof(struct ivtv_sg_host_element) * s->sg_pending_size);
528 	s->sg_processing_size = s->sg_pending_size;
529 	s->sg_pending_size = 0;
530 	s->sg_processed = 0;
531 
532 	IVTV_DEBUG_HI_DMA("start DMA for %s\n", s->name);
533 	itv->dma_retries = 0;
534 	ivtv_dma_dec_start_xfer(s);
535 	set_bit(IVTV_F_I_DMA, &itv->i_flags);
536 	itv->cur_dma_stream = s->type;
537 }
538 
539 static void ivtv_irq_dma_read(struct ivtv *itv)
540 {
541 	struct ivtv_stream *s = NULL;
542 	struct ivtv_buffer *buf;
543 	int hw_stream_type = 0;
544 
545 	IVTV_DEBUG_HI_IRQ("DEC DMA READ\n");
546 
547 	del_timer(&itv->dma_timer);
548 
549 	if (!test_bit(IVTV_F_I_UDMA, &itv->i_flags) && itv->cur_dma_stream < 0)
550 		return;
551 
552 	if (!test_bit(IVTV_F_I_UDMA, &itv->i_flags)) {
553 		s = &itv->streams[itv->cur_dma_stream];
554 		ivtv_stream_sync_for_cpu(s);
555 
556 		if (read_reg(IVTV_REG_DMASTATUS) & 0x14) {
557 			IVTV_DEBUG_WARN("DEC DMA ERROR %x (xfer %d of %d, retry %d)\n",
558 					read_reg(IVTV_REG_DMASTATUS),
559 					s->sg_processed, s->sg_processing_size, itv->dma_retries);
560 			write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
561 			if (itv->dma_retries == 3) {
562 				/* Too many retries, give up on this frame */
563 				itv->dma_retries = 0;
564 				s->sg_processed = s->sg_processing_size;
565 			}
566 			else {
567 				/* Retry, starting with the first xfer segment.
568 				   Just retrying the current segment is not sufficient. */
569 				s->sg_processed = 0;
570 				itv->dma_retries++;
571 			}
572 		}
573 		if (s->sg_processed < s->sg_processing_size) {
574 			/* DMA next buffer */
575 			ivtv_dma_dec_start_xfer(s);
576 			return;
577 		}
578 		if (s->type == IVTV_DEC_STREAM_TYPE_YUV)
579 			hw_stream_type = 2;
580 		IVTV_DEBUG_HI_DMA("DEC DATA READ %s: %d\n", s->name, s->q_dma.bytesused);
581 
582 		/* For some reason must kick the firmware, like PIO mode,
583 		   I think this tells the firmware we are done and the size
584 		   of the xfer so it can calculate what we need next.
585 		   I think we can do this part ourselves but would have to
586 		   fully calculate xfer info ourselves and not use interrupts
587 		 */
588 		ivtv_vapi(itv, CX2341X_DEC_SCHED_DMA_FROM_HOST, 3, 0, s->q_dma.bytesused,
589 				hw_stream_type);
590 
591 		/* Free last DMA call */
592 		while ((buf = ivtv_dequeue(s, &s->q_dma)) != NULL) {
593 			ivtv_buf_sync_for_cpu(s, buf);
594 			ivtv_enqueue(s, buf, &s->q_free);
595 		}
596 		wake_up(&s->waitq);
597 	}
598 	clear_bit(IVTV_F_I_UDMA, &itv->i_flags);
599 	clear_bit(IVTV_F_I_DMA, &itv->i_flags);
600 	itv->cur_dma_stream = -1;
601 	wake_up(&itv->dma_waitq);
602 }
603 
604 static void ivtv_irq_enc_dma_complete(struct ivtv *itv)
605 {
606 	u32 data[CX2341X_MBOX_MAX_DATA];
607 	struct ivtv_stream *s;
608 
609 	ivtv_api_get_data(&itv->enc_mbox, IVTV_MBOX_DMA_END, 2, data);
610 	IVTV_DEBUG_HI_IRQ("ENC DMA COMPLETE %x %d (%d)\n", data[0], data[1], itv->cur_dma_stream);
611 
612 	del_timer(&itv->dma_timer);
613 
614 	if (itv->cur_dma_stream < 0)
615 		return;
616 
617 	s = &itv->streams[itv->cur_dma_stream];
618 	ivtv_stream_sync_for_cpu(s);
619 
620 	if (data[0] & 0x18) {
621 		IVTV_DEBUG_WARN("ENC DMA ERROR %x (offset %08x, xfer %d of %d, retry %d)\n", data[0],
622 			s->dma_offset, s->sg_processed, s->sg_processing_size, itv->dma_retries);
623 		write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
624 		if (itv->dma_retries == 3) {
625 			/* Too many retries, give up on this frame */
626 			itv->dma_retries = 0;
627 			s->sg_processed = s->sg_processing_size;
628 		}
629 		else {
630 			/* Retry, starting with the first xfer segment.
631 			   Just retrying the current segment is not sufficient. */
632 			s->sg_processed = 0;
633 			itv->dma_retries++;
634 		}
635 	}
636 	if (s->sg_processed < s->sg_processing_size) {
637 		/* DMA next buffer */
638 		ivtv_dma_enc_start_xfer(s);
639 		return;
640 	}
641 	clear_bit(IVTV_F_I_DMA, &itv->i_flags);
642 	itv->cur_dma_stream = -1;
643 	dma_post(s);
644 	if (test_and_clear_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags)) {
645 		s = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
646 		dma_post(s);
647 	}
648 	s->sg_processing_size = 0;
649 	s->sg_processed = 0;
650 	wake_up(&itv->dma_waitq);
651 }
652 
653 static void ivtv_irq_enc_pio_complete(struct ivtv *itv)
654 {
655 	struct ivtv_stream *s;
656 
657 	if (itv->cur_pio_stream < 0 || itv->cur_pio_stream >= IVTV_MAX_STREAMS) {
658 		itv->cur_pio_stream = -1;
659 		return;
660 	}
661 	s = &itv->streams[itv->cur_pio_stream];
662 	IVTV_DEBUG_HI_IRQ("ENC PIO COMPLETE %s\n", s->name);
663 	clear_bit(IVTV_F_I_PIO, &itv->i_flags);
664 	itv->cur_pio_stream = -1;
665 	dma_post(s);
666 	if (s->type == IVTV_ENC_STREAM_TYPE_MPG)
667 		ivtv_vapi(itv, CX2341X_ENC_SCHED_DMA_TO_HOST, 3, 0, 0, 0);
668 	else if (s->type == IVTV_ENC_STREAM_TYPE_YUV)
669 		ivtv_vapi(itv, CX2341X_ENC_SCHED_DMA_TO_HOST, 3, 0, 0, 1);
670 	else if (s->type == IVTV_ENC_STREAM_TYPE_PCM)
671 		ivtv_vapi(itv, CX2341X_ENC_SCHED_DMA_TO_HOST, 3, 0, 0, 2);
672 	clear_bit(IVTV_F_I_PIO, &itv->i_flags);
673 	if (test_and_clear_bit(IVTV_F_S_DMA_HAS_VBI, &s->s_flags)) {
674 		s = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
675 		dma_post(s);
676 	}
677 	wake_up(&itv->dma_waitq);
678 }
679 
680 static void ivtv_irq_dma_err(struct ivtv *itv)
681 {
682 	u32 data[CX2341X_MBOX_MAX_DATA];
683 	u32 status;
684 
685 	del_timer(&itv->dma_timer);
686 
687 	ivtv_api_get_data(&itv->enc_mbox, IVTV_MBOX_DMA_END, 2, data);
688 	status = read_reg(IVTV_REG_DMASTATUS);
689 	IVTV_DEBUG_WARN("DMA ERROR %08x %08x %08x %d\n", data[0], data[1],
690 				status, itv->cur_dma_stream);
691 	/*
692 	 * We do *not* write back to the IVTV_REG_DMASTATUS register to
693 	 * clear the error status, if either the encoder write (0x02) or
694 	 * decoder read (0x01) bus master DMA operation do not indicate
695 	 * completed.  We can race with the DMA engine, which may have
696 	 * transitioned to completed status *after* we read the register.
697 	 * Setting a IVTV_REG_DMASTATUS flag back to "busy" status, after the
698 	 * DMA engine has completed, will cause the DMA engine to stop working.
699 	 */
700 	status &= 0x3;
701 	if (status == 0x3)
702 		write_reg(status, IVTV_REG_DMASTATUS);
703 
704 	if (!test_bit(IVTV_F_I_UDMA, &itv->i_flags) &&
705 	    itv->cur_dma_stream >= 0 && itv->cur_dma_stream < IVTV_MAX_STREAMS) {
706 		struct ivtv_stream *s = &itv->streams[itv->cur_dma_stream];
707 
708 		if (s->type >= IVTV_DEC_STREAM_TYPE_MPG) {
709 			/* retry */
710 			/*
711 			 * FIXME - handle cases of DMA error similar to
712 			 * encoder below, except conditioned on status & 0x1
713 			 */
714 			ivtv_dma_dec_start(s);
715 			return;
716 		} else {
717 			if ((status & 0x2) == 0) {
718 				/*
719 				 * CX2341x Bus Master DMA write is ongoing.
720 				 * Reset the timer and let it complete.
721 				 */
722 				itv->dma_timer.expires =
723 						jiffies + msecs_to_jiffies(600);
724 				add_timer(&itv->dma_timer);
725 				return;
726 			}
727 
728 			if (itv->dma_retries < 3) {
729 				/*
730 				 * CX2341x Bus Master DMA write has ended.
731 				 * Retry the write, starting with the first
732 				 * xfer segment. Just retrying the current
733 				 * segment is not sufficient.
734 				 */
735 				s->sg_processed = 0;
736 				itv->dma_retries++;
737 				ivtv_dma_enc_start_xfer(s);
738 				return;
739 			}
740 			/* Too many retries, give up on this one */
741 		}
742 
743 	}
744 	if (test_bit(IVTV_F_I_UDMA, &itv->i_flags)) {
745 		ivtv_udma_start(itv);
746 		return;
747 	}
748 	clear_bit(IVTV_F_I_UDMA, &itv->i_flags);
749 	clear_bit(IVTV_F_I_DMA, &itv->i_flags);
750 	itv->cur_dma_stream = -1;
751 	wake_up(&itv->dma_waitq);
752 }
753 
754 static void ivtv_irq_enc_start_cap(struct ivtv *itv)
755 {
756 	u32 data[CX2341X_MBOX_MAX_DATA];
757 	struct ivtv_stream *s;
758 
759 	/* Get DMA destination and size arguments from card */
760 	ivtv_api_get_data(&itv->enc_mbox, IVTV_MBOX_DMA, 7, data);
761 	IVTV_DEBUG_HI_IRQ("ENC START CAP %d: %08x %08x\n", data[0], data[1], data[2]);
762 
763 	if (data[0] > 2 || data[1] == 0 || data[2] == 0) {
764 		IVTV_DEBUG_WARN("Unknown input: %08x %08x %08x\n",
765 				data[0], data[1], data[2]);
766 		return;
767 	}
768 	s = &itv->streams[ivtv_stream_map[data[0]]];
769 	if (!stream_enc_dma_append(s, data)) {
770 		set_bit(ivtv_use_pio(s) ? IVTV_F_S_PIO_PENDING : IVTV_F_S_DMA_PENDING, &s->s_flags);
771 	}
772 }
773 
774 static void ivtv_irq_enc_vbi_cap(struct ivtv *itv)
775 {
776 	u32 data[CX2341X_MBOX_MAX_DATA];
777 	struct ivtv_stream *s;
778 
779 	IVTV_DEBUG_HI_IRQ("ENC START VBI CAP\n");
780 	s = &itv->streams[IVTV_ENC_STREAM_TYPE_VBI];
781 
782 	if (!stream_enc_dma_append(s, data))
783 		set_bit(ivtv_use_pio(s) ? IVTV_F_S_PIO_PENDING : IVTV_F_S_DMA_PENDING, &s->s_flags);
784 }
785 
786 static void ivtv_irq_dec_vbi_reinsert(struct ivtv *itv)
787 {
788 	u32 data[CX2341X_MBOX_MAX_DATA];
789 	struct ivtv_stream *s = &itv->streams[IVTV_DEC_STREAM_TYPE_VBI];
790 
791 	IVTV_DEBUG_HI_IRQ("DEC VBI REINSERT\n");
792 	if (test_bit(IVTV_F_S_CLAIMED, &s->s_flags) &&
793 			!stream_enc_dma_append(s, data)) {
794 		set_bit(IVTV_F_S_PIO_PENDING, &s->s_flags);
795 	}
796 }
797 
798 static void ivtv_irq_dec_data_req(struct ivtv *itv)
799 {
800 	u32 data[CX2341X_MBOX_MAX_DATA];
801 	struct ivtv_stream *s;
802 
803 	/* YUV or MPG */
804 
805 	if (test_bit(IVTV_F_I_DEC_YUV, &itv->i_flags)) {
806 		ivtv_api_get_data(&itv->dec_mbox, IVTV_MBOX_DMA, 2, data);
807 		itv->dma_data_req_size =
808 				 1080 * ((itv->yuv_info.v4l2_src_h + 31) & ~31);
809 		itv->dma_data_req_offset = data[1];
810 		if (atomic_read(&itv->yuv_info.next_dma_frame) >= 0)
811 			ivtv_yuv_frame_complete(itv);
812 		s = &itv->streams[IVTV_DEC_STREAM_TYPE_YUV];
813 	}
814 	else {
815 		ivtv_api_get_data(&itv->dec_mbox, IVTV_MBOX_DMA, 3, data);
816 		itv->dma_data_req_size = min_t(u32, data[2], 0x10000);
817 		itv->dma_data_req_offset = data[1];
818 		s = &itv->streams[IVTV_DEC_STREAM_TYPE_MPG];
819 	}
820 	IVTV_DEBUG_HI_IRQ("DEC DATA REQ %s: %d %08x %u\n", s->name, s->q_full.bytesused,
821 		       itv->dma_data_req_offset, itv->dma_data_req_size);
822 	if (itv->dma_data_req_size == 0 || s->q_full.bytesused < itv->dma_data_req_size) {
823 		set_bit(IVTV_F_S_NEEDS_DATA, &s->s_flags);
824 	}
825 	else {
826 		if (test_bit(IVTV_F_I_DEC_YUV, &itv->i_flags))
827 			ivtv_yuv_setup_stream_frame(itv);
828 		clear_bit(IVTV_F_S_NEEDS_DATA, &s->s_flags);
829 		ivtv_queue_move(s, &s->q_full, NULL, &s->q_predma, itv->dma_data_req_size);
830 		ivtv_dma_stream_dec_prepare(s, itv->dma_data_req_offset + IVTV_DECODER_OFFSET, 0);
831 	}
832 }
833 
834 static void ivtv_irq_vsync(struct ivtv *itv)
835 {
836 	/* The vsync interrupt is unusual in that it won't clear until
837 	 * the end of the first line for the current field, at which
838 	 * point it clears itself. This can result in repeated vsync
839 	 * interrupts, or a missed vsync. Read some of the registers
840 	 * to determine the line being displayed and ensure we handle
841 	 * one vsync per frame.
842 	 */
843 	unsigned int frame = read_reg(IVTV_REG_DEC_LINE_FIELD) & 1;
844 	struct yuv_playback_info *yi = &itv->yuv_info;
845 	int last_dma_frame = atomic_read(&yi->next_dma_frame);
846 	struct yuv_frame_info *f = &yi->new_frame_info[last_dma_frame];
847 
848 	if (0) IVTV_DEBUG_IRQ("DEC VSYNC\n");
849 
850 	if (((frame ^ f->sync_field) == 0 &&
851 		((itv->last_vsync_field & 1) ^ f->sync_field)) ||
852 			(frame != (itv->last_vsync_field & 1) && !f->interlaced)) {
853 		int next_dma_frame = last_dma_frame;
854 
855 		if (!(f->interlaced && f->delay && yi->fields_lapsed < 1)) {
856 			if (next_dma_frame >= 0 && next_dma_frame != atomic_read(&yi->next_fill_frame)) {
857 				write_reg(yuv_offset[next_dma_frame] >> 4, 0x82c);
858 				write_reg((yuv_offset[next_dma_frame] + IVTV_YUV_BUFFER_UV_OFFSET) >> 4, 0x830);
859 				write_reg(yuv_offset[next_dma_frame] >> 4, 0x834);
860 				write_reg((yuv_offset[next_dma_frame] + IVTV_YUV_BUFFER_UV_OFFSET) >> 4, 0x838);
861 				next_dma_frame = (next_dma_frame + 1) % IVTV_YUV_BUFFERS;
862 				atomic_set(&yi->next_dma_frame, next_dma_frame);
863 				yi->fields_lapsed = -1;
864 				yi->running = 1;
865 			}
866 		}
867 	}
868 	if (frame != (itv->last_vsync_field & 1)) {
869 		static const struct v4l2_event evtop = {
870 			.type = V4L2_EVENT_VSYNC,
871 			.u.vsync.field = V4L2_FIELD_TOP,
872 		};
873 		static const struct v4l2_event evbottom = {
874 			.type = V4L2_EVENT_VSYNC,
875 			.u.vsync.field = V4L2_FIELD_BOTTOM,
876 		};
877 		struct ivtv_stream *s = ivtv_get_output_stream(itv);
878 
879 		itv->last_vsync_field += 1;
880 		if (frame == 0) {
881 			clear_bit(IVTV_F_I_VALID_DEC_TIMINGS, &itv->i_flags);
882 			clear_bit(IVTV_F_I_EV_VSYNC_FIELD, &itv->i_flags);
883 		}
884 		else {
885 			set_bit(IVTV_F_I_EV_VSYNC_FIELD, &itv->i_flags);
886 		}
887 		if (test_bit(IVTV_F_I_EV_VSYNC_ENABLED, &itv->i_flags)) {
888 			set_bit(IVTV_F_I_EV_VSYNC, &itv->i_flags);
889 			wake_up(&itv->event_waitq);
890 			if (s)
891 				wake_up(&s->waitq);
892 		}
893 		if (s && s->vdev.v4l2_dev)
894 			v4l2_event_queue(&s->vdev, frame ? &evtop : &evbottom);
895 		wake_up(&itv->vsync_waitq);
896 
897 		/* Send VBI to saa7127 */
898 		if (frame && (itv->output_mode == OUT_PASSTHROUGH ||
899 			test_bit(IVTV_F_I_UPDATE_WSS, &itv->i_flags) ||
900 			test_bit(IVTV_F_I_UPDATE_VPS, &itv->i_flags) ||
901 			test_bit(IVTV_F_I_UPDATE_CC, &itv->i_flags))) {
902 			set_bit(IVTV_F_I_WORK_HANDLER_VBI, &itv->i_flags);
903 			set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
904 		}
905 
906 		/* Check if we need to update the yuv registers */
907 		if (yi->running && (yi->yuv_forced_update || f->update)) {
908 			if (!f->update) {
909 				last_dma_frame =
910 					(u8)(atomic_read(&yi->next_dma_frame) -
911 						 1) % IVTV_YUV_BUFFERS;
912 				f = &yi->new_frame_info[last_dma_frame];
913 			}
914 
915 			if (f->src_w) {
916 				yi->update_frame = last_dma_frame;
917 				f->update = 0;
918 				yi->yuv_forced_update = 0;
919 				set_bit(IVTV_F_I_WORK_HANDLER_YUV, &itv->i_flags);
920 				set_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags);
921 			}
922 		}
923 
924 		yi->fields_lapsed++;
925 	}
926 }
927 
928 #define IVTV_IRQ_DMA (IVTV_IRQ_DMA_READ | IVTV_IRQ_ENC_DMA_COMPLETE | IVTV_IRQ_DMA_ERR | IVTV_IRQ_ENC_START_CAP | IVTV_IRQ_ENC_VBI_CAP | IVTV_IRQ_DEC_DATA_REQ | IVTV_IRQ_DEC_VBI_RE_INSERT)
929 
930 irqreturn_t ivtv_irq_handler(int irq, void *dev_id)
931 {
932 	struct ivtv *itv = (struct ivtv *)dev_id;
933 	u32 combo;
934 	u32 stat;
935 	int i;
936 	u8 vsync_force = 0;
937 
938 	spin_lock(&itv->dma_reg_lock);
939 	/* get contents of irq status register */
940 	stat = read_reg(IVTV_REG_IRQSTATUS);
941 
942 	combo = ~itv->irqmask & stat;
943 
944 	/* Clear out IRQ */
945 	if (combo) write_reg(combo, IVTV_REG_IRQSTATUS);
946 
947 	if (0 == combo) {
948 		/* The vsync interrupt is unusual and clears itself. If we
949 		 * took too long, we may have missed it. Do some checks
950 		 */
951 		if (~itv->irqmask & IVTV_IRQ_DEC_VSYNC) {
952 			/* vsync is enabled, see if we're in a new field */
953 			if ((itv->last_vsync_field & 1) !=
954 			    (read_reg(IVTV_REG_DEC_LINE_FIELD) & 1)) {
955 				/* New field, looks like we missed it */
956 				IVTV_DEBUG_YUV("VSync interrupt missed %d\n",
957 				       read_reg(IVTV_REG_DEC_LINE_FIELD) >> 16);
958 				vsync_force = 1;
959 			}
960 		}
961 
962 		if (!vsync_force) {
963 			/* No Vsync expected, wasn't for us */
964 			spin_unlock(&itv->dma_reg_lock);
965 			return IRQ_NONE;
966 		}
967 	}
968 
969 	/* Exclude interrupts noted below from the output, otherwise the log is flooded with
970 	   these messages */
971 	if (combo & ~0xff6d0400)
972 		IVTV_DEBUG_HI_IRQ("======= valid IRQ bits: 0x%08x ======\n", combo);
973 
974 	if (combo & IVTV_IRQ_DEC_DMA_COMPLETE) {
975 		IVTV_DEBUG_HI_IRQ("DEC DMA COMPLETE\n");
976 	}
977 
978 	if (combo & IVTV_IRQ_DMA_READ) {
979 		ivtv_irq_dma_read(itv);
980 	}
981 
982 	if (combo & IVTV_IRQ_ENC_DMA_COMPLETE) {
983 		ivtv_irq_enc_dma_complete(itv);
984 	}
985 
986 	if (combo & IVTV_IRQ_ENC_PIO_COMPLETE) {
987 		ivtv_irq_enc_pio_complete(itv);
988 	}
989 
990 	if (combo & IVTV_IRQ_DMA_ERR) {
991 		ivtv_irq_dma_err(itv);
992 	}
993 
994 	if (combo & IVTV_IRQ_ENC_START_CAP) {
995 		ivtv_irq_enc_start_cap(itv);
996 	}
997 
998 	if (combo & IVTV_IRQ_ENC_VBI_CAP) {
999 		ivtv_irq_enc_vbi_cap(itv);
1000 	}
1001 
1002 	if (combo & IVTV_IRQ_DEC_VBI_RE_INSERT) {
1003 		ivtv_irq_dec_vbi_reinsert(itv);
1004 	}
1005 
1006 	if (combo & IVTV_IRQ_ENC_EOS) {
1007 		IVTV_DEBUG_IRQ("ENC EOS\n");
1008 		set_bit(IVTV_F_I_EOS, &itv->i_flags);
1009 		wake_up(&itv->eos_waitq);
1010 	}
1011 
1012 	if (combo & IVTV_IRQ_DEC_DATA_REQ) {
1013 		ivtv_irq_dec_data_req(itv);
1014 	}
1015 
1016 	/* Decoder Vertical Sync - We can't rely on 'combo', so check if vsync enabled */
1017 	if (~itv->irqmask & IVTV_IRQ_DEC_VSYNC) {
1018 		ivtv_irq_vsync(itv);
1019 	}
1020 
1021 	if (combo & IVTV_IRQ_ENC_VIM_RST) {
1022 		IVTV_DEBUG_IRQ("VIM RST\n");
1023 		/*ivtv_vapi(itv, CX2341X_ENC_REFRESH_INPUT, 0); */
1024 	}
1025 
1026 	if (combo & IVTV_IRQ_DEC_AUD_MODE_CHG) {
1027 		IVTV_DEBUG_INFO("Stereo mode changed\n");
1028 	}
1029 
1030 	if ((combo & IVTV_IRQ_DMA) && !test_bit(IVTV_F_I_DMA, &itv->i_flags)) {
1031 		itv->irq_rr_idx++;
1032 		for (i = 0; i < IVTV_MAX_STREAMS; i++) {
1033 			int idx = (i + itv->irq_rr_idx) % IVTV_MAX_STREAMS;
1034 			struct ivtv_stream *s = &itv->streams[idx];
1035 
1036 			if (!test_and_clear_bit(IVTV_F_S_DMA_PENDING, &s->s_flags))
1037 				continue;
1038 			if (s->type >= IVTV_DEC_STREAM_TYPE_MPG)
1039 				ivtv_dma_dec_start(s);
1040 			else
1041 				ivtv_dma_enc_start(s);
1042 			break;
1043 		}
1044 
1045 		if (i == IVTV_MAX_STREAMS &&
1046 		    test_bit(IVTV_F_I_UDMA_PENDING, &itv->i_flags))
1047 			ivtv_udma_start(itv);
1048 	}
1049 
1050 	if ((combo & IVTV_IRQ_DMA) && !test_bit(IVTV_F_I_PIO, &itv->i_flags)) {
1051 		itv->irq_rr_idx++;
1052 		for (i = 0; i < IVTV_MAX_STREAMS; i++) {
1053 			int idx = (i + itv->irq_rr_idx) % IVTV_MAX_STREAMS;
1054 			struct ivtv_stream *s = &itv->streams[idx];
1055 
1056 			if (!test_and_clear_bit(IVTV_F_S_PIO_PENDING, &s->s_flags))
1057 				continue;
1058 			if (s->type == IVTV_DEC_STREAM_TYPE_VBI || s->type < IVTV_DEC_STREAM_TYPE_MPG)
1059 				ivtv_dma_enc_start(s);
1060 			break;
1061 		}
1062 	}
1063 
1064 	if (test_and_clear_bit(IVTV_F_I_HAVE_WORK, &itv->i_flags)) {
1065 		kthread_queue_work(&itv->irq_worker, &itv->irq_work);
1066 	}
1067 
1068 	spin_unlock(&itv->dma_reg_lock);
1069 
1070 	/* If we've just handled a 'forced' vsync, it's safest to say it
1071 	 * wasn't ours. Another device may have triggered it at just
1072 	 * the right time.
1073 	 */
1074 	return vsync_force ? IRQ_NONE : IRQ_HANDLED;
1075 }
1076 
1077 void ivtv_unfinished_dma(struct timer_list *t)
1078 {
1079 	struct ivtv *itv = from_timer(itv, t, dma_timer);
1080 
1081 	if (!test_bit(IVTV_F_I_DMA, &itv->i_flags))
1082 		return;
1083 	IVTV_ERR("DMA TIMEOUT %08x %d\n", read_reg(IVTV_REG_DMASTATUS), itv->cur_dma_stream);
1084 
1085 	write_reg(read_reg(IVTV_REG_DMASTATUS) & 3, IVTV_REG_DMASTATUS);
1086 	clear_bit(IVTV_F_I_UDMA, &itv->i_flags);
1087 	clear_bit(IVTV_F_I_DMA, &itv->i_flags);
1088 	itv->cur_dma_stream = -1;
1089 	wake_up(&itv->dma_waitq);
1090 }
1091