1 /* 2 buffer queues. 3 Copyright (C) 2003-2004 Kevin Thayer <nufan_wfk at yahoo.com> 4 Copyright (C) 2004 Chris Kennedy <c@groovy.org> 5 Copyright (C) 2005-2007 Hans Verkuil <hverkuil@xs4all.nl> 6 7 This program is free software; you can redistribute it and/or modify 8 it under the terms of the GNU General Public License as published by 9 the Free Software Foundation; either version 2 of the License, or 10 (at your option) any later version. 11 12 This program is distributed in the hope that it will be useful, 13 but WITHOUT ANY WARRANTY; without even the implied warranty of 14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 15 GNU General Public License for more details. 16 17 You should have received a copy of the GNU General Public License 18 along with this program; if not, write to the Free Software 19 Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA 20 */ 21 22 #include "ivtv-driver.h" 23 #include "ivtv-queue.h" 24 25 int ivtv_buf_copy_from_user(struct ivtv_stream *s, struct ivtv_buffer *buf, const char __user *src, int copybytes) 26 { 27 if (s->buf_size - buf->bytesused < copybytes) 28 copybytes = s->buf_size - buf->bytesused; 29 if (copy_from_user(buf->buf + buf->bytesused, src, copybytes)) { 30 return -EFAULT; 31 } 32 buf->bytesused += copybytes; 33 return copybytes; 34 } 35 36 void ivtv_buf_swap(struct ivtv_buffer *buf) 37 { 38 int i; 39 40 for (i = 0; i < buf->bytesused; i += 4) 41 swab32s((u32 *)(buf->buf + i)); 42 } 43 44 void ivtv_queue_init(struct ivtv_queue *q) 45 { 46 INIT_LIST_HEAD(&q->list); 47 q->buffers = 0; 48 q->length = 0; 49 q->bytesused = 0; 50 } 51 52 void ivtv_enqueue(struct ivtv_stream *s, struct ivtv_buffer *buf, struct ivtv_queue *q) 53 { 54 unsigned long flags; 55 56 /* clear the buffer if it is going to be enqueued to the free queue */ 57 if (q == &s->q_free) { 58 buf->bytesused = 0; 59 buf->readpos = 0; 60 buf->b_flags = 0; 61 buf->dma_xfer_cnt = 0; 62 } 63 spin_lock_irqsave(&s->qlock, flags); 64 list_add_tail(&buf->list, &q->list); 65 q->buffers++; 66 q->length += s->buf_size; 67 q->bytesused += buf->bytesused - buf->readpos; 68 spin_unlock_irqrestore(&s->qlock, flags); 69 } 70 71 struct ivtv_buffer *ivtv_dequeue(struct ivtv_stream *s, struct ivtv_queue *q) 72 { 73 struct ivtv_buffer *buf = NULL; 74 unsigned long flags; 75 76 spin_lock_irqsave(&s->qlock, flags); 77 if (!list_empty(&q->list)) { 78 buf = list_entry(q->list.next, struct ivtv_buffer, list); 79 list_del_init(q->list.next); 80 q->buffers--; 81 q->length -= s->buf_size; 82 q->bytesused -= buf->bytesused - buf->readpos; 83 } 84 spin_unlock_irqrestore(&s->qlock, flags); 85 return buf; 86 } 87 88 static void ivtv_queue_move_buf(struct ivtv_stream *s, struct ivtv_queue *from, 89 struct ivtv_queue *to, int clear) 90 { 91 struct ivtv_buffer *buf = list_entry(from->list.next, struct ivtv_buffer, list); 92 93 list_move_tail(from->list.next, &to->list); 94 from->buffers--; 95 from->length -= s->buf_size; 96 from->bytesused -= buf->bytesused - buf->readpos; 97 /* special handling for q_free */ 98 if (clear) 99 buf->bytesused = buf->readpos = buf->b_flags = buf->dma_xfer_cnt = 0; 100 to->buffers++; 101 to->length += s->buf_size; 102 to->bytesused += buf->bytesused - buf->readpos; 103 } 104 105 /* Move 'needed_bytes' worth of buffers from queue 'from' into queue 'to'. 106 If 'needed_bytes' == 0, then move all buffers from 'from' into 'to'. 107 If 'steal' != NULL, then buffers may also taken from that queue if 108 needed, but only if 'from' is the free queue. 109 110 The buffer is automatically cleared if it goes to the free queue. It is 111 also cleared if buffers need to be taken from the 'steal' queue and 112 the 'from' queue is the free queue. 113 114 When 'from' is q_free, then needed_bytes is compared to the total 115 available buffer length, otherwise needed_bytes is compared to the 116 bytesused value. For the 'steal' queue the total available buffer 117 length is always used. 118 119 -ENOMEM is returned if the buffers could not be obtained, 0 if all 120 buffers where obtained from the 'from' list and if non-zero then 121 the number of stolen buffers is returned. */ 122 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, 123 struct ivtv_queue *to, int needed_bytes) 124 { 125 unsigned long flags; 126 int rc = 0; 127 int from_free = from == &s->q_free; 128 int to_free = to == &s->q_free; 129 int bytes_available, bytes_steal; 130 131 spin_lock_irqsave(&s->qlock, flags); 132 if (needed_bytes == 0) { 133 from_free = 1; 134 needed_bytes = from->length; 135 } 136 137 bytes_available = from_free ? from->length : from->bytesused; 138 bytes_steal = (from_free && steal) ? steal->length : 0; 139 140 if (bytes_available + bytes_steal < needed_bytes) { 141 spin_unlock_irqrestore(&s->qlock, flags); 142 return -ENOMEM; 143 } 144 while (steal && bytes_available < needed_bytes) { 145 struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); 146 u16 dma_xfer_cnt = buf->dma_xfer_cnt; 147 148 /* move buffers from the tail of the 'steal' queue to the tail of the 149 'from' queue. Always copy all the buffers with the same dma_xfer_cnt 150 value, this ensures that you do not end up with partial frame data 151 if one frame is stored in multiple buffers. */ 152 while (dma_xfer_cnt == buf->dma_xfer_cnt) { 153 list_move_tail(steal->list.prev, &from->list); 154 rc++; 155 steal->buffers--; 156 steal->length -= s->buf_size; 157 steal->bytesused -= buf->bytesused - buf->readpos; 158 buf->bytesused = buf->readpos = buf->b_flags = buf->dma_xfer_cnt = 0; 159 from->buffers++; 160 from->length += s->buf_size; 161 bytes_available += s->buf_size; 162 if (list_empty(&steal->list)) 163 break; 164 buf = list_entry(steal->list.prev, struct ivtv_buffer, list); 165 } 166 } 167 if (from_free) { 168 u32 old_length = to->length; 169 170 while (to->length - old_length < needed_bytes) { 171 ivtv_queue_move_buf(s, from, to, 1); 172 } 173 } 174 else { 175 u32 old_bytesused = to->bytesused; 176 177 while (to->bytesused - old_bytesused < needed_bytes) { 178 ivtv_queue_move_buf(s, from, to, to_free); 179 } 180 } 181 spin_unlock_irqrestore(&s->qlock, flags); 182 return rc; 183 } 184 185 void ivtv_flush_queues(struct ivtv_stream *s) 186 { 187 ivtv_queue_move(s, &s->q_io, NULL, &s->q_free, 0); 188 ivtv_queue_move(s, &s->q_full, NULL, &s->q_free, 0); 189 ivtv_queue_move(s, &s->q_dma, NULL, &s->q_free, 0); 190 ivtv_queue_move(s, &s->q_predma, NULL, &s->q_free, 0); 191 } 192 193 int ivtv_stream_alloc(struct ivtv_stream *s) 194 { 195 struct ivtv *itv = s->itv; 196 int SGsize = sizeof(struct ivtv_sg_host_element) * s->buffers; 197 int i; 198 199 if (s->buffers == 0) 200 return 0; 201 202 IVTV_DEBUG_INFO("Allocate %s%s stream: %d x %d buffers (%dkB total)\n", 203 s->dma != PCI_DMA_NONE ? "DMA " : "", 204 s->name, s->buffers, s->buf_size, s->buffers * s->buf_size / 1024); 205 206 s->sg_pending = kzalloc(SGsize, GFP_KERNEL|__GFP_NOWARN); 207 if (s->sg_pending == NULL) { 208 IVTV_ERR("Could not allocate sg_pending for %s stream\n", s->name); 209 return -ENOMEM; 210 } 211 s->sg_pending_size = 0; 212 213 s->sg_processing = kzalloc(SGsize, GFP_KERNEL|__GFP_NOWARN); 214 if (s->sg_processing == NULL) { 215 IVTV_ERR("Could not allocate sg_processing for %s stream\n", s->name); 216 kfree(s->sg_pending); 217 s->sg_pending = NULL; 218 return -ENOMEM; 219 } 220 s->sg_processing_size = 0; 221 222 s->sg_dma = kzalloc(sizeof(struct ivtv_sg_element), 223 GFP_KERNEL|__GFP_NOWARN); 224 if (s->sg_dma == NULL) { 225 IVTV_ERR("Could not allocate sg_dma for %s stream\n", s->name); 226 kfree(s->sg_pending); 227 s->sg_pending = NULL; 228 kfree(s->sg_processing); 229 s->sg_processing = NULL; 230 return -ENOMEM; 231 } 232 if (ivtv_might_use_dma(s)) { 233 s->sg_handle = pci_map_single(itv->pdev, s->sg_dma, 234 sizeof(struct ivtv_sg_element), PCI_DMA_TODEVICE); 235 ivtv_stream_sync_for_cpu(s); 236 } 237 238 /* allocate stream buffers. Initially all buffers are in q_free. */ 239 for (i = 0; i < s->buffers; i++) { 240 struct ivtv_buffer *buf = kzalloc(sizeof(struct ivtv_buffer), 241 GFP_KERNEL|__GFP_NOWARN); 242 243 if (buf == NULL) 244 break; 245 buf->buf = kmalloc(s->buf_size + 256, GFP_KERNEL|__GFP_NOWARN); 246 if (buf->buf == NULL) { 247 kfree(buf); 248 break; 249 } 250 INIT_LIST_HEAD(&buf->list); 251 if (ivtv_might_use_dma(s)) { 252 buf->dma_handle = pci_map_single(s->itv->pdev, 253 buf->buf, s->buf_size + 256, s->dma); 254 ivtv_buf_sync_for_cpu(s, buf); 255 } 256 ivtv_enqueue(s, buf, &s->q_free); 257 } 258 if (i == s->buffers) 259 return 0; 260 IVTV_ERR("Couldn't allocate buffers for %s stream\n", s->name); 261 ivtv_stream_free(s); 262 return -ENOMEM; 263 } 264 265 void ivtv_stream_free(struct ivtv_stream *s) 266 { 267 struct ivtv_buffer *buf; 268 269 /* move all buffers to q_free */ 270 ivtv_flush_queues(s); 271 272 /* empty q_free */ 273 while ((buf = ivtv_dequeue(s, &s->q_free))) { 274 if (ivtv_might_use_dma(s)) 275 pci_unmap_single(s->itv->pdev, buf->dma_handle, 276 s->buf_size + 256, s->dma); 277 kfree(buf->buf); 278 kfree(buf); 279 } 280 281 /* Free SG Array/Lists */ 282 if (s->sg_dma != NULL) { 283 if (s->sg_handle != IVTV_DMA_UNMAPPED) { 284 pci_unmap_single(s->itv->pdev, s->sg_handle, 285 sizeof(struct ivtv_sg_element), PCI_DMA_TODEVICE); 286 s->sg_handle = IVTV_DMA_UNMAPPED; 287 } 288 kfree(s->sg_pending); 289 kfree(s->sg_processing); 290 kfree(s->sg_dma); 291 s->sg_pending = NULL; 292 s->sg_processing = NULL; 293 s->sg_dma = NULL; 294 s->sg_pending_size = 0; 295 s->sg_processing_size = 0; 296 } 297 } 298