1 /* 2 * Copyright 2013-2015 Analog Devices Inc. 3 * Author: Lars-Peter Clausen <lars@metafoo.de> 4 * 5 * Licensed under the GPL-2. 6 */ 7 8 #include <linux/slab.h> 9 #include <linux/kernel.h> 10 #include <linux/module.h> 11 #include <linux/device.h> 12 #include <linux/workqueue.h> 13 #include <linux/mutex.h> 14 #include <linux/sched.h> 15 #include <linux/poll.h> 16 #include <linux/iio/buffer.h> 17 #include <linux/iio/buffer-dma.h> 18 #include <linux/dma-mapping.h> 19 #include <linux/sizes.h> 20 21 /* 22 * For DMA buffers the storage is sub-divided into so called blocks. Each block 23 * has its own memory buffer. The size of the block is the granularity at which 24 * memory is exchanged between the hardware and the application. Increasing the 25 * basic unit of data exchange from one sample to one block decreases the 26 * management overhead that is associated with each sample. E.g. if we say the 27 * management overhead for one exchange is x and the unit of exchange is one 28 * sample the overhead will be x for each sample. Whereas when using a block 29 * which contains n samples the overhead per sample is reduced to x/n. This 30 * allows to achieve much higher samplerates than what can be sustained with 31 * the one sample approach. 32 * 33 * Blocks are exchanged between the DMA controller and the application via the 34 * means of two queues. The incoming queue and the outgoing queue. Blocks on the 35 * incoming queue are waiting for the DMA controller to pick them up and fill 36 * them with data. Block on the outgoing queue have been filled with data and 37 * are waiting for the application to dequeue them and read the data. 38 * 39 * A block can be in one of the following states: 40 * * Owned by the application. In this state the application can read data from 41 * the block. 42 * * On the incoming list: Blocks on the incoming list are queued up to be 43 * processed by the DMA controller. 44 * * Owned by the DMA controller: The DMA controller is processing the block 45 * and filling it with data. 46 * * On the outgoing list: Blocks on the outgoing list have been successfully 47 * processed by the DMA controller and contain data. They can be dequeued by 48 * the application. 49 * * Dead: A block that is dead has been marked as to be freed. It might still 50 * be owned by either the application or the DMA controller at the moment. 51 * But once they are done processing it instead of going to either the 52 * incoming or outgoing queue the block will be freed. 53 * 54 * In addition to this blocks are reference counted and the memory associated 55 * with both the block structure as well as the storage memory for the block 56 * will be freed when the last reference to the block is dropped. This means a 57 * block must not be accessed without holding a reference. 58 * 59 * The iio_dma_buffer implementation provides a generic infrastructure for 60 * managing the blocks. 61 * 62 * A driver for a specific piece of hardware that has DMA capabilities need to 63 * implement the submit() callback from the iio_dma_buffer_ops structure. This 64 * callback is supposed to initiate the DMA transfer copying data from the 65 * converter to the memory region of the block. Once the DMA transfer has been 66 * completed the driver must call iio_dma_buffer_block_done() for the completed 67 * block. 68 * 69 * Prior to this it must set the bytes_used field of the block contains 70 * the actual number of bytes in the buffer. Typically this will be equal to the 71 * size of the block, but if the DMA hardware has certain alignment requirements 72 * for the transfer length it might choose to use less than the full size. In 73 * either case it is expected that bytes_used is a multiple of the bytes per 74 * datum, i.e. the block must not contain partial samples. 75 * 76 * The driver must call iio_dma_buffer_block_done() for each block it has 77 * received through its submit_block() callback, even if it does not actually 78 * perform a DMA transfer for the block, e.g. because the buffer was disabled 79 * before the block transfer was started. In this case it should set bytes_used 80 * to 0. 81 * 82 * In addition it is recommended that a driver implements the abort() callback. 83 * It will be called when the buffer is disabled and can be used to cancel 84 * pending and stop active transfers. 85 * 86 * The specific driver implementation should use the default callback 87 * implementations provided by this module for the iio_buffer_access_funcs 88 * struct. It may overload some callbacks with custom variants if the hardware 89 * has special requirements that are not handled by the generic functions. If a 90 * driver chooses to overload a callback it has to ensure that the generic 91 * callback is called from within the custom callback. 92 */ 93 94 static void iio_buffer_block_release(struct kref *kref) 95 { 96 struct iio_dma_buffer_block *block = container_of(kref, 97 struct iio_dma_buffer_block, kref); 98 99 WARN_ON(block->state != IIO_BLOCK_STATE_DEAD); 100 101 dma_free_coherent(block->queue->dev, PAGE_ALIGN(block->size), 102 block->vaddr, block->phys_addr); 103 104 iio_buffer_put(&block->queue->buffer); 105 kfree(block); 106 } 107 108 static void iio_buffer_block_get(struct iio_dma_buffer_block *block) 109 { 110 kref_get(&block->kref); 111 } 112 113 static void iio_buffer_block_put(struct iio_dma_buffer_block *block) 114 { 115 kref_put(&block->kref, iio_buffer_block_release); 116 } 117 118 /* 119 * dma_free_coherent can sleep, hence we need to take some special care to be 120 * able to drop a reference from an atomic context. 121 */ 122 static LIST_HEAD(iio_dma_buffer_dead_blocks); 123 static DEFINE_SPINLOCK(iio_dma_buffer_dead_blocks_lock); 124 125 static void iio_dma_buffer_cleanup_worker(struct work_struct *work) 126 { 127 struct iio_dma_buffer_block *block, *_block; 128 LIST_HEAD(block_list); 129 130 spin_lock_irq(&iio_dma_buffer_dead_blocks_lock); 131 list_splice_tail_init(&iio_dma_buffer_dead_blocks, &block_list); 132 spin_unlock_irq(&iio_dma_buffer_dead_blocks_lock); 133 134 list_for_each_entry_safe(block, _block, &block_list, head) 135 iio_buffer_block_release(&block->kref); 136 } 137 static DECLARE_WORK(iio_dma_buffer_cleanup_work, iio_dma_buffer_cleanup_worker); 138 139 static void iio_buffer_block_release_atomic(struct kref *kref) 140 { 141 struct iio_dma_buffer_block *block; 142 unsigned long flags; 143 144 block = container_of(kref, struct iio_dma_buffer_block, kref); 145 146 spin_lock_irqsave(&iio_dma_buffer_dead_blocks_lock, flags); 147 list_add_tail(&block->head, &iio_dma_buffer_dead_blocks); 148 spin_unlock_irqrestore(&iio_dma_buffer_dead_blocks_lock, flags); 149 150 schedule_work(&iio_dma_buffer_cleanup_work); 151 } 152 153 /* 154 * Version of iio_buffer_block_put() that can be called from atomic context 155 */ 156 static void iio_buffer_block_put_atomic(struct iio_dma_buffer_block *block) 157 { 158 kref_put(&block->kref, iio_buffer_block_release_atomic); 159 } 160 161 static struct iio_dma_buffer_queue *iio_buffer_to_queue(struct iio_buffer *buf) 162 { 163 return container_of(buf, struct iio_dma_buffer_queue, buffer); 164 } 165 166 static struct iio_dma_buffer_block *iio_dma_buffer_alloc_block( 167 struct iio_dma_buffer_queue *queue, size_t size) 168 { 169 struct iio_dma_buffer_block *block; 170 171 block = kzalloc(sizeof(*block), GFP_KERNEL); 172 if (!block) 173 return NULL; 174 175 block->vaddr = dma_alloc_coherent(queue->dev, PAGE_ALIGN(size), 176 &block->phys_addr, GFP_KERNEL); 177 if (!block->vaddr) { 178 kfree(block); 179 return NULL; 180 } 181 182 block->size = size; 183 block->state = IIO_BLOCK_STATE_DEQUEUED; 184 block->queue = queue; 185 INIT_LIST_HEAD(&block->head); 186 kref_init(&block->kref); 187 188 iio_buffer_get(&queue->buffer); 189 190 return block; 191 } 192 193 static void _iio_dma_buffer_block_done(struct iio_dma_buffer_block *block) 194 { 195 struct iio_dma_buffer_queue *queue = block->queue; 196 197 /* 198 * The buffer has already been freed by the application, just drop the 199 * reference. 200 */ 201 if (block->state != IIO_BLOCK_STATE_DEAD) { 202 block->state = IIO_BLOCK_STATE_DONE; 203 list_add_tail(&block->head, &queue->outgoing); 204 } 205 } 206 207 /** 208 * iio_dma_buffer_block_done() - Indicate that a block has been completed 209 * @block: The completed block 210 * 211 * Should be called when the DMA controller has finished handling the block to 212 * pass back ownership of the block to the queue. 213 */ 214 void iio_dma_buffer_block_done(struct iio_dma_buffer_block *block) 215 { 216 struct iio_dma_buffer_queue *queue = block->queue; 217 unsigned long flags; 218 219 spin_lock_irqsave(&queue->list_lock, flags); 220 _iio_dma_buffer_block_done(block); 221 spin_unlock_irqrestore(&queue->list_lock, flags); 222 223 iio_buffer_block_put_atomic(block); 224 wake_up_interruptible_poll(&queue->buffer.pollq, POLLIN | POLLRDNORM); 225 } 226 EXPORT_SYMBOL_GPL(iio_dma_buffer_block_done); 227 228 /** 229 * iio_dma_buffer_block_list_abort() - Indicate that a list block has been 230 * aborted 231 * @queue: Queue for which to complete blocks. 232 * @list: List of aborted blocks. All blocks in this list must be from @queue. 233 * 234 * Typically called from the abort() callback after the DMA controller has been 235 * stopped. This will set bytes_used to 0 for each block in the list and then 236 * hand the blocks back to the queue. 237 */ 238 void iio_dma_buffer_block_list_abort(struct iio_dma_buffer_queue *queue, 239 struct list_head *list) 240 { 241 struct iio_dma_buffer_block *block, *_block; 242 unsigned long flags; 243 244 spin_lock_irqsave(&queue->list_lock, flags); 245 list_for_each_entry_safe(block, _block, list, head) { 246 list_del(&block->head); 247 block->bytes_used = 0; 248 _iio_dma_buffer_block_done(block); 249 iio_buffer_block_put_atomic(block); 250 } 251 spin_unlock_irqrestore(&queue->list_lock, flags); 252 253 wake_up_interruptible_poll(&queue->buffer.pollq, POLLIN | POLLRDNORM); 254 } 255 EXPORT_SYMBOL_GPL(iio_dma_buffer_block_list_abort); 256 257 static bool iio_dma_block_reusable(struct iio_dma_buffer_block *block) 258 { 259 /* 260 * If the core owns the block it can be re-used. This should be the 261 * default case when enabling the buffer, unless the DMA controller does 262 * not support abort and has not given back the block yet. 263 */ 264 switch (block->state) { 265 case IIO_BLOCK_STATE_DEQUEUED: 266 case IIO_BLOCK_STATE_QUEUED: 267 case IIO_BLOCK_STATE_DONE: 268 return true; 269 default: 270 return false; 271 } 272 } 273 274 /** 275 * iio_dma_buffer_request_update() - DMA buffer request_update callback 276 * @buffer: The buffer which to request an update 277 * 278 * Should be used as the iio_dma_buffer_request_update() callback for 279 * iio_buffer_access_ops struct for DMA buffers. 280 */ 281 int iio_dma_buffer_request_update(struct iio_buffer *buffer) 282 { 283 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); 284 struct iio_dma_buffer_block *block; 285 bool try_reuse = false; 286 size_t size; 287 int ret = 0; 288 int i; 289 290 /* 291 * Split the buffer into two even parts. This is used as a double 292 * buffering scheme with usually one block at a time being used by the 293 * DMA and the other one by the application. 294 */ 295 size = DIV_ROUND_UP(queue->buffer.bytes_per_datum * 296 queue->buffer.length, 2); 297 298 mutex_lock(&queue->lock); 299 300 /* Allocations are page aligned */ 301 if (PAGE_ALIGN(queue->fileio.block_size) == PAGE_ALIGN(size)) 302 try_reuse = true; 303 304 queue->fileio.block_size = size; 305 queue->fileio.active_block = NULL; 306 307 spin_lock_irq(&queue->list_lock); 308 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { 309 block = queue->fileio.blocks[i]; 310 311 /* If we can't re-use it free it */ 312 if (block && (!iio_dma_block_reusable(block) || !try_reuse)) 313 block->state = IIO_BLOCK_STATE_DEAD; 314 } 315 316 /* 317 * At this point all blocks are either owned by the core or marked as 318 * dead. This means we can reset the lists without having to fear 319 * corrution. 320 */ 321 INIT_LIST_HEAD(&queue->outgoing); 322 spin_unlock_irq(&queue->list_lock); 323 324 INIT_LIST_HEAD(&queue->incoming); 325 326 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { 327 if (queue->fileio.blocks[i]) { 328 block = queue->fileio.blocks[i]; 329 if (block->state == IIO_BLOCK_STATE_DEAD) { 330 /* Could not reuse it */ 331 iio_buffer_block_put(block); 332 block = NULL; 333 } else { 334 block->size = size; 335 } 336 } else { 337 block = NULL; 338 } 339 340 if (!block) { 341 block = iio_dma_buffer_alloc_block(queue, size); 342 if (!block) { 343 ret = -ENOMEM; 344 goto out_unlock; 345 } 346 queue->fileio.blocks[i] = block; 347 } 348 349 block->state = IIO_BLOCK_STATE_QUEUED; 350 list_add_tail(&block->head, &queue->incoming); 351 } 352 353 out_unlock: 354 mutex_unlock(&queue->lock); 355 356 return ret; 357 } 358 EXPORT_SYMBOL_GPL(iio_dma_buffer_request_update); 359 360 static void iio_dma_buffer_submit_block(struct iio_dma_buffer_queue *queue, 361 struct iio_dma_buffer_block *block) 362 { 363 int ret; 364 365 /* 366 * If the hardware has already been removed we put the block into 367 * limbo. It will neither be on the incoming nor outgoing list, nor will 368 * it ever complete. It will just wait to be freed eventually. 369 */ 370 if (!queue->ops) 371 return; 372 373 block->state = IIO_BLOCK_STATE_ACTIVE; 374 iio_buffer_block_get(block); 375 ret = queue->ops->submit(queue, block); 376 if (ret) { 377 /* 378 * This is a bit of a problem and there is not much we can do 379 * other then wait for the buffer to be disabled and re-enabled 380 * and try again. But it should not really happen unless we run 381 * out of memory or something similar. 382 * 383 * TODO: Implement support in the IIO core to allow buffers to 384 * notify consumers that something went wrong and the buffer 385 * should be disabled. 386 */ 387 iio_buffer_block_put(block); 388 } 389 } 390 391 /** 392 * iio_dma_buffer_enable() - Enable DMA buffer 393 * @buffer: IIO buffer to enable 394 * @indio_dev: IIO device the buffer is attached to 395 * 396 * Needs to be called when the device that the buffer is attached to starts 397 * sampling. Typically should be the iio_buffer_access_ops enable callback. 398 * 399 * This will allocate the DMA buffers and start the DMA transfers. 400 */ 401 int iio_dma_buffer_enable(struct iio_buffer *buffer, 402 struct iio_dev *indio_dev) 403 { 404 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); 405 struct iio_dma_buffer_block *block, *_block; 406 407 mutex_lock(&queue->lock); 408 queue->active = true; 409 list_for_each_entry_safe(block, _block, &queue->incoming, head) { 410 list_del(&block->head); 411 iio_dma_buffer_submit_block(queue, block); 412 } 413 mutex_unlock(&queue->lock); 414 415 return 0; 416 } 417 EXPORT_SYMBOL_GPL(iio_dma_buffer_enable); 418 419 /** 420 * iio_dma_buffer_disable() - Disable DMA buffer 421 * @buffer: IIO DMA buffer to disable 422 * @indio_dev: IIO device the buffer is attached to 423 * 424 * Needs to be called when the device that the buffer is attached to stops 425 * sampling. Typically should be the iio_buffer_access_ops disable callback. 426 */ 427 int iio_dma_buffer_disable(struct iio_buffer *buffer, 428 struct iio_dev *indio_dev) 429 { 430 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); 431 432 mutex_lock(&queue->lock); 433 queue->active = false; 434 435 if (queue->ops && queue->ops->abort) 436 queue->ops->abort(queue); 437 mutex_unlock(&queue->lock); 438 439 return 0; 440 } 441 EXPORT_SYMBOL_GPL(iio_dma_buffer_disable); 442 443 static void iio_dma_buffer_enqueue(struct iio_dma_buffer_queue *queue, 444 struct iio_dma_buffer_block *block) 445 { 446 if (block->state == IIO_BLOCK_STATE_DEAD) { 447 iio_buffer_block_put(block); 448 } else if (queue->active) { 449 iio_dma_buffer_submit_block(queue, block); 450 } else { 451 block->state = IIO_BLOCK_STATE_QUEUED; 452 list_add_tail(&block->head, &queue->incoming); 453 } 454 } 455 456 static struct iio_dma_buffer_block *iio_dma_buffer_dequeue( 457 struct iio_dma_buffer_queue *queue) 458 { 459 struct iio_dma_buffer_block *block; 460 461 spin_lock_irq(&queue->list_lock); 462 block = list_first_entry_or_null(&queue->outgoing, struct 463 iio_dma_buffer_block, head); 464 if (block != NULL) { 465 list_del(&block->head); 466 block->state = IIO_BLOCK_STATE_DEQUEUED; 467 } 468 spin_unlock_irq(&queue->list_lock); 469 470 return block; 471 } 472 473 /** 474 * iio_dma_buffer_read() - DMA buffer read callback 475 * @buffer: Buffer to read form 476 * @n: Number of bytes to read 477 * @user_buffer: Userspace buffer to copy the data to 478 * 479 * Should be used as the read_first_n callback for iio_buffer_access_ops 480 * struct for DMA buffers. 481 */ 482 int iio_dma_buffer_read(struct iio_buffer *buffer, size_t n, 483 char __user *user_buffer) 484 { 485 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer); 486 struct iio_dma_buffer_block *block; 487 int ret; 488 489 if (n < buffer->bytes_per_datum) 490 return -EINVAL; 491 492 mutex_lock(&queue->lock); 493 494 if (!queue->fileio.active_block) { 495 block = iio_dma_buffer_dequeue(queue); 496 if (block == NULL) { 497 ret = 0; 498 goto out_unlock; 499 } 500 queue->fileio.pos = 0; 501 queue->fileio.active_block = block; 502 } else { 503 block = queue->fileio.active_block; 504 } 505 506 n = rounddown(n, buffer->bytes_per_datum); 507 if (n > block->bytes_used - queue->fileio.pos) 508 n = block->bytes_used - queue->fileio.pos; 509 510 if (copy_to_user(user_buffer, block->vaddr + queue->fileio.pos, n)) { 511 ret = -EFAULT; 512 goto out_unlock; 513 } 514 515 queue->fileio.pos += n; 516 517 if (queue->fileio.pos == block->bytes_used) { 518 queue->fileio.active_block = NULL; 519 iio_dma_buffer_enqueue(queue, block); 520 } 521 522 ret = n; 523 524 out_unlock: 525 mutex_unlock(&queue->lock); 526 527 return ret; 528 } 529 EXPORT_SYMBOL_GPL(iio_dma_buffer_read); 530 531 /** 532 * iio_dma_buffer_data_available() - DMA buffer data_available callback 533 * @buf: Buffer to check for data availability 534 * 535 * Should be used as the data_available callback for iio_buffer_access_ops 536 * struct for DMA buffers. 537 */ 538 size_t iio_dma_buffer_data_available(struct iio_buffer *buf) 539 { 540 struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buf); 541 struct iio_dma_buffer_block *block; 542 size_t data_available = 0; 543 544 /* 545 * For counting the available bytes we'll use the size of the block not 546 * the number of actual bytes available in the block. Otherwise it is 547 * possible that we end up with a value that is lower than the watermark 548 * but won't increase since all blocks are in use. 549 */ 550 551 mutex_lock(&queue->lock); 552 if (queue->fileio.active_block) 553 data_available += queue->fileio.active_block->size; 554 555 spin_lock_irq(&queue->list_lock); 556 list_for_each_entry(block, &queue->outgoing, head) 557 data_available += block->size; 558 spin_unlock_irq(&queue->list_lock); 559 mutex_unlock(&queue->lock); 560 561 return data_available; 562 } 563 EXPORT_SYMBOL_GPL(iio_dma_buffer_data_available); 564 565 /** 566 * iio_dma_buffer_set_bytes_per_datum() - DMA buffer set_bytes_per_datum callback 567 * @buffer: Buffer to set the bytes-per-datum for 568 * @bpd: The new bytes-per-datum value 569 * 570 * Should be used as the set_bytes_per_datum callback for iio_buffer_access_ops 571 * struct for DMA buffers. 572 */ 573 int iio_dma_buffer_set_bytes_per_datum(struct iio_buffer *buffer, size_t bpd) 574 { 575 buffer->bytes_per_datum = bpd; 576 577 return 0; 578 } 579 EXPORT_SYMBOL_GPL(iio_dma_buffer_set_bytes_per_datum); 580 581 /** 582 * iio_dma_buffer_set_length - DMA buffer set_length callback 583 * @buffer: Buffer to set the length for 584 * @length: The new buffer length 585 * 586 * Should be used as the set_length callback for iio_buffer_access_ops 587 * struct for DMA buffers. 588 */ 589 int iio_dma_buffer_set_length(struct iio_buffer *buffer, int length) 590 { 591 /* Avoid an invalid state */ 592 if (length < 2) 593 length = 2; 594 buffer->length = length; 595 buffer->watermark = length / 2; 596 597 return 0; 598 } 599 EXPORT_SYMBOL_GPL(iio_dma_buffer_set_length); 600 601 /** 602 * iio_dma_buffer_init() - Initialize DMA buffer queue 603 * @queue: Buffer to initialize 604 * @dev: DMA device 605 * @ops: DMA buffer queue callback operations 606 * 607 * The DMA device will be used by the queue to do DMA memory allocations. So it 608 * should refer to the device that will perform the DMA to ensure that 609 * allocations are done from a memory region that can be accessed by the device. 610 */ 611 int iio_dma_buffer_init(struct iio_dma_buffer_queue *queue, 612 struct device *dev, const struct iio_dma_buffer_ops *ops) 613 { 614 iio_buffer_init(&queue->buffer); 615 queue->buffer.length = PAGE_SIZE; 616 queue->buffer.watermark = queue->buffer.length / 2; 617 queue->dev = dev; 618 queue->ops = ops; 619 620 INIT_LIST_HEAD(&queue->incoming); 621 INIT_LIST_HEAD(&queue->outgoing); 622 623 mutex_init(&queue->lock); 624 spin_lock_init(&queue->list_lock); 625 626 return 0; 627 } 628 EXPORT_SYMBOL_GPL(iio_dma_buffer_init); 629 630 /** 631 * iio_dma_buffer_exit() - Cleanup DMA buffer queue 632 * @queue: Buffer to cleanup 633 * 634 * After this function has completed it is safe to free any resources that are 635 * associated with the buffer and are accessed inside the callback operations. 636 */ 637 void iio_dma_buffer_exit(struct iio_dma_buffer_queue *queue) 638 { 639 unsigned int i; 640 641 mutex_lock(&queue->lock); 642 643 spin_lock_irq(&queue->list_lock); 644 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { 645 if (!queue->fileio.blocks[i]) 646 continue; 647 queue->fileio.blocks[i]->state = IIO_BLOCK_STATE_DEAD; 648 } 649 INIT_LIST_HEAD(&queue->outgoing); 650 spin_unlock_irq(&queue->list_lock); 651 652 INIT_LIST_HEAD(&queue->incoming); 653 654 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { 655 if (!queue->fileio.blocks[i]) 656 continue; 657 iio_buffer_block_put(queue->fileio.blocks[i]); 658 queue->fileio.blocks[i] = NULL; 659 } 660 queue->fileio.active_block = NULL; 661 queue->ops = NULL; 662 663 mutex_unlock(&queue->lock); 664 } 665 EXPORT_SYMBOL_GPL(iio_dma_buffer_exit); 666 667 /** 668 * iio_dma_buffer_release() - Release final buffer resources 669 * @queue: Buffer to release 670 * 671 * Frees resources that can't yet be freed in iio_dma_buffer_exit(). Should be 672 * called in the buffers release callback implementation right before freeing 673 * the memory associated with the buffer. 674 */ 675 void iio_dma_buffer_release(struct iio_dma_buffer_queue *queue) 676 { 677 mutex_destroy(&queue->lock); 678 } 679 EXPORT_SYMBOL_GPL(iio_dma_buffer_release); 680 681 MODULE_AUTHOR("Lars-Peter Clausen <lars@metafoo.de>"); 682 MODULE_DESCRIPTION("DMA buffer for the IIO framework"); 683 MODULE_LICENSE("GPL v2"); 684