1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Digital Audio (PCM) abstract layer 4 * Copyright (c) by Jaroslav Kysela <perex@perex.cz> 5 * Abramo Bagnara <abramo@alsa-project.org> 6 */ 7 8 #include <linux/slab.h> 9 #include <linux/sched/signal.h> 10 #include <linux/time.h> 11 #include <linux/math64.h> 12 #include <linux/export.h> 13 #include <sound/core.h> 14 #include <sound/control.h> 15 #include <sound/tlv.h> 16 #include <sound/info.h> 17 #include <sound/pcm.h> 18 #include <sound/pcm_params.h> 19 #include <sound/timer.h> 20 21 #include "pcm_local.h" 22 23 #ifdef CONFIG_SND_PCM_XRUN_DEBUG 24 #define CREATE_TRACE_POINTS 25 #include "pcm_trace.h" 26 #else 27 #define trace_hwptr(substream, pos, in_interrupt) 28 #define trace_xrun(substream) 29 #define trace_hw_ptr_error(substream, reason) 30 #define trace_applptr(substream, prev, curr) 31 #endif 32 33 static int fill_silence_frames(struct snd_pcm_substream *substream, 34 snd_pcm_uframes_t off, snd_pcm_uframes_t frames); 35 36 /* 37 * fill ring buffer with silence 38 * runtime->silence_start: starting pointer to silence area 39 * runtime->silence_filled: size filled with silence 40 * runtime->silence_threshold: threshold from application 41 * runtime->silence_size: maximal size from application 42 * 43 * when runtime->silence_size >= runtime->boundary - fill processed area with silence immediately 44 */ 45 void snd_pcm_playback_silence(struct snd_pcm_substream *substream, snd_pcm_uframes_t new_hw_ptr) 46 { 47 struct snd_pcm_runtime *runtime = substream->runtime; 48 snd_pcm_uframes_t frames, ofs, transfer; 49 int err; 50 51 if (runtime->silence_size < runtime->boundary) { 52 snd_pcm_sframes_t noise_dist, n; 53 snd_pcm_uframes_t appl_ptr = READ_ONCE(runtime->control->appl_ptr); 54 if (runtime->silence_start != appl_ptr) { 55 n = appl_ptr - runtime->silence_start; 56 if (n < 0) 57 n += runtime->boundary; 58 if ((snd_pcm_uframes_t)n < runtime->silence_filled) 59 runtime->silence_filled -= n; 60 else 61 runtime->silence_filled = 0; 62 runtime->silence_start = appl_ptr; 63 } 64 if (runtime->silence_filled >= runtime->buffer_size) 65 return; 66 noise_dist = snd_pcm_playback_hw_avail(runtime) + runtime->silence_filled; 67 if (noise_dist >= (snd_pcm_sframes_t) runtime->silence_threshold) 68 return; 69 frames = runtime->silence_threshold - noise_dist; 70 if (frames > runtime->silence_size) 71 frames = runtime->silence_size; 72 } else { 73 if (new_hw_ptr == ULONG_MAX) { /* initialization */ 74 snd_pcm_sframes_t avail = snd_pcm_playback_hw_avail(runtime); 75 if (avail > runtime->buffer_size) 76 avail = runtime->buffer_size; 77 runtime->silence_filled = avail > 0 ? avail : 0; 78 runtime->silence_start = (runtime->status->hw_ptr + 79 runtime->silence_filled) % 80 runtime->boundary; 81 } else { 82 ofs = runtime->status->hw_ptr; 83 frames = new_hw_ptr - ofs; 84 if ((snd_pcm_sframes_t)frames < 0) 85 frames += runtime->boundary; 86 runtime->silence_filled -= frames; 87 if ((snd_pcm_sframes_t)runtime->silence_filled < 0) { 88 runtime->silence_filled = 0; 89 runtime->silence_start = new_hw_ptr; 90 } else { 91 runtime->silence_start = ofs; 92 } 93 } 94 frames = runtime->buffer_size - runtime->silence_filled; 95 } 96 if (snd_BUG_ON(frames > runtime->buffer_size)) 97 return; 98 if (frames == 0) 99 return; 100 ofs = runtime->silence_start % runtime->buffer_size; 101 while (frames > 0) { 102 transfer = ofs + frames > runtime->buffer_size ? runtime->buffer_size - ofs : frames; 103 err = fill_silence_frames(substream, ofs, transfer); 104 snd_BUG_ON(err < 0); 105 runtime->silence_filled += transfer; 106 frames -= transfer; 107 ofs = 0; 108 } 109 snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_DEVICE); 110 } 111 112 #ifdef CONFIG_SND_DEBUG 113 void snd_pcm_debug_name(struct snd_pcm_substream *substream, 114 char *name, size_t len) 115 { 116 snprintf(name, len, "pcmC%dD%d%c:%d", 117 substream->pcm->card->number, 118 substream->pcm->device, 119 substream->stream ? 'c' : 'p', 120 substream->number); 121 } 122 EXPORT_SYMBOL(snd_pcm_debug_name); 123 #endif 124 125 #define XRUN_DEBUG_BASIC (1<<0) 126 #define XRUN_DEBUG_STACK (1<<1) /* dump also stack */ 127 #define XRUN_DEBUG_JIFFIESCHECK (1<<2) /* do jiffies check */ 128 129 #ifdef CONFIG_SND_PCM_XRUN_DEBUG 130 131 #define xrun_debug(substream, mask) \ 132 ((substream)->pstr->xrun_debug & (mask)) 133 #else 134 #define xrun_debug(substream, mask) 0 135 #endif 136 137 #define dump_stack_on_xrun(substream) do { \ 138 if (xrun_debug(substream, XRUN_DEBUG_STACK)) \ 139 dump_stack(); \ 140 } while (0) 141 142 /* call with stream lock held */ 143 void __snd_pcm_xrun(struct snd_pcm_substream *substream) 144 { 145 struct snd_pcm_runtime *runtime = substream->runtime; 146 147 trace_xrun(substream); 148 if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { 149 struct timespec64 tstamp; 150 151 snd_pcm_gettime(runtime, &tstamp); 152 runtime->status->tstamp.tv_sec = tstamp.tv_sec; 153 runtime->status->tstamp.tv_nsec = tstamp.tv_nsec; 154 } 155 snd_pcm_stop(substream, SNDRV_PCM_STATE_XRUN); 156 if (xrun_debug(substream, XRUN_DEBUG_BASIC)) { 157 char name[16]; 158 snd_pcm_debug_name(substream, name, sizeof(name)); 159 pcm_warn(substream->pcm, "XRUN: %s\n", name); 160 dump_stack_on_xrun(substream); 161 } 162 } 163 164 #ifdef CONFIG_SND_PCM_XRUN_DEBUG 165 #define hw_ptr_error(substream, in_interrupt, reason, fmt, args...) \ 166 do { \ 167 trace_hw_ptr_error(substream, reason); \ 168 if (xrun_debug(substream, XRUN_DEBUG_BASIC)) { \ 169 pr_err_ratelimited("ALSA: PCM: [%c] " reason ": " fmt, \ 170 (in_interrupt) ? 'Q' : 'P', ##args); \ 171 dump_stack_on_xrun(substream); \ 172 } \ 173 } while (0) 174 175 #else /* ! CONFIG_SND_PCM_XRUN_DEBUG */ 176 177 #define hw_ptr_error(substream, fmt, args...) do { } while (0) 178 179 #endif 180 181 int snd_pcm_update_state(struct snd_pcm_substream *substream, 182 struct snd_pcm_runtime *runtime) 183 { 184 snd_pcm_uframes_t avail; 185 186 avail = snd_pcm_avail(substream); 187 if (avail > runtime->avail_max) 188 runtime->avail_max = avail; 189 if (runtime->status->state == SNDRV_PCM_STATE_DRAINING) { 190 if (avail >= runtime->buffer_size) { 191 snd_pcm_drain_done(substream); 192 return -EPIPE; 193 } 194 } else { 195 if (avail >= runtime->stop_threshold) { 196 __snd_pcm_xrun(substream); 197 return -EPIPE; 198 } 199 } 200 if (runtime->twake) { 201 if (avail >= runtime->twake) 202 wake_up(&runtime->tsleep); 203 } else if (avail >= runtime->control->avail_min) 204 wake_up(&runtime->sleep); 205 return 0; 206 } 207 208 static void update_audio_tstamp(struct snd_pcm_substream *substream, 209 struct timespec64 *curr_tstamp, 210 struct timespec64 *audio_tstamp) 211 { 212 struct snd_pcm_runtime *runtime = substream->runtime; 213 u64 audio_frames, audio_nsecs; 214 struct timespec64 driver_tstamp; 215 216 if (runtime->tstamp_mode != SNDRV_PCM_TSTAMP_ENABLE) 217 return; 218 219 if (!(substream->ops->get_time_info) || 220 (runtime->audio_tstamp_report.actual_type == 221 SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT)) { 222 223 /* 224 * provide audio timestamp derived from pointer position 225 * add delay only if requested 226 */ 227 228 audio_frames = runtime->hw_ptr_wrap + runtime->status->hw_ptr; 229 230 if (runtime->audio_tstamp_config.report_delay) { 231 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 232 audio_frames -= runtime->delay; 233 else 234 audio_frames += runtime->delay; 235 } 236 audio_nsecs = div_u64(audio_frames * 1000000000LL, 237 runtime->rate); 238 *audio_tstamp = ns_to_timespec64(audio_nsecs); 239 } 240 241 if (runtime->status->audio_tstamp.tv_sec != audio_tstamp->tv_sec || 242 runtime->status->audio_tstamp.tv_nsec != audio_tstamp->tv_nsec) { 243 runtime->status->audio_tstamp.tv_sec = audio_tstamp->tv_sec; 244 runtime->status->audio_tstamp.tv_nsec = audio_tstamp->tv_nsec; 245 runtime->status->tstamp.tv_sec = curr_tstamp->tv_sec; 246 runtime->status->tstamp.tv_nsec = curr_tstamp->tv_nsec; 247 } 248 249 250 /* 251 * re-take a driver timestamp to let apps detect if the reference tstamp 252 * read by low-level hardware was provided with a delay 253 */ 254 snd_pcm_gettime(substream->runtime, &driver_tstamp); 255 runtime->driver_tstamp = driver_tstamp; 256 } 257 258 static int snd_pcm_update_hw_ptr0(struct snd_pcm_substream *substream, 259 unsigned int in_interrupt) 260 { 261 struct snd_pcm_runtime *runtime = substream->runtime; 262 snd_pcm_uframes_t pos; 263 snd_pcm_uframes_t old_hw_ptr, new_hw_ptr, hw_base; 264 snd_pcm_sframes_t hdelta, delta; 265 unsigned long jdelta; 266 unsigned long curr_jiffies; 267 struct timespec64 curr_tstamp; 268 struct timespec64 audio_tstamp; 269 int crossed_boundary = 0; 270 271 old_hw_ptr = runtime->status->hw_ptr; 272 273 /* 274 * group pointer, time and jiffies reads to allow for more 275 * accurate correlations/corrections. 276 * The values are stored at the end of this routine after 277 * corrections for hw_ptr position 278 */ 279 pos = substream->ops->pointer(substream); 280 curr_jiffies = jiffies; 281 if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { 282 if ((substream->ops->get_time_info) && 283 (runtime->audio_tstamp_config.type_requested != SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT)) { 284 substream->ops->get_time_info(substream, &curr_tstamp, 285 &audio_tstamp, 286 &runtime->audio_tstamp_config, 287 &runtime->audio_tstamp_report); 288 289 /* re-test in case tstamp type is not supported in hardware and was demoted to DEFAULT */ 290 if (runtime->audio_tstamp_report.actual_type == SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT) 291 snd_pcm_gettime(runtime, &curr_tstamp); 292 } else 293 snd_pcm_gettime(runtime, &curr_tstamp); 294 } 295 296 if (pos == SNDRV_PCM_POS_XRUN) { 297 __snd_pcm_xrun(substream); 298 return -EPIPE; 299 } 300 if (pos >= runtime->buffer_size) { 301 if (printk_ratelimit()) { 302 char name[16]; 303 snd_pcm_debug_name(substream, name, sizeof(name)); 304 pcm_err(substream->pcm, 305 "invalid position: %s, pos = %ld, buffer size = %ld, period size = %ld\n", 306 name, pos, runtime->buffer_size, 307 runtime->period_size); 308 } 309 pos = 0; 310 } 311 pos -= pos % runtime->min_align; 312 trace_hwptr(substream, pos, in_interrupt); 313 hw_base = runtime->hw_ptr_base; 314 new_hw_ptr = hw_base + pos; 315 if (in_interrupt) { 316 /* we know that one period was processed */ 317 /* delta = "expected next hw_ptr" for in_interrupt != 0 */ 318 delta = runtime->hw_ptr_interrupt + runtime->period_size; 319 if (delta > new_hw_ptr) { 320 /* check for double acknowledged interrupts */ 321 hdelta = curr_jiffies - runtime->hw_ptr_jiffies; 322 if (hdelta > runtime->hw_ptr_buffer_jiffies/2 + 1) { 323 hw_base += runtime->buffer_size; 324 if (hw_base >= runtime->boundary) { 325 hw_base = 0; 326 crossed_boundary++; 327 } 328 new_hw_ptr = hw_base + pos; 329 goto __delta; 330 } 331 } 332 } 333 /* new_hw_ptr might be lower than old_hw_ptr in case when */ 334 /* pointer crosses the end of the ring buffer */ 335 if (new_hw_ptr < old_hw_ptr) { 336 hw_base += runtime->buffer_size; 337 if (hw_base >= runtime->boundary) { 338 hw_base = 0; 339 crossed_boundary++; 340 } 341 new_hw_ptr = hw_base + pos; 342 } 343 __delta: 344 delta = new_hw_ptr - old_hw_ptr; 345 if (delta < 0) 346 delta += runtime->boundary; 347 348 if (runtime->no_period_wakeup) { 349 snd_pcm_sframes_t xrun_threshold; 350 /* 351 * Without regular period interrupts, we have to check 352 * the elapsed time to detect xruns. 353 */ 354 jdelta = curr_jiffies - runtime->hw_ptr_jiffies; 355 if (jdelta < runtime->hw_ptr_buffer_jiffies / 2) 356 goto no_delta_check; 357 hdelta = jdelta - delta * HZ / runtime->rate; 358 xrun_threshold = runtime->hw_ptr_buffer_jiffies / 2 + 1; 359 while (hdelta > xrun_threshold) { 360 delta += runtime->buffer_size; 361 hw_base += runtime->buffer_size; 362 if (hw_base >= runtime->boundary) { 363 hw_base = 0; 364 crossed_boundary++; 365 } 366 new_hw_ptr = hw_base + pos; 367 hdelta -= runtime->hw_ptr_buffer_jiffies; 368 } 369 goto no_delta_check; 370 } 371 372 /* something must be really wrong */ 373 if (delta >= runtime->buffer_size + runtime->period_size) { 374 hw_ptr_error(substream, in_interrupt, "Unexpected hw_ptr", 375 "(stream=%i, pos=%ld, new_hw_ptr=%ld, old_hw_ptr=%ld)\n", 376 substream->stream, (long)pos, 377 (long)new_hw_ptr, (long)old_hw_ptr); 378 return 0; 379 } 380 381 /* Do jiffies check only in xrun_debug mode */ 382 if (!xrun_debug(substream, XRUN_DEBUG_JIFFIESCHECK)) 383 goto no_jiffies_check; 384 385 /* Skip the jiffies check for hardwares with BATCH flag. 386 * Such hardware usually just increases the position at each IRQ, 387 * thus it can't give any strange position. 388 */ 389 if (runtime->hw.info & SNDRV_PCM_INFO_BATCH) 390 goto no_jiffies_check; 391 hdelta = delta; 392 if (hdelta < runtime->delay) 393 goto no_jiffies_check; 394 hdelta -= runtime->delay; 395 jdelta = curr_jiffies - runtime->hw_ptr_jiffies; 396 if (((hdelta * HZ) / runtime->rate) > jdelta + HZ/100) { 397 delta = jdelta / 398 (((runtime->period_size * HZ) / runtime->rate) 399 + HZ/100); 400 /* move new_hw_ptr according jiffies not pos variable */ 401 new_hw_ptr = old_hw_ptr; 402 hw_base = delta; 403 /* use loop to avoid checks for delta overflows */ 404 /* the delta value is small or zero in most cases */ 405 while (delta > 0) { 406 new_hw_ptr += runtime->period_size; 407 if (new_hw_ptr >= runtime->boundary) { 408 new_hw_ptr -= runtime->boundary; 409 crossed_boundary--; 410 } 411 delta--; 412 } 413 /* align hw_base to buffer_size */ 414 hw_ptr_error(substream, in_interrupt, "hw_ptr skipping", 415 "(pos=%ld, delta=%ld, period=%ld, jdelta=%lu/%lu/%lu, hw_ptr=%ld/%ld)\n", 416 (long)pos, (long)hdelta, 417 (long)runtime->period_size, jdelta, 418 ((hdelta * HZ) / runtime->rate), hw_base, 419 (unsigned long)old_hw_ptr, 420 (unsigned long)new_hw_ptr); 421 /* reset values to proper state */ 422 delta = 0; 423 hw_base = new_hw_ptr - (new_hw_ptr % runtime->buffer_size); 424 } 425 no_jiffies_check: 426 if (delta > runtime->period_size + runtime->period_size / 2) { 427 hw_ptr_error(substream, in_interrupt, 428 "Lost interrupts?", 429 "(stream=%i, delta=%ld, new_hw_ptr=%ld, old_hw_ptr=%ld)\n", 430 substream->stream, (long)delta, 431 (long)new_hw_ptr, 432 (long)old_hw_ptr); 433 } 434 435 no_delta_check: 436 if (runtime->status->hw_ptr == new_hw_ptr) { 437 runtime->hw_ptr_jiffies = curr_jiffies; 438 update_audio_tstamp(substream, &curr_tstamp, &audio_tstamp); 439 return 0; 440 } 441 442 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 443 runtime->silence_size > 0) 444 snd_pcm_playback_silence(substream, new_hw_ptr); 445 446 if (in_interrupt) { 447 delta = new_hw_ptr - runtime->hw_ptr_interrupt; 448 if (delta < 0) 449 delta += runtime->boundary; 450 delta -= (snd_pcm_uframes_t)delta % runtime->period_size; 451 runtime->hw_ptr_interrupt += delta; 452 if (runtime->hw_ptr_interrupt >= runtime->boundary) 453 runtime->hw_ptr_interrupt -= runtime->boundary; 454 } 455 runtime->hw_ptr_base = hw_base; 456 runtime->status->hw_ptr = new_hw_ptr; 457 runtime->hw_ptr_jiffies = curr_jiffies; 458 if (crossed_boundary) { 459 snd_BUG_ON(crossed_boundary != 1); 460 runtime->hw_ptr_wrap += runtime->boundary; 461 } 462 463 update_audio_tstamp(substream, &curr_tstamp, &audio_tstamp); 464 465 return snd_pcm_update_state(substream, runtime); 466 } 467 468 /* CAUTION: call it with irq disabled */ 469 int snd_pcm_update_hw_ptr(struct snd_pcm_substream *substream) 470 { 471 return snd_pcm_update_hw_ptr0(substream, 0); 472 } 473 474 /** 475 * snd_pcm_set_ops - set the PCM operators 476 * @pcm: the pcm instance 477 * @direction: stream direction, SNDRV_PCM_STREAM_XXX 478 * @ops: the operator table 479 * 480 * Sets the given PCM operators to the pcm instance. 481 */ 482 void snd_pcm_set_ops(struct snd_pcm *pcm, int direction, 483 const struct snd_pcm_ops *ops) 484 { 485 struct snd_pcm_str *stream = &pcm->streams[direction]; 486 struct snd_pcm_substream *substream; 487 488 for (substream = stream->substream; substream != NULL; substream = substream->next) 489 substream->ops = ops; 490 } 491 EXPORT_SYMBOL(snd_pcm_set_ops); 492 493 /** 494 * snd_pcm_set_sync - set the PCM sync id 495 * @substream: the pcm substream 496 * 497 * Sets the PCM sync identifier for the card. 498 */ 499 void snd_pcm_set_sync(struct snd_pcm_substream *substream) 500 { 501 struct snd_pcm_runtime *runtime = substream->runtime; 502 503 runtime->sync.id32[0] = substream->pcm->card->number; 504 runtime->sync.id32[1] = -1; 505 runtime->sync.id32[2] = -1; 506 runtime->sync.id32[3] = -1; 507 } 508 EXPORT_SYMBOL(snd_pcm_set_sync); 509 510 /* 511 * Standard ioctl routine 512 */ 513 514 static inline unsigned int div32(unsigned int a, unsigned int b, 515 unsigned int *r) 516 { 517 if (b == 0) { 518 *r = 0; 519 return UINT_MAX; 520 } 521 *r = a % b; 522 return a / b; 523 } 524 525 static inline unsigned int div_down(unsigned int a, unsigned int b) 526 { 527 if (b == 0) 528 return UINT_MAX; 529 return a / b; 530 } 531 532 static inline unsigned int div_up(unsigned int a, unsigned int b) 533 { 534 unsigned int r; 535 unsigned int q; 536 if (b == 0) 537 return UINT_MAX; 538 q = div32(a, b, &r); 539 if (r) 540 ++q; 541 return q; 542 } 543 544 static inline unsigned int mul(unsigned int a, unsigned int b) 545 { 546 if (a == 0) 547 return 0; 548 if (div_down(UINT_MAX, a) < b) 549 return UINT_MAX; 550 return a * b; 551 } 552 553 static inline unsigned int muldiv32(unsigned int a, unsigned int b, 554 unsigned int c, unsigned int *r) 555 { 556 u_int64_t n = (u_int64_t) a * b; 557 if (c == 0) { 558 *r = 0; 559 return UINT_MAX; 560 } 561 n = div_u64_rem(n, c, r); 562 if (n >= UINT_MAX) { 563 *r = 0; 564 return UINT_MAX; 565 } 566 return n; 567 } 568 569 /** 570 * snd_interval_refine - refine the interval value of configurator 571 * @i: the interval value to refine 572 * @v: the interval value to refer to 573 * 574 * Refines the interval value with the reference value. 575 * The interval is changed to the range satisfying both intervals. 576 * The interval status (min, max, integer, etc.) are evaluated. 577 * 578 * Return: Positive if the value is changed, zero if it's not changed, or a 579 * negative error code. 580 */ 581 int snd_interval_refine(struct snd_interval *i, const struct snd_interval *v) 582 { 583 int changed = 0; 584 if (snd_BUG_ON(snd_interval_empty(i))) 585 return -EINVAL; 586 if (i->min < v->min) { 587 i->min = v->min; 588 i->openmin = v->openmin; 589 changed = 1; 590 } else if (i->min == v->min && !i->openmin && v->openmin) { 591 i->openmin = 1; 592 changed = 1; 593 } 594 if (i->max > v->max) { 595 i->max = v->max; 596 i->openmax = v->openmax; 597 changed = 1; 598 } else if (i->max == v->max && !i->openmax && v->openmax) { 599 i->openmax = 1; 600 changed = 1; 601 } 602 if (!i->integer && v->integer) { 603 i->integer = 1; 604 changed = 1; 605 } 606 if (i->integer) { 607 if (i->openmin) { 608 i->min++; 609 i->openmin = 0; 610 } 611 if (i->openmax) { 612 i->max--; 613 i->openmax = 0; 614 } 615 } else if (!i->openmin && !i->openmax && i->min == i->max) 616 i->integer = 1; 617 if (snd_interval_checkempty(i)) { 618 snd_interval_none(i); 619 return -EINVAL; 620 } 621 return changed; 622 } 623 EXPORT_SYMBOL(snd_interval_refine); 624 625 static int snd_interval_refine_first(struct snd_interval *i) 626 { 627 const unsigned int last_max = i->max; 628 629 if (snd_BUG_ON(snd_interval_empty(i))) 630 return -EINVAL; 631 if (snd_interval_single(i)) 632 return 0; 633 i->max = i->min; 634 if (i->openmin) 635 i->max++; 636 /* only exclude max value if also excluded before refine */ 637 i->openmax = (i->openmax && i->max >= last_max); 638 return 1; 639 } 640 641 static int snd_interval_refine_last(struct snd_interval *i) 642 { 643 const unsigned int last_min = i->min; 644 645 if (snd_BUG_ON(snd_interval_empty(i))) 646 return -EINVAL; 647 if (snd_interval_single(i)) 648 return 0; 649 i->min = i->max; 650 if (i->openmax) 651 i->min--; 652 /* only exclude min value if also excluded before refine */ 653 i->openmin = (i->openmin && i->min <= last_min); 654 return 1; 655 } 656 657 void snd_interval_mul(const struct snd_interval *a, const struct snd_interval *b, struct snd_interval *c) 658 { 659 if (a->empty || b->empty) { 660 snd_interval_none(c); 661 return; 662 } 663 c->empty = 0; 664 c->min = mul(a->min, b->min); 665 c->openmin = (a->openmin || b->openmin); 666 c->max = mul(a->max, b->max); 667 c->openmax = (a->openmax || b->openmax); 668 c->integer = (a->integer && b->integer); 669 } 670 671 /** 672 * snd_interval_div - refine the interval value with division 673 * @a: dividend 674 * @b: divisor 675 * @c: quotient 676 * 677 * c = a / b 678 * 679 * Returns non-zero if the value is changed, zero if not changed. 680 */ 681 void snd_interval_div(const struct snd_interval *a, const struct snd_interval *b, struct snd_interval *c) 682 { 683 unsigned int r; 684 if (a->empty || b->empty) { 685 snd_interval_none(c); 686 return; 687 } 688 c->empty = 0; 689 c->min = div32(a->min, b->max, &r); 690 c->openmin = (r || a->openmin || b->openmax); 691 if (b->min > 0) { 692 c->max = div32(a->max, b->min, &r); 693 if (r) { 694 c->max++; 695 c->openmax = 1; 696 } else 697 c->openmax = (a->openmax || b->openmin); 698 } else { 699 c->max = UINT_MAX; 700 c->openmax = 0; 701 } 702 c->integer = 0; 703 } 704 705 /** 706 * snd_interval_muldivk - refine the interval value 707 * @a: dividend 1 708 * @b: dividend 2 709 * @k: divisor (as integer) 710 * @c: result 711 * 712 * c = a * b / k 713 * 714 * Returns non-zero if the value is changed, zero if not changed. 715 */ 716 void snd_interval_muldivk(const struct snd_interval *a, const struct snd_interval *b, 717 unsigned int k, struct snd_interval *c) 718 { 719 unsigned int r; 720 if (a->empty || b->empty) { 721 snd_interval_none(c); 722 return; 723 } 724 c->empty = 0; 725 c->min = muldiv32(a->min, b->min, k, &r); 726 c->openmin = (r || a->openmin || b->openmin); 727 c->max = muldiv32(a->max, b->max, k, &r); 728 if (r) { 729 c->max++; 730 c->openmax = 1; 731 } else 732 c->openmax = (a->openmax || b->openmax); 733 c->integer = 0; 734 } 735 736 /** 737 * snd_interval_mulkdiv - refine the interval value 738 * @a: dividend 1 739 * @k: dividend 2 (as integer) 740 * @b: divisor 741 * @c: result 742 * 743 * c = a * k / b 744 * 745 * Returns non-zero if the value is changed, zero if not changed. 746 */ 747 void snd_interval_mulkdiv(const struct snd_interval *a, unsigned int k, 748 const struct snd_interval *b, struct snd_interval *c) 749 { 750 unsigned int r; 751 if (a->empty || b->empty) { 752 snd_interval_none(c); 753 return; 754 } 755 c->empty = 0; 756 c->min = muldiv32(a->min, k, b->max, &r); 757 c->openmin = (r || a->openmin || b->openmax); 758 if (b->min > 0) { 759 c->max = muldiv32(a->max, k, b->min, &r); 760 if (r) { 761 c->max++; 762 c->openmax = 1; 763 } else 764 c->openmax = (a->openmax || b->openmin); 765 } else { 766 c->max = UINT_MAX; 767 c->openmax = 0; 768 } 769 c->integer = 0; 770 } 771 772 /* ---- */ 773 774 775 /** 776 * snd_interval_ratnum - refine the interval value 777 * @i: interval to refine 778 * @rats_count: number of ratnum_t 779 * @rats: ratnum_t array 780 * @nump: pointer to store the resultant numerator 781 * @denp: pointer to store the resultant denominator 782 * 783 * Return: Positive if the value is changed, zero if it's not changed, or a 784 * negative error code. 785 */ 786 int snd_interval_ratnum(struct snd_interval *i, 787 unsigned int rats_count, const struct snd_ratnum *rats, 788 unsigned int *nump, unsigned int *denp) 789 { 790 unsigned int best_num, best_den; 791 int best_diff; 792 unsigned int k; 793 struct snd_interval t; 794 int err; 795 unsigned int result_num, result_den; 796 int result_diff; 797 798 best_num = best_den = best_diff = 0; 799 for (k = 0; k < rats_count; ++k) { 800 unsigned int num = rats[k].num; 801 unsigned int den; 802 unsigned int q = i->min; 803 int diff; 804 if (q == 0) 805 q = 1; 806 den = div_up(num, q); 807 if (den < rats[k].den_min) 808 continue; 809 if (den > rats[k].den_max) 810 den = rats[k].den_max; 811 else { 812 unsigned int r; 813 r = (den - rats[k].den_min) % rats[k].den_step; 814 if (r != 0) 815 den -= r; 816 } 817 diff = num - q * den; 818 if (diff < 0) 819 diff = -diff; 820 if (best_num == 0 || 821 diff * best_den < best_diff * den) { 822 best_diff = diff; 823 best_den = den; 824 best_num = num; 825 } 826 } 827 if (best_den == 0) { 828 i->empty = 1; 829 return -EINVAL; 830 } 831 t.min = div_down(best_num, best_den); 832 t.openmin = !!(best_num % best_den); 833 834 result_num = best_num; 835 result_diff = best_diff; 836 result_den = best_den; 837 best_num = best_den = best_diff = 0; 838 for (k = 0; k < rats_count; ++k) { 839 unsigned int num = rats[k].num; 840 unsigned int den; 841 unsigned int q = i->max; 842 int diff; 843 if (q == 0) { 844 i->empty = 1; 845 return -EINVAL; 846 } 847 den = div_down(num, q); 848 if (den > rats[k].den_max) 849 continue; 850 if (den < rats[k].den_min) 851 den = rats[k].den_min; 852 else { 853 unsigned int r; 854 r = (den - rats[k].den_min) % rats[k].den_step; 855 if (r != 0) 856 den += rats[k].den_step - r; 857 } 858 diff = q * den - num; 859 if (diff < 0) 860 diff = -diff; 861 if (best_num == 0 || 862 diff * best_den < best_diff * den) { 863 best_diff = diff; 864 best_den = den; 865 best_num = num; 866 } 867 } 868 if (best_den == 0) { 869 i->empty = 1; 870 return -EINVAL; 871 } 872 t.max = div_up(best_num, best_den); 873 t.openmax = !!(best_num % best_den); 874 t.integer = 0; 875 err = snd_interval_refine(i, &t); 876 if (err < 0) 877 return err; 878 879 if (snd_interval_single(i)) { 880 if (best_diff * result_den < result_diff * best_den) { 881 result_num = best_num; 882 result_den = best_den; 883 } 884 if (nump) 885 *nump = result_num; 886 if (denp) 887 *denp = result_den; 888 } 889 return err; 890 } 891 EXPORT_SYMBOL(snd_interval_ratnum); 892 893 /** 894 * snd_interval_ratden - refine the interval value 895 * @i: interval to refine 896 * @rats_count: number of struct ratden 897 * @rats: struct ratden array 898 * @nump: pointer to store the resultant numerator 899 * @denp: pointer to store the resultant denominator 900 * 901 * Return: Positive if the value is changed, zero if it's not changed, or a 902 * negative error code. 903 */ 904 static int snd_interval_ratden(struct snd_interval *i, 905 unsigned int rats_count, 906 const struct snd_ratden *rats, 907 unsigned int *nump, unsigned int *denp) 908 { 909 unsigned int best_num, best_diff, best_den; 910 unsigned int k; 911 struct snd_interval t; 912 int err; 913 914 best_num = best_den = best_diff = 0; 915 for (k = 0; k < rats_count; ++k) { 916 unsigned int num; 917 unsigned int den = rats[k].den; 918 unsigned int q = i->min; 919 int diff; 920 num = mul(q, den); 921 if (num > rats[k].num_max) 922 continue; 923 if (num < rats[k].num_min) 924 num = rats[k].num_max; 925 else { 926 unsigned int r; 927 r = (num - rats[k].num_min) % rats[k].num_step; 928 if (r != 0) 929 num += rats[k].num_step - r; 930 } 931 diff = num - q * den; 932 if (best_num == 0 || 933 diff * best_den < best_diff * den) { 934 best_diff = diff; 935 best_den = den; 936 best_num = num; 937 } 938 } 939 if (best_den == 0) { 940 i->empty = 1; 941 return -EINVAL; 942 } 943 t.min = div_down(best_num, best_den); 944 t.openmin = !!(best_num % best_den); 945 946 best_num = best_den = best_diff = 0; 947 for (k = 0; k < rats_count; ++k) { 948 unsigned int num; 949 unsigned int den = rats[k].den; 950 unsigned int q = i->max; 951 int diff; 952 num = mul(q, den); 953 if (num < rats[k].num_min) 954 continue; 955 if (num > rats[k].num_max) 956 num = rats[k].num_max; 957 else { 958 unsigned int r; 959 r = (num - rats[k].num_min) % rats[k].num_step; 960 if (r != 0) 961 num -= r; 962 } 963 diff = q * den - num; 964 if (best_num == 0 || 965 diff * best_den < best_diff * den) { 966 best_diff = diff; 967 best_den = den; 968 best_num = num; 969 } 970 } 971 if (best_den == 0) { 972 i->empty = 1; 973 return -EINVAL; 974 } 975 t.max = div_up(best_num, best_den); 976 t.openmax = !!(best_num % best_den); 977 t.integer = 0; 978 err = snd_interval_refine(i, &t); 979 if (err < 0) 980 return err; 981 982 if (snd_interval_single(i)) { 983 if (nump) 984 *nump = best_num; 985 if (denp) 986 *denp = best_den; 987 } 988 return err; 989 } 990 991 /** 992 * snd_interval_list - refine the interval value from the list 993 * @i: the interval value to refine 994 * @count: the number of elements in the list 995 * @list: the value list 996 * @mask: the bit-mask to evaluate 997 * 998 * Refines the interval value from the list. 999 * When mask is non-zero, only the elements corresponding to bit 1 are 1000 * evaluated. 1001 * 1002 * Return: Positive if the value is changed, zero if it's not changed, or a 1003 * negative error code. 1004 */ 1005 int snd_interval_list(struct snd_interval *i, unsigned int count, 1006 const unsigned int *list, unsigned int mask) 1007 { 1008 unsigned int k; 1009 struct snd_interval list_range; 1010 1011 if (!count) { 1012 i->empty = 1; 1013 return -EINVAL; 1014 } 1015 snd_interval_any(&list_range); 1016 list_range.min = UINT_MAX; 1017 list_range.max = 0; 1018 for (k = 0; k < count; k++) { 1019 if (mask && !(mask & (1 << k))) 1020 continue; 1021 if (!snd_interval_test(i, list[k])) 1022 continue; 1023 list_range.min = min(list_range.min, list[k]); 1024 list_range.max = max(list_range.max, list[k]); 1025 } 1026 return snd_interval_refine(i, &list_range); 1027 } 1028 EXPORT_SYMBOL(snd_interval_list); 1029 1030 /** 1031 * snd_interval_ranges - refine the interval value from the list of ranges 1032 * @i: the interval value to refine 1033 * @count: the number of elements in the list of ranges 1034 * @ranges: the ranges list 1035 * @mask: the bit-mask to evaluate 1036 * 1037 * Refines the interval value from the list of ranges. 1038 * When mask is non-zero, only the elements corresponding to bit 1 are 1039 * evaluated. 1040 * 1041 * Return: Positive if the value is changed, zero if it's not changed, or a 1042 * negative error code. 1043 */ 1044 int snd_interval_ranges(struct snd_interval *i, unsigned int count, 1045 const struct snd_interval *ranges, unsigned int mask) 1046 { 1047 unsigned int k; 1048 struct snd_interval range_union; 1049 struct snd_interval range; 1050 1051 if (!count) { 1052 snd_interval_none(i); 1053 return -EINVAL; 1054 } 1055 snd_interval_any(&range_union); 1056 range_union.min = UINT_MAX; 1057 range_union.max = 0; 1058 for (k = 0; k < count; k++) { 1059 if (mask && !(mask & (1 << k))) 1060 continue; 1061 snd_interval_copy(&range, &ranges[k]); 1062 if (snd_interval_refine(&range, i) < 0) 1063 continue; 1064 if (snd_interval_empty(&range)) 1065 continue; 1066 1067 if (range.min < range_union.min) { 1068 range_union.min = range.min; 1069 range_union.openmin = 1; 1070 } 1071 if (range.min == range_union.min && !range.openmin) 1072 range_union.openmin = 0; 1073 if (range.max > range_union.max) { 1074 range_union.max = range.max; 1075 range_union.openmax = 1; 1076 } 1077 if (range.max == range_union.max && !range.openmax) 1078 range_union.openmax = 0; 1079 } 1080 return snd_interval_refine(i, &range_union); 1081 } 1082 EXPORT_SYMBOL(snd_interval_ranges); 1083 1084 static int snd_interval_step(struct snd_interval *i, unsigned int step) 1085 { 1086 unsigned int n; 1087 int changed = 0; 1088 n = i->min % step; 1089 if (n != 0 || i->openmin) { 1090 i->min += step - n; 1091 i->openmin = 0; 1092 changed = 1; 1093 } 1094 n = i->max % step; 1095 if (n != 0 || i->openmax) { 1096 i->max -= n; 1097 i->openmax = 0; 1098 changed = 1; 1099 } 1100 if (snd_interval_checkempty(i)) { 1101 i->empty = 1; 1102 return -EINVAL; 1103 } 1104 return changed; 1105 } 1106 1107 /* Info constraints helpers */ 1108 1109 /** 1110 * snd_pcm_hw_rule_add - add the hw-constraint rule 1111 * @runtime: the pcm runtime instance 1112 * @cond: condition bits 1113 * @var: the variable to evaluate 1114 * @func: the evaluation function 1115 * @private: the private data pointer passed to function 1116 * @dep: the dependent variables 1117 * 1118 * Return: Zero if successful, or a negative error code on failure. 1119 */ 1120 int snd_pcm_hw_rule_add(struct snd_pcm_runtime *runtime, unsigned int cond, 1121 int var, 1122 snd_pcm_hw_rule_func_t func, void *private, 1123 int dep, ...) 1124 { 1125 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 1126 struct snd_pcm_hw_rule *c; 1127 unsigned int k; 1128 va_list args; 1129 va_start(args, dep); 1130 if (constrs->rules_num >= constrs->rules_all) { 1131 struct snd_pcm_hw_rule *new; 1132 unsigned int new_rules = constrs->rules_all + 16; 1133 new = krealloc_array(constrs->rules, new_rules, 1134 sizeof(*c), GFP_KERNEL); 1135 if (!new) { 1136 va_end(args); 1137 return -ENOMEM; 1138 } 1139 constrs->rules = new; 1140 constrs->rules_all = new_rules; 1141 } 1142 c = &constrs->rules[constrs->rules_num]; 1143 c->cond = cond; 1144 c->func = func; 1145 c->var = var; 1146 c->private = private; 1147 k = 0; 1148 while (1) { 1149 if (snd_BUG_ON(k >= ARRAY_SIZE(c->deps))) { 1150 va_end(args); 1151 return -EINVAL; 1152 } 1153 c->deps[k++] = dep; 1154 if (dep < 0) 1155 break; 1156 dep = va_arg(args, int); 1157 } 1158 constrs->rules_num++; 1159 va_end(args); 1160 return 0; 1161 } 1162 EXPORT_SYMBOL(snd_pcm_hw_rule_add); 1163 1164 /** 1165 * snd_pcm_hw_constraint_mask - apply the given bitmap mask constraint 1166 * @runtime: PCM runtime instance 1167 * @var: hw_params variable to apply the mask 1168 * @mask: the bitmap mask 1169 * 1170 * Apply the constraint of the given bitmap mask to a 32-bit mask parameter. 1171 * 1172 * Return: Zero if successful, or a negative error code on failure. 1173 */ 1174 int snd_pcm_hw_constraint_mask(struct snd_pcm_runtime *runtime, snd_pcm_hw_param_t var, 1175 u_int32_t mask) 1176 { 1177 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 1178 struct snd_mask *maskp = constrs_mask(constrs, var); 1179 *maskp->bits &= mask; 1180 memset(maskp->bits + 1, 0, (SNDRV_MASK_MAX-32) / 8); /* clear rest */ 1181 if (*maskp->bits == 0) 1182 return -EINVAL; 1183 return 0; 1184 } 1185 1186 /** 1187 * snd_pcm_hw_constraint_mask64 - apply the given bitmap mask constraint 1188 * @runtime: PCM runtime instance 1189 * @var: hw_params variable to apply the mask 1190 * @mask: the 64bit bitmap mask 1191 * 1192 * Apply the constraint of the given bitmap mask to a 64-bit mask parameter. 1193 * 1194 * Return: Zero if successful, or a negative error code on failure. 1195 */ 1196 int snd_pcm_hw_constraint_mask64(struct snd_pcm_runtime *runtime, snd_pcm_hw_param_t var, 1197 u_int64_t mask) 1198 { 1199 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 1200 struct snd_mask *maskp = constrs_mask(constrs, var); 1201 maskp->bits[0] &= (u_int32_t)mask; 1202 maskp->bits[1] &= (u_int32_t)(mask >> 32); 1203 memset(maskp->bits + 2, 0, (SNDRV_MASK_MAX-64) / 8); /* clear rest */ 1204 if (! maskp->bits[0] && ! maskp->bits[1]) 1205 return -EINVAL; 1206 return 0; 1207 } 1208 EXPORT_SYMBOL(snd_pcm_hw_constraint_mask64); 1209 1210 /** 1211 * snd_pcm_hw_constraint_integer - apply an integer constraint to an interval 1212 * @runtime: PCM runtime instance 1213 * @var: hw_params variable to apply the integer constraint 1214 * 1215 * Apply the constraint of integer to an interval parameter. 1216 * 1217 * Return: Positive if the value is changed, zero if it's not changed, or a 1218 * negative error code. 1219 */ 1220 int snd_pcm_hw_constraint_integer(struct snd_pcm_runtime *runtime, snd_pcm_hw_param_t var) 1221 { 1222 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 1223 return snd_interval_setinteger(constrs_interval(constrs, var)); 1224 } 1225 EXPORT_SYMBOL(snd_pcm_hw_constraint_integer); 1226 1227 /** 1228 * snd_pcm_hw_constraint_minmax - apply a min/max range constraint to an interval 1229 * @runtime: PCM runtime instance 1230 * @var: hw_params variable to apply the range 1231 * @min: the minimal value 1232 * @max: the maximal value 1233 * 1234 * Apply the min/max range constraint to an interval parameter. 1235 * 1236 * Return: Positive if the value is changed, zero if it's not changed, or a 1237 * negative error code. 1238 */ 1239 int snd_pcm_hw_constraint_minmax(struct snd_pcm_runtime *runtime, snd_pcm_hw_param_t var, 1240 unsigned int min, unsigned int max) 1241 { 1242 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 1243 struct snd_interval t; 1244 t.min = min; 1245 t.max = max; 1246 t.openmin = t.openmax = 0; 1247 t.integer = 0; 1248 return snd_interval_refine(constrs_interval(constrs, var), &t); 1249 } 1250 EXPORT_SYMBOL(snd_pcm_hw_constraint_minmax); 1251 1252 static int snd_pcm_hw_rule_list(struct snd_pcm_hw_params *params, 1253 struct snd_pcm_hw_rule *rule) 1254 { 1255 struct snd_pcm_hw_constraint_list *list = rule->private; 1256 return snd_interval_list(hw_param_interval(params, rule->var), list->count, list->list, list->mask); 1257 } 1258 1259 1260 /** 1261 * snd_pcm_hw_constraint_list - apply a list of constraints to a parameter 1262 * @runtime: PCM runtime instance 1263 * @cond: condition bits 1264 * @var: hw_params variable to apply the list constraint 1265 * @l: list 1266 * 1267 * Apply the list of constraints to an interval parameter. 1268 * 1269 * Return: Zero if successful, or a negative error code on failure. 1270 */ 1271 int snd_pcm_hw_constraint_list(struct snd_pcm_runtime *runtime, 1272 unsigned int cond, 1273 snd_pcm_hw_param_t var, 1274 const struct snd_pcm_hw_constraint_list *l) 1275 { 1276 return snd_pcm_hw_rule_add(runtime, cond, var, 1277 snd_pcm_hw_rule_list, (void *)l, 1278 var, -1); 1279 } 1280 EXPORT_SYMBOL(snd_pcm_hw_constraint_list); 1281 1282 static int snd_pcm_hw_rule_ranges(struct snd_pcm_hw_params *params, 1283 struct snd_pcm_hw_rule *rule) 1284 { 1285 struct snd_pcm_hw_constraint_ranges *r = rule->private; 1286 return snd_interval_ranges(hw_param_interval(params, rule->var), 1287 r->count, r->ranges, r->mask); 1288 } 1289 1290 1291 /** 1292 * snd_pcm_hw_constraint_ranges - apply list of range constraints to a parameter 1293 * @runtime: PCM runtime instance 1294 * @cond: condition bits 1295 * @var: hw_params variable to apply the list of range constraints 1296 * @r: ranges 1297 * 1298 * Apply the list of range constraints to an interval parameter. 1299 * 1300 * Return: Zero if successful, or a negative error code on failure. 1301 */ 1302 int snd_pcm_hw_constraint_ranges(struct snd_pcm_runtime *runtime, 1303 unsigned int cond, 1304 snd_pcm_hw_param_t var, 1305 const struct snd_pcm_hw_constraint_ranges *r) 1306 { 1307 return snd_pcm_hw_rule_add(runtime, cond, var, 1308 snd_pcm_hw_rule_ranges, (void *)r, 1309 var, -1); 1310 } 1311 EXPORT_SYMBOL(snd_pcm_hw_constraint_ranges); 1312 1313 static int snd_pcm_hw_rule_ratnums(struct snd_pcm_hw_params *params, 1314 struct snd_pcm_hw_rule *rule) 1315 { 1316 const struct snd_pcm_hw_constraint_ratnums *r = rule->private; 1317 unsigned int num = 0, den = 0; 1318 int err; 1319 err = snd_interval_ratnum(hw_param_interval(params, rule->var), 1320 r->nrats, r->rats, &num, &den); 1321 if (err >= 0 && den && rule->var == SNDRV_PCM_HW_PARAM_RATE) { 1322 params->rate_num = num; 1323 params->rate_den = den; 1324 } 1325 return err; 1326 } 1327 1328 /** 1329 * snd_pcm_hw_constraint_ratnums - apply ratnums constraint to a parameter 1330 * @runtime: PCM runtime instance 1331 * @cond: condition bits 1332 * @var: hw_params variable to apply the ratnums constraint 1333 * @r: struct snd_ratnums constriants 1334 * 1335 * Return: Zero if successful, or a negative error code on failure. 1336 */ 1337 int snd_pcm_hw_constraint_ratnums(struct snd_pcm_runtime *runtime, 1338 unsigned int cond, 1339 snd_pcm_hw_param_t var, 1340 const struct snd_pcm_hw_constraint_ratnums *r) 1341 { 1342 return snd_pcm_hw_rule_add(runtime, cond, var, 1343 snd_pcm_hw_rule_ratnums, (void *)r, 1344 var, -1); 1345 } 1346 EXPORT_SYMBOL(snd_pcm_hw_constraint_ratnums); 1347 1348 static int snd_pcm_hw_rule_ratdens(struct snd_pcm_hw_params *params, 1349 struct snd_pcm_hw_rule *rule) 1350 { 1351 const struct snd_pcm_hw_constraint_ratdens *r = rule->private; 1352 unsigned int num = 0, den = 0; 1353 int err = snd_interval_ratden(hw_param_interval(params, rule->var), 1354 r->nrats, r->rats, &num, &den); 1355 if (err >= 0 && den && rule->var == SNDRV_PCM_HW_PARAM_RATE) { 1356 params->rate_num = num; 1357 params->rate_den = den; 1358 } 1359 return err; 1360 } 1361 1362 /** 1363 * snd_pcm_hw_constraint_ratdens - apply ratdens constraint to a parameter 1364 * @runtime: PCM runtime instance 1365 * @cond: condition bits 1366 * @var: hw_params variable to apply the ratdens constraint 1367 * @r: struct snd_ratdens constriants 1368 * 1369 * Return: Zero if successful, or a negative error code on failure. 1370 */ 1371 int snd_pcm_hw_constraint_ratdens(struct snd_pcm_runtime *runtime, 1372 unsigned int cond, 1373 snd_pcm_hw_param_t var, 1374 const struct snd_pcm_hw_constraint_ratdens *r) 1375 { 1376 return snd_pcm_hw_rule_add(runtime, cond, var, 1377 snd_pcm_hw_rule_ratdens, (void *)r, 1378 var, -1); 1379 } 1380 EXPORT_SYMBOL(snd_pcm_hw_constraint_ratdens); 1381 1382 static int snd_pcm_hw_rule_msbits(struct snd_pcm_hw_params *params, 1383 struct snd_pcm_hw_rule *rule) 1384 { 1385 unsigned int l = (unsigned long) rule->private; 1386 int width = l & 0xffff; 1387 unsigned int msbits = l >> 16; 1388 const struct snd_interval *i = 1389 hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_SAMPLE_BITS); 1390 1391 if (!snd_interval_single(i)) 1392 return 0; 1393 1394 if ((snd_interval_value(i) == width) || 1395 (width == 0 && snd_interval_value(i) > msbits)) 1396 params->msbits = min_not_zero(params->msbits, msbits); 1397 1398 return 0; 1399 } 1400 1401 /** 1402 * snd_pcm_hw_constraint_msbits - add a hw constraint msbits rule 1403 * @runtime: PCM runtime instance 1404 * @cond: condition bits 1405 * @width: sample bits width 1406 * @msbits: msbits width 1407 * 1408 * This constraint will set the number of most significant bits (msbits) if a 1409 * sample format with the specified width has been select. If width is set to 0 1410 * the msbits will be set for any sample format with a width larger than the 1411 * specified msbits. 1412 * 1413 * Return: Zero if successful, or a negative error code on failure. 1414 */ 1415 int snd_pcm_hw_constraint_msbits(struct snd_pcm_runtime *runtime, 1416 unsigned int cond, 1417 unsigned int width, 1418 unsigned int msbits) 1419 { 1420 unsigned long l = (msbits << 16) | width; 1421 return snd_pcm_hw_rule_add(runtime, cond, -1, 1422 snd_pcm_hw_rule_msbits, 1423 (void*) l, 1424 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 1425 } 1426 EXPORT_SYMBOL(snd_pcm_hw_constraint_msbits); 1427 1428 static int snd_pcm_hw_rule_step(struct snd_pcm_hw_params *params, 1429 struct snd_pcm_hw_rule *rule) 1430 { 1431 unsigned long step = (unsigned long) rule->private; 1432 return snd_interval_step(hw_param_interval(params, rule->var), step); 1433 } 1434 1435 /** 1436 * snd_pcm_hw_constraint_step - add a hw constraint step rule 1437 * @runtime: PCM runtime instance 1438 * @cond: condition bits 1439 * @var: hw_params variable to apply the step constraint 1440 * @step: step size 1441 * 1442 * Return: Zero if successful, or a negative error code on failure. 1443 */ 1444 int snd_pcm_hw_constraint_step(struct snd_pcm_runtime *runtime, 1445 unsigned int cond, 1446 snd_pcm_hw_param_t var, 1447 unsigned long step) 1448 { 1449 return snd_pcm_hw_rule_add(runtime, cond, var, 1450 snd_pcm_hw_rule_step, (void *) step, 1451 var, -1); 1452 } 1453 EXPORT_SYMBOL(snd_pcm_hw_constraint_step); 1454 1455 static int snd_pcm_hw_rule_pow2(struct snd_pcm_hw_params *params, struct snd_pcm_hw_rule *rule) 1456 { 1457 static const unsigned int pow2_sizes[] = { 1458 1<<0, 1<<1, 1<<2, 1<<3, 1<<4, 1<<5, 1<<6, 1<<7, 1459 1<<8, 1<<9, 1<<10, 1<<11, 1<<12, 1<<13, 1<<14, 1<<15, 1460 1<<16, 1<<17, 1<<18, 1<<19, 1<<20, 1<<21, 1<<22, 1<<23, 1461 1<<24, 1<<25, 1<<26, 1<<27, 1<<28, 1<<29, 1<<30 1462 }; 1463 return snd_interval_list(hw_param_interval(params, rule->var), 1464 ARRAY_SIZE(pow2_sizes), pow2_sizes, 0); 1465 } 1466 1467 /** 1468 * snd_pcm_hw_constraint_pow2 - add a hw constraint power-of-2 rule 1469 * @runtime: PCM runtime instance 1470 * @cond: condition bits 1471 * @var: hw_params variable to apply the power-of-2 constraint 1472 * 1473 * Return: Zero if successful, or a negative error code on failure. 1474 */ 1475 int snd_pcm_hw_constraint_pow2(struct snd_pcm_runtime *runtime, 1476 unsigned int cond, 1477 snd_pcm_hw_param_t var) 1478 { 1479 return snd_pcm_hw_rule_add(runtime, cond, var, 1480 snd_pcm_hw_rule_pow2, NULL, 1481 var, -1); 1482 } 1483 EXPORT_SYMBOL(snd_pcm_hw_constraint_pow2); 1484 1485 static int snd_pcm_hw_rule_noresample_func(struct snd_pcm_hw_params *params, 1486 struct snd_pcm_hw_rule *rule) 1487 { 1488 unsigned int base_rate = (unsigned int)(uintptr_t)rule->private; 1489 struct snd_interval *rate; 1490 1491 rate = hw_param_interval(params, SNDRV_PCM_HW_PARAM_RATE); 1492 return snd_interval_list(rate, 1, &base_rate, 0); 1493 } 1494 1495 /** 1496 * snd_pcm_hw_rule_noresample - add a rule to allow disabling hw resampling 1497 * @runtime: PCM runtime instance 1498 * @base_rate: the rate at which the hardware does not resample 1499 * 1500 * Return: Zero if successful, or a negative error code on failure. 1501 */ 1502 int snd_pcm_hw_rule_noresample(struct snd_pcm_runtime *runtime, 1503 unsigned int base_rate) 1504 { 1505 return snd_pcm_hw_rule_add(runtime, SNDRV_PCM_HW_PARAMS_NORESAMPLE, 1506 SNDRV_PCM_HW_PARAM_RATE, 1507 snd_pcm_hw_rule_noresample_func, 1508 (void *)(uintptr_t)base_rate, 1509 SNDRV_PCM_HW_PARAM_RATE, -1); 1510 } 1511 EXPORT_SYMBOL(snd_pcm_hw_rule_noresample); 1512 1513 static void _snd_pcm_hw_param_any(struct snd_pcm_hw_params *params, 1514 snd_pcm_hw_param_t var) 1515 { 1516 if (hw_is_mask(var)) { 1517 snd_mask_any(hw_param_mask(params, var)); 1518 params->cmask |= 1 << var; 1519 params->rmask |= 1 << var; 1520 return; 1521 } 1522 if (hw_is_interval(var)) { 1523 snd_interval_any(hw_param_interval(params, var)); 1524 params->cmask |= 1 << var; 1525 params->rmask |= 1 << var; 1526 return; 1527 } 1528 snd_BUG(); 1529 } 1530 1531 void _snd_pcm_hw_params_any(struct snd_pcm_hw_params *params) 1532 { 1533 unsigned int k; 1534 memset(params, 0, sizeof(*params)); 1535 for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) 1536 _snd_pcm_hw_param_any(params, k); 1537 for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) 1538 _snd_pcm_hw_param_any(params, k); 1539 params->info = ~0U; 1540 } 1541 EXPORT_SYMBOL(_snd_pcm_hw_params_any); 1542 1543 /** 1544 * snd_pcm_hw_param_value - return @params field @var value 1545 * @params: the hw_params instance 1546 * @var: parameter to retrieve 1547 * @dir: pointer to the direction (-1,0,1) or %NULL 1548 * 1549 * Return: The value for field @var if it's fixed in configuration space 1550 * defined by @params. -%EINVAL otherwise. 1551 */ 1552 int snd_pcm_hw_param_value(const struct snd_pcm_hw_params *params, 1553 snd_pcm_hw_param_t var, int *dir) 1554 { 1555 if (hw_is_mask(var)) { 1556 const struct snd_mask *mask = hw_param_mask_c(params, var); 1557 if (!snd_mask_single(mask)) 1558 return -EINVAL; 1559 if (dir) 1560 *dir = 0; 1561 return snd_mask_value(mask); 1562 } 1563 if (hw_is_interval(var)) { 1564 const struct snd_interval *i = hw_param_interval_c(params, var); 1565 if (!snd_interval_single(i)) 1566 return -EINVAL; 1567 if (dir) 1568 *dir = i->openmin; 1569 return snd_interval_value(i); 1570 } 1571 return -EINVAL; 1572 } 1573 EXPORT_SYMBOL(snd_pcm_hw_param_value); 1574 1575 void _snd_pcm_hw_param_setempty(struct snd_pcm_hw_params *params, 1576 snd_pcm_hw_param_t var) 1577 { 1578 if (hw_is_mask(var)) { 1579 snd_mask_none(hw_param_mask(params, var)); 1580 params->cmask |= 1 << var; 1581 params->rmask |= 1 << var; 1582 } else if (hw_is_interval(var)) { 1583 snd_interval_none(hw_param_interval(params, var)); 1584 params->cmask |= 1 << var; 1585 params->rmask |= 1 << var; 1586 } else { 1587 snd_BUG(); 1588 } 1589 } 1590 EXPORT_SYMBOL(_snd_pcm_hw_param_setempty); 1591 1592 static int _snd_pcm_hw_param_first(struct snd_pcm_hw_params *params, 1593 snd_pcm_hw_param_t var) 1594 { 1595 int changed; 1596 if (hw_is_mask(var)) 1597 changed = snd_mask_refine_first(hw_param_mask(params, var)); 1598 else if (hw_is_interval(var)) 1599 changed = snd_interval_refine_first(hw_param_interval(params, var)); 1600 else 1601 return -EINVAL; 1602 if (changed > 0) { 1603 params->cmask |= 1 << var; 1604 params->rmask |= 1 << var; 1605 } 1606 return changed; 1607 } 1608 1609 1610 /** 1611 * snd_pcm_hw_param_first - refine config space and return minimum value 1612 * @pcm: PCM instance 1613 * @params: the hw_params instance 1614 * @var: parameter to retrieve 1615 * @dir: pointer to the direction (-1,0,1) or %NULL 1616 * 1617 * Inside configuration space defined by @params remove from @var all 1618 * values > minimum. Reduce configuration space accordingly. 1619 * 1620 * Return: The minimum, or a negative error code on failure. 1621 */ 1622 int snd_pcm_hw_param_first(struct snd_pcm_substream *pcm, 1623 struct snd_pcm_hw_params *params, 1624 snd_pcm_hw_param_t var, int *dir) 1625 { 1626 int changed = _snd_pcm_hw_param_first(params, var); 1627 if (changed < 0) 1628 return changed; 1629 if (params->rmask) { 1630 int err = snd_pcm_hw_refine(pcm, params); 1631 if (err < 0) 1632 return err; 1633 } 1634 return snd_pcm_hw_param_value(params, var, dir); 1635 } 1636 EXPORT_SYMBOL(snd_pcm_hw_param_first); 1637 1638 static int _snd_pcm_hw_param_last(struct snd_pcm_hw_params *params, 1639 snd_pcm_hw_param_t var) 1640 { 1641 int changed; 1642 if (hw_is_mask(var)) 1643 changed = snd_mask_refine_last(hw_param_mask(params, var)); 1644 else if (hw_is_interval(var)) 1645 changed = snd_interval_refine_last(hw_param_interval(params, var)); 1646 else 1647 return -EINVAL; 1648 if (changed > 0) { 1649 params->cmask |= 1 << var; 1650 params->rmask |= 1 << var; 1651 } 1652 return changed; 1653 } 1654 1655 1656 /** 1657 * snd_pcm_hw_param_last - refine config space and return maximum value 1658 * @pcm: PCM instance 1659 * @params: the hw_params instance 1660 * @var: parameter to retrieve 1661 * @dir: pointer to the direction (-1,0,1) or %NULL 1662 * 1663 * Inside configuration space defined by @params remove from @var all 1664 * values < maximum. Reduce configuration space accordingly. 1665 * 1666 * Return: The maximum, or a negative error code on failure. 1667 */ 1668 int snd_pcm_hw_param_last(struct snd_pcm_substream *pcm, 1669 struct snd_pcm_hw_params *params, 1670 snd_pcm_hw_param_t var, int *dir) 1671 { 1672 int changed = _snd_pcm_hw_param_last(params, var); 1673 if (changed < 0) 1674 return changed; 1675 if (params->rmask) { 1676 int err = snd_pcm_hw_refine(pcm, params); 1677 if (err < 0) 1678 return err; 1679 } 1680 return snd_pcm_hw_param_value(params, var, dir); 1681 } 1682 EXPORT_SYMBOL(snd_pcm_hw_param_last); 1683 1684 static int snd_pcm_lib_ioctl_reset(struct snd_pcm_substream *substream, 1685 void *arg) 1686 { 1687 struct snd_pcm_runtime *runtime = substream->runtime; 1688 unsigned long flags; 1689 snd_pcm_stream_lock_irqsave(substream, flags); 1690 if (snd_pcm_running(substream) && 1691 snd_pcm_update_hw_ptr(substream) >= 0) 1692 runtime->status->hw_ptr %= runtime->buffer_size; 1693 else { 1694 runtime->status->hw_ptr = 0; 1695 runtime->hw_ptr_wrap = 0; 1696 } 1697 snd_pcm_stream_unlock_irqrestore(substream, flags); 1698 return 0; 1699 } 1700 1701 static int snd_pcm_lib_ioctl_channel_info(struct snd_pcm_substream *substream, 1702 void *arg) 1703 { 1704 struct snd_pcm_channel_info *info = arg; 1705 struct snd_pcm_runtime *runtime = substream->runtime; 1706 int width; 1707 if (!(runtime->info & SNDRV_PCM_INFO_MMAP)) { 1708 info->offset = -1; 1709 return 0; 1710 } 1711 width = snd_pcm_format_physical_width(runtime->format); 1712 if (width < 0) 1713 return width; 1714 info->offset = 0; 1715 switch (runtime->access) { 1716 case SNDRV_PCM_ACCESS_MMAP_INTERLEAVED: 1717 case SNDRV_PCM_ACCESS_RW_INTERLEAVED: 1718 info->first = info->channel * width; 1719 info->step = runtime->channels * width; 1720 break; 1721 case SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED: 1722 case SNDRV_PCM_ACCESS_RW_NONINTERLEAVED: 1723 { 1724 size_t size = runtime->dma_bytes / runtime->channels; 1725 info->first = info->channel * size * 8; 1726 info->step = width; 1727 break; 1728 } 1729 default: 1730 snd_BUG(); 1731 break; 1732 } 1733 return 0; 1734 } 1735 1736 static int snd_pcm_lib_ioctl_fifo_size(struct snd_pcm_substream *substream, 1737 void *arg) 1738 { 1739 struct snd_pcm_hw_params *params = arg; 1740 snd_pcm_format_t format; 1741 int channels; 1742 ssize_t frame_size; 1743 1744 params->fifo_size = substream->runtime->hw.fifo_size; 1745 if (!(substream->runtime->hw.info & SNDRV_PCM_INFO_FIFO_IN_FRAMES)) { 1746 format = params_format(params); 1747 channels = params_channels(params); 1748 frame_size = snd_pcm_format_size(format, channels); 1749 if (frame_size > 0) 1750 params->fifo_size /= frame_size; 1751 } 1752 return 0; 1753 } 1754 1755 /** 1756 * snd_pcm_lib_ioctl - a generic PCM ioctl callback 1757 * @substream: the pcm substream instance 1758 * @cmd: ioctl command 1759 * @arg: ioctl argument 1760 * 1761 * Processes the generic ioctl commands for PCM. 1762 * Can be passed as the ioctl callback for PCM ops. 1763 * 1764 * Return: Zero if successful, or a negative error code on failure. 1765 */ 1766 int snd_pcm_lib_ioctl(struct snd_pcm_substream *substream, 1767 unsigned int cmd, void *arg) 1768 { 1769 switch (cmd) { 1770 case SNDRV_PCM_IOCTL1_RESET: 1771 return snd_pcm_lib_ioctl_reset(substream, arg); 1772 case SNDRV_PCM_IOCTL1_CHANNEL_INFO: 1773 return snd_pcm_lib_ioctl_channel_info(substream, arg); 1774 case SNDRV_PCM_IOCTL1_FIFO_SIZE: 1775 return snd_pcm_lib_ioctl_fifo_size(substream, arg); 1776 } 1777 return -ENXIO; 1778 } 1779 EXPORT_SYMBOL(snd_pcm_lib_ioctl); 1780 1781 /** 1782 * snd_pcm_period_elapsed_under_stream_lock() - update the status of runtime for the next period 1783 * under acquired lock of PCM substream. 1784 * @substream: the instance of pcm substream. 1785 * 1786 * This function is called when the batch of audio data frames as the same size as the period of 1787 * buffer is already processed in audio data transmission. 1788 * 1789 * The call of function updates the status of runtime with the latest position of audio data 1790 * transmission, checks overrun and underrun over buffer, awaken user processes from waiting for 1791 * available audio data frames, sampling audio timestamp, and performs stop or drain the PCM 1792 * substream according to configured threshold. 1793 * 1794 * The function is intended to use for the case that PCM driver operates audio data frames under 1795 * acquired lock of PCM substream; e.g. in callback of any operation of &snd_pcm_ops in process 1796 * context. In any interrupt context, it's preferrable to use ``snd_pcm_period_elapsed()`` instead 1797 * since lock of PCM substream should be acquired in advance. 1798 * 1799 * Developer should pay enough attention that some callbacks in &snd_pcm_ops are done by the call of 1800 * function: 1801 * 1802 * - .pointer - to retrieve current position of audio data transmission by frame count or XRUN state. 1803 * - .trigger - with SNDRV_PCM_TRIGGER_STOP at XRUN or DRAINING state. 1804 * - .get_time_info - to retrieve audio time stamp if needed. 1805 * 1806 * Even if more than one periods have elapsed since the last call, you have to call this only once. 1807 */ 1808 void snd_pcm_period_elapsed_under_stream_lock(struct snd_pcm_substream *substream) 1809 { 1810 struct snd_pcm_runtime *runtime; 1811 1812 if (PCM_RUNTIME_CHECK(substream)) 1813 return; 1814 runtime = substream->runtime; 1815 1816 if (!snd_pcm_running(substream) || 1817 snd_pcm_update_hw_ptr0(substream, 1) < 0) 1818 goto _end; 1819 1820 #ifdef CONFIG_SND_PCM_TIMER 1821 if (substream->timer_running) 1822 snd_timer_interrupt(substream->timer, 1); 1823 #endif 1824 _end: 1825 snd_kill_fasync(runtime->fasync, SIGIO, POLL_IN); 1826 } 1827 EXPORT_SYMBOL(snd_pcm_period_elapsed_under_stream_lock); 1828 1829 /** 1830 * snd_pcm_period_elapsed() - update the status of runtime for the next period by acquiring lock of 1831 * PCM substream. 1832 * @substream: the instance of PCM substream. 1833 * 1834 * This function is mostly similar to ``snd_pcm_period_elapsed_under_stream_lock()`` except for 1835 * acquiring lock of PCM substream voluntarily. 1836 * 1837 * It's typically called by any type of IRQ handler when hardware IRQ occurs to notify event that 1838 * the batch of audio data frames as the same size as the period of buffer is already processed in 1839 * audio data transmission. 1840 */ 1841 void snd_pcm_period_elapsed(struct snd_pcm_substream *substream) 1842 { 1843 unsigned long flags; 1844 1845 if (snd_BUG_ON(!substream)) 1846 return; 1847 1848 snd_pcm_stream_lock_irqsave(substream, flags); 1849 snd_pcm_period_elapsed_under_stream_lock(substream); 1850 snd_pcm_stream_unlock_irqrestore(substream, flags); 1851 } 1852 EXPORT_SYMBOL(snd_pcm_period_elapsed); 1853 1854 /* 1855 * Wait until avail_min data becomes available 1856 * Returns a negative error code if any error occurs during operation. 1857 * The available space is stored on availp. When err = 0 and avail = 0 1858 * on the capture stream, it indicates the stream is in DRAINING state. 1859 */ 1860 static int wait_for_avail(struct snd_pcm_substream *substream, 1861 snd_pcm_uframes_t *availp) 1862 { 1863 struct snd_pcm_runtime *runtime = substream->runtime; 1864 int is_playback = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; 1865 wait_queue_entry_t wait; 1866 int err = 0; 1867 snd_pcm_uframes_t avail = 0; 1868 long wait_time, tout; 1869 1870 init_waitqueue_entry(&wait, current); 1871 set_current_state(TASK_INTERRUPTIBLE); 1872 add_wait_queue(&runtime->tsleep, &wait); 1873 1874 if (runtime->no_period_wakeup) 1875 wait_time = MAX_SCHEDULE_TIMEOUT; 1876 else { 1877 /* use wait time from substream if available */ 1878 if (substream->wait_time) { 1879 wait_time = substream->wait_time; 1880 } else { 1881 wait_time = 10; 1882 1883 if (runtime->rate) { 1884 long t = runtime->period_size * 2 / 1885 runtime->rate; 1886 wait_time = max(t, wait_time); 1887 } 1888 wait_time = msecs_to_jiffies(wait_time * 1000); 1889 } 1890 } 1891 1892 for (;;) { 1893 if (signal_pending(current)) { 1894 err = -ERESTARTSYS; 1895 break; 1896 } 1897 1898 /* 1899 * We need to check if space became available already 1900 * (and thus the wakeup happened already) first to close 1901 * the race of space already having become available. 1902 * This check must happen after been added to the waitqueue 1903 * and having current state be INTERRUPTIBLE. 1904 */ 1905 avail = snd_pcm_avail(substream); 1906 if (avail >= runtime->twake) 1907 break; 1908 snd_pcm_stream_unlock_irq(substream); 1909 1910 tout = schedule_timeout(wait_time); 1911 1912 snd_pcm_stream_lock_irq(substream); 1913 set_current_state(TASK_INTERRUPTIBLE); 1914 switch (runtime->status->state) { 1915 case SNDRV_PCM_STATE_SUSPENDED: 1916 err = -ESTRPIPE; 1917 goto _endloop; 1918 case SNDRV_PCM_STATE_XRUN: 1919 err = -EPIPE; 1920 goto _endloop; 1921 case SNDRV_PCM_STATE_DRAINING: 1922 if (is_playback) 1923 err = -EPIPE; 1924 else 1925 avail = 0; /* indicate draining */ 1926 goto _endloop; 1927 case SNDRV_PCM_STATE_OPEN: 1928 case SNDRV_PCM_STATE_SETUP: 1929 case SNDRV_PCM_STATE_DISCONNECTED: 1930 err = -EBADFD; 1931 goto _endloop; 1932 case SNDRV_PCM_STATE_PAUSED: 1933 continue; 1934 } 1935 if (!tout) { 1936 pcm_dbg(substream->pcm, 1937 "%s write error (DMA or IRQ trouble?)\n", 1938 is_playback ? "playback" : "capture"); 1939 err = -EIO; 1940 break; 1941 } 1942 } 1943 _endloop: 1944 set_current_state(TASK_RUNNING); 1945 remove_wait_queue(&runtime->tsleep, &wait); 1946 *availp = avail; 1947 return err; 1948 } 1949 1950 typedef int (*pcm_transfer_f)(struct snd_pcm_substream *substream, 1951 int channel, unsigned long hwoff, 1952 void *buf, unsigned long bytes); 1953 1954 typedef int (*pcm_copy_f)(struct snd_pcm_substream *, snd_pcm_uframes_t, void *, 1955 snd_pcm_uframes_t, snd_pcm_uframes_t, pcm_transfer_f); 1956 1957 /* calculate the target DMA-buffer position to be written/read */ 1958 static void *get_dma_ptr(struct snd_pcm_runtime *runtime, 1959 int channel, unsigned long hwoff) 1960 { 1961 return runtime->dma_area + hwoff + 1962 channel * (runtime->dma_bytes / runtime->channels); 1963 } 1964 1965 /* default copy_user ops for write; used for both interleaved and non- modes */ 1966 static int default_write_copy(struct snd_pcm_substream *substream, 1967 int channel, unsigned long hwoff, 1968 void *buf, unsigned long bytes) 1969 { 1970 if (copy_from_user(get_dma_ptr(substream->runtime, channel, hwoff), 1971 (void __user *)buf, bytes)) 1972 return -EFAULT; 1973 return 0; 1974 } 1975 1976 /* default copy_kernel ops for write */ 1977 static int default_write_copy_kernel(struct snd_pcm_substream *substream, 1978 int channel, unsigned long hwoff, 1979 void *buf, unsigned long bytes) 1980 { 1981 memcpy(get_dma_ptr(substream->runtime, channel, hwoff), buf, bytes); 1982 return 0; 1983 } 1984 1985 /* fill silence instead of copy data; called as a transfer helper 1986 * from __snd_pcm_lib_write() or directly from noninterleaved_copy() when 1987 * a NULL buffer is passed 1988 */ 1989 static int fill_silence(struct snd_pcm_substream *substream, int channel, 1990 unsigned long hwoff, void *buf, unsigned long bytes) 1991 { 1992 struct snd_pcm_runtime *runtime = substream->runtime; 1993 1994 if (substream->stream != SNDRV_PCM_STREAM_PLAYBACK) 1995 return 0; 1996 if (substream->ops->fill_silence) 1997 return substream->ops->fill_silence(substream, channel, 1998 hwoff, bytes); 1999 2000 snd_pcm_format_set_silence(runtime->format, 2001 get_dma_ptr(runtime, channel, hwoff), 2002 bytes_to_samples(runtime, bytes)); 2003 return 0; 2004 } 2005 2006 /* default copy_user ops for read; used for both interleaved and non- modes */ 2007 static int default_read_copy(struct snd_pcm_substream *substream, 2008 int channel, unsigned long hwoff, 2009 void *buf, unsigned long bytes) 2010 { 2011 if (copy_to_user((void __user *)buf, 2012 get_dma_ptr(substream->runtime, channel, hwoff), 2013 bytes)) 2014 return -EFAULT; 2015 return 0; 2016 } 2017 2018 /* default copy_kernel ops for read */ 2019 static int default_read_copy_kernel(struct snd_pcm_substream *substream, 2020 int channel, unsigned long hwoff, 2021 void *buf, unsigned long bytes) 2022 { 2023 memcpy(buf, get_dma_ptr(substream->runtime, channel, hwoff), bytes); 2024 return 0; 2025 } 2026 2027 /* call transfer function with the converted pointers and sizes; 2028 * for interleaved mode, it's one shot for all samples 2029 */ 2030 static int interleaved_copy(struct snd_pcm_substream *substream, 2031 snd_pcm_uframes_t hwoff, void *data, 2032 snd_pcm_uframes_t off, 2033 snd_pcm_uframes_t frames, 2034 pcm_transfer_f transfer) 2035 { 2036 struct snd_pcm_runtime *runtime = substream->runtime; 2037 2038 /* convert to bytes */ 2039 hwoff = frames_to_bytes(runtime, hwoff); 2040 off = frames_to_bytes(runtime, off); 2041 frames = frames_to_bytes(runtime, frames); 2042 return transfer(substream, 0, hwoff, data + off, frames); 2043 } 2044 2045 /* call transfer function with the converted pointers and sizes for each 2046 * non-interleaved channel; when buffer is NULL, silencing instead of copying 2047 */ 2048 static int noninterleaved_copy(struct snd_pcm_substream *substream, 2049 snd_pcm_uframes_t hwoff, void *data, 2050 snd_pcm_uframes_t off, 2051 snd_pcm_uframes_t frames, 2052 pcm_transfer_f transfer) 2053 { 2054 struct snd_pcm_runtime *runtime = substream->runtime; 2055 int channels = runtime->channels; 2056 void **bufs = data; 2057 int c, err; 2058 2059 /* convert to bytes; note that it's not frames_to_bytes() here. 2060 * in non-interleaved mode, we copy for each channel, thus 2061 * each copy is n_samples bytes x channels = whole frames. 2062 */ 2063 off = samples_to_bytes(runtime, off); 2064 frames = samples_to_bytes(runtime, frames); 2065 hwoff = samples_to_bytes(runtime, hwoff); 2066 for (c = 0; c < channels; ++c, ++bufs) { 2067 if (!data || !*bufs) 2068 err = fill_silence(substream, c, hwoff, NULL, frames); 2069 else 2070 err = transfer(substream, c, hwoff, *bufs + off, 2071 frames); 2072 if (err < 0) 2073 return err; 2074 } 2075 return 0; 2076 } 2077 2078 /* fill silence on the given buffer position; 2079 * called from snd_pcm_playback_silence() 2080 */ 2081 static int fill_silence_frames(struct snd_pcm_substream *substream, 2082 snd_pcm_uframes_t off, snd_pcm_uframes_t frames) 2083 { 2084 if (substream->runtime->access == SNDRV_PCM_ACCESS_RW_INTERLEAVED || 2085 substream->runtime->access == SNDRV_PCM_ACCESS_MMAP_INTERLEAVED) 2086 return interleaved_copy(substream, off, NULL, 0, frames, 2087 fill_silence); 2088 else 2089 return noninterleaved_copy(substream, off, NULL, 0, frames, 2090 fill_silence); 2091 } 2092 2093 /* sanity-check for read/write methods */ 2094 static int pcm_sanity_check(struct snd_pcm_substream *substream) 2095 { 2096 struct snd_pcm_runtime *runtime; 2097 if (PCM_RUNTIME_CHECK(substream)) 2098 return -ENXIO; 2099 runtime = substream->runtime; 2100 if (snd_BUG_ON(!substream->ops->copy_user && !runtime->dma_area)) 2101 return -EINVAL; 2102 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 2103 return -EBADFD; 2104 return 0; 2105 } 2106 2107 static int pcm_accessible_state(struct snd_pcm_runtime *runtime) 2108 { 2109 switch (runtime->status->state) { 2110 case SNDRV_PCM_STATE_PREPARED: 2111 case SNDRV_PCM_STATE_RUNNING: 2112 case SNDRV_PCM_STATE_PAUSED: 2113 return 0; 2114 case SNDRV_PCM_STATE_XRUN: 2115 return -EPIPE; 2116 case SNDRV_PCM_STATE_SUSPENDED: 2117 return -ESTRPIPE; 2118 default: 2119 return -EBADFD; 2120 } 2121 } 2122 2123 /* update to the given appl_ptr and call ack callback if needed; 2124 * when an error is returned, take back to the original value 2125 */ 2126 int pcm_lib_apply_appl_ptr(struct snd_pcm_substream *substream, 2127 snd_pcm_uframes_t appl_ptr) 2128 { 2129 struct snd_pcm_runtime *runtime = substream->runtime; 2130 snd_pcm_uframes_t old_appl_ptr = runtime->control->appl_ptr; 2131 snd_pcm_sframes_t diff; 2132 int ret; 2133 2134 if (old_appl_ptr == appl_ptr) 2135 return 0; 2136 2137 if (appl_ptr >= runtime->boundary) 2138 return -EINVAL; 2139 /* 2140 * check if a rewind is requested by the application 2141 */ 2142 if (substream->runtime->info & SNDRV_PCM_INFO_NO_REWINDS) { 2143 diff = appl_ptr - old_appl_ptr; 2144 if (diff >= 0) { 2145 if (diff > runtime->buffer_size) 2146 return -EINVAL; 2147 } else { 2148 if (runtime->boundary + diff > runtime->buffer_size) 2149 return -EINVAL; 2150 } 2151 } 2152 2153 runtime->control->appl_ptr = appl_ptr; 2154 if (substream->ops->ack) { 2155 ret = substream->ops->ack(substream); 2156 if (ret < 0) { 2157 runtime->control->appl_ptr = old_appl_ptr; 2158 return ret; 2159 } 2160 } 2161 2162 trace_applptr(substream, old_appl_ptr, appl_ptr); 2163 2164 return 0; 2165 } 2166 2167 /* the common loop for read/write data */ 2168 snd_pcm_sframes_t __snd_pcm_lib_xfer(struct snd_pcm_substream *substream, 2169 void *data, bool interleaved, 2170 snd_pcm_uframes_t size, bool in_kernel) 2171 { 2172 struct snd_pcm_runtime *runtime = substream->runtime; 2173 snd_pcm_uframes_t xfer = 0; 2174 snd_pcm_uframes_t offset = 0; 2175 snd_pcm_uframes_t avail; 2176 pcm_copy_f writer; 2177 pcm_transfer_f transfer; 2178 bool nonblock; 2179 bool is_playback; 2180 int err; 2181 2182 err = pcm_sanity_check(substream); 2183 if (err < 0) 2184 return err; 2185 2186 is_playback = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; 2187 if (interleaved) { 2188 if (runtime->access != SNDRV_PCM_ACCESS_RW_INTERLEAVED && 2189 runtime->channels > 1) 2190 return -EINVAL; 2191 writer = interleaved_copy; 2192 } else { 2193 if (runtime->access != SNDRV_PCM_ACCESS_RW_NONINTERLEAVED) 2194 return -EINVAL; 2195 writer = noninterleaved_copy; 2196 } 2197 2198 if (!data) { 2199 if (is_playback) 2200 transfer = fill_silence; 2201 else 2202 return -EINVAL; 2203 } else if (in_kernel) { 2204 if (substream->ops->copy_kernel) 2205 transfer = substream->ops->copy_kernel; 2206 else 2207 transfer = is_playback ? 2208 default_write_copy_kernel : default_read_copy_kernel; 2209 } else { 2210 if (substream->ops->copy_user) 2211 transfer = (pcm_transfer_f)substream->ops->copy_user; 2212 else 2213 transfer = is_playback ? 2214 default_write_copy : default_read_copy; 2215 } 2216 2217 if (size == 0) 2218 return 0; 2219 2220 nonblock = !!(substream->f_flags & O_NONBLOCK); 2221 2222 snd_pcm_stream_lock_irq(substream); 2223 err = pcm_accessible_state(runtime); 2224 if (err < 0) 2225 goto _end_unlock; 2226 2227 runtime->twake = runtime->control->avail_min ? : 1; 2228 if (runtime->status->state == SNDRV_PCM_STATE_RUNNING) 2229 snd_pcm_update_hw_ptr(substream); 2230 2231 /* 2232 * If size < start_threshold, wait indefinitely. Another 2233 * thread may start capture 2234 */ 2235 if (!is_playback && 2236 runtime->status->state == SNDRV_PCM_STATE_PREPARED && 2237 size >= runtime->start_threshold) { 2238 err = snd_pcm_start(substream); 2239 if (err < 0) 2240 goto _end_unlock; 2241 } 2242 2243 avail = snd_pcm_avail(substream); 2244 2245 while (size > 0) { 2246 snd_pcm_uframes_t frames, appl_ptr, appl_ofs; 2247 snd_pcm_uframes_t cont; 2248 if (!avail) { 2249 if (!is_playback && 2250 runtime->status->state == SNDRV_PCM_STATE_DRAINING) { 2251 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 2252 goto _end_unlock; 2253 } 2254 if (nonblock) { 2255 err = -EAGAIN; 2256 goto _end_unlock; 2257 } 2258 runtime->twake = min_t(snd_pcm_uframes_t, size, 2259 runtime->control->avail_min ? : 1); 2260 err = wait_for_avail(substream, &avail); 2261 if (err < 0) 2262 goto _end_unlock; 2263 if (!avail) 2264 continue; /* draining */ 2265 } 2266 frames = size > avail ? avail : size; 2267 appl_ptr = READ_ONCE(runtime->control->appl_ptr); 2268 appl_ofs = appl_ptr % runtime->buffer_size; 2269 cont = runtime->buffer_size - appl_ofs; 2270 if (frames > cont) 2271 frames = cont; 2272 if (snd_BUG_ON(!frames)) { 2273 err = -EINVAL; 2274 goto _end_unlock; 2275 } 2276 if (!atomic_inc_unless_negative(&runtime->buffer_accessing)) { 2277 err = -EBUSY; 2278 goto _end_unlock; 2279 } 2280 snd_pcm_stream_unlock_irq(substream); 2281 if (!is_playback) 2282 snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_CPU); 2283 err = writer(substream, appl_ofs, data, offset, frames, 2284 transfer); 2285 if (is_playback) 2286 snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_DEVICE); 2287 snd_pcm_stream_lock_irq(substream); 2288 atomic_dec(&runtime->buffer_accessing); 2289 if (err < 0) 2290 goto _end_unlock; 2291 err = pcm_accessible_state(runtime); 2292 if (err < 0) 2293 goto _end_unlock; 2294 appl_ptr += frames; 2295 if (appl_ptr >= runtime->boundary) 2296 appl_ptr -= runtime->boundary; 2297 err = pcm_lib_apply_appl_ptr(substream, appl_ptr); 2298 if (err < 0) 2299 goto _end_unlock; 2300 2301 offset += frames; 2302 size -= frames; 2303 xfer += frames; 2304 avail -= frames; 2305 if (is_playback && 2306 runtime->status->state == SNDRV_PCM_STATE_PREPARED && 2307 snd_pcm_playback_hw_avail(runtime) >= (snd_pcm_sframes_t)runtime->start_threshold) { 2308 err = snd_pcm_start(substream); 2309 if (err < 0) 2310 goto _end_unlock; 2311 } 2312 } 2313 _end_unlock: 2314 runtime->twake = 0; 2315 if (xfer > 0 && err >= 0) 2316 snd_pcm_update_state(substream, runtime); 2317 snd_pcm_stream_unlock_irq(substream); 2318 return xfer > 0 ? (snd_pcm_sframes_t)xfer : err; 2319 } 2320 EXPORT_SYMBOL(__snd_pcm_lib_xfer); 2321 2322 /* 2323 * standard channel mapping helpers 2324 */ 2325 2326 /* default channel maps for multi-channel playbacks, up to 8 channels */ 2327 const struct snd_pcm_chmap_elem snd_pcm_std_chmaps[] = { 2328 { .channels = 1, 2329 .map = { SNDRV_CHMAP_MONO } }, 2330 { .channels = 2, 2331 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR } }, 2332 { .channels = 4, 2333 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR, 2334 SNDRV_CHMAP_RL, SNDRV_CHMAP_RR } }, 2335 { .channels = 6, 2336 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR, 2337 SNDRV_CHMAP_RL, SNDRV_CHMAP_RR, 2338 SNDRV_CHMAP_FC, SNDRV_CHMAP_LFE } }, 2339 { .channels = 8, 2340 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR, 2341 SNDRV_CHMAP_RL, SNDRV_CHMAP_RR, 2342 SNDRV_CHMAP_FC, SNDRV_CHMAP_LFE, 2343 SNDRV_CHMAP_SL, SNDRV_CHMAP_SR } }, 2344 { } 2345 }; 2346 EXPORT_SYMBOL_GPL(snd_pcm_std_chmaps); 2347 2348 /* alternative channel maps with CLFE <-> surround swapped for 6/8 channels */ 2349 const struct snd_pcm_chmap_elem snd_pcm_alt_chmaps[] = { 2350 { .channels = 1, 2351 .map = { SNDRV_CHMAP_MONO } }, 2352 { .channels = 2, 2353 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR } }, 2354 { .channels = 4, 2355 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR, 2356 SNDRV_CHMAP_RL, SNDRV_CHMAP_RR } }, 2357 { .channels = 6, 2358 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR, 2359 SNDRV_CHMAP_FC, SNDRV_CHMAP_LFE, 2360 SNDRV_CHMAP_RL, SNDRV_CHMAP_RR } }, 2361 { .channels = 8, 2362 .map = { SNDRV_CHMAP_FL, SNDRV_CHMAP_FR, 2363 SNDRV_CHMAP_FC, SNDRV_CHMAP_LFE, 2364 SNDRV_CHMAP_RL, SNDRV_CHMAP_RR, 2365 SNDRV_CHMAP_SL, SNDRV_CHMAP_SR } }, 2366 { } 2367 }; 2368 EXPORT_SYMBOL_GPL(snd_pcm_alt_chmaps); 2369 2370 static bool valid_chmap_channels(const struct snd_pcm_chmap *info, int ch) 2371 { 2372 if (ch > info->max_channels) 2373 return false; 2374 return !info->channel_mask || (info->channel_mask & (1U << ch)); 2375 } 2376 2377 static int pcm_chmap_ctl_info(struct snd_kcontrol *kcontrol, 2378 struct snd_ctl_elem_info *uinfo) 2379 { 2380 struct snd_pcm_chmap *info = snd_kcontrol_chip(kcontrol); 2381 2382 uinfo->type = SNDRV_CTL_ELEM_TYPE_INTEGER; 2383 uinfo->count = info->max_channels; 2384 uinfo->value.integer.min = 0; 2385 uinfo->value.integer.max = SNDRV_CHMAP_LAST; 2386 return 0; 2387 } 2388 2389 /* get callback for channel map ctl element 2390 * stores the channel position firstly matching with the current channels 2391 */ 2392 static int pcm_chmap_ctl_get(struct snd_kcontrol *kcontrol, 2393 struct snd_ctl_elem_value *ucontrol) 2394 { 2395 struct snd_pcm_chmap *info = snd_kcontrol_chip(kcontrol); 2396 unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); 2397 struct snd_pcm_substream *substream; 2398 const struct snd_pcm_chmap_elem *map; 2399 2400 if (!info->chmap) 2401 return -EINVAL; 2402 substream = snd_pcm_chmap_substream(info, idx); 2403 if (!substream) 2404 return -ENODEV; 2405 memset(ucontrol->value.integer.value, 0, 2406 sizeof(long) * info->max_channels); 2407 if (!substream->runtime) 2408 return 0; /* no channels set */ 2409 for (map = info->chmap; map->channels; map++) { 2410 int i; 2411 if (map->channels == substream->runtime->channels && 2412 valid_chmap_channels(info, map->channels)) { 2413 for (i = 0; i < map->channels; i++) 2414 ucontrol->value.integer.value[i] = map->map[i]; 2415 return 0; 2416 } 2417 } 2418 return -EINVAL; 2419 } 2420 2421 /* tlv callback for channel map ctl element 2422 * expands the pre-defined channel maps in a form of TLV 2423 */ 2424 static int pcm_chmap_ctl_tlv(struct snd_kcontrol *kcontrol, int op_flag, 2425 unsigned int size, unsigned int __user *tlv) 2426 { 2427 struct snd_pcm_chmap *info = snd_kcontrol_chip(kcontrol); 2428 const struct snd_pcm_chmap_elem *map; 2429 unsigned int __user *dst; 2430 int c, count = 0; 2431 2432 if (!info->chmap) 2433 return -EINVAL; 2434 if (size < 8) 2435 return -ENOMEM; 2436 if (put_user(SNDRV_CTL_TLVT_CONTAINER, tlv)) 2437 return -EFAULT; 2438 size -= 8; 2439 dst = tlv + 2; 2440 for (map = info->chmap; map->channels; map++) { 2441 int chs_bytes = map->channels * 4; 2442 if (!valid_chmap_channels(info, map->channels)) 2443 continue; 2444 if (size < 8) 2445 return -ENOMEM; 2446 if (put_user(SNDRV_CTL_TLVT_CHMAP_FIXED, dst) || 2447 put_user(chs_bytes, dst + 1)) 2448 return -EFAULT; 2449 dst += 2; 2450 size -= 8; 2451 count += 8; 2452 if (size < chs_bytes) 2453 return -ENOMEM; 2454 size -= chs_bytes; 2455 count += chs_bytes; 2456 for (c = 0; c < map->channels; c++) { 2457 if (put_user(map->map[c], dst)) 2458 return -EFAULT; 2459 dst++; 2460 } 2461 } 2462 if (put_user(count, tlv + 1)) 2463 return -EFAULT; 2464 return 0; 2465 } 2466 2467 static void pcm_chmap_ctl_private_free(struct snd_kcontrol *kcontrol) 2468 { 2469 struct snd_pcm_chmap *info = snd_kcontrol_chip(kcontrol); 2470 info->pcm->streams[info->stream].chmap_kctl = NULL; 2471 kfree(info); 2472 } 2473 2474 /** 2475 * snd_pcm_add_chmap_ctls - create channel-mapping control elements 2476 * @pcm: the assigned PCM instance 2477 * @stream: stream direction 2478 * @chmap: channel map elements (for query) 2479 * @max_channels: the max number of channels for the stream 2480 * @private_value: the value passed to each kcontrol's private_value field 2481 * @info_ret: store struct snd_pcm_chmap instance if non-NULL 2482 * 2483 * Create channel-mapping control elements assigned to the given PCM stream(s). 2484 * Return: Zero if successful, or a negative error value. 2485 */ 2486 int snd_pcm_add_chmap_ctls(struct snd_pcm *pcm, int stream, 2487 const struct snd_pcm_chmap_elem *chmap, 2488 int max_channels, 2489 unsigned long private_value, 2490 struct snd_pcm_chmap **info_ret) 2491 { 2492 struct snd_pcm_chmap *info; 2493 struct snd_kcontrol_new knew = { 2494 .iface = SNDRV_CTL_ELEM_IFACE_PCM, 2495 .access = SNDRV_CTL_ELEM_ACCESS_READ | 2496 SNDRV_CTL_ELEM_ACCESS_TLV_READ | 2497 SNDRV_CTL_ELEM_ACCESS_TLV_CALLBACK, 2498 .info = pcm_chmap_ctl_info, 2499 .get = pcm_chmap_ctl_get, 2500 .tlv.c = pcm_chmap_ctl_tlv, 2501 }; 2502 int err; 2503 2504 if (WARN_ON(pcm->streams[stream].chmap_kctl)) 2505 return -EBUSY; 2506 info = kzalloc(sizeof(*info), GFP_KERNEL); 2507 if (!info) 2508 return -ENOMEM; 2509 info->pcm = pcm; 2510 info->stream = stream; 2511 info->chmap = chmap; 2512 info->max_channels = max_channels; 2513 if (stream == SNDRV_PCM_STREAM_PLAYBACK) 2514 knew.name = "Playback Channel Map"; 2515 else 2516 knew.name = "Capture Channel Map"; 2517 knew.device = pcm->device; 2518 knew.count = pcm->streams[stream].substream_count; 2519 knew.private_value = private_value; 2520 info->kctl = snd_ctl_new1(&knew, info); 2521 if (!info->kctl) { 2522 kfree(info); 2523 return -ENOMEM; 2524 } 2525 info->kctl->private_free = pcm_chmap_ctl_private_free; 2526 err = snd_ctl_add(pcm->card, info->kctl); 2527 if (err < 0) 2528 return err; 2529 pcm->streams[stream].chmap_kctl = info->kctl; 2530 if (info_ret) 2531 *info_ret = info; 2532 return 0; 2533 } 2534 EXPORT_SYMBOL_GPL(snd_pcm_add_chmap_ctls); 2535