1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Digital Audio (PCM) abstract layer 4 * Copyright (c) by Jaroslav Kysela <perex@perex.cz> 5 */ 6 7 #include <linux/compat.h> 8 #include <linux/mm.h> 9 #include <linux/module.h> 10 #include <linux/file.h> 11 #include <linux/slab.h> 12 #include <linux/sched/signal.h> 13 #include <linux/time.h> 14 #include <linux/pm_qos.h> 15 #include <linux/io.h> 16 #include <linux/dma-mapping.h> 17 #include <linux/vmalloc.h> 18 #include <sound/core.h> 19 #include <sound/control.h> 20 #include <sound/info.h> 21 #include <sound/pcm.h> 22 #include <sound/pcm_params.h> 23 #include <sound/timer.h> 24 #include <sound/minors.h> 25 #include <linux/uio.h> 26 #include <linux/delay.h> 27 28 #include "pcm_local.h" 29 30 #ifdef CONFIG_SND_DEBUG 31 #define CREATE_TRACE_POINTS 32 #include "pcm_param_trace.h" 33 #else 34 #define trace_hw_mask_param_enabled() 0 35 #define trace_hw_interval_param_enabled() 0 36 #define trace_hw_mask_param(substream, type, index, prev, curr) 37 #define trace_hw_interval_param(substream, type, index, prev, curr) 38 #endif 39 40 /* 41 * Compatibility 42 */ 43 44 struct snd_pcm_hw_params_old { 45 unsigned int flags; 46 unsigned int masks[SNDRV_PCM_HW_PARAM_SUBFORMAT - 47 SNDRV_PCM_HW_PARAM_ACCESS + 1]; 48 struct snd_interval intervals[SNDRV_PCM_HW_PARAM_TICK_TIME - 49 SNDRV_PCM_HW_PARAM_SAMPLE_BITS + 1]; 50 unsigned int rmask; 51 unsigned int cmask; 52 unsigned int info; 53 unsigned int msbits; 54 unsigned int rate_num; 55 unsigned int rate_den; 56 snd_pcm_uframes_t fifo_size; 57 unsigned char reserved[64]; 58 }; 59 60 #ifdef CONFIG_SND_SUPPORT_OLD_API 61 #define SNDRV_PCM_IOCTL_HW_REFINE_OLD _IOWR('A', 0x10, struct snd_pcm_hw_params_old) 62 #define SNDRV_PCM_IOCTL_HW_PARAMS_OLD _IOWR('A', 0x11, struct snd_pcm_hw_params_old) 63 64 static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream, 65 struct snd_pcm_hw_params_old __user * _oparams); 66 static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream, 67 struct snd_pcm_hw_params_old __user * _oparams); 68 #endif 69 static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream); 70 71 /* 72 * 73 */ 74 75 static DECLARE_RWSEM(snd_pcm_link_rwsem); 76 77 void snd_pcm_group_init(struct snd_pcm_group *group) 78 { 79 spin_lock_init(&group->lock); 80 mutex_init(&group->mutex); 81 INIT_LIST_HEAD(&group->substreams); 82 refcount_set(&group->refs, 1); 83 } 84 85 /* define group lock helpers */ 86 #define DEFINE_PCM_GROUP_LOCK(action, mutex_action) \ 87 static void snd_pcm_group_ ## action(struct snd_pcm_group *group, bool nonatomic) \ 88 { \ 89 if (nonatomic) \ 90 mutex_ ## mutex_action(&group->mutex); \ 91 else \ 92 spin_ ## action(&group->lock); \ 93 } 94 95 DEFINE_PCM_GROUP_LOCK(lock, lock); 96 DEFINE_PCM_GROUP_LOCK(unlock, unlock); 97 DEFINE_PCM_GROUP_LOCK(lock_irq, lock); 98 DEFINE_PCM_GROUP_LOCK(unlock_irq, unlock); 99 100 /** 101 * snd_pcm_stream_lock - Lock the PCM stream 102 * @substream: PCM substream 103 * 104 * This locks the PCM stream's spinlock or mutex depending on the nonatomic 105 * flag of the given substream. This also takes the global link rw lock 106 * (or rw sem), too, for avoiding the race with linked streams. 107 */ 108 void snd_pcm_stream_lock(struct snd_pcm_substream *substream) 109 { 110 snd_pcm_group_lock(&substream->self_group, substream->pcm->nonatomic); 111 } 112 EXPORT_SYMBOL_GPL(snd_pcm_stream_lock); 113 114 /** 115 * snd_pcm_stream_unlock - Unlock the PCM stream 116 * @substream: PCM substream 117 * 118 * This unlocks the PCM stream that has been locked via snd_pcm_stream_lock(). 119 */ 120 void snd_pcm_stream_unlock(struct snd_pcm_substream *substream) 121 { 122 snd_pcm_group_unlock(&substream->self_group, substream->pcm->nonatomic); 123 } 124 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock); 125 126 /** 127 * snd_pcm_stream_lock_irq - Lock the PCM stream 128 * @substream: PCM substream 129 * 130 * This locks the PCM stream like snd_pcm_stream_lock() and disables the local 131 * IRQ (only when nonatomic is false). In nonatomic case, this is identical 132 * as snd_pcm_stream_lock(). 133 */ 134 void snd_pcm_stream_lock_irq(struct snd_pcm_substream *substream) 135 { 136 snd_pcm_group_lock_irq(&substream->self_group, 137 substream->pcm->nonatomic); 138 } 139 EXPORT_SYMBOL_GPL(snd_pcm_stream_lock_irq); 140 141 static void snd_pcm_stream_lock_nested(struct snd_pcm_substream *substream) 142 { 143 struct snd_pcm_group *group = &substream->self_group; 144 145 if (substream->pcm->nonatomic) 146 mutex_lock_nested(&group->mutex, SINGLE_DEPTH_NESTING); 147 else 148 spin_lock_nested(&group->lock, SINGLE_DEPTH_NESTING); 149 } 150 151 /** 152 * snd_pcm_stream_unlock_irq - Unlock the PCM stream 153 * @substream: PCM substream 154 * 155 * This is a counter-part of snd_pcm_stream_lock_irq(). 156 */ 157 void snd_pcm_stream_unlock_irq(struct snd_pcm_substream *substream) 158 { 159 snd_pcm_group_unlock_irq(&substream->self_group, 160 substream->pcm->nonatomic); 161 } 162 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irq); 163 164 unsigned long _snd_pcm_stream_lock_irqsave(struct snd_pcm_substream *substream) 165 { 166 unsigned long flags = 0; 167 if (substream->pcm->nonatomic) 168 mutex_lock(&substream->self_group.mutex); 169 else 170 spin_lock_irqsave(&substream->self_group.lock, flags); 171 return flags; 172 } 173 EXPORT_SYMBOL_GPL(_snd_pcm_stream_lock_irqsave); 174 175 /** 176 * snd_pcm_stream_unlock_irqrestore - Unlock the PCM stream 177 * @substream: PCM substream 178 * @flags: irq flags 179 * 180 * This is a counter-part of snd_pcm_stream_lock_irqsave(). 181 */ 182 void snd_pcm_stream_unlock_irqrestore(struct snd_pcm_substream *substream, 183 unsigned long flags) 184 { 185 if (substream->pcm->nonatomic) 186 mutex_unlock(&substream->self_group.mutex); 187 else 188 spin_unlock_irqrestore(&substream->self_group.lock, flags); 189 } 190 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irqrestore); 191 192 /* Run PCM ioctl ops */ 193 static int snd_pcm_ops_ioctl(struct snd_pcm_substream *substream, 194 unsigned cmd, void *arg) 195 { 196 if (substream->ops->ioctl) 197 return substream->ops->ioctl(substream, cmd, arg); 198 else 199 return snd_pcm_lib_ioctl(substream, cmd, arg); 200 } 201 202 int snd_pcm_info(struct snd_pcm_substream *substream, struct snd_pcm_info *info) 203 { 204 struct snd_pcm *pcm = substream->pcm; 205 struct snd_pcm_str *pstr = substream->pstr; 206 207 memset(info, 0, sizeof(*info)); 208 info->card = pcm->card->number; 209 info->device = pcm->device; 210 info->stream = substream->stream; 211 info->subdevice = substream->number; 212 strscpy(info->id, pcm->id, sizeof(info->id)); 213 strscpy(info->name, pcm->name, sizeof(info->name)); 214 info->dev_class = pcm->dev_class; 215 info->dev_subclass = pcm->dev_subclass; 216 info->subdevices_count = pstr->substream_count; 217 info->subdevices_avail = pstr->substream_count - pstr->substream_opened; 218 strscpy(info->subname, substream->name, sizeof(info->subname)); 219 220 return 0; 221 } 222 223 int snd_pcm_info_user(struct snd_pcm_substream *substream, 224 struct snd_pcm_info __user * _info) 225 { 226 struct snd_pcm_info *info; 227 int err; 228 229 info = kmalloc(sizeof(*info), GFP_KERNEL); 230 if (! info) 231 return -ENOMEM; 232 err = snd_pcm_info(substream, info); 233 if (err >= 0) { 234 if (copy_to_user(_info, info, sizeof(*info))) 235 err = -EFAULT; 236 } 237 kfree(info); 238 return err; 239 } 240 241 /* macro for simplified cast */ 242 #define PARAM_MASK_BIT(b) (1U << (__force int)(b)) 243 244 static bool hw_support_mmap(struct snd_pcm_substream *substream) 245 { 246 if (!(substream->runtime->hw.info & SNDRV_PCM_INFO_MMAP)) 247 return false; 248 249 if (substream->ops->mmap || substream->ops->page) 250 return true; 251 252 switch (substream->dma_buffer.dev.type) { 253 case SNDRV_DMA_TYPE_UNKNOWN: 254 /* we can't know the device, so just assume that the driver does 255 * everything right 256 */ 257 return true; 258 case SNDRV_DMA_TYPE_CONTINUOUS: 259 case SNDRV_DMA_TYPE_VMALLOC: 260 return true; 261 default: 262 return dma_can_mmap(substream->dma_buffer.dev.dev); 263 } 264 } 265 266 static int constrain_mask_params(struct snd_pcm_substream *substream, 267 struct snd_pcm_hw_params *params) 268 { 269 struct snd_pcm_hw_constraints *constrs = 270 &substream->runtime->hw_constraints; 271 struct snd_mask *m; 272 unsigned int k; 273 struct snd_mask old_mask; 274 int changed; 275 276 for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) { 277 m = hw_param_mask(params, k); 278 if (snd_mask_empty(m)) 279 return -EINVAL; 280 281 /* This parameter is not requested to change by a caller. */ 282 if (!(params->rmask & PARAM_MASK_BIT(k))) 283 continue; 284 285 if (trace_hw_mask_param_enabled()) 286 old_mask = *m; 287 288 changed = snd_mask_refine(m, constrs_mask(constrs, k)); 289 if (changed < 0) 290 return changed; 291 if (changed == 0) 292 continue; 293 294 /* Set corresponding flag so that the caller gets it. */ 295 trace_hw_mask_param(substream, k, 0, &old_mask, m); 296 params->cmask |= PARAM_MASK_BIT(k); 297 } 298 299 return 0; 300 } 301 302 static int constrain_interval_params(struct snd_pcm_substream *substream, 303 struct snd_pcm_hw_params *params) 304 { 305 struct snd_pcm_hw_constraints *constrs = 306 &substream->runtime->hw_constraints; 307 struct snd_interval *i; 308 unsigned int k; 309 struct snd_interval old_interval; 310 int changed; 311 312 for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) { 313 i = hw_param_interval(params, k); 314 if (snd_interval_empty(i)) 315 return -EINVAL; 316 317 /* This parameter is not requested to change by a caller. */ 318 if (!(params->rmask & PARAM_MASK_BIT(k))) 319 continue; 320 321 if (trace_hw_interval_param_enabled()) 322 old_interval = *i; 323 324 changed = snd_interval_refine(i, constrs_interval(constrs, k)); 325 if (changed < 0) 326 return changed; 327 if (changed == 0) 328 continue; 329 330 /* Set corresponding flag so that the caller gets it. */ 331 trace_hw_interval_param(substream, k, 0, &old_interval, i); 332 params->cmask |= PARAM_MASK_BIT(k); 333 } 334 335 return 0; 336 } 337 338 static int constrain_params_by_rules(struct snd_pcm_substream *substream, 339 struct snd_pcm_hw_params *params) 340 { 341 struct snd_pcm_hw_constraints *constrs = 342 &substream->runtime->hw_constraints; 343 unsigned int k; 344 unsigned int *rstamps; 345 unsigned int vstamps[SNDRV_PCM_HW_PARAM_LAST_INTERVAL + 1]; 346 unsigned int stamp; 347 struct snd_pcm_hw_rule *r; 348 unsigned int d; 349 struct snd_mask old_mask; 350 struct snd_interval old_interval; 351 bool again; 352 int changed, err = 0; 353 354 /* 355 * Each application of rule has own sequence number. 356 * 357 * Each member of 'rstamps' array represents the sequence number of 358 * recent application of corresponding rule. 359 */ 360 rstamps = kcalloc(constrs->rules_num, sizeof(unsigned int), GFP_KERNEL); 361 if (!rstamps) 362 return -ENOMEM; 363 364 /* 365 * Each member of 'vstamps' array represents the sequence number of 366 * recent application of rule in which corresponding parameters were 367 * changed. 368 * 369 * In initial state, elements corresponding to parameters requested by 370 * a caller is 1. For unrequested parameters, corresponding members 371 * have 0 so that the parameters are never changed anymore. 372 */ 373 for (k = 0; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) 374 vstamps[k] = (params->rmask & PARAM_MASK_BIT(k)) ? 1 : 0; 375 376 /* Due to the above design, actual sequence number starts at 2. */ 377 stamp = 2; 378 retry: 379 /* Apply all rules in order. */ 380 again = false; 381 for (k = 0; k < constrs->rules_num; k++) { 382 r = &constrs->rules[k]; 383 384 /* 385 * Check condition bits of this rule. When the rule has 386 * some condition bits, parameter without the bits is 387 * never processed. SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP 388 * is an example of the condition bits. 389 */ 390 if (r->cond && !(r->cond & params->flags)) 391 continue; 392 393 /* 394 * The 'deps' array includes maximum four dependencies 395 * to SNDRV_PCM_HW_PARAM_XXXs for this rule. The fifth 396 * member of this array is a sentinel and should be 397 * negative value. 398 * 399 * This rule should be processed in this time when dependent 400 * parameters were changed at former applications of the other 401 * rules. 402 */ 403 for (d = 0; r->deps[d] >= 0; d++) { 404 if (vstamps[r->deps[d]] > rstamps[k]) 405 break; 406 } 407 if (r->deps[d] < 0) 408 continue; 409 410 if (trace_hw_mask_param_enabled()) { 411 if (hw_is_mask(r->var)) 412 old_mask = *hw_param_mask(params, r->var); 413 } 414 if (trace_hw_interval_param_enabled()) { 415 if (hw_is_interval(r->var)) 416 old_interval = *hw_param_interval(params, r->var); 417 } 418 419 changed = r->func(params, r); 420 if (changed < 0) { 421 err = changed; 422 goto out; 423 } 424 425 /* 426 * When the parameter is changed, notify it to the caller 427 * by corresponding returned bit, then preparing for next 428 * iteration. 429 */ 430 if (changed && r->var >= 0) { 431 if (hw_is_mask(r->var)) { 432 trace_hw_mask_param(substream, r->var, 433 k + 1, &old_mask, 434 hw_param_mask(params, r->var)); 435 } 436 if (hw_is_interval(r->var)) { 437 trace_hw_interval_param(substream, r->var, 438 k + 1, &old_interval, 439 hw_param_interval(params, r->var)); 440 } 441 442 params->cmask |= PARAM_MASK_BIT(r->var); 443 vstamps[r->var] = stamp; 444 again = true; 445 } 446 447 rstamps[k] = stamp++; 448 } 449 450 /* Iterate to evaluate all rules till no parameters are changed. */ 451 if (again) 452 goto retry; 453 454 out: 455 kfree(rstamps); 456 return err; 457 } 458 459 static int fixup_unreferenced_params(struct snd_pcm_substream *substream, 460 struct snd_pcm_hw_params *params) 461 { 462 const struct snd_interval *i; 463 const struct snd_mask *m; 464 int err; 465 466 if (!params->msbits) { 467 i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_SAMPLE_BITS); 468 if (snd_interval_single(i)) 469 params->msbits = snd_interval_value(i); 470 } 471 472 if (!params->rate_den) { 473 i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_RATE); 474 if (snd_interval_single(i)) { 475 params->rate_num = snd_interval_value(i); 476 params->rate_den = 1; 477 } 478 } 479 480 if (!params->fifo_size) { 481 m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT); 482 i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_CHANNELS); 483 if (snd_mask_single(m) && snd_interval_single(i)) { 484 err = snd_pcm_ops_ioctl(substream, 485 SNDRV_PCM_IOCTL1_FIFO_SIZE, 486 params); 487 if (err < 0) 488 return err; 489 } 490 } 491 492 if (!params->info) { 493 params->info = substream->runtime->hw.info; 494 params->info &= ~(SNDRV_PCM_INFO_FIFO_IN_FRAMES | 495 SNDRV_PCM_INFO_DRAIN_TRIGGER); 496 if (!hw_support_mmap(substream)) 497 params->info &= ~(SNDRV_PCM_INFO_MMAP | 498 SNDRV_PCM_INFO_MMAP_VALID); 499 } 500 501 return 0; 502 } 503 504 int snd_pcm_hw_refine(struct snd_pcm_substream *substream, 505 struct snd_pcm_hw_params *params) 506 { 507 int err; 508 509 params->info = 0; 510 params->fifo_size = 0; 511 if (params->rmask & PARAM_MASK_BIT(SNDRV_PCM_HW_PARAM_SAMPLE_BITS)) 512 params->msbits = 0; 513 if (params->rmask & PARAM_MASK_BIT(SNDRV_PCM_HW_PARAM_RATE)) { 514 params->rate_num = 0; 515 params->rate_den = 0; 516 } 517 518 err = constrain_mask_params(substream, params); 519 if (err < 0) 520 return err; 521 522 err = constrain_interval_params(substream, params); 523 if (err < 0) 524 return err; 525 526 err = constrain_params_by_rules(substream, params); 527 if (err < 0) 528 return err; 529 530 params->rmask = 0; 531 532 return 0; 533 } 534 EXPORT_SYMBOL(snd_pcm_hw_refine); 535 536 static int snd_pcm_hw_refine_user(struct snd_pcm_substream *substream, 537 struct snd_pcm_hw_params __user * _params) 538 { 539 struct snd_pcm_hw_params *params; 540 int err; 541 542 params = memdup_user(_params, sizeof(*params)); 543 if (IS_ERR(params)) 544 return PTR_ERR(params); 545 546 err = snd_pcm_hw_refine(substream, params); 547 if (err < 0) 548 goto end; 549 550 err = fixup_unreferenced_params(substream, params); 551 if (err < 0) 552 goto end; 553 554 if (copy_to_user(_params, params, sizeof(*params))) 555 err = -EFAULT; 556 end: 557 kfree(params); 558 return err; 559 } 560 561 static int period_to_usecs(struct snd_pcm_runtime *runtime) 562 { 563 int usecs; 564 565 if (! runtime->rate) 566 return -1; /* invalid */ 567 568 /* take 75% of period time as the deadline */ 569 usecs = (750000 / runtime->rate) * runtime->period_size; 570 usecs += ((750000 % runtime->rate) * runtime->period_size) / 571 runtime->rate; 572 573 return usecs; 574 } 575 576 static void snd_pcm_set_state(struct snd_pcm_substream *substream, 577 snd_pcm_state_t state) 578 { 579 snd_pcm_stream_lock_irq(substream); 580 if (substream->runtime->status->state != SNDRV_PCM_STATE_DISCONNECTED) 581 substream->runtime->status->state = state; 582 snd_pcm_stream_unlock_irq(substream); 583 } 584 585 static inline void snd_pcm_timer_notify(struct snd_pcm_substream *substream, 586 int event) 587 { 588 #ifdef CONFIG_SND_PCM_TIMER 589 if (substream->timer) 590 snd_timer_notify(substream->timer, event, 591 &substream->runtime->trigger_tstamp); 592 #endif 593 } 594 595 void snd_pcm_sync_stop(struct snd_pcm_substream *substream, bool sync_irq) 596 { 597 if (substream->runtime && substream->runtime->stop_operating) { 598 substream->runtime->stop_operating = false; 599 if (substream->ops && substream->ops->sync_stop) 600 substream->ops->sync_stop(substream); 601 else if (sync_irq && substream->pcm->card->sync_irq > 0) 602 synchronize_irq(substream->pcm->card->sync_irq); 603 } 604 } 605 606 /** 607 * snd_pcm_hw_params_choose - choose a configuration defined by @params 608 * @pcm: PCM instance 609 * @params: the hw_params instance 610 * 611 * Choose one configuration from configuration space defined by @params. 612 * The configuration chosen is that obtained fixing in this order: 613 * first access, first format, first subformat, min channels, 614 * min rate, min period time, max buffer size, min tick time 615 * 616 * Return: Zero if successful, or a negative error code on failure. 617 */ 618 static int snd_pcm_hw_params_choose(struct snd_pcm_substream *pcm, 619 struct snd_pcm_hw_params *params) 620 { 621 static const int vars[] = { 622 SNDRV_PCM_HW_PARAM_ACCESS, 623 SNDRV_PCM_HW_PARAM_FORMAT, 624 SNDRV_PCM_HW_PARAM_SUBFORMAT, 625 SNDRV_PCM_HW_PARAM_CHANNELS, 626 SNDRV_PCM_HW_PARAM_RATE, 627 SNDRV_PCM_HW_PARAM_PERIOD_TIME, 628 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 629 SNDRV_PCM_HW_PARAM_TICK_TIME, 630 -1 631 }; 632 const int *v; 633 struct snd_mask old_mask; 634 struct snd_interval old_interval; 635 int changed; 636 637 for (v = vars; *v != -1; v++) { 638 /* Keep old parameter to trace. */ 639 if (trace_hw_mask_param_enabled()) { 640 if (hw_is_mask(*v)) 641 old_mask = *hw_param_mask(params, *v); 642 } 643 if (trace_hw_interval_param_enabled()) { 644 if (hw_is_interval(*v)) 645 old_interval = *hw_param_interval(params, *v); 646 } 647 if (*v != SNDRV_PCM_HW_PARAM_BUFFER_SIZE) 648 changed = snd_pcm_hw_param_first(pcm, params, *v, NULL); 649 else 650 changed = snd_pcm_hw_param_last(pcm, params, *v, NULL); 651 if (changed < 0) 652 return changed; 653 if (changed == 0) 654 continue; 655 656 /* Trace the changed parameter. */ 657 if (hw_is_mask(*v)) { 658 trace_hw_mask_param(pcm, *v, 0, &old_mask, 659 hw_param_mask(params, *v)); 660 } 661 if (hw_is_interval(*v)) { 662 trace_hw_interval_param(pcm, *v, 0, &old_interval, 663 hw_param_interval(params, *v)); 664 } 665 } 666 667 return 0; 668 } 669 670 static int snd_pcm_hw_params(struct snd_pcm_substream *substream, 671 struct snd_pcm_hw_params *params) 672 { 673 struct snd_pcm_runtime *runtime; 674 int err, usecs; 675 unsigned int bits; 676 snd_pcm_uframes_t frames; 677 678 if (PCM_RUNTIME_CHECK(substream)) 679 return -ENXIO; 680 runtime = substream->runtime; 681 snd_pcm_stream_lock_irq(substream); 682 switch (runtime->status->state) { 683 case SNDRV_PCM_STATE_OPEN: 684 case SNDRV_PCM_STATE_SETUP: 685 case SNDRV_PCM_STATE_PREPARED: 686 break; 687 default: 688 snd_pcm_stream_unlock_irq(substream); 689 return -EBADFD; 690 } 691 snd_pcm_stream_unlock_irq(substream); 692 #if IS_ENABLED(CONFIG_SND_PCM_OSS) 693 if (!substream->oss.oss) 694 #endif 695 if (atomic_read(&substream->mmap_count)) 696 return -EBADFD; 697 698 snd_pcm_sync_stop(substream, true); 699 700 params->rmask = ~0U; 701 err = snd_pcm_hw_refine(substream, params); 702 if (err < 0) 703 goto _error; 704 705 err = snd_pcm_hw_params_choose(substream, params); 706 if (err < 0) 707 goto _error; 708 709 err = fixup_unreferenced_params(substream, params); 710 if (err < 0) 711 goto _error; 712 713 if (substream->managed_buffer_alloc) { 714 err = snd_pcm_lib_malloc_pages(substream, 715 params_buffer_bytes(params)); 716 if (err < 0) 717 goto _error; 718 runtime->buffer_changed = err > 0; 719 } 720 721 if (substream->ops->hw_params != NULL) { 722 err = substream->ops->hw_params(substream, params); 723 if (err < 0) 724 goto _error; 725 } 726 727 runtime->access = params_access(params); 728 runtime->format = params_format(params); 729 runtime->subformat = params_subformat(params); 730 runtime->channels = params_channels(params); 731 runtime->rate = params_rate(params); 732 runtime->period_size = params_period_size(params); 733 runtime->periods = params_periods(params); 734 runtime->buffer_size = params_buffer_size(params); 735 runtime->info = params->info; 736 runtime->rate_num = params->rate_num; 737 runtime->rate_den = params->rate_den; 738 runtime->no_period_wakeup = 739 (params->info & SNDRV_PCM_INFO_NO_PERIOD_WAKEUP) && 740 (params->flags & SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP); 741 742 bits = snd_pcm_format_physical_width(runtime->format); 743 runtime->sample_bits = bits; 744 bits *= runtime->channels; 745 runtime->frame_bits = bits; 746 frames = 1; 747 while (bits % 8 != 0) { 748 bits *= 2; 749 frames *= 2; 750 } 751 runtime->byte_align = bits / 8; 752 runtime->min_align = frames; 753 754 /* Default sw params */ 755 runtime->tstamp_mode = SNDRV_PCM_TSTAMP_NONE; 756 runtime->period_step = 1; 757 runtime->control->avail_min = runtime->period_size; 758 runtime->start_threshold = 1; 759 runtime->stop_threshold = runtime->buffer_size; 760 runtime->silence_threshold = 0; 761 runtime->silence_size = 0; 762 runtime->boundary = runtime->buffer_size; 763 while (runtime->boundary * 2 <= LONG_MAX - runtime->buffer_size) 764 runtime->boundary *= 2; 765 766 /* clear the buffer for avoiding possible kernel info leaks */ 767 if (runtime->dma_area && !substream->ops->copy_user) { 768 size_t size = runtime->dma_bytes; 769 770 if (runtime->info & SNDRV_PCM_INFO_MMAP) 771 size = PAGE_ALIGN(size); 772 memset(runtime->dma_area, 0, size); 773 } 774 775 snd_pcm_timer_resolution_change(substream); 776 snd_pcm_set_state(substream, SNDRV_PCM_STATE_SETUP); 777 778 if (cpu_latency_qos_request_active(&substream->latency_pm_qos_req)) 779 cpu_latency_qos_remove_request(&substream->latency_pm_qos_req); 780 usecs = period_to_usecs(runtime); 781 if (usecs >= 0) 782 cpu_latency_qos_add_request(&substream->latency_pm_qos_req, 783 usecs); 784 return 0; 785 _error: 786 /* hardware might be unusable from this time, 787 so we force application to retry to set 788 the correct hardware parameter settings */ 789 snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN); 790 if (substream->ops->hw_free != NULL) 791 substream->ops->hw_free(substream); 792 if (substream->managed_buffer_alloc) 793 snd_pcm_lib_free_pages(substream); 794 return err; 795 } 796 797 static int snd_pcm_hw_params_user(struct snd_pcm_substream *substream, 798 struct snd_pcm_hw_params __user * _params) 799 { 800 struct snd_pcm_hw_params *params; 801 int err; 802 803 params = memdup_user(_params, sizeof(*params)); 804 if (IS_ERR(params)) 805 return PTR_ERR(params); 806 807 err = snd_pcm_hw_params(substream, params); 808 if (err < 0) 809 goto end; 810 811 if (copy_to_user(_params, params, sizeof(*params))) 812 err = -EFAULT; 813 end: 814 kfree(params); 815 return err; 816 } 817 818 static int do_hw_free(struct snd_pcm_substream *substream) 819 { 820 int result = 0; 821 822 snd_pcm_sync_stop(substream, true); 823 if (substream->ops->hw_free) 824 result = substream->ops->hw_free(substream); 825 if (substream->managed_buffer_alloc) 826 snd_pcm_lib_free_pages(substream); 827 return result; 828 } 829 830 static int snd_pcm_hw_free(struct snd_pcm_substream *substream) 831 { 832 struct snd_pcm_runtime *runtime; 833 int result; 834 835 if (PCM_RUNTIME_CHECK(substream)) 836 return -ENXIO; 837 runtime = substream->runtime; 838 snd_pcm_stream_lock_irq(substream); 839 switch (runtime->status->state) { 840 case SNDRV_PCM_STATE_SETUP: 841 case SNDRV_PCM_STATE_PREPARED: 842 break; 843 default: 844 snd_pcm_stream_unlock_irq(substream); 845 return -EBADFD; 846 } 847 snd_pcm_stream_unlock_irq(substream); 848 if (atomic_read(&substream->mmap_count)) 849 return -EBADFD; 850 result = do_hw_free(substream); 851 snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN); 852 cpu_latency_qos_remove_request(&substream->latency_pm_qos_req); 853 return result; 854 } 855 856 static int snd_pcm_sw_params(struct snd_pcm_substream *substream, 857 struct snd_pcm_sw_params *params) 858 { 859 struct snd_pcm_runtime *runtime; 860 int err; 861 862 if (PCM_RUNTIME_CHECK(substream)) 863 return -ENXIO; 864 runtime = substream->runtime; 865 snd_pcm_stream_lock_irq(substream); 866 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) { 867 snd_pcm_stream_unlock_irq(substream); 868 return -EBADFD; 869 } 870 snd_pcm_stream_unlock_irq(substream); 871 872 if (params->tstamp_mode < 0 || 873 params->tstamp_mode > SNDRV_PCM_TSTAMP_LAST) 874 return -EINVAL; 875 if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12) && 876 params->tstamp_type > SNDRV_PCM_TSTAMP_TYPE_LAST) 877 return -EINVAL; 878 if (params->avail_min == 0) 879 return -EINVAL; 880 if (params->silence_size >= runtime->boundary) { 881 if (params->silence_threshold != 0) 882 return -EINVAL; 883 } else { 884 if (params->silence_size > params->silence_threshold) 885 return -EINVAL; 886 if (params->silence_threshold > runtime->buffer_size) 887 return -EINVAL; 888 } 889 err = 0; 890 snd_pcm_stream_lock_irq(substream); 891 runtime->tstamp_mode = params->tstamp_mode; 892 if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12)) 893 runtime->tstamp_type = params->tstamp_type; 894 runtime->period_step = params->period_step; 895 runtime->control->avail_min = params->avail_min; 896 runtime->start_threshold = params->start_threshold; 897 runtime->stop_threshold = params->stop_threshold; 898 runtime->silence_threshold = params->silence_threshold; 899 runtime->silence_size = params->silence_size; 900 params->boundary = runtime->boundary; 901 if (snd_pcm_running(substream)) { 902 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 903 runtime->silence_size > 0) 904 snd_pcm_playback_silence(substream, ULONG_MAX); 905 err = snd_pcm_update_state(substream, runtime); 906 } 907 snd_pcm_stream_unlock_irq(substream); 908 return err; 909 } 910 911 static int snd_pcm_sw_params_user(struct snd_pcm_substream *substream, 912 struct snd_pcm_sw_params __user * _params) 913 { 914 struct snd_pcm_sw_params params; 915 int err; 916 if (copy_from_user(¶ms, _params, sizeof(params))) 917 return -EFAULT; 918 err = snd_pcm_sw_params(substream, ¶ms); 919 if (copy_to_user(_params, ¶ms, sizeof(params))) 920 return -EFAULT; 921 return err; 922 } 923 924 static inline snd_pcm_uframes_t 925 snd_pcm_calc_delay(struct snd_pcm_substream *substream) 926 { 927 snd_pcm_uframes_t delay; 928 929 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 930 delay = snd_pcm_playback_hw_avail(substream->runtime); 931 else 932 delay = snd_pcm_capture_avail(substream->runtime); 933 return delay + substream->runtime->delay; 934 } 935 936 int snd_pcm_status64(struct snd_pcm_substream *substream, 937 struct snd_pcm_status64 *status) 938 { 939 struct snd_pcm_runtime *runtime = substream->runtime; 940 941 snd_pcm_stream_lock_irq(substream); 942 943 snd_pcm_unpack_audio_tstamp_config(status->audio_tstamp_data, 944 &runtime->audio_tstamp_config); 945 946 /* backwards compatible behavior */ 947 if (runtime->audio_tstamp_config.type_requested == 948 SNDRV_PCM_AUDIO_TSTAMP_TYPE_COMPAT) { 949 if (runtime->hw.info & SNDRV_PCM_INFO_HAS_WALL_CLOCK) 950 runtime->audio_tstamp_config.type_requested = 951 SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK; 952 else 953 runtime->audio_tstamp_config.type_requested = 954 SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT; 955 runtime->audio_tstamp_report.valid = 0; 956 } else 957 runtime->audio_tstamp_report.valid = 1; 958 959 status->state = runtime->status->state; 960 status->suspended_state = runtime->status->suspended_state; 961 if (status->state == SNDRV_PCM_STATE_OPEN) 962 goto _end; 963 status->trigger_tstamp_sec = runtime->trigger_tstamp.tv_sec; 964 status->trigger_tstamp_nsec = runtime->trigger_tstamp.tv_nsec; 965 if (snd_pcm_running(substream)) { 966 snd_pcm_update_hw_ptr(substream); 967 if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { 968 status->tstamp_sec = runtime->status->tstamp.tv_sec; 969 status->tstamp_nsec = 970 runtime->status->tstamp.tv_nsec; 971 status->driver_tstamp_sec = 972 runtime->driver_tstamp.tv_sec; 973 status->driver_tstamp_nsec = 974 runtime->driver_tstamp.tv_nsec; 975 status->audio_tstamp_sec = 976 runtime->status->audio_tstamp.tv_sec; 977 status->audio_tstamp_nsec = 978 runtime->status->audio_tstamp.tv_nsec; 979 if (runtime->audio_tstamp_report.valid == 1) 980 /* backwards compatibility, no report provided in COMPAT mode */ 981 snd_pcm_pack_audio_tstamp_report(&status->audio_tstamp_data, 982 &status->audio_tstamp_accuracy, 983 &runtime->audio_tstamp_report); 984 985 goto _tstamp_end; 986 } 987 } else { 988 /* get tstamp only in fallback mode and only if enabled */ 989 if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { 990 struct timespec64 tstamp; 991 992 snd_pcm_gettime(runtime, &tstamp); 993 status->tstamp_sec = tstamp.tv_sec; 994 status->tstamp_nsec = tstamp.tv_nsec; 995 } 996 } 997 _tstamp_end: 998 status->appl_ptr = runtime->control->appl_ptr; 999 status->hw_ptr = runtime->status->hw_ptr; 1000 status->avail = snd_pcm_avail(substream); 1001 status->delay = snd_pcm_running(substream) ? 1002 snd_pcm_calc_delay(substream) : 0; 1003 status->avail_max = runtime->avail_max; 1004 status->overrange = runtime->overrange; 1005 runtime->avail_max = 0; 1006 runtime->overrange = 0; 1007 _end: 1008 snd_pcm_stream_unlock_irq(substream); 1009 return 0; 1010 } 1011 1012 static int snd_pcm_status_user64(struct snd_pcm_substream *substream, 1013 struct snd_pcm_status64 __user * _status, 1014 bool ext) 1015 { 1016 struct snd_pcm_status64 status; 1017 int res; 1018 1019 memset(&status, 0, sizeof(status)); 1020 /* 1021 * with extension, parameters are read/write, 1022 * get audio_tstamp_data from user, 1023 * ignore rest of status structure 1024 */ 1025 if (ext && get_user(status.audio_tstamp_data, 1026 (u32 __user *)(&_status->audio_tstamp_data))) 1027 return -EFAULT; 1028 res = snd_pcm_status64(substream, &status); 1029 if (res < 0) 1030 return res; 1031 if (copy_to_user(_status, &status, sizeof(status))) 1032 return -EFAULT; 1033 return 0; 1034 } 1035 1036 static int snd_pcm_status_user32(struct snd_pcm_substream *substream, 1037 struct snd_pcm_status32 __user * _status, 1038 bool ext) 1039 { 1040 struct snd_pcm_status64 status64; 1041 struct snd_pcm_status32 status32; 1042 int res; 1043 1044 memset(&status64, 0, sizeof(status64)); 1045 memset(&status32, 0, sizeof(status32)); 1046 /* 1047 * with extension, parameters are read/write, 1048 * get audio_tstamp_data from user, 1049 * ignore rest of status structure 1050 */ 1051 if (ext && get_user(status64.audio_tstamp_data, 1052 (u32 __user *)(&_status->audio_tstamp_data))) 1053 return -EFAULT; 1054 res = snd_pcm_status64(substream, &status64); 1055 if (res < 0) 1056 return res; 1057 1058 status32 = (struct snd_pcm_status32) { 1059 .state = status64.state, 1060 .trigger_tstamp_sec = status64.trigger_tstamp_sec, 1061 .trigger_tstamp_nsec = status64.trigger_tstamp_nsec, 1062 .tstamp_sec = status64.tstamp_sec, 1063 .tstamp_nsec = status64.tstamp_nsec, 1064 .appl_ptr = status64.appl_ptr, 1065 .hw_ptr = status64.hw_ptr, 1066 .delay = status64.delay, 1067 .avail = status64.avail, 1068 .avail_max = status64.avail_max, 1069 .overrange = status64.overrange, 1070 .suspended_state = status64.suspended_state, 1071 .audio_tstamp_data = status64.audio_tstamp_data, 1072 .audio_tstamp_sec = status64.audio_tstamp_sec, 1073 .audio_tstamp_nsec = status64.audio_tstamp_nsec, 1074 .driver_tstamp_sec = status64.audio_tstamp_sec, 1075 .driver_tstamp_nsec = status64.audio_tstamp_nsec, 1076 .audio_tstamp_accuracy = status64.audio_tstamp_accuracy, 1077 }; 1078 1079 if (copy_to_user(_status, &status32, sizeof(status32))) 1080 return -EFAULT; 1081 1082 return 0; 1083 } 1084 1085 static int snd_pcm_channel_info(struct snd_pcm_substream *substream, 1086 struct snd_pcm_channel_info * info) 1087 { 1088 struct snd_pcm_runtime *runtime; 1089 unsigned int channel; 1090 1091 channel = info->channel; 1092 runtime = substream->runtime; 1093 snd_pcm_stream_lock_irq(substream); 1094 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) { 1095 snd_pcm_stream_unlock_irq(substream); 1096 return -EBADFD; 1097 } 1098 snd_pcm_stream_unlock_irq(substream); 1099 if (channel >= runtime->channels) 1100 return -EINVAL; 1101 memset(info, 0, sizeof(*info)); 1102 info->channel = channel; 1103 return snd_pcm_ops_ioctl(substream, SNDRV_PCM_IOCTL1_CHANNEL_INFO, info); 1104 } 1105 1106 static int snd_pcm_channel_info_user(struct snd_pcm_substream *substream, 1107 struct snd_pcm_channel_info __user * _info) 1108 { 1109 struct snd_pcm_channel_info info; 1110 int res; 1111 1112 if (copy_from_user(&info, _info, sizeof(info))) 1113 return -EFAULT; 1114 res = snd_pcm_channel_info(substream, &info); 1115 if (res < 0) 1116 return res; 1117 if (copy_to_user(_info, &info, sizeof(info))) 1118 return -EFAULT; 1119 return 0; 1120 } 1121 1122 static void snd_pcm_trigger_tstamp(struct snd_pcm_substream *substream) 1123 { 1124 struct snd_pcm_runtime *runtime = substream->runtime; 1125 if (runtime->trigger_master == NULL) 1126 return; 1127 if (runtime->trigger_master == substream) { 1128 if (!runtime->trigger_tstamp_latched) 1129 snd_pcm_gettime(runtime, &runtime->trigger_tstamp); 1130 } else { 1131 snd_pcm_trigger_tstamp(runtime->trigger_master); 1132 runtime->trigger_tstamp = runtime->trigger_master->runtime->trigger_tstamp; 1133 } 1134 runtime->trigger_master = NULL; 1135 } 1136 1137 #define ACTION_ARG_IGNORE (__force snd_pcm_state_t)0 1138 1139 struct action_ops { 1140 int (*pre_action)(struct snd_pcm_substream *substream, 1141 snd_pcm_state_t state); 1142 int (*do_action)(struct snd_pcm_substream *substream, 1143 snd_pcm_state_t state); 1144 void (*undo_action)(struct snd_pcm_substream *substream, 1145 snd_pcm_state_t state); 1146 void (*post_action)(struct snd_pcm_substream *substream, 1147 snd_pcm_state_t state); 1148 }; 1149 1150 /* 1151 * this functions is core for handling of linked stream 1152 * Note: the stream state might be changed also on failure 1153 * Note2: call with calling stream lock + link lock 1154 */ 1155 static int snd_pcm_action_group(const struct action_ops *ops, 1156 struct snd_pcm_substream *substream, 1157 snd_pcm_state_t state, 1158 bool do_lock) 1159 { 1160 struct snd_pcm_substream *s = NULL; 1161 struct snd_pcm_substream *s1; 1162 int res = 0, depth = 1; 1163 1164 snd_pcm_group_for_each_entry(s, substream) { 1165 if (do_lock && s != substream) { 1166 if (s->pcm->nonatomic) 1167 mutex_lock_nested(&s->self_group.mutex, depth); 1168 else 1169 spin_lock_nested(&s->self_group.lock, depth); 1170 depth++; 1171 } 1172 res = ops->pre_action(s, state); 1173 if (res < 0) 1174 goto _unlock; 1175 } 1176 snd_pcm_group_for_each_entry(s, substream) { 1177 res = ops->do_action(s, state); 1178 if (res < 0) { 1179 if (ops->undo_action) { 1180 snd_pcm_group_for_each_entry(s1, substream) { 1181 if (s1 == s) /* failed stream */ 1182 break; 1183 ops->undo_action(s1, state); 1184 } 1185 } 1186 s = NULL; /* unlock all */ 1187 goto _unlock; 1188 } 1189 } 1190 snd_pcm_group_for_each_entry(s, substream) { 1191 ops->post_action(s, state); 1192 } 1193 _unlock: 1194 if (do_lock) { 1195 /* unlock streams */ 1196 snd_pcm_group_for_each_entry(s1, substream) { 1197 if (s1 != substream) { 1198 if (s1->pcm->nonatomic) 1199 mutex_unlock(&s1->self_group.mutex); 1200 else 1201 spin_unlock(&s1->self_group.lock); 1202 } 1203 if (s1 == s) /* end */ 1204 break; 1205 } 1206 } 1207 return res; 1208 } 1209 1210 /* 1211 * Note: call with stream lock 1212 */ 1213 static int snd_pcm_action_single(const struct action_ops *ops, 1214 struct snd_pcm_substream *substream, 1215 snd_pcm_state_t state) 1216 { 1217 int res; 1218 1219 res = ops->pre_action(substream, state); 1220 if (res < 0) 1221 return res; 1222 res = ops->do_action(substream, state); 1223 if (res == 0) 1224 ops->post_action(substream, state); 1225 else if (ops->undo_action) 1226 ops->undo_action(substream, state); 1227 return res; 1228 } 1229 1230 static void snd_pcm_group_assign(struct snd_pcm_substream *substream, 1231 struct snd_pcm_group *new_group) 1232 { 1233 substream->group = new_group; 1234 list_move(&substream->link_list, &new_group->substreams); 1235 } 1236 1237 /* 1238 * Unref and unlock the group, but keep the stream lock; 1239 * when the group becomes empty and no longer referred, destroy itself 1240 */ 1241 static void snd_pcm_group_unref(struct snd_pcm_group *group, 1242 struct snd_pcm_substream *substream) 1243 { 1244 bool do_free; 1245 1246 if (!group) 1247 return; 1248 do_free = refcount_dec_and_test(&group->refs); 1249 snd_pcm_group_unlock(group, substream->pcm->nonatomic); 1250 if (do_free) 1251 kfree(group); 1252 } 1253 1254 /* 1255 * Lock the group inside a stream lock and reference it; 1256 * return the locked group object, or NULL if not linked 1257 */ 1258 static struct snd_pcm_group * 1259 snd_pcm_stream_group_ref(struct snd_pcm_substream *substream) 1260 { 1261 bool nonatomic = substream->pcm->nonatomic; 1262 struct snd_pcm_group *group; 1263 bool trylock; 1264 1265 for (;;) { 1266 if (!snd_pcm_stream_linked(substream)) 1267 return NULL; 1268 group = substream->group; 1269 /* block freeing the group object */ 1270 refcount_inc(&group->refs); 1271 1272 trylock = nonatomic ? mutex_trylock(&group->mutex) : 1273 spin_trylock(&group->lock); 1274 if (trylock) 1275 break; /* OK */ 1276 1277 /* re-lock for avoiding ABBA deadlock */ 1278 snd_pcm_stream_unlock(substream); 1279 snd_pcm_group_lock(group, nonatomic); 1280 snd_pcm_stream_lock(substream); 1281 1282 /* check the group again; the above opens a small race window */ 1283 if (substream->group == group) 1284 break; /* OK */ 1285 /* group changed, try again */ 1286 snd_pcm_group_unref(group, substream); 1287 } 1288 return group; 1289 } 1290 1291 /* 1292 * Note: call with stream lock 1293 */ 1294 static int snd_pcm_action(const struct action_ops *ops, 1295 struct snd_pcm_substream *substream, 1296 snd_pcm_state_t state) 1297 { 1298 struct snd_pcm_group *group; 1299 int res; 1300 1301 group = snd_pcm_stream_group_ref(substream); 1302 if (group) 1303 res = snd_pcm_action_group(ops, substream, state, true); 1304 else 1305 res = snd_pcm_action_single(ops, substream, state); 1306 snd_pcm_group_unref(group, substream); 1307 return res; 1308 } 1309 1310 /* 1311 * Note: don't use any locks before 1312 */ 1313 static int snd_pcm_action_lock_irq(const struct action_ops *ops, 1314 struct snd_pcm_substream *substream, 1315 snd_pcm_state_t state) 1316 { 1317 int res; 1318 1319 snd_pcm_stream_lock_irq(substream); 1320 res = snd_pcm_action(ops, substream, state); 1321 snd_pcm_stream_unlock_irq(substream); 1322 return res; 1323 } 1324 1325 /* 1326 */ 1327 static int snd_pcm_action_nonatomic(const struct action_ops *ops, 1328 struct snd_pcm_substream *substream, 1329 snd_pcm_state_t state) 1330 { 1331 int res; 1332 1333 /* Guarantee the group members won't change during non-atomic action */ 1334 down_read(&snd_pcm_link_rwsem); 1335 if (snd_pcm_stream_linked(substream)) 1336 res = snd_pcm_action_group(ops, substream, state, false); 1337 else 1338 res = snd_pcm_action_single(ops, substream, state); 1339 up_read(&snd_pcm_link_rwsem); 1340 return res; 1341 } 1342 1343 /* 1344 * start callbacks 1345 */ 1346 static int snd_pcm_pre_start(struct snd_pcm_substream *substream, 1347 snd_pcm_state_t state) 1348 { 1349 struct snd_pcm_runtime *runtime = substream->runtime; 1350 if (runtime->status->state != SNDRV_PCM_STATE_PREPARED) 1351 return -EBADFD; 1352 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 1353 !snd_pcm_playback_data(substream)) 1354 return -EPIPE; 1355 runtime->trigger_tstamp_latched = false; 1356 runtime->trigger_master = substream; 1357 return 0; 1358 } 1359 1360 static int snd_pcm_do_start(struct snd_pcm_substream *substream, 1361 snd_pcm_state_t state) 1362 { 1363 if (substream->runtime->trigger_master != substream) 1364 return 0; 1365 return substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_START); 1366 } 1367 1368 static void snd_pcm_undo_start(struct snd_pcm_substream *substream, 1369 snd_pcm_state_t state) 1370 { 1371 if (substream->runtime->trigger_master == substream) 1372 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP); 1373 } 1374 1375 static void snd_pcm_post_start(struct snd_pcm_substream *substream, 1376 snd_pcm_state_t state) 1377 { 1378 struct snd_pcm_runtime *runtime = substream->runtime; 1379 snd_pcm_trigger_tstamp(substream); 1380 runtime->hw_ptr_jiffies = jiffies; 1381 runtime->hw_ptr_buffer_jiffies = (runtime->buffer_size * HZ) / 1382 runtime->rate; 1383 runtime->status->state = state; 1384 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 1385 runtime->silence_size > 0) 1386 snd_pcm_playback_silence(substream, ULONG_MAX); 1387 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSTART); 1388 } 1389 1390 static const struct action_ops snd_pcm_action_start = { 1391 .pre_action = snd_pcm_pre_start, 1392 .do_action = snd_pcm_do_start, 1393 .undo_action = snd_pcm_undo_start, 1394 .post_action = snd_pcm_post_start 1395 }; 1396 1397 /** 1398 * snd_pcm_start - start all linked streams 1399 * @substream: the PCM substream instance 1400 * 1401 * Return: Zero if successful, or a negative error code. 1402 * The stream lock must be acquired before calling this function. 1403 */ 1404 int snd_pcm_start(struct snd_pcm_substream *substream) 1405 { 1406 return snd_pcm_action(&snd_pcm_action_start, substream, 1407 SNDRV_PCM_STATE_RUNNING); 1408 } 1409 1410 /* take the stream lock and start the streams */ 1411 static int snd_pcm_start_lock_irq(struct snd_pcm_substream *substream) 1412 { 1413 return snd_pcm_action_lock_irq(&snd_pcm_action_start, substream, 1414 SNDRV_PCM_STATE_RUNNING); 1415 } 1416 1417 /* 1418 * stop callbacks 1419 */ 1420 static int snd_pcm_pre_stop(struct snd_pcm_substream *substream, 1421 snd_pcm_state_t state) 1422 { 1423 struct snd_pcm_runtime *runtime = substream->runtime; 1424 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 1425 return -EBADFD; 1426 runtime->trigger_master = substream; 1427 return 0; 1428 } 1429 1430 static int snd_pcm_do_stop(struct snd_pcm_substream *substream, 1431 snd_pcm_state_t state) 1432 { 1433 if (substream->runtime->trigger_master == substream && 1434 snd_pcm_running(substream)) { 1435 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP); 1436 substream->runtime->stop_operating = true; 1437 } 1438 return 0; /* unconditionally stop all substreams */ 1439 } 1440 1441 static void snd_pcm_post_stop(struct snd_pcm_substream *substream, 1442 snd_pcm_state_t state) 1443 { 1444 struct snd_pcm_runtime *runtime = substream->runtime; 1445 if (runtime->status->state != state) { 1446 snd_pcm_trigger_tstamp(substream); 1447 runtime->status->state = state; 1448 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSTOP); 1449 } 1450 wake_up(&runtime->sleep); 1451 wake_up(&runtime->tsleep); 1452 } 1453 1454 static const struct action_ops snd_pcm_action_stop = { 1455 .pre_action = snd_pcm_pre_stop, 1456 .do_action = snd_pcm_do_stop, 1457 .post_action = snd_pcm_post_stop 1458 }; 1459 1460 /** 1461 * snd_pcm_stop - try to stop all running streams in the substream group 1462 * @substream: the PCM substream instance 1463 * @state: PCM state after stopping the stream 1464 * 1465 * The state of each stream is then changed to the given state unconditionally. 1466 * 1467 * Return: Zero if successful, or a negative error code. 1468 */ 1469 int snd_pcm_stop(struct snd_pcm_substream *substream, snd_pcm_state_t state) 1470 { 1471 return snd_pcm_action(&snd_pcm_action_stop, substream, state); 1472 } 1473 EXPORT_SYMBOL(snd_pcm_stop); 1474 1475 /** 1476 * snd_pcm_drain_done - stop the DMA only when the given stream is playback 1477 * @substream: the PCM substream 1478 * 1479 * After stopping, the state is changed to SETUP. 1480 * Unlike snd_pcm_stop(), this affects only the given stream. 1481 * 1482 * Return: Zero if successful, or a negative error code. 1483 */ 1484 int snd_pcm_drain_done(struct snd_pcm_substream *substream) 1485 { 1486 return snd_pcm_action_single(&snd_pcm_action_stop, substream, 1487 SNDRV_PCM_STATE_SETUP); 1488 } 1489 1490 /** 1491 * snd_pcm_stop_xrun - stop the running streams as XRUN 1492 * @substream: the PCM substream instance 1493 * 1494 * This stops the given running substream (and all linked substreams) as XRUN. 1495 * Unlike snd_pcm_stop(), this function takes the substream lock by itself. 1496 * 1497 * Return: Zero if successful, or a negative error code. 1498 */ 1499 int snd_pcm_stop_xrun(struct snd_pcm_substream *substream) 1500 { 1501 unsigned long flags; 1502 1503 snd_pcm_stream_lock_irqsave(substream, flags); 1504 if (substream->runtime && snd_pcm_running(substream)) 1505 __snd_pcm_xrun(substream); 1506 snd_pcm_stream_unlock_irqrestore(substream, flags); 1507 return 0; 1508 } 1509 EXPORT_SYMBOL_GPL(snd_pcm_stop_xrun); 1510 1511 /* 1512 * pause callbacks: pass boolean (to start pause or resume) as state argument 1513 */ 1514 #define pause_pushed(state) (__force bool)(state) 1515 1516 static int snd_pcm_pre_pause(struct snd_pcm_substream *substream, 1517 snd_pcm_state_t state) 1518 { 1519 struct snd_pcm_runtime *runtime = substream->runtime; 1520 if (!(runtime->info & SNDRV_PCM_INFO_PAUSE)) 1521 return -ENOSYS; 1522 if (pause_pushed(state)) { 1523 if (runtime->status->state != SNDRV_PCM_STATE_RUNNING) 1524 return -EBADFD; 1525 } else if (runtime->status->state != SNDRV_PCM_STATE_PAUSED) 1526 return -EBADFD; 1527 runtime->trigger_master = substream; 1528 return 0; 1529 } 1530 1531 static int snd_pcm_do_pause(struct snd_pcm_substream *substream, 1532 snd_pcm_state_t state) 1533 { 1534 if (substream->runtime->trigger_master != substream) 1535 return 0; 1536 /* some drivers might use hw_ptr to recover from the pause - 1537 update the hw_ptr now */ 1538 if (pause_pushed(state)) 1539 snd_pcm_update_hw_ptr(substream); 1540 /* The jiffies check in snd_pcm_update_hw_ptr*() is done by 1541 * a delta between the current jiffies, this gives a large enough 1542 * delta, effectively to skip the check once. 1543 */ 1544 substream->runtime->hw_ptr_jiffies = jiffies - HZ * 1000; 1545 return substream->ops->trigger(substream, 1546 pause_pushed(state) ? 1547 SNDRV_PCM_TRIGGER_PAUSE_PUSH : 1548 SNDRV_PCM_TRIGGER_PAUSE_RELEASE); 1549 } 1550 1551 static void snd_pcm_undo_pause(struct snd_pcm_substream *substream, 1552 snd_pcm_state_t state) 1553 { 1554 if (substream->runtime->trigger_master == substream) 1555 substream->ops->trigger(substream, 1556 pause_pushed(state) ? 1557 SNDRV_PCM_TRIGGER_PAUSE_RELEASE : 1558 SNDRV_PCM_TRIGGER_PAUSE_PUSH); 1559 } 1560 1561 static void snd_pcm_post_pause(struct snd_pcm_substream *substream, 1562 snd_pcm_state_t state) 1563 { 1564 struct snd_pcm_runtime *runtime = substream->runtime; 1565 snd_pcm_trigger_tstamp(substream); 1566 if (pause_pushed(state)) { 1567 runtime->status->state = SNDRV_PCM_STATE_PAUSED; 1568 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MPAUSE); 1569 wake_up(&runtime->sleep); 1570 wake_up(&runtime->tsleep); 1571 } else { 1572 runtime->status->state = SNDRV_PCM_STATE_RUNNING; 1573 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MCONTINUE); 1574 } 1575 } 1576 1577 static const struct action_ops snd_pcm_action_pause = { 1578 .pre_action = snd_pcm_pre_pause, 1579 .do_action = snd_pcm_do_pause, 1580 .undo_action = snd_pcm_undo_pause, 1581 .post_action = snd_pcm_post_pause 1582 }; 1583 1584 /* 1585 * Push/release the pause for all linked streams. 1586 */ 1587 static int snd_pcm_pause(struct snd_pcm_substream *substream, bool push) 1588 { 1589 return snd_pcm_action(&snd_pcm_action_pause, substream, 1590 (__force snd_pcm_state_t)push); 1591 } 1592 1593 static int snd_pcm_pause_lock_irq(struct snd_pcm_substream *substream, 1594 bool push) 1595 { 1596 return snd_pcm_action_lock_irq(&snd_pcm_action_pause, substream, 1597 (__force snd_pcm_state_t)push); 1598 } 1599 1600 #ifdef CONFIG_PM 1601 /* suspend callback: state argument ignored */ 1602 1603 static int snd_pcm_pre_suspend(struct snd_pcm_substream *substream, 1604 snd_pcm_state_t state) 1605 { 1606 struct snd_pcm_runtime *runtime = substream->runtime; 1607 switch (runtime->status->state) { 1608 case SNDRV_PCM_STATE_SUSPENDED: 1609 return -EBUSY; 1610 /* unresumable PCM state; return -EBUSY for skipping suspend */ 1611 case SNDRV_PCM_STATE_OPEN: 1612 case SNDRV_PCM_STATE_SETUP: 1613 case SNDRV_PCM_STATE_DISCONNECTED: 1614 return -EBUSY; 1615 } 1616 runtime->trigger_master = substream; 1617 return 0; 1618 } 1619 1620 static int snd_pcm_do_suspend(struct snd_pcm_substream *substream, 1621 snd_pcm_state_t state) 1622 { 1623 struct snd_pcm_runtime *runtime = substream->runtime; 1624 if (runtime->trigger_master != substream) 1625 return 0; 1626 if (! snd_pcm_running(substream)) 1627 return 0; 1628 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND); 1629 runtime->stop_operating = true; 1630 return 0; /* suspend unconditionally */ 1631 } 1632 1633 static void snd_pcm_post_suspend(struct snd_pcm_substream *substream, 1634 snd_pcm_state_t state) 1635 { 1636 struct snd_pcm_runtime *runtime = substream->runtime; 1637 snd_pcm_trigger_tstamp(substream); 1638 runtime->status->suspended_state = runtime->status->state; 1639 runtime->status->state = SNDRV_PCM_STATE_SUSPENDED; 1640 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSUSPEND); 1641 wake_up(&runtime->sleep); 1642 wake_up(&runtime->tsleep); 1643 } 1644 1645 static const struct action_ops snd_pcm_action_suspend = { 1646 .pre_action = snd_pcm_pre_suspend, 1647 .do_action = snd_pcm_do_suspend, 1648 .post_action = snd_pcm_post_suspend 1649 }; 1650 1651 /* 1652 * snd_pcm_suspend - trigger SUSPEND to all linked streams 1653 * @substream: the PCM substream 1654 * 1655 * After this call, all streams are changed to SUSPENDED state. 1656 * 1657 * Return: Zero if successful, or a negative error code. 1658 */ 1659 static int snd_pcm_suspend(struct snd_pcm_substream *substream) 1660 { 1661 int err; 1662 unsigned long flags; 1663 1664 snd_pcm_stream_lock_irqsave(substream, flags); 1665 err = snd_pcm_action(&snd_pcm_action_suspend, substream, 1666 ACTION_ARG_IGNORE); 1667 snd_pcm_stream_unlock_irqrestore(substream, flags); 1668 return err; 1669 } 1670 1671 /** 1672 * snd_pcm_suspend_all - trigger SUSPEND to all substreams in the given pcm 1673 * @pcm: the PCM instance 1674 * 1675 * After this call, all streams are changed to SUSPENDED state. 1676 * 1677 * Return: Zero if successful (or @pcm is %NULL), or a negative error code. 1678 */ 1679 int snd_pcm_suspend_all(struct snd_pcm *pcm) 1680 { 1681 struct snd_pcm_substream *substream; 1682 int stream, err = 0; 1683 1684 if (! pcm) 1685 return 0; 1686 1687 for_each_pcm_substream(pcm, stream, substream) { 1688 /* FIXME: the open/close code should lock this as well */ 1689 if (!substream->runtime) 1690 continue; 1691 1692 /* 1693 * Skip BE dai link PCM's that are internal and may 1694 * not have their substream ops set. 1695 */ 1696 if (!substream->ops) 1697 continue; 1698 1699 err = snd_pcm_suspend(substream); 1700 if (err < 0 && err != -EBUSY) 1701 return err; 1702 } 1703 1704 for_each_pcm_substream(pcm, stream, substream) 1705 snd_pcm_sync_stop(substream, false); 1706 1707 return 0; 1708 } 1709 EXPORT_SYMBOL(snd_pcm_suspend_all); 1710 1711 /* resume callbacks: state argument ignored */ 1712 1713 static int snd_pcm_pre_resume(struct snd_pcm_substream *substream, 1714 snd_pcm_state_t state) 1715 { 1716 struct snd_pcm_runtime *runtime = substream->runtime; 1717 if (!(runtime->info & SNDRV_PCM_INFO_RESUME)) 1718 return -ENOSYS; 1719 runtime->trigger_master = substream; 1720 return 0; 1721 } 1722 1723 static int snd_pcm_do_resume(struct snd_pcm_substream *substream, 1724 snd_pcm_state_t state) 1725 { 1726 struct snd_pcm_runtime *runtime = substream->runtime; 1727 if (runtime->trigger_master != substream) 1728 return 0; 1729 /* DMA not running previously? */ 1730 if (runtime->status->suspended_state != SNDRV_PCM_STATE_RUNNING && 1731 (runtime->status->suspended_state != SNDRV_PCM_STATE_DRAINING || 1732 substream->stream != SNDRV_PCM_STREAM_PLAYBACK)) 1733 return 0; 1734 return substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_RESUME); 1735 } 1736 1737 static void snd_pcm_undo_resume(struct snd_pcm_substream *substream, 1738 snd_pcm_state_t state) 1739 { 1740 if (substream->runtime->trigger_master == substream && 1741 snd_pcm_running(substream)) 1742 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND); 1743 } 1744 1745 static void snd_pcm_post_resume(struct snd_pcm_substream *substream, 1746 snd_pcm_state_t state) 1747 { 1748 struct snd_pcm_runtime *runtime = substream->runtime; 1749 snd_pcm_trigger_tstamp(substream); 1750 runtime->status->state = runtime->status->suspended_state; 1751 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MRESUME); 1752 } 1753 1754 static const struct action_ops snd_pcm_action_resume = { 1755 .pre_action = snd_pcm_pre_resume, 1756 .do_action = snd_pcm_do_resume, 1757 .undo_action = snd_pcm_undo_resume, 1758 .post_action = snd_pcm_post_resume 1759 }; 1760 1761 static int snd_pcm_resume(struct snd_pcm_substream *substream) 1762 { 1763 return snd_pcm_action_lock_irq(&snd_pcm_action_resume, substream, 1764 ACTION_ARG_IGNORE); 1765 } 1766 1767 #else 1768 1769 static int snd_pcm_resume(struct snd_pcm_substream *substream) 1770 { 1771 return -ENOSYS; 1772 } 1773 1774 #endif /* CONFIG_PM */ 1775 1776 /* 1777 * xrun ioctl 1778 * 1779 * Change the RUNNING stream(s) to XRUN state. 1780 */ 1781 static int snd_pcm_xrun(struct snd_pcm_substream *substream) 1782 { 1783 struct snd_pcm_runtime *runtime = substream->runtime; 1784 int result; 1785 1786 snd_pcm_stream_lock_irq(substream); 1787 switch (runtime->status->state) { 1788 case SNDRV_PCM_STATE_XRUN: 1789 result = 0; /* already there */ 1790 break; 1791 case SNDRV_PCM_STATE_RUNNING: 1792 __snd_pcm_xrun(substream); 1793 result = 0; 1794 break; 1795 default: 1796 result = -EBADFD; 1797 } 1798 snd_pcm_stream_unlock_irq(substream); 1799 return result; 1800 } 1801 1802 /* 1803 * reset ioctl 1804 */ 1805 /* reset callbacks: state argument ignored */ 1806 static int snd_pcm_pre_reset(struct snd_pcm_substream *substream, 1807 snd_pcm_state_t state) 1808 { 1809 struct snd_pcm_runtime *runtime = substream->runtime; 1810 switch (runtime->status->state) { 1811 case SNDRV_PCM_STATE_RUNNING: 1812 case SNDRV_PCM_STATE_PREPARED: 1813 case SNDRV_PCM_STATE_PAUSED: 1814 case SNDRV_PCM_STATE_SUSPENDED: 1815 return 0; 1816 default: 1817 return -EBADFD; 1818 } 1819 } 1820 1821 static int snd_pcm_do_reset(struct snd_pcm_substream *substream, 1822 snd_pcm_state_t state) 1823 { 1824 struct snd_pcm_runtime *runtime = substream->runtime; 1825 int err = snd_pcm_ops_ioctl(substream, SNDRV_PCM_IOCTL1_RESET, NULL); 1826 if (err < 0) 1827 return err; 1828 runtime->hw_ptr_base = 0; 1829 runtime->hw_ptr_interrupt = runtime->status->hw_ptr - 1830 runtime->status->hw_ptr % runtime->period_size; 1831 runtime->silence_start = runtime->status->hw_ptr; 1832 runtime->silence_filled = 0; 1833 return 0; 1834 } 1835 1836 static void snd_pcm_post_reset(struct snd_pcm_substream *substream, 1837 snd_pcm_state_t state) 1838 { 1839 struct snd_pcm_runtime *runtime = substream->runtime; 1840 runtime->control->appl_ptr = runtime->status->hw_ptr; 1841 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 1842 runtime->silence_size > 0) 1843 snd_pcm_playback_silence(substream, ULONG_MAX); 1844 } 1845 1846 static const struct action_ops snd_pcm_action_reset = { 1847 .pre_action = snd_pcm_pre_reset, 1848 .do_action = snd_pcm_do_reset, 1849 .post_action = snd_pcm_post_reset 1850 }; 1851 1852 static int snd_pcm_reset(struct snd_pcm_substream *substream) 1853 { 1854 return snd_pcm_action_nonatomic(&snd_pcm_action_reset, substream, 1855 ACTION_ARG_IGNORE); 1856 } 1857 1858 /* 1859 * prepare ioctl 1860 */ 1861 /* pass f_flags as state argument */ 1862 static int snd_pcm_pre_prepare(struct snd_pcm_substream *substream, 1863 snd_pcm_state_t state) 1864 { 1865 struct snd_pcm_runtime *runtime = substream->runtime; 1866 int f_flags = (__force int)state; 1867 1868 if (runtime->status->state == SNDRV_PCM_STATE_OPEN || 1869 runtime->status->state == SNDRV_PCM_STATE_DISCONNECTED) 1870 return -EBADFD; 1871 if (snd_pcm_running(substream)) 1872 return -EBUSY; 1873 substream->f_flags = f_flags; 1874 return 0; 1875 } 1876 1877 static int snd_pcm_do_prepare(struct snd_pcm_substream *substream, 1878 snd_pcm_state_t state) 1879 { 1880 int err; 1881 snd_pcm_sync_stop(substream, true); 1882 err = substream->ops->prepare(substream); 1883 if (err < 0) 1884 return err; 1885 return snd_pcm_do_reset(substream, state); 1886 } 1887 1888 static void snd_pcm_post_prepare(struct snd_pcm_substream *substream, 1889 snd_pcm_state_t state) 1890 { 1891 struct snd_pcm_runtime *runtime = substream->runtime; 1892 runtime->control->appl_ptr = runtime->status->hw_ptr; 1893 snd_pcm_set_state(substream, SNDRV_PCM_STATE_PREPARED); 1894 } 1895 1896 static const struct action_ops snd_pcm_action_prepare = { 1897 .pre_action = snd_pcm_pre_prepare, 1898 .do_action = snd_pcm_do_prepare, 1899 .post_action = snd_pcm_post_prepare 1900 }; 1901 1902 /** 1903 * snd_pcm_prepare - prepare the PCM substream to be triggerable 1904 * @substream: the PCM substream instance 1905 * @file: file to refer f_flags 1906 * 1907 * Return: Zero if successful, or a negative error code. 1908 */ 1909 static int snd_pcm_prepare(struct snd_pcm_substream *substream, 1910 struct file *file) 1911 { 1912 int f_flags; 1913 1914 if (file) 1915 f_flags = file->f_flags; 1916 else 1917 f_flags = substream->f_flags; 1918 1919 snd_pcm_stream_lock_irq(substream); 1920 switch (substream->runtime->status->state) { 1921 case SNDRV_PCM_STATE_PAUSED: 1922 snd_pcm_pause(substream, false); 1923 fallthrough; 1924 case SNDRV_PCM_STATE_SUSPENDED: 1925 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 1926 break; 1927 } 1928 snd_pcm_stream_unlock_irq(substream); 1929 1930 return snd_pcm_action_nonatomic(&snd_pcm_action_prepare, 1931 substream, 1932 (__force snd_pcm_state_t)f_flags); 1933 } 1934 1935 /* 1936 * drain ioctl 1937 */ 1938 1939 /* drain init callbacks: state argument ignored */ 1940 static int snd_pcm_pre_drain_init(struct snd_pcm_substream *substream, 1941 snd_pcm_state_t state) 1942 { 1943 struct snd_pcm_runtime *runtime = substream->runtime; 1944 switch (runtime->status->state) { 1945 case SNDRV_PCM_STATE_OPEN: 1946 case SNDRV_PCM_STATE_DISCONNECTED: 1947 case SNDRV_PCM_STATE_SUSPENDED: 1948 return -EBADFD; 1949 } 1950 runtime->trigger_master = substream; 1951 return 0; 1952 } 1953 1954 static int snd_pcm_do_drain_init(struct snd_pcm_substream *substream, 1955 snd_pcm_state_t state) 1956 { 1957 struct snd_pcm_runtime *runtime = substream->runtime; 1958 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { 1959 switch (runtime->status->state) { 1960 case SNDRV_PCM_STATE_PREPARED: 1961 /* start playback stream if possible */ 1962 if (! snd_pcm_playback_empty(substream)) { 1963 snd_pcm_do_start(substream, SNDRV_PCM_STATE_DRAINING); 1964 snd_pcm_post_start(substream, SNDRV_PCM_STATE_DRAINING); 1965 } else { 1966 runtime->status->state = SNDRV_PCM_STATE_SETUP; 1967 } 1968 break; 1969 case SNDRV_PCM_STATE_RUNNING: 1970 runtime->status->state = SNDRV_PCM_STATE_DRAINING; 1971 break; 1972 case SNDRV_PCM_STATE_XRUN: 1973 runtime->status->state = SNDRV_PCM_STATE_SETUP; 1974 break; 1975 default: 1976 break; 1977 } 1978 } else { 1979 /* stop running stream */ 1980 if (runtime->status->state == SNDRV_PCM_STATE_RUNNING) { 1981 snd_pcm_state_t new_state; 1982 1983 new_state = snd_pcm_capture_avail(runtime) > 0 ? 1984 SNDRV_PCM_STATE_DRAINING : SNDRV_PCM_STATE_SETUP; 1985 snd_pcm_do_stop(substream, new_state); 1986 snd_pcm_post_stop(substream, new_state); 1987 } 1988 } 1989 1990 if (runtime->status->state == SNDRV_PCM_STATE_DRAINING && 1991 runtime->trigger_master == substream && 1992 (runtime->hw.info & SNDRV_PCM_INFO_DRAIN_TRIGGER)) 1993 return substream->ops->trigger(substream, 1994 SNDRV_PCM_TRIGGER_DRAIN); 1995 1996 return 0; 1997 } 1998 1999 static void snd_pcm_post_drain_init(struct snd_pcm_substream *substream, 2000 snd_pcm_state_t state) 2001 { 2002 } 2003 2004 static const struct action_ops snd_pcm_action_drain_init = { 2005 .pre_action = snd_pcm_pre_drain_init, 2006 .do_action = snd_pcm_do_drain_init, 2007 .post_action = snd_pcm_post_drain_init 2008 }; 2009 2010 /* 2011 * Drain the stream(s). 2012 * When the substream is linked, sync until the draining of all playback streams 2013 * is finished. 2014 * After this call, all streams are supposed to be either SETUP or DRAINING 2015 * (capture only) state. 2016 */ 2017 static int snd_pcm_drain(struct snd_pcm_substream *substream, 2018 struct file *file) 2019 { 2020 struct snd_card *card; 2021 struct snd_pcm_runtime *runtime; 2022 struct snd_pcm_substream *s; 2023 struct snd_pcm_group *group; 2024 wait_queue_entry_t wait; 2025 int result = 0; 2026 int nonblock = 0; 2027 2028 card = substream->pcm->card; 2029 runtime = substream->runtime; 2030 2031 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 2032 return -EBADFD; 2033 2034 if (file) { 2035 if (file->f_flags & O_NONBLOCK) 2036 nonblock = 1; 2037 } else if (substream->f_flags & O_NONBLOCK) 2038 nonblock = 1; 2039 2040 snd_pcm_stream_lock_irq(substream); 2041 /* resume pause */ 2042 if (runtime->status->state == SNDRV_PCM_STATE_PAUSED) 2043 snd_pcm_pause(substream, false); 2044 2045 /* pre-start/stop - all running streams are changed to DRAINING state */ 2046 result = snd_pcm_action(&snd_pcm_action_drain_init, substream, 2047 ACTION_ARG_IGNORE); 2048 if (result < 0) 2049 goto unlock; 2050 /* in non-blocking, we don't wait in ioctl but let caller poll */ 2051 if (nonblock) { 2052 result = -EAGAIN; 2053 goto unlock; 2054 } 2055 2056 for (;;) { 2057 long tout; 2058 struct snd_pcm_runtime *to_check; 2059 if (signal_pending(current)) { 2060 result = -ERESTARTSYS; 2061 break; 2062 } 2063 /* find a substream to drain */ 2064 to_check = NULL; 2065 group = snd_pcm_stream_group_ref(substream); 2066 snd_pcm_group_for_each_entry(s, substream) { 2067 if (s->stream != SNDRV_PCM_STREAM_PLAYBACK) 2068 continue; 2069 runtime = s->runtime; 2070 if (runtime->status->state == SNDRV_PCM_STATE_DRAINING) { 2071 to_check = runtime; 2072 break; 2073 } 2074 } 2075 snd_pcm_group_unref(group, substream); 2076 if (!to_check) 2077 break; /* all drained */ 2078 init_waitqueue_entry(&wait, current); 2079 set_current_state(TASK_INTERRUPTIBLE); 2080 add_wait_queue(&to_check->sleep, &wait); 2081 snd_pcm_stream_unlock_irq(substream); 2082 if (runtime->no_period_wakeup) 2083 tout = MAX_SCHEDULE_TIMEOUT; 2084 else { 2085 tout = 10; 2086 if (runtime->rate) { 2087 long t = runtime->period_size * 2 / runtime->rate; 2088 tout = max(t, tout); 2089 } 2090 tout = msecs_to_jiffies(tout * 1000); 2091 } 2092 tout = schedule_timeout(tout); 2093 2094 snd_pcm_stream_lock_irq(substream); 2095 group = snd_pcm_stream_group_ref(substream); 2096 snd_pcm_group_for_each_entry(s, substream) { 2097 if (s->runtime == to_check) { 2098 remove_wait_queue(&to_check->sleep, &wait); 2099 break; 2100 } 2101 } 2102 snd_pcm_group_unref(group, substream); 2103 2104 if (card->shutdown) { 2105 result = -ENODEV; 2106 break; 2107 } 2108 if (tout == 0) { 2109 if (substream->runtime->status->state == SNDRV_PCM_STATE_SUSPENDED) 2110 result = -ESTRPIPE; 2111 else { 2112 dev_dbg(substream->pcm->card->dev, 2113 "playback drain error (DMA or IRQ trouble?)\n"); 2114 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 2115 result = -EIO; 2116 } 2117 break; 2118 } 2119 } 2120 2121 unlock: 2122 snd_pcm_stream_unlock_irq(substream); 2123 2124 return result; 2125 } 2126 2127 /* 2128 * drop ioctl 2129 * 2130 * Immediately put all linked substreams into SETUP state. 2131 */ 2132 static int snd_pcm_drop(struct snd_pcm_substream *substream) 2133 { 2134 struct snd_pcm_runtime *runtime; 2135 int result = 0; 2136 2137 if (PCM_RUNTIME_CHECK(substream)) 2138 return -ENXIO; 2139 runtime = substream->runtime; 2140 2141 if (runtime->status->state == SNDRV_PCM_STATE_OPEN || 2142 runtime->status->state == SNDRV_PCM_STATE_DISCONNECTED) 2143 return -EBADFD; 2144 2145 snd_pcm_stream_lock_irq(substream); 2146 /* resume pause */ 2147 if (runtime->status->state == SNDRV_PCM_STATE_PAUSED) 2148 snd_pcm_pause(substream, false); 2149 2150 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 2151 /* runtime->control->appl_ptr = runtime->status->hw_ptr; */ 2152 snd_pcm_stream_unlock_irq(substream); 2153 2154 return result; 2155 } 2156 2157 2158 static bool is_pcm_file(struct file *file) 2159 { 2160 struct inode *inode = file_inode(file); 2161 struct snd_pcm *pcm; 2162 unsigned int minor; 2163 2164 if (!S_ISCHR(inode->i_mode) || imajor(inode) != snd_major) 2165 return false; 2166 minor = iminor(inode); 2167 pcm = snd_lookup_minor_data(minor, SNDRV_DEVICE_TYPE_PCM_PLAYBACK); 2168 if (!pcm) 2169 pcm = snd_lookup_minor_data(minor, SNDRV_DEVICE_TYPE_PCM_CAPTURE); 2170 if (!pcm) 2171 return false; 2172 snd_card_unref(pcm->card); 2173 return true; 2174 } 2175 2176 /* 2177 * PCM link handling 2178 */ 2179 static int snd_pcm_link(struct snd_pcm_substream *substream, int fd) 2180 { 2181 int res = 0; 2182 struct snd_pcm_file *pcm_file; 2183 struct snd_pcm_substream *substream1; 2184 struct snd_pcm_group *group, *target_group; 2185 bool nonatomic = substream->pcm->nonatomic; 2186 struct fd f = fdget(fd); 2187 2188 if (!f.file) 2189 return -EBADFD; 2190 if (!is_pcm_file(f.file)) { 2191 res = -EBADFD; 2192 goto _badf; 2193 } 2194 pcm_file = f.file->private_data; 2195 substream1 = pcm_file->substream; 2196 2197 if (substream == substream1) { 2198 res = -EINVAL; 2199 goto _badf; 2200 } 2201 2202 group = kzalloc(sizeof(*group), GFP_KERNEL); 2203 if (!group) { 2204 res = -ENOMEM; 2205 goto _nolock; 2206 } 2207 snd_pcm_group_init(group); 2208 2209 down_write(&snd_pcm_link_rwsem); 2210 if (substream->runtime->status->state == SNDRV_PCM_STATE_OPEN || 2211 substream->runtime->status->state != substream1->runtime->status->state || 2212 substream->pcm->nonatomic != substream1->pcm->nonatomic) { 2213 res = -EBADFD; 2214 goto _end; 2215 } 2216 if (snd_pcm_stream_linked(substream1)) { 2217 res = -EALREADY; 2218 goto _end; 2219 } 2220 2221 snd_pcm_stream_lock_irq(substream); 2222 if (!snd_pcm_stream_linked(substream)) { 2223 snd_pcm_group_assign(substream, group); 2224 group = NULL; /* assigned, don't free this one below */ 2225 } 2226 target_group = substream->group; 2227 snd_pcm_stream_unlock_irq(substream); 2228 2229 snd_pcm_group_lock_irq(target_group, nonatomic); 2230 snd_pcm_stream_lock_nested(substream1); 2231 snd_pcm_group_assign(substream1, target_group); 2232 refcount_inc(&target_group->refs); 2233 snd_pcm_stream_unlock(substream1); 2234 snd_pcm_group_unlock_irq(target_group, nonatomic); 2235 _end: 2236 up_write(&snd_pcm_link_rwsem); 2237 _nolock: 2238 kfree(group); 2239 _badf: 2240 fdput(f); 2241 return res; 2242 } 2243 2244 static void relink_to_local(struct snd_pcm_substream *substream) 2245 { 2246 snd_pcm_stream_lock_nested(substream); 2247 snd_pcm_group_assign(substream, &substream->self_group); 2248 snd_pcm_stream_unlock(substream); 2249 } 2250 2251 static int snd_pcm_unlink(struct snd_pcm_substream *substream) 2252 { 2253 struct snd_pcm_group *group; 2254 bool nonatomic = substream->pcm->nonatomic; 2255 bool do_free = false; 2256 int res = 0; 2257 2258 down_write(&snd_pcm_link_rwsem); 2259 2260 if (!snd_pcm_stream_linked(substream)) { 2261 res = -EALREADY; 2262 goto _end; 2263 } 2264 2265 group = substream->group; 2266 snd_pcm_group_lock_irq(group, nonatomic); 2267 2268 relink_to_local(substream); 2269 refcount_dec(&group->refs); 2270 2271 /* detach the last stream, too */ 2272 if (list_is_singular(&group->substreams)) { 2273 relink_to_local(list_first_entry(&group->substreams, 2274 struct snd_pcm_substream, 2275 link_list)); 2276 do_free = refcount_dec_and_test(&group->refs); 2277 } 2278 2279 snd_pcm_group_unlock_irq(group, nonatomic); 2280 if (do_free) 2281 kfree(group); 2282 2283 _end: 2284 up_write(&snd_pcm_link_rwsem); 2285 return res; 2286 } 2287 2288 /* 2289 * hw configurator 2290 */ 2291 static int snd_pcm_hw_rule_mul(struct snd_pcm_hw_params *params, 2292 struct snd_pcm_hw_rule *rule) 2293 { 2294 struct snd_interval t; 2295 snd_interval_mul(hw_param_interval_c(params, rule->deps[0]), 2296 hw_param_interval_c(params, rule->deps[1]), &t); 2297 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2298 } 2299 2300 static int snd_pcm_hw_rule_div(struct snd_pcm_hw_params *params, 2301 struct snd_pcm_hw_rule *rule) 2302 { 2303 struct snd_interval t; 2304 snd_interval_div(hw_param_interval_c(params, rule->deps[0]), 2305 hw_param_interval_c(params, rule->deps[1]), &t); 2306 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2307 } 2308 2309 static int snd_pcm_hw_rule_muldivk(struct snd_pcm_hw_params *params, 2310 struct snd_pcm_hw_rule *rule) 2311 { 2312 struct snd_interval t; 2313 snd_interval_muldivk(hw_param_interval_c(params, rule->deps[0]), 2314 hw_param_interval_c(params, rule->deps[1]), 2315 (unsigned long) rule->private, &t); 2316 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2317 } 2318 2319 static int snd_pcm_hw_rule_mulkdiv(struct snd_pcm_hw_params *params, 2320 struct snd_pcm_hw_rule *rule) 2321 { 2322 struct snd_interval t; 2323 snd_interval_mulkdiv(hw_param_interval_c(params, rule->deps[0]), 2324 (unsigned long) rule->private, 2325 hw_param_interval_c(params, rule->deps[1]), &t); 2326 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2327 } 2328 2329 static int snd_pcm_hw_rule_format(struct snd_pcm_hw_params *params, 2330 struct snd_pcm_hw_rule *rule) 2331 { 2332 snd_pcm_format_t k; 2333 const struct snd_interval *i = 2334 hw_param_interval_c(params, rule->deps[0]); 2335 struct snd_mask m; 2336 struct snd_mask *mask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT); 2337 snd_mask_any(&m); 2338 pcm_for_each_format(k) { 2339 int bits; 2340 if (!snd_mask_test_format(mask, k)) 2341 continue; 2342 bits = snd_pcm_format_physical_width(k); 2343 if (bits <= 0) 2344 continue; /* ignore invalid formats */ 2345 if ((unsigned)bits < i->min || (unsigned)bits > i->max) 2346 snd_mask_reset(&m, (__force unsigned)k); 2347 } 2348 return snd_mask_refine(mask, &m); 2349 } 2350 2351 static int snd_pcm_hw_rule_sample_bits(struct snd_pcm_hw_params *params, 2352 struct snd_pcm_hw_rule *rule) 2353 { 2354 struct snd_interval t; 2355 snd_pcm_format_t k; 2356 2357 t.min = UINT_MAX; 2358 t.max = 0; 2359 t.openmin = 0; 2360 t.openmax = 0; 2361 pcm_for_each_format(k) { 2362 int bits; 2363 if (!snd_mask_test_format(hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT), k)) 2364 continue; 2365 bits = snd_pcm_format_physical_width(k); 2366 if (bits <= 0) 2367 continue; /* ignore invalid formats */ 2368 if (t.min > (unsigned)bits) 2369 t.min = bits; 2370 if (t.max < (unsigned)bits) 2371 t.max = bits; 2372 } 2373 t.integer = 1; 2374 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2375 } 2376 2377 #if SNDRV_PCM_RATE_5512 != 1 << 0 || SNDRV_PCM_RATE_192000 != 1 << 12 2378 #error "Change this table" 2379 #endif 2380 2381 static const unsigned int rates[] = { 2382 5512, 8000, 11025, 16000, 22050, 32000, 44100, 2383 48000, 64000, 88200, 96000, 176400, 192000, 352800, 384000 2384 }; 2385 2386 const struct snd_pcm_hw_constraint_list snd_pcm_known_rates = { 2387 .count = ARRAY_SIZE(rates), 2388 .list = rates, 2389 }; 2390 2391 static int snd_pcm_hw_rule_rate(struct snd_pcm_hw_params *params, 2392 struct snd_pcm_hw_rule *rule) 2393 { 2394 struct snd_pcm_hardware *hw = rule->private; 2395 return snd_interval_list(hw_param_interval(params, rule->var), 2396 snd_pcm_known_rates.count, 2397 snd_pcm_known_rates.list, hw->rates); 2398 } 2399 2400 static int snd_pcm_hw_rule_buffer_bytes_max(struct snd_pcm_hw_params *params, 2401 struct snd_pcm_hw_rule *rule) 2402 { 2403 struct snd_interval t; 2404 struct snd_pcm_substream *substream = rule->private; 2405 t.min = 0; 2406 t.max = substream->buffer_bytes_max; 2407 t.openmin = 0; 2408 t.openmax = 0; 2409 t.integer = 1; 2410 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2411 } 2412 2413 static int snd_pcm_hw_constraints_init(struct snd_pcm_substream *substream) 2414 { 2415 struct snd_pcm_runtime *runtime = substream->runtime; 2416 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 2417 int k, err; 2418 2419 for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) { 2420 snd_mask_any(constrs_mask(constrs, k)); 2421 } 2422 2423 for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) { 2424 snd_interval_any(constrs_interval(constrs, k)); 2425 } 2426 2427 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_CHANNELS)); 2428 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_SIZE)); 2429 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_BYTES)); 2430 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_SAMPLE_BITS)); 2431 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_FRAME_BITS)); 2432 2433 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FORMAT, 2434 snd_pcm_hw_rule_format, NULL, 2435 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 2436 if (err < 0) 2437 return err; 2438 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, 2439 snd_pcm_hw_rule_sample_bits, NULL, 2440 SNDRV_PCM_HW_PARAM_FORMAT, 2441 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 2442 if (err < 0) 2443 return err; 2444 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, 2445 snd_pcm_hw_rule_div, NULL, 2446 SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1); 2447 if (err < 0) 2448 return err; 2449 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, 2450 snd_pcm_hw_rule_mul, NULL, 2451 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1); 2452 if (err < 0) 2453 return err; 2454 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, 2455 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2456 SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1); 2457 if (err < 0) 2458 return err; 2459 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, 2460 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2461 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, -1); 2462 if (err < 0) 2463 return err; 2464 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_CHANNELS, 2465 snd_pcm_hw_rule_div, NULL, 2466 SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 2467 if (err < 0) 2468 return err; 2469 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE, 2470 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2471 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_TIME, -1); 2472 if (err < 0) 2473 return err; 2474 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE, 2475 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2476 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_BUFFER_TIME, -1); 2477 if (err < 0) 2478 return err; 2479 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIODS, 2480 snd_pcm_hw_rule_div, NULL, 2481 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1); 2482 if (err < 0) 2483 return err; 2484 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, 2485 snd_pcm_hw_rule_div, NULL, 2486 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1); 2487 if (err < 0) 2488 return err; 2489 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, 2490 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2491 SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2492 if (err < 0) 2493 return err; 2494 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, 2495 snd_pcm_hw_rule_muldivk, (void*) 1000000, 2496 SNDRV_PCM_HW_PARAM_PERIOD_TIME, SNDRV_PCM_HW_PARAM_RATE, -1); 2497 if (err < 0) 2498 return err; 2499 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 2500 snd_pcm_hw_rule_mul, NULL, 2501 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1); 2502 if (err < 0) 2503 return err; 2504 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 2505 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2506 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2507 if (err < 0) 2508 return err; 2509 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 2510 snd_pcm_hw_rule_muldivk, (void*) 1000000, 2511 SNDRV_PCM_HW_PARAM_BUFFER_TIME, SNDRV_PCM_HW_PARAM_RATE, -1); 2512 if (err < 0) 2513 return err; 2514 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, 2515 snd_pcm_hw_rule_muldivk, (void*) 8, 2516 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2517 if (err < 0) 2518 return err; 2519 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 2520 snd_pcm_hw_rule_muldivk, (void*) 8, 2521 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2522 if (err < 0) 2523 return err; 2524 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_TIME, 2525 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2526 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1); 2527 if (err < 0) 2528 return err; 2529 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_TIME, 2530 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2531 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1); 2532 if (err < 0) 2533 return err; 2534 return 0; 2535 } 2536 2537 static int snd_pcm_hw_constraints_complete(struct snd_pcm_substream *substream) 2538 { 2539 struct snd_pcm_runtime *runtime = substream->runtime; 2540 struct snd_pcm_hardware *hw = &runtime->hw; 2541 int err; 2542 unsigned int mask = 0; 2543 2544 if (hw->info & SNDRV_PCM_INFO_INTERLEAVED) 2545 mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_RW_INTERLEAVED); 2546 if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED) 2547 mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_RW_NONINTERLEAVED); 2548 if (hw_support_mmap(substream)) { 2549 if (hw->info & SNDRV_PCM_INFO_INTERLEAVED) 2550 mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_INTERLEAVED); 2551 if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED) 2552 mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED); 2553 if (hw->info & SNDRV_PCM_INFO_COMPLEX) 2554 mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_COMPLEX); 2555 } 2556 err = snd_pcm_hw_constraint_mask(runtime, SNDRV_PCM_HW_PARAM_ACCESS, mask); 2557 if (err < 0) 2558 return err; 2559 2560 err = snd_pcm_hw_constraint_mask64(runtime, SNDRV_PCM_HW_PARAM_FORMAT, hw->formats); 2561 if (err < 0) 2562 return err; 2563 2564 err = snd_pcm_hw_constraint_mask(runtime, SNDRV_PCM_HW_PARAM_SUBFORMAT, 2565 PARAM_MASK_BIT(SNDRV_PCM_SUBFORMAT_STD)); 2566 if (err < 0) 2567 return err; 2568 2569 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_CHANNELS, 2570 hw->channels_min, hw->channels_max); 2571 if (err < 0) 2572 return err; 2573 2574 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_RATE, 2575 hw->rate_min, hw->rate_max); 2576 if (err < 0) 2577 return err; 2578 2579 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, 2580 hw->period_bytes_min, hw->period_bytes_max); 2581 if (err < 0) 2582 return err; 2583 2584 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIODS, 2585 hw->periods_min, hw->periods_max); 2586 if (err < 0) 2587 return err; 2588 2589 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 2590 hw->period_bytes_min, hw->buffer_bytes_max); 2591 if (err < 0) 2592 return err; 2593 2594 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 2595 snd_pcm_hw_rule_buffer_bytes_max, substream, 2596 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, -1); 2597 if (err < 0) 2598 return err; 2599 2600 /* FIXME: remove */ 2601 if (runtime->dma_bytes) { 2602 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 0, runtime->dma_bytes); 2603 if (err < 0) 2604 return err; 2605 } 2606 2607 if (!(hw->rates & (SNDRV_PCM_RATE_KNOT | SNDRV_PCM_RATE_CONTINUOUS))) { 2608 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE, 2609 snd_pcm_hw_rule_rate, hw, 2610 SNDRV_PCM_HW_PARAM_RATE, -1); 2611 if (err < 0) 2612 return err; 2613 } 2614 2615 /* FIXME: this belong to lowlevel */ 2616 snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIOD_SIZE); 2617 2618 return 0; 2619 } 2620 2621 static void pcm_release_private(struct snd_pcm_substream *substream) 2622 { 2623 if (snd_pcm_stream_linked(substream)) 2624 snd_pcm_unlink(substream); 2625 } 2626 2627 void snd_pcm_release_substream(struct snd_pcm_substream *substream) 2628 { 2629 substream->ref_count--; 2630 if (substream->ref_count > 0) 2631 return; 2632 2633 snd_pcm_drop(substream); 2634 if (substream->hw_opened) { 2635 if (substream->runtime->status->state != SNDRV_PCM_STATE_OPEN) 2636 do_hw_free(substream); 2637 substream->ops->close(substream); 2638 substream->hw_opened = 0; 2639 } 2640 if (cpu_latency_qos_request_active(&substream->latency_pm_qos_req)) 2641 cpu_latency_qos_remove_request(&substream->latency_pm_qos_req); 2642 if (substream->pcm_release) { 2643 substream->pcm_release(substream); 2644 substream->pcm_release = NULL; 2645 } 2646 snd_pcm_detach_substream(substream); 2647 } 2648 EXPORT_SYMBOL(snd_pcm_release_substream); 2649 2650 int snd_pcm_open_substream(struct snd_pcm *pcm, int stream, 2651 struct file *file, 2652 struct snd_pcm_substream **rsubstream) 2653 { 2654 struct snd_pcm_substream *substream; 2655 int err; 2656 2657 err = snd_pcm_attach_substream(pcm, stream, file, &substream); 2658 if (err < 0) 2659 return err; 2660 if (substream->ref_count > 1) { 2661 *rsubstream = substream; 2662 return 0; 2663 } 2664 2665 err = snd_pcm_hw_constraints_init(substream); 2666 if (err < 0) { 2667 pcm_dbg(pcm, "snd_pcm_hw_constraints_init failed\n"); 2668 goto error; 2669 } 2670 2671 err = substream->ops->open(substream); 2672 if (err < 0) 2673 goto error; 2674 2675 substream->hw_opened = 1; 2676 2677 err = snd_pcm_hw_constraints_complete(substream); 2678 if (err < 0) { 2679 pcm_dbg(pcm, "snd_pcm_hw_constraints_complete failed\n"); 2680 goto error; 2681 } 2682 2683 *rsubstream = substream; 2684 return 0; 2685 2686 error: 2687 snd_pcm_release_substream(substream); 2688 return err; 2689 } 2690 EXPORT_SYMBOL(snd_pcm_open_substream); 2691 2692 static int snd_pcm_open_file(struct file *file, 2693 struct snd_pcm *pcm, 2694 int stream) 2695 { 2696 struct snd_pcm_file *pcm_file; 2697 struct snd_pcm_substream *substream; 2698 int err; 2699 2700 err = snd_pcm_open_substream(pcm, stream, file, &substream); 2701 if (err < 0) 2702 return err; 2703 2704 pcm_file = kzalloc(sizeof(*pcm_file), GFP_KERNEL); 2705 if (pcm_file == NULL) { 2706 snd_pcm_release_substream(substream); 2707 return -ENOMEM; 2708 } 2709 pcm_file->substream = substream; 2710 if (substream->ref_count == 1) 2711 substream->pcm_release = pcm_release_private; 2712 file->private_data = pcm_file; 2713 2714 return 0; 2715 } 2716 2717 static int snd_pcm_playback_open(struct inode *inode, struct file *file) 2718 { 2719 struct snd_pcm *pcm; 2720 int err = nonseekable_open(inode, file); 2721 if (err < 0) 2722 return err; 2723 pcm = snd_lookup_minor_data(iminor(inode), 2724 SNDRV_DEVICE_TYPE_PCM_PLAYBACK); 2725 err = snd_pcm_open(file, pcm, SNDRV_PCM_STREAM_PLAYBACK); 2726 if (pcm) 2727 snd_card_unref(pcm->card); 2728 return err; 2729 } 2730 2731 static int snd_pcm_capture_open(struct inode *inode, struct file *file) 2732 { 2733 struct snd_pcm *pcm; 2734 int err = nonseekable_open(inode, file); 2735 if (err < 0) 2736 return err; 2737 pcm = snd_lookup_minor_data(iminor(inode), 2738 SNDRV_DEVICE_TYPE_PCM_CAPTURE); 2739 err = snd_pcm_open(file, pcm, SNDRV_PCM_STREAM_CAPTURE); 2740 if (pcm) 2741 snd_card_unref(pcm->card); 2742 return err; 2743 } 2744 2745 static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream) 2746 { 2747 int err; 2748 wait_queue_entry_t wait; 2749 2750 if (pcm == NULL) { 2751 err = -ENODEV; 2752 goto __error1; 2753 } 2754 err = snd_card_file_add(pcm->card, file); 2755 if (err < 0) 2756 goto __error1; 2757 if (!try_module_get(pcm->card->module)) { 2758 err = -EFAULT; 2759 goto __error2; 2760 } 2761 init_waitqueue_entry(&wait, current); 2762 add_wait_queue(&pcm->open_wait, &wait); 2763 mutex_lock(&pcm->open_mutex); 2764 while (1) { 2765 err = snd_pcm_open_file(file, pcm, stream); 2766 if (err >= 0) 2767 break; 2768 if (err == -EAGAIN) { 2769 if (file->f_flags & O_NONBLOCK) { 2770 err = -EBUSY; 2771 break; 2772 } 2773 } else 2774 break; 2775 set_current_state(TASK_INTERRUPTIBLE); 2776 mutex_unlock(&pcm->open_mutex); 2777 schedule(); 2778 mutex_lock(&pcm->open_mutex); 2779 if (pcm->card->shutdown) { 2780 err = -ENODEV; 2781 break; 2782 } 2783 if (signal_pending(current)) { 2784 err = -ERESTARTSYS; 2785 break; 2786 } 2787 } 2788 remove_wait_queue(&pcm->open_wait, &wait); 2789 mutex_unlock(&pcm->open_mutex); 2790 if (err < 0) 2791 goto __error; 2792 return err; 2793 2794 __error: 2795 module_put(pcm->card->module); 2796 __error2: 2797 snd_card_file_remove(pcm->card, file); 2798 __error1: 2799 return err; 2800 } 2801 2802 static int snd_pcm_release(struct inode *inode, struct file *file) 2803 { 2804 struct snd_pcm *pcm; 2805 struct snd_pcm_substream *substream; 2806 struct snd_pcm_file *pcm_file; 2807 2808 pcm_file = file->private_data; 2809 substream = pcm_file->substream; 2810 if (snd_BUG_ON(!substream)) 2811 return -ENXIO; 2812 pcm = substream->pcm; 2813 2814 /* block until the device gets woken up as it may touch the hardware */ 2815 snd_power_wait(pcm->card); 2816 2817 mutex_lock(&pcm->open_mutex); 2818 snd_pcm_release_substream(substream); 2819 kfree(pcm_file); 2820 mutex_unlock(&pcm->open_mutex); 2821 wake_up(&pcm->open_wait); 2822 module_put(pcm->card->module); 2823 snd_card_file_remove(pcm->card, file); 2824 return 0; 2825 } 2826 2827 /* check and update PCM state; return 0 or a negative error 2828 * call this inside PCM lock 2829 */ 2830 static int do_pcm_hwsync(struct snd_pcm_substream *substream) 2831 { 2832 switch (substream->runtime->status->state) { 2833 case SNDRV_PCM_STATE_DRAINING: 2834 if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) 2835 return -EBADFD; 2836 fallthrough; 2837 case SNDRV_PCM_STATE_RUNNING: 2838 return snd_pcm_update_hw_ptr(substream); 2839 case SNDRV_PCM_STATE_PREPARED: 2840 case SNDRV_PCM_STATE_PAUSED: 2841 return 0; 2842 case SNDRV_PCM_STATE_SUSPENDED: 2843 return -ESTRPIPE; 2844 case SNDRV_PCM_STATE_XRUN: 2845 return -EPIPE; 2846 default: 2847 return -EBADFD; 2848 } 2849 } 2850 2851 /* increase the appl_ptr; returns the processed frames or a negative error */ 2852 static snd_pcm_sframes_t forward_appl_ptr(struct snd_pcm_substream *substream, 2853 snd_pcm_uframes_t frames, 2854 snd_pcm_sframes_t avail) 2855 { 2856 struct snd_pcm_runtime *runtime = substream->runtime; 2857 snd_pcm_sframes_t appl_ptr; 2858 int ret; 2859 2860 if (avail <= 0) 2861 return 0; 2862 if (frames > (snd_pcm_uframes_t)avail) 2863 frames = avail; 2864 appl_ptr = runtime->control->appl_ptr + frames; 2865 if (appl_ptr >= (snd_pcm_sframes_t)runtime->boundary) 2866 appl_ptr -= runtime->boundary; 2867 ret = pcm_lib_apply_appl_ptr(substream, appl_ptr); 2868 return ret < 0 ? ret : frames; 2869 } 2870 2871 /* decrease the appl_ptr; returns the processed frames or zero for error */ 2872 static snd_pcm_sframes_t rewind_appl_ptr(struct snd_pcm_substream *substream, 2873 snd_pcm_uframes_t frames, 2874 snd_pcm_sframes_t avail) 2875 { 2876 struct snd_pcm_runtime *runtime = substream->runtime; 2877 snd_pcm_sframes_t appl_ptr; 2878 int ret; 2879 2880 if (avail <= 0) 2881 return 0; 2882 if (frames > (snd_pcm_uframes_t)avail) 2883 frames = avail; 2884 appl_ptr = runtime->control->appl_ptr - frames; 2885 if (appl_ptr < 0) 2886 appl_ptr += runtime->boundary; 2887 ret = pcm_lib_apply_appl_ptr(substream, appl_ptr); 2888 /* NOTE: we return zero for errors because PulseAudio gets depressed 2889 * upon receiving an error from rewind ioctl and stops processing 2890 * any longer. Returning zero means that no rewind is done, so 2891 * it's not absolutely wrong to answer like that. 2892 */ 2893 return ret < 0 ? 0 : frames; 2894 } 2895 2896 static snd_pcm_sframes_t snd_pcm_rewind(struct snd_pcm_substream *substream, 2897 snd_pcm_uframes_t frames) 2898 { 2899 snd_pcm_sframes_t ret; 2900 2901 if (frames == 0) 2902 return 0; 2903 2904 snd_pcm_stream_lock_irq(substream); 2905 ret = do_pcm_hwsync(substream); 2906 if (!ret) 2907 ret = rewind_appl_ptr(substream, frames, 2908 snd_pcm_hw_avail(substream)); 2909 snd_pcm_stream_unlock_irq(substream); 2910 return ret; 2911 } 2912 2913 static snd_pcm_sframes_t snd_pcm_forward(struct snd_pcm_substream *substream, 2914 snd_pcm_uframes_t frames) 2915 { 2916 snd_pcm_sframes_t ret; 2917 2918 if (frames == 0) 2919 return 0; 2920 2921 snd_pcm_stream_lock_irq(substream); 2922 ret = do_pcm_hwsync(substream); 2923 if (!ret) 2924 ret = forward_appl_ptr(substream, frames, 2925 snd_pcm_avail(substream)); 2926 snd_pcm_stream_unlock_irq(substream); 2927 return ret; 2928 } 2929 2930 static int snd_pcm_hwsync(struct snd_pcm_substream *substream) 2931 { 2932 int err; 2933 2934 snd_pcm_stream_lock_irq(substream); 2935 err = do_pcm_hwsync(substream); 2936 snd_pcm_stream_unlock_irq(substream); 2937 return err; 2938 } 2939 2940 static int snd_pcm_delay(struct snd_pcm_substream *substream, 2941 snd_pcm_sframes_t *delay) 2942 { 2943 int err; 2944 snd_pcm_sframes_t n = 0; 2945 2946 snd_pcm_stream_lock_irq(substream); 2947 err = do_pcm_hwsync(substream); 2948 if (!err) 2949 n = snd_pcm_calc_delay(substream); 2950 snd_pcm_stream_unlock_irq(substream); 2951 if (!err) 2952 *delay = n; 2953 return err; 2954 } 2955 2956 static int snd_pcm_sync_ptr(struct snd_pcm_substream *substream, 2957 struct snd_pcm_sync_ptr __user *_sync_ptr) 2958 { 2959 struct snd_pcm_runtime *runtime = substream->runtime; 2960 struct snd_pcm_sync_ptr sync_ptr; 2961 volatile struct snd_pcm_mmap_status *status; 2962 volatile struct snd_pcm_mmap_control *control; 2963 int err; 2964 2965 memset(&sync_ptr, 0, sizeof(sync_ptr)); 2966 if (get_user(sync_ptr.flags, (unsigned __user *)&(_sync_ptr->flags))) 2967 return -EFAULT; 2968 if (copy_from_user(&sync_ptr.c.control, &(_sync_ptr->c.control), sizeof(struct snd_pcm_mmap_control))) 2969 return -EFAULT; 2970 status = runtime->status; 2971 control = runtime->control; 2972 if (sync_ptr.flags & SNDRV_PCM_SYNC_PTR_HWSYNC) { 2973 err = snd_pcm_hwsync(substream); 2974 if (err < 0) 2975 return err; 2976 } 2977 snd_pcm_stream_lock_irq(substream); 2978 if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_APPL)) { 2979 err = pcm_lib_apply_appl_ptr(substream, 2980 sync_ptr.c.control.appl_ptr); 2981 if (err < 0) { 2982 snd_pcm_stream_unlock_irq(substream); 2983 return err; 2984 } 2985 } else { 2986 sync_ptr.c.control.appl_ptr = control->appl_ptr; 2987 } 2988 if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN)) 2989 control->avail_min = sync_ptr.c.control.avail_min; 2990 else 2991 sync_ptr.c.control.avail_min = control->avail_min; 2992 sync_ptr.s.status.state = status->state; 2993 sync_ptr.s.status.hw_ptr = status->hw_ptr; 2994 sync_ptr.s.status.tstamp = status->tstamp; 2995 sync_ptr.s.status.suspended_state = status->suspended_state; 2996 sync_ptr.s.status.audio_tstamp = status->audio_tstamp; 2997 snd_pcm_stream_unlock_irq(substream); 2998 if (copy_to_user(_sync_ptr, &sync_ptr, sizeof(sync_ptr))) 2999 return -EFAULT; 3000 return 0; 3001 } 3002 3003 struct snd_pcm_mmap_status32 { 3004 snd_pcm_state_t state; 3005 s32 pad1; 3006 u32 hw_ptr; 3007 s32 tstamp_sec; 3008 s32 tstamp_nsec; 3009 snd_pcm_state_t suspended_state; 3010 s32 audio_tstamp_sec; 3011 s32 audio_tstamp_nsec; 3012 } __attribute__((packed)); 3013 3014 struct snd_pcm_mmap_control32 { 3015 u32 appl_ptr; 3016 u32 avail_min; 3017 }; 3018 3019 struct snd_pcm_sync_ptr32 { 3020 u32 flags; 3021 union { 3022 struct snd_pcm_mmap_status32 status; 3023 unsigned char reserved[64]; 3024 } s; 3025 union { 3026 struct snd_pcm_mmap_control32 control; 3027 unsigned char reserved[64]; 3028 } c; 3029 } __attribute__((packed)); 3030 3031 /* recalcuate the boundary within 32bit */ 3032 static snd_pcm_uframes_t recalculate_boundary(struct snd_pcm_runtime *runtime) 3033 { 3034 snd_pcm_uframes_t boundary; 3035 3036 if (! runtime->buffer_size) 3037 return 0; 3038 boundary = runtime->buffer_size; 3039 while (boundary * 2 <= 0x7fffffffUL - runtime->buffer_size) 3040 boundary *= 2; 3041 return boundary; 3042 } 3043 3044 static int snd_pcm_ioctl_sync_ptr_compat(struct snd_pcm_substream *substream, 3045 struct snd_pcm_sync_ptr32 __user *src) 3046 { 3047 struct snd_pcm_runtime *runtime = substream->runtime; 3048 volatile struct snd_pcm_mmap_status *status; 3049 volatile struct snd_pcm_mmap_control *control; 3050 u32 sflags; 3051 struct snd_pcm_mmap_control scontrol; 3052 struct snd_pcm_mmap_status sstatus; 3053 snd_pcm_uframes_t boundary; 3054 int err; 3055 3056 if (snd_BUG_ON(!runtime)) 3057 return -EINVAL; 3058 3059 if (get_user(sflags, &src->flags) || 3060 get_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || 3061 get_user(scontrol.avail_min, &src->c.control.avail_min)) 3062 return -EFAULT; 3063 if (sflags & SNDRV_PCM_SYNC_PTR_HWSYNC) { 3064 err = snd_pcm_hwsync(substream); 3065 if (err < 0) 3066 return err; 3067 } 3068 status = runtime->status; 3069 control = runtime->control; 3070 boundary = recalculate_boundary(runtime); 3071 if (! boundary) 3072 boundary = 0x7fffffff; 3073 snd_pcm_stream_lock_irq(substream); 3074 /* FIXME: we should consider the boundary for the sync from app */ 3075 if (!(sflags & SNDRV_PCM_SYNC_PTR_APPL)) { 3076 err = pcm_lib_apply_appl_ptr(substream, 3077 scontrol.appl_ptr); 3078 if (err < 0) { 3079 snd_pcm_stream_unlock_irq(substream); 3080 return err; 3081 } 3082 } else 3083 scontrol.appl_ptr = control->appl_ptr % boundary; 3084 if (!(sflags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN)) 3085 control->avail_min = scontrol.avail_min; 3086 else 3087 scontrol.avail_min = control->avail_min; 3088 sstatus.state = status->state; 3089 sstatus.hw_ptr = status->hw_ptr % boundary; 3090 sstatus.tstamp = status->tstamp; 3091 sstatus.suspended_state = status->suspended_state; 3092 sstatus.audio_tstamp = status->audio_tstamp; 3093 snd_pcm_stream_unlock_irq(substream); 3094 if (put_user(sstatus.state, &src->s.status.state) || 3095 put_user(sstatus.hw_ptr, &src->s.status.hw_ptr) || 3096 put_user(sstatus.tstamp.tv_sec, &src->s.status.tstamp_sec) || 3097 put_user(sstatus.tstamp.tv_nsec, &src->s.status.tstamp_nsec) || 3098 put_user(sstatus.suspended_state, &src->s.status.suspended_state) || 3099 put_user(sstatus.audio_tstamp.tv_sec, &src->s.status.audio_tstamp_sec) || 3100 put_user(sstatus.audio_tstamp.tv_nsec, &src->s.status.audio_tstamp_nsec) || 3101 put_user(scontrol.appl_ptr, &src->c.control.appl_ptr) || 3102 put_user(scontrol.avail_min, &src->c.control.avail_min)) 3103 return -EFAULT; 3104 3105 return 0; 3106 } 3107 #define __SNDRV_PCM_IOCTL_SYNC_PTR32 _IOWR('A', 0x23, struct snd_pcm_sync_ptr32) 3108 3109 static int snd_pcm_tstamp(struct snd_pcm_substream *substream, int __user *_arg) 3110 { 3111 struct snd_pcm_runtime *runtime = substream->runtime; 3112 int arg; 3113 3114 if (get_user(arg, _arg)) 3115 return -EFAULT; 3116 if (arg < 0 || arg > SNDRV_PCM_TSTAMP_TYPE_LAST) 3117 return -EINVAL; 3118 runtime->tstamp_type = arg; 3119 return 0; 3120 } 3121 3122 static int snd_pcm_xferi_frames_ioctl(struct snd_pcm_substream *substream, 3123 struct snd_xferi __user *_xferi) 3124 { 3125 struct snd_xferi xferi; 3126 struct snd_pcm_runtime *runtime = substream->runtime; 3127 snd_pcm_sframes_t result; 3128 3129 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3130 return -EBADFD; 3131 if (put_user(0, &_xferi->result)) 3132 return -EFAULT; 3133 if (copy_from_user(&xferi, _xferi, sizeof(xferi))) 3134 return -EFAULT; 3135 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 3136 result = snd_pcm_lib_write(substream, xferi.buf, xferi.frames); 3137 else 3138 result = snd_pcm_lib_read(substream, xferi.buf, xferi.frames); 3139 if (put_user(result, &_xferi->result)) 3140 return -EFAULT; 3141 return result < 0 ? result : 0; 3142 } 3143 3144 static int snd_pcm_xfern_frames_ioctl(struct snd_pcm_substream *substream, 3145 struct snd_xfern __user *_xfern) 3146 { 3147 struct snd_xfern xfern; 3148 struct snd_pcm_runtime *runtime = substream->runtime; 3149 void *bufs; 3150 snd_pcm_sframes_t result; 3151 3152 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3153 return -EBADFD; 3154 if (runtime->channels > 128) 3155 return -EINVAL; 3156 if (put_user(0, &_xfern->result)) 3157 return -EFAULT; 3158 if (copy_from_user(&xfern, _xfern, sizeof(xfern))) 3159 return -EFAULT; 3160 3161 bufs = memdup_user(xfern.bufs, sizeof(void *) * runtime->channels); 3162 if (IS_ERR(bufs)) 3163 return PTR_ERR(bufs); 3164 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 3165 result = snd_pcm_lib_writev(substream, bufs, xfern.frames); 3166 else 3167 result = snd_pcm_lib_readv(substream, bufs, xfern.frames); 3168 kfree(bufs); 3169 if (put_user(result, &_xfern->result)) 3170 return -EFAULT; 3171 return result < 0 ? result : 0; 3172 } 3173 3174 static int snd_pcm_rewind_ioctl(struct snd_pcm_substream *substream, 3175 snd_pcm_uframes_t __user *_frames) 3176 { 3177 snd_pcm_uframes_t frames; 3178 snd_pcm_sframes_t result; 3179 3180 if (get_user(frames, _frames)) 3181 return -EFAULT; 3182 if (put_user(0, _frames)) 3183 return -EFAULT; 3184 result = snd_pcm_rewind(substream, frames); 3185 if (put_user(result, _frames)) 3186 return -EFAULT; 3187 return result < 0 ? result : 0; 3188 } 3189 3190 static int snd_pcm_forward_ioctl(struct snd_pcm_substream *substream, 3191 snd_pcm_uframes_t __user *_frames) 3192 { 3193 snd_pcm_uframes_t frames; 3194 snd_pcm_sframes_t result; 3195 3196 if (get_user(frames, _frames)) 3197 return -EFAULT; 3198 if (put_user(0, _frames)) 3199 return -EFAULT; 3200 result = snd_pcm_forward(substream, frames); 3201 if (put_user(result, _frames)) 3202 return -EFAULT; 3203 return result < 0 ? result : 0; 3204 } 3205 3206 static int snd_pcm_common_ioctl(struct file *file, 3207 struct snd_pcm_substream *substream, 3208 unsigned int cmd, void __user *arg) 3209 { 3210 struct snd_pcm_file *pcm_file = file->private_data; 3211 int res; 3212 3213 if (PCM_RUNTIME_CHECK(substream)) 3214 return -ENXIO; 3215 3216 res = snd_power_wait(substream->pcm->card); 3217 if (res < 0) 3218 return res; 3219 3220 switch (cmd) { 3221 case SNDRV_PCM_IOCTL_PVERSION: 3222 return put_user(SNDRV_PCM_VERSION, (int __user *)arg) ? -EFAULT : 0; 3223 case SNDRV_PCM_IOCTL_INFO: 3224 return snd_pcm_info_user(substream, arg); 3225 case SNDRV_PCM_IOCTL_TSTAMP: /* just for compatibility */ 3226 return 0; 3227 case SNDRV_PCM_IOCTL_TTSTAMP: 3228 return snd_pcm_tstamp(substream, arg); 3229 case SNDRV_PCM_IOCTL_USER_PVERSION: 3230 if (get_user(pcm_file->user_pversion, 3231 (unsigned int __user *)arg)) 3232 return -EFAULT; 3233 return 0; 3234 case SNDRV_PCM_IOCTL_HW_REFINE: 3235 return snd_pcm_hw_refine_user(substream, arg); 3236 case SNDRV_PCM_IOCTL_HW_PARAMS: 3237 return snd_pcm_hw_params_user(substream, arg); 3238 case SNDRV_PCM_IOCTL_HW_FREE: 3239 return snd_pcm_hw_free(substream); 3240 case SNDRV_PCM_IOCTL_SW_PARAMS: 3241 return snd_pcm_sw_params_user(substream, arg); 3242 case SNDRV_PCM_IOCTL_STATUS32: 3243 return snd_pcm_status_user32(substream, arg, false); 3244 case SNDRV_PCM_IOCTL_STATUS_EXT32: 3245 return snd_pcm_status_user32(substream, arg, true); 3246 case SNDRV_PCM_IOCTL_STATUS64: 3247 return snd_pcm_status_user64(substream, arg, false); 3248 case SNDRV_PCM_IOCTL_STATUS_EXT64: 3249 return snd_pcm_status_user64(substream, arg, true); 3250 case SNDRV_PCM_IOCTL_CHANNEL_INFO: 3251 return snd_pcm_channel_info_user(substream, arg); 3252 case SNDRV_PCM_IOCTL_PREPARE: 3253 return snd_pcm_prepare(substream, file); 3254 case SNDRV_PCM_IOCTL_RESET: 3255 return snd_pcm_reset(substream); 3256 case SNDRV_PCM_IOCTL_START: 3257 return snd_pcm_start_lock_irq(substream); 3258 case SNDRV_PCM_IOCTL_LINK: 3259 return snd_pcm_link(substream, (int)(unsigned long) arg); 3260 case SNDRV_PCM_IOCTL_UNLINK: 3261 return snd_pcm_unlink(substream); 3262 case SNDRV_PCM_IOCTL_RESUME: 3263 return snd_pcm_resume(substream); 3264 case SNDRV_PCM_IOCTL_XRUN: 3265 return snd_pcm_xrun(substream); 3266 case SNDRV_PCM_IOCTL_HWSYNC: 3267 return snd_pcm_hwsync(substream); 3268 case SNDRV_PCM_IOCTL_DELAY: 3269 { 3270 snd_pcm_sframes_t delay; 3271 snd_pcm_sframes_t __user *res = arg; 3272 int err; 3273 3274 err = snd_pcm_delay(substream, &delay); 3275 if (err) 3276 return err; 3277 if (put_user(delay, res)) 3278 return -EFAULT; 3279 return 0; 3280 } 3281 case __SNDRV_PCM_IOCTL_SYNC_PTR32: 3282 return snd_pcm_ioctl_sync_ptr_compat(substream, arg); 3283 case __SNDRV_PCM_IOCTL_SYNC_PTR64: 3284 return snd_pcm_sync_ptr(substream, arg); 3285 #ifdef CONFIG_SND_SUPPORT_OLD_API 3286 case SNDRV_PCM_IOCTL_HW_REFINE_OLD: 3287 return snd_pcm_hw_refine_old_user(substream, arg); 3288 case SNDRV_PCM_IOCTL_HW_PARAMS_OLD: 3289 return snd_pcm_hw_params_old_user(substream, arg); 3290 #endif 3291 case SNDRV_PCM_IOCTL_DRAIN: 3292 return snd_pcm_drain(substream, file); 3293 case SNDRV_PCM_IOCTL_DROP: 3294 return snd_pcm_drop(substream); 3295 case SNDRV_PCM_IOCTL_PAUSE: 3296 return snd_pcm_pause_lock_irq(substream, (unsigned long)arg); 3297 case SNDRV_PCM_IOCTL_WRITEI_FRAMES: 3298 case SNDRV_PCM_IOCTL_READI_FRAMES: 3299 return snd_pcm_xferi_frames_ioctl(substream, arg); 3300 case SNDRV_PCM_IOCTL_WRITEN_FRAMES: 3301 case SNDRV_PCM_IOCTL_READN_FRAMES: 3302 return snd_pcm_xfern_frames_ioctl(substream, arg); 3303 case SNDRV_PCM_IOCTL_REWIND: 3304 return snd_pcm_rewind_ioctl(substream, arg); 3305 case SNDRV_PCM_IOCTL_FORWARD: 3306 return snd_pcm_forward_ioctl(substream, arg); 3307 } 3308 pcm_dbg(substream->pcm, "unknown ioctl = 0x%x\n", cmd); 3309 return -ENOTTY; 3310 } 3311 3312 static long snd_pcm_ioctl(struct file *file, unsigned int cmd, 3313 unsigned long arg) 3314 { 3315 struct snd_pcm_file *pcm_file; 3316 3317 pcm_file = file->private_data; 3318 3319 if (((cmd >> 8) & 0xff) != 'A') 3320 return -ENOTTY; 3321 3322 return snd_pcm_common_ioctl(file, pcm_file->substream, cmd, 3323 (void __user *)arg); 3324 } 3325 3326 /** 3327 * snd_pcm_kernel_ioctl - Execute PCM ioctl in the kernel-space 3328 * @substream: PCM substream 3329 * @cmd: IOCTL cmd 3330 * @arg: IOCTL argument 3331 * 3332 * The function is provided primarily for OSS layer and USB gadget drivers, 3333 * and it allows only the limited set of ioctls (hw_params, sw_params, 3334 * prepare, start, drain, drop, forward). 3335 */ 3336 int snd_pcm_kernel_ioctl(struct snd_pcm_substream *substream, 3337 unsigned int cmd, void *arg) 3338 { 3339 snd_pcm_uframes_t *frames = arg; 3340 snd_pcm_sframes_t result; 3341 3342 switch (cmd) { 3343 case SNDRV_PCM_IOCTL_FORWARD: 3344 { 3345 /* provided only for OSS; capture-only and no value returned */ 3346 if (substream->stream != SNDRV_PCM_STREAM_CAPTURE) 3347 return -EINVAL; 3348 result = snd_pcm_forward(substream, *frames); 3349 return result < 0 ? result : 0; 3350 } 3351 case SNDRV_PCM_IOCTL_HW_PARAMS: 3352 return snd_pcm_hw_params(substream, arg); 3353 case SNDRV_PCM_IOCTL_SW_PARAMS: 3354 return snd_pcm_sw_params(substream, arg); 3355 case SNDRV_PCM_IOCTL_PREPARE: 3356 return snd_pcm_prepare(substream, NULL); 3357 case SNDRV_PCM_IOCTL_START: 3358 return snd_pcm_start_lock_irq(substream); 3359 case SNDRV_PCM_IOCTL_DRAIN: 3360 return snd_pcm_drain(substream, NULL); 3361 case SNDRV_PCM_IOCTL_DROP: 3362 return snd_pcm_drop(substream); 3363 case SNDRV_PCM_IOCTL_DELAY: 3364 return snd_pcm_delay(substream, frames); 3365 default: 3366 return -EINVAL; 3367 } 3368 } 3369 EXPORT_SYMBOL(snd_pcm_kernel_ioctl); 3370 3371 static ssize_t snd_pcm_read(struct file *file, char __user *buf, size_t count, 3372 loff_t * offset) 3373 { 3374 struct snd_pcm_file *pcm_file; 3375 struct snd_pcm_substream *substream; 3376 struct snd_pcm_runtime *runtime; 3377 snd_pcm_sframes_t result; 3378 3379 pcm_file = file->private_data; 3380 substream = pcm_file->substream; 3381 if (PCM_RUNTIME_CHECK(substream)) 3382 return -ENXIO; 3383 runtime = substream->runtime; 3384 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3385 return -EBADFD; 3386 if (!frame_aligned(runtime, count)) 3387 return -EINVAL; 3388 count = bytes_to_frames(runtime, count); 3389 result = snd_pcm_lib_read(substream, buf, count); 3390 if (result > 0) 3391 result = frames_to_bytes(runtime, result); 3392 return result; 3393 } 3394 3395 static ssize_t snd_pcm_write(struct file *file, const char __user *buf, 3396 size_t count, loff_t * offset) 3397 { 3398 struct snd_pcm_file *pcm_file; 3399 struct snd_pcm_substream *substream; 3400 struct snd_pcm_runtime *runtime; 3401 snd_pcm_sframes_t result; 3402 3403 pcm_file = file->private_data; 3404 substream = pcm_file->substream; 3405 if (PCM_RUNTIME_CHECK(substream)) 3406 return -ENXIO; 3407 runtime = substream->runtime; 3408 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3409 return -EBADFD; 3410 if (!frame_aligned(runtime, count)) 3411 return -EINVAL; 3412 count = bytes_to_frames(runtime, count); 3413 result = snd_pcm_lib_write(substream, buf, count); 3414 if (result > 0) 3415 result = frames_to_bytes(runtime, result); 3416 return result; 3417 } 3418 3419 static ssize_t snd_pcm_readv(struct kiocb *iocb, struct iov_iter *to) 3420 { 3421 struct snd_pcm_file *pcm_file; 3422 struct snd_pcm_substream *substream; 3423 struct snd_pcm_runtime *runtime; 3424 snd_pcm_sframes_t result; 3425 unsigned long i; 3426 void __user **bufs; 3427 snd_pcm_uframes_t frames; 3428 3429 pcm_file = iocb->ki_filp->private_data; 3430 substream = pcm_file->substream; 3431 if (PCM_RUNTIME_CHECK(substream)) 3432 return -ENXIO; 3433 runtime = substream->runtime; 3434 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3435 return -EBADFD; 3436 if (!iter_is_iovec(to)) 3437 return -EINVAL; 3438 if (to->nr_segs > 1024 || to->nr_segs != runtime->channels) 3439 return -EINVAL; 3440 if (!frame_aligned(runtime, to->iov->iov_len)) 3441 return -EINVAL; 3442 frames = bytes_to_samples(runtime, to->iov->iov_len); 3443 bufs = kmalloc_array(to->nr_segs, sizeof(void *), GFP_KERNEL); 3444 if (bufs == NULL) 3445 return -ENOMEM; 3446 for (i = 0; i < to->nr_segs; ++i) 3447 bufs[i] = to->iov[i].iov_base; 3448 result = snd_pcm_lib_readv(substream, bufs, frames); 3449 if (result > 0) 3450 result = frames_to_bytes(runtime, result); 3451 kfree(bufs); 3452 return result; 3453 } 3454 3455 static ssize_t snd_pcm_writev(struct kiocb *iocb, struct iov_iter *from) 3456 { 3457 struct snd_pcm_file *pcm_file; 3458 struct snd_pcm_substream *substream; 3459 struct snd_pcm_runtime *runtime; 3460 snd_pcm_sframes_t result; 3461 unsigned long i; 3462 void __user **bufs; 3463 snd_pcm_uframes_t frames; 3464 3465 pcm_file = iocb->ki_filp->private_data; 3466 substream = pcm_file->substream; 3467 if (PCM_RUNTIME_CHECK(substream)) 3468 return -ENXIO; 3469 runtime = substream->runtime; 3470 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3471 return -EBADFD; 3472 if (!iter_is_iovec(from)) 3473 return -EINVAL; 3474 if (from->nr_segs > 128 || from->nr_segs != runtime->channels || 3475 !frame_aligned(runtime, from->iov->iov_len)) 3476 return -EINVAL; 3477 frames = bytes_to_samples(runtime, from->iov->iov_len); 3478 bufs = kmalloc_array(from->nr_segs, sizeof(void *), GFP_KERNEL); 3479 if (bufs == NULL) 3480 return -ENOMEM; 3481 for (i = 0; i < from->nr_segs; ++i) 3482 bufs[i] = from->iov[i].iov_base; 3483 result = snd_pcm_lib_writev(substream, bufs, frames); 3484 if (result > 0) 3485 result = frames_to_bytes(runtime, result); 3486 kfree(bufs); 3487 return result; 3488 } 3489 3490 static __poll_t snd_pcm_poll(struct file *file, poll_table *wait) 3491 { 3492 struct snd_pcm_file *pcm_file; 3493 struct snd_pcm_substream *substream; 3494 struct snd_pcm_runtime *runtime; 3495 __poll_t mask, ok; 3496 snd_pcm_uframes_t avail; 3497 3498 pcm_file = file->private_data; 3499 3500 substream = pcm_file->substream; 3501 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 3502 ok = EPOLLOUT | EPOLLWRNORM; 3503 else 3504 ok = EPOLLIN | EPOLLRDNORM; 3505 if (PCM_RUNTIME_CHECK(substream)) 3506 return ok | EPOLLERR; 3507 3508 runtime = substream->runtime; 3509 poll_wait(file, &runtime->sleep, wait); 3510 3511 mask = 0; 3512 snd_pcm_stream_lock_irq(substream); 3513 avail = snd_pcm_avail(substream); 3514 switch (runtime->status->state) { 3515 case SNDRV_PCM_STATE_RUNNING: 3516 case SNDRV_PCM_STATE_PREPARED: 3517 case SNDRV_PCM_STATE_PAUSED: 3518 if (avail >= runtime->control->avail_min) 3519 mask = ok; 3520 break; 3521 case SNDRV_PCM_STATE_DRAINING: 3522 if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) { 3523 mask = ok; 3524 if (!avail) 3525 mask |= EPOLLERR; 3526 } 3527 break; 3528 default: 3529 mask = ok | EPOLLERR; 3530 break; 3531 } 3532 snd_pcm_stream_unlock_irq(substream); 3533 return mask; 3534 } 3535 3536 /* 3537 * mmap support 3538 */ 3539 3540 /* 3541 * Only on coherent architectures, we can mmap the status and the control records 3542 * for effcient data transfer. On others, we have to use HWSYNC ioctl... 3543 */ 3544 #if defined(CONFIG_X86) || defined(CONFIG_PPC) || defined(CONFIG_ALPHA) 3545 /* 3546 * mmap status record 3547 */ 3548 static vm_fault_t snd_pcm_mmap_status_fault(struct vm_fault *vmf) 3549 { 3550 struct snd_pcm_substream *substream = vmf->vma->vm_private_data; 3551 struct snd_pcm_runtime *runtime; 3552 3553 if (substream == NULL) 3554 return VM_FAULT_SIGBUS; 3555 runtime = substream->runtime; 3556 vmf->page = virt_to_page(runtime->status); 3557 get_page(vmf->page); 3558 return 0; 3559 } 3560 3561 static const struct vm_operations_struct snd_pcm_vm_ops_status = 3562 { 3563 .fault = snd_pcm_mmap_status_fault, 3564 }; 3565 3566 static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file, 3567 struct vm_area_struct *area) 3568 { 3569 long size; 3570 if (!(area->vm_flags & VM_READ)) 3571 return -EINVAL; 3572 size = area->vm_end - area->vm_start; 3573 if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_status))) 3574 return -EINVAL; 3575 area->vm_ops = &snd_pcm_vm_ops_status; 3576 area->vm_private_data = substream; 3577 area->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; 3578 return 0; 3579 } 3580 3581 /* 3582 * mmap control record 3583 */ 3584 static vm_fault_t snd_pcm_mmap_control_fault(struct vm_fault *vmf) 3585 { 3586 struct snd_pcm_substream *substream = vmf->vma->vm_private_data; 3587 struct snd_pcm_runtime *runtime; 3588 3589 if (substream == NULL) 3590 return VM_FAULT_SIGBUS; 3591 runtime = substream->runtime; 3592 vmf->page = virt_to_page(runtime->control); 3593 get_page(vmf->page); 3594 return 0; 3595 } 3596 3597 static const struct vm_operations_struct snd_pcm_vm_ops_control = 3598 { 3599 .fault = snd_pcm_mmap_control_fault, 3600 }; 3601 3602 static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file, 3603 struct vm_area_struct *area) 3604 { 3605 long size; 3606 if (!(area->vm_flags & VM_READ)) 3607 return -EINVAL; 3608 size = area->vm_end - area->vm_start; 3609 if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_control))) 3610 return -EINVAL; 3611 area->vm_ops = &snd_pcm_vm_ops_control; 3612 area->vm_private_data = substream; 3613 area->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; 3614 return 0; 3615 } 3616 3617 static bool pcm_status_mmap_allowed(struct snd_pcm_file *pcm_file) 3618 { 3619 /* See pcm_control_mmap_allowed() below. 3620 * Since older alsa-lib requires both status and control mmaps to be 3621 * coupled, we have to disable the status mmap for old alsa-lib, too. 3622 */ 3623 if (pcm_file->user_pversion < SNDRV_PROTOCOL_VERSION(2, 0, 14) && 3624 (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR)) 3625 return false; 3626 return true; 3627 } 3628 3629 static bool pcm_control_mmap_allowed(struct snd_pcm_file *pcm_file) 3630 { 3631 if (pcm_file->no_compat_mmap) 3632 return false; 3633 /* Disallow the control mmap when SYNC_APPLPTR flag is set; 3634 * it enforces the user-space to fall back to snd_pcm_sync_ptr(), 3635 * thus it effectively assures the manual update of appl_ptr. 3636 */ 3637 if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR) 3638 return false; 3639 return true; 3640 } 3641 3642 #else /* ! coherent mmap */ 3643 /* 3644 * don't support mmap for status and control records. 3645 */ 3646 #define pcm_status_mmap_allowed(pcm_file) false 3647 #define pcm_control_mmap_allowed(pcm_file) false 3648 3649 static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file, 3650 struct vm_area_struct *area) 3651 { 3652 return -ENXIO; 3653 } 3654 static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file, 3655 struct vm_area_struct *area) 3656 { 3657 return -ENXIO; 3658 } 3659 #endif /* coherent mmap */ 3660 3661 /* 3662 * fault callback for mmapping a RAM page 3663 */ 3664 static vm_fault_t snd_pcm_mmap_data_fault(struct vm_fault *vmf) 3665 { 3666 struct snd_pcm_substream *substream = vmf->vma->vm_private_data; 3667 struct snd_pcm_runtime *runtime; 3668 unsigned long offset; 3669 struct page * page; 3670 size_t dma_bytes; 3671 3672 if (substream == NULL) 3673 return VM_FAULT_SIGBUS; 3674 runtime = substream->runtime; 3675 offset = vmf->pgoff << PAGE_SHIFT; 3676 dma_bytes = PAGE_ALIGN(runtime->dma_bytes); 3677 if (offset > dma_bytes - PAGE_SIZE) 3678 return VM_FAULT_SIGBUS; 3679 if (substream->ops->page) 3680 page = substream->ops->page(substream, offset); 3681 else if (!snd_pcm_get_dma_buf(substream)) 3682 page = virt_to_page(runtime->dma_area + offset); 3683 else 3684 page = snd_sgbuf_get_page(snd_pcm_get_dma_buf(substream), offset); 3685 if (!page) 3686 return VM_FAULT_SIGBUS; 3687 get_page(page); 3688 vmf->page = page; 3689 return 0; 3690 } 3691 3692 static const struct vm_operations_struct snd_pcm_vm_ops_data = { 3693 .open = snd_pcm_mmap_data_open, 3694 .close = snd_pcm_mmap_data_close, 3695 }; 3696 3697 static const struct vm_operations_struct snd_pcm_vm_ops_data_fault = { 3698 .open = snd_pcm_mmap_data_open, 3699 .close = snd_pcm_mmap_data_close, 3700 .fault = snd_pcm_mmap_data_fault, 3701 }; 3702 3703 /* 3704 * mmap the DMA buffer on RAM 3705 */ 3706 3707 /** 3708 * snd_pcm_lib_default_mmap - Default PCM data mmap function 3709 * @substream: PCM substream 3710 * @area: VMA 3711 * 3712 * This is the default mmap handler for PCM data. When mmap pcm_ops is NULL, 3713 * this function is invoked implicitly. 3714 */ 3715 int snd_pcm_lib_default_mmap(struct snd_pcm_substream *substream, 3716 struct vm_area_struct *area) 3717 { 3718 area->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; 3719 if (!substream->ops->page && 3720 !snd_dma_buffer_mmap(snd_pcm_get_dma_buf(substream), area)) 3721 return 0; 3722 /* mmap with fault handler */ 3723 area->vm_ops = &snd_pcm_vm_ops_data_fault; 3724 return 0; 3725 } 3726 EXPORT_SYMBOL_GPL(snd_pcm_lib_default_mmap); 3727 3728 /* 3729 * mmap the DMA buffer on I/O memory area 3730 */ 3731 #if SNDRV_PCM_INFO_MMAP_IOMEM 3732 /** 3733 * snd_pcm_lib_mmap_iomem - Default PCM data mmap function for I/O mem 3734 * @substream: PCM substream 3735 * @area: VMA 3736 * 3737 * When your hardware uses the iomapped pages as the hardware buffer and 3738 * wants to mmap it, pass this function as mmap pcm_ops. Note that this 3739 * is supposed to work only on limited architectures. 3740 */ 3741 int snd_pcm_lib_mmap_iomem(struct snd_pcm_substream *substream, 3742 struct vm_area_struct *area) 3743 { 3744 struct snd_pcm_runtime *runtime = substream->runtime; 3745 3746 area->vm_page_prot = pgprot_noncached(area->vm_page_prot); 3747 return vm_iomap_memory(area, runtime->dma_addr, runtime->dma_bytes); 3748 } 3749 EXPORT_SYMBOL(snd_pcm_lib_mmap_iomem); 3750 #endif /* SNDRV_PCM_INFO_MMAP */ 3751 3752 /* 3753 * mmap DMA buffer 3754 */ 3755 int snd_pcm_mmap_data(struct snd_pcm_substream *substream, struct file *file, 3756 struct vm_area_struct *area) 3757 { 3758 struct snd_pcm_runtime *runtime; 3759 long size; 3760 unsigned long offset; 3761 size_t dma_bytes; 3762 int err; 3763 3764 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { 3765 if (!(area->vm_flags & (VM_WRITE|VM_READ))) 3766 return -EINVAL; 3767 } else { 3768 if (!(area->vm_flags & VM_READ)) 3769 return -EINVAL; 3770 } 3771 runtime = substream->runtime; 3772 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3773 return -EBADFD; 3774 if (!(runtime->info & SNDRV_PCM_INFO_MMAP)) 3775 return -ENXIO; 3776 if (runtime->access == SNDRV_PCM_ACCESS_RW_INTERLEAVED || 3777 runtime->access == SNDRV_PCM_ACCESS_RW_NONINTERLEAVED) 3778 return -EINVAL; 3779 size = area->vm_end - area->vm_start; 3780 offset = area->vm_pgoff << PAGE_SHIFT; 3781 dma_bytes = PAGE_ALIGN(runtime->dma_bytes); 3782 if ((size_t)size > dma_bytes) 3783 return -EINVAL; 3784 if (offset > dma_bytes - size) 3785 return -EINVAL; 3786 3787 area->vm_ops = &snd_pcm_vm_ops_data; 3788 area->vm_private_data = substream; 3789 if (substream->ops->mmap) 3790 err = substream->ops->mmap(substream, area); 3791 else 3792 err = snd_pcm_lib_default_mmap(substream, area); 3793 if (!err) 3794 atomic_inc(&substream->mmap_count); 3795 return err; 3796 } 3797 EXPORT_SYMBOL(snd_pcm_mmap_data); 3798 3799 static int snd_pcm_mmap(struct file *file, struct vm_area_struct *area) 3800 { 3801 struct snd_pcm_file * pcm_file; 3802 struct snd_pcm_substream *substream; 3803 unsigned long offset; 3804 3805 pcm_file = file->private_data; 3806 substream = pcm_file->substream; 3807 if (PCM_RUNTIME_CHECK(substream)) 3808 return -ENXIO; 3809 3810 offset = area->vm_pgoff << PAGE_SHIFT; 3811 switch (offset) { 3812 case SNDRV_PCM_MMAP_OFFSET_STATUS_OLD: 3813 if (pcm_file->no_compat_mmap || !IS_ENABLED(CONFIG_64BIT)) 3814 return -ENXIO; 3815 fallthrough; 3816 case SNDRV_PCM_MMAP_OFFSET_STATUS_NEW: 3817 if (!pcm_status_mmap_allowed(pcm_file)) 3818 return -ENXIO; 3819 return snd_pcm_mmap_status(substream, file, area); 3820 case SNDRV_PCM_MMAP_OFFSET_CONTROL_OLD: 3821 if (pcm_file->no_compat_mmap || !IS_ENABLED(CONFIG_64BIT)) 3822 return -ENXIO; 3823 fallthrough; 3824 case SNDRV_PCM_MMAP_OFFSET_CONTROL_NEW: 3825 if (!pcm_control_mmap_allowed(pcm_file)) 3826 return -ENXIO; 3827 return snd_pcm_mmap_control(substream, file, area); 3828 default: 3829 return snd_pcm_mmap_data(substream, file, area); 3830 } 3831 return 0; 3832 } 3833 3834 static int snd_pcm_fasync(int fd, struct file * file, int on) 3835 { 3836 struct snd_pcm_file * pcm_file; 3837 struct snd_pcm_substream *substream; 3838 struct snd_pcm_runtime *runtime; 3839 3840 pcm_file = file->private_data; 3841 substream = pcm_file->substream; 3842 if (PCM_RUNTIME_CHECK(substream)) 3843 return -ENXIO; 3844 runtime = substream->runtime; 3845 return fasync_helper(fd, file, on, &runtime->fasync); 3846 } 3847 3848 /* 3849 * ioctl32 compat 3850 */ 3851 #ifdef CONFIG_COMPAT 3852 #include "pcm_compat.c" 3853 #else 3854 #define snd_pcm_ioctl_compat NULL 3855 #endif 3856 3857 /* 3858 * To be removed helpers to keep binary compatibility 3859 */ 3860 3861 #ifdef CONFIG_SND_SUPPORT_OLD_API 3862 #define __OLD_TO_NEW_MASK(x) ((x&7)|((x&0x07fffff8)<<5)) 3863 #define __NEW_TO_OLD_MASK(x) ((x&7)|((x&0xffffff00)>>5)) 3864 3865 static void snd_pcm_hw_convert_from_old_params(struct snd_pcm_hw_params *params, 3866 struct snd_pcm_hw_params_old *oparams) 3867 { 3868 unsigned int i; 3869 3870 memset(params, 0, sizeof(*params)); 3871 params->flags = oparams->flags; 3872 for (i = 0; i < ARRAY_SIZE(oparams->masks); i++) 3873 params->masks[i].bits[0] = oparams->masks[i]; 3874 memcpy(params->intervals, oparams->intervals, sizeof(oparams->intervals)); 3875 params->rmask = __OLD_TO_NEW_MASK(oparams->rmask); 3876 params->cmask = __OLD_TO_NEW_MASK(oparams->cmask); 3877 params->info = oparams->info; 3878 params->msbits = oparams->msbits; 3879 params->rate_num = oparams->rate_num; 3880 params->rate_den = oparams->rate_den; 3881 params->fifo_size = oparams->fifo_size; 3882 } 3883 3884 static void snd_pcm_hw_convert_to_old_params(struct snd_pcm_hw_params_old *oparams, 3885 struct snd_pcm_hw_params *params) 3886 { 3887 unsigned int i; 3888 3889 memset(oparams, 0, sizeof(*oparams)); 3890 oparams->flags = params->flags; 3891 for (i = 0; i < ARRAY_SIZE(oparams->masks); i++) 3892 oparams->masks[i] = params->masks[i].bits[0]; 3893 memcpy(oparams->intervals, params->intervals, sizeof(oparams->intervals)); 3894 oparams->rmask = __NEW_TO_OLD_MASK(params->rmask); 3895 oparams->cmask = __NEW_TO_OLD_MASK(params->cmask); 3896 oparams->info = params->info; 3897 oparams->msbits = params->msbits; 3898 oparams->rate_num = params->rate_num; 3899 oparams->rate_den = params->rate_den; 3900 oparams->fifo_size = params->fifo_size; 3901 } 3902 3903 static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream, 3904 struct snd_pcm_hw_params_old __user * _oparams) 3905 { 3906 struct snd_pcm_hw_params *params; 3907 struct snd_pcm_hw_params_old *oparams = NULL; 3908 int err; 3909 3910 params = kmalloc(sizeof(*params), GFP_KERNEL); 3911 if (!params) 3912 return -ENOMEM; 3913 3914 oparams = memdup_user(_oparams, sizeof(*oparams)); 3915 if (IS_ERR(oparams)) { 3916 err = PTR_ERR(oparams); 3917 goto out; 3918 } 3919 snd_pcm_hw_convert_from_old_params(params, oparams); 3920 err = snd_pcm_hw_refine(substream, params); 3921 if (err < 0) 3922 goto out_old; 3923 3924 err = fixup_unreferenced_params(substream, params); 3925 if (err < 0) 3926 goto out_old; 3927 3928 snd_pcm_hw_convert_to_old_params(oparams, params); 3929 if (copy_to_user(_oparams, oparams, sizeof(*oparams))) 3930 err = -EFAULT; 3931 out_old: 3932 kfree(oparams); 3933 out: 3934 kfree(params); 3935 return err; 3936 } 3937 3938 static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream, 3939 struct snd_pcm_hw_params_old __user * _oparams) 3940 { 3941 struct snd_pcm_hw_params *params; 3942 struct snd_pcm_hw_params_old *oparams = NULL; 3943 int err; 3944 3945 params = kmalloc(sizeof(*params), GFP_KERNEL); 3946 if (!params) 3947 return -ENOMEM; 3948 3949 oparams = memdup_user(_oparams, sizeof(*oparams)); 3950 if (IS_ERR(oparams)) { 3951 err = PTR_ERR(oparams); 3952 goto out; 3953 } 3954 3955 snd_pcm_hw_convert_from_old_params(params, oparams); 3956 err = snd_pcm_hw_params(substream, params); 3957 if (err < 0) 3958 goto out_old; 3959 3960 snd_pcm_hw_convert_to_old_params(oparams, params); 3961 if (copy_to_user(_oparams, oparams, sizeof(*oparams))) 3962 err = -EFAULT; 3963 out_old: 3964 kfree(oparams); 3965 out: 3966 kfree(params); 3967 return err; 3968 } 3969 #endif /* CONFIG_SND_SUPPORT_OLD_API */ 3970 3971 #ifndef CONFIG_MMU 3972 static unsigned long snd_pcm_get_unmapped_area(struct file *file, 3973 unsigned long addr, 3974 unsigned long len, 3975 unsigned long pgoff, 3976 unsigned long flags) 3977 { 3978 struct snd_pcm_file *pcm_file = file->private_data; 3979 struct snd_pcm_substream *substream = pcm_file->substream; 3980 struct snd_pcm_runtime *runtime = substream->runtime; 3981 unsigned long offset = pgoff << PAGE_SHIFT; 3982 3983 switch (offset) { 3984 case SNDRV_PCM_MMAP_OFFSET_STATUS_NEW: 3985 return (unsigned long)runtime->status; 3986 case SNDRV_PCM_MMAP_OFFSET_CONTROL_NEW: 3987 return (unsigned long)runtime->control; 3988 default: 3989 return (unsigned long)runtime->dma_area + offset; 3990 } 3991 } 3992 #else 3993 # define snd_pcm_get_unmapped_area NULL 3994 #endif 3995 3996 /* 3997 * Register section 3998 */ 3999 4000 const struct file_operations snd_pcm_f_ops[2] = { 4001 { 4002 .owner = THIS_MODULE, 4003 .write = snd_pcm_write, 4004 .write_iter = snd_pcm_writev, 4005 .open = snd_pcm_playback_open, 4006 .release = snd_pcm_release, 4007 .llseek = no_llseek, 4008 .poll = snd_pcm_poll, 4009 .unlocked_ioctl = snd_pcm_ioctl, 4010 .compat_ioctl = snd_pcm_ioctl_compat, 4011 .mmap = snd_pcm_mmap, 4012 .fasync = snd_pcm_fasync, 4013 .get_unmapped_area = snd_pcm_get_unmapped_area, 4014 }, 4015 { 4016 .owner = THIS_MODULE, 4017 .read = snd_pcm_read, 4018 .read_iter = snd_pcm_readv, 4019 .open = snd_pcm_capture_open, 4020 .release = snd_pcm_release, 4021 .llseek = no_llseek, 4022 .poll = snd_pcm_poll, 4023 .unlocked_ioctl = snd_pcm_ioctl, 4024 .compat_ioctl = snd_pcm_ioctl_compat, 4025 .mmap = snd_pcm_mmap, 4026 .fasync = snd_pcm_fasync, 4027 .get_unmapped_area = snd_pcm_get_unmapped_area, 4028 } 4029 }; 4030