1 /* 2 * Digital Audio (PCM) abstract layer 3 * Copyright (c) by Jaroslav Kysela <perex@perex.cz> 4 * 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation; either version 2 of the License, or 9 * (at your option) any later version. 10 * 11 * This program is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14 * GNU General Public License for more details. 15 * 16 * You should have received a copy of the GNU General Public License 17 * along with this program; if not, write to the Free Software 18 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA 19 * 20 */ 21 22 #include <linux/mm.h> 23 #include <linux/module.h> 24 #include <linux/file.h> 25 #include <linux/slab.h> 26 #include <linux/sched/signal.h> 27 #include <linux/time.h> 28 #include <linux/pm_qos.h> 29 #include <linux/io.h> 30 #include <linux/dma-mapping.h> 31 #include <sound/core.h> 32 #include <sound/control.h> 33 #include <sound/info.h> 34 #include <sound/pcm.h> 35 #include <sound/pcm_params.h> 36 #include <sound/timer.h> 37 #include <sound/minors.h> 38 #include <linux/uio.h> 39 40 #include "pcm_local.h" 41 42 #ifdef CONFIG_SND_DEBUG 43 #define CREATE_TRACE_POINTS 44 #include "pcm_param_trace.h" 45 #else 46 #define trace_hw_mask_param_enabled() 0 47 #define trace_hw_interval_param_enabled() 0 48 #define trace_hw_mask_param(substream, type, index, prev, curr) 49 #define trace_hw_interval_param(substream, type, index, prev, curr) 50 #endif 51 52 /* 53 * Compatibility 54 */ 55 56 struct snd_pcm_hw_params_old { 57 unsigned int flags; 58 unsigned int masks[SNDRV_PCM_HW_PARAM_SUBFORMAT - 59 SNDRV_PCM_HW_PARAM_ACCESS + 1]; 60 struct snd_interval intervals[SNDRV_PCM_HW_PARAM_TICK_TIME - 61 SNDRV_PCM_HW_PARAM_SAMPLE_BITS + 1]; 62 unsigned int rmask; 63 unsigned int cmask; 64 unsigned int info; 65 unsigned int msbits; 66 unsigned int rate_num; 67 unsigned int rate_den; 68 snd_pcm_uframes_t fifo_size; 69 unsigned char reserved[64]; 70 }; 71 72 #ifdef CONFIG_SND_SUPPORT_OLD_API 73 #define SNDRV_PCM_IOCTL_HW_REFINE_OLD _IOWR('A', 0x10, struct snd_pcm_hw_params_old) 74 #define SNDRV_PCM_IOCTL_HW_PARAMS_OLD _IOWR('A', 0x11, struct snd_pcm_hw_params_old) 75 76 static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream, 77 struct snd_pcm_hw_params_old __user * _oparams); 78 static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream, 79 struct snd_pcm_hw_params_old __user * _oparams); 80 #endif 81 static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream); 82 83 /* 84 * 85 */ 86 87 static DEFINE_RWLOCK(snd_pcm_link_rwlock); 88 static DECLARE_RWSEM(snd_pcm_link_rwsem); 89 90 /* Writer in rwsem may block readers even during its waiting in queue, 91 * and this may lead to a deadlock when the code path takes read sem 92 * twice (e.g. one in snd_pcm_action_nonatomic() and another in 93 * snd_pcm_stream_lock()). As a (suboptimal) workaround, let writer to 94 * spin until it gets the lock. 95 */ 96 static inline void down_write_nonblock(struct rw_semaphore *lock) 97 { 98 while (!down_write_trylock(lock)) 99 cond_resched(); 100 } 101 102 #define PCM_LOCK_DEFAULT 0 103 #define PCM_LOCK_IRQ 1 104 #define PCM_LOCK_IRQSAVE 2 105 106 static unsigned long __snd_pcm_stream_lock_mode(struct snd_pcm_substream *substream, 107 unsigned int mode) 108 { 109 unsigned long flags = 0; 110 if (substream->pcm->nonatomic) { 111 down_read_nested(&snd_pcm_link_rwsem, SINGLE_DEPTH_NESTING); 112 mutex_lock(&substream->self_group.mutex); 113 } else { 114 switch (mode) { 115 case PCM_LOCK_DEFAULT: 116 read_lock(&snd_pcm_link_rwlock); 117 break; 118 case PCM_LOCK_IRQ: 119 read_lock_irq(&snd_pcm_link_rwlock); 120 break; 121 case PCM_LOCK_IRQSAVE: 122 read_lock_irqsave(&snd_pcm_link_rwlock, flags); 123 break; 124 } 125 spin_lock(&substream->self_group.lock); 126 } 127 return flags; 128 } 129 130 static void __snd_pcm_stream_unlock_mode(struct snd_pcm_substream *substream, 131 unsigned int mode, unsigned long flags) 132 { 133 if (substream->pcm->nonatomic) { 134 mutex_unlock(&substream->self_group.mutex); 135 up_read(&snd_pcm_link_rwsem); 136 } else { 137 spin_unlock(&substream->self_group.lock); 138 139 switch (mode) { 140 case PCM_LOCK_DEFAULT: 141 read_unlock(&snd_pcm_link_rwlock); 142 break; 143 case PCM_LOCK_IRQ: 144 read_unlock_irq(&snd_pcm_link_rwlock); 145 break; 146 case PCM_LOCK_IRQSAVE: 147 read_unlock_irqrestore(&snd_pcm_link_rwlock, flags); 148 break; 149 } 150 } 151 } 152 153 /** 154 * snd_pcm_stream_lock - Lock the PCM stream 155 * @substream: PCM substream 156 * 157 * This locks the PCM stream's spinlock or mutex depending on the nonatomic 158 * flag of the given substream. This also takes the global link rw lock 159 * (or rw sem), too, for avoiding the race with linked streams. 160 */ 161 void snd_pcm_stream_lock(struct snd_pcm_substream *substream) 162 { 163 __snd_pcm_stream_lock_mode(substream, PCM_LOCK_DEFAULT); 164 } 165 EXPORT_SYMBOL_GPL(snd_pcm_stream_lock); 166 167 /** 168 * snd_pcm_stream_lock - Unlock the PCM stream 169 * @substream: PCM substream 170 * 171 * This unlocks the PCM stream that has been locked via snd_pcm_stream_lock(). 172 */ 173 void snd_pcm_stream_unlock(struct snd_pcm_substream *substream) 174 { 175 __snd_pcm_stream_unlock_mode(substream, PCM_LOCK_DEFAULT, 0); 176 } 177 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock); 178 179 /** 180 * snd_pcm_stream_lock_irq - Lock the PCM stream 181 * @substream: PCM substream 182 * 183 * This locks the PCM stream like snd_pcm_stream_lock() and disables the local 184 * IRQ (only when nonatomic is false). In nonatomic case, this is identical 185 * as snd_pcm_stream_lock(). 186 */ 187 void snd_pcm_stream_lock_irq(struct snd_pcm_substream *substream) 188 { 189 __snd_pcm_stream_lock_mode(substream, PCM_LOCK_IRQ); 190 } 191 EXPORT_SYMBOL_GPL(snd_pcm_stream_lock_irq); 192 193 /** 194 * snd_pcm_stream_unlock_irq - Unlock the PCM stream 195 * @substream: PCM substream 196 * 197 * This is a counter-part of snd_pcm_stream_lock_irq(). 198 */ 199 void snd_pcm_stream_unlock_irq(struct snd_pcm_substream *substream) 200 { 201 __snd_pcm_stream_unlock_mode(substream, PCM_LOCK_IRQ, 0); 202 } 203 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irq); 204 205 unsigned long _snd_pcm_stream_lock_irqsave(struct snd_pcm_substream *substream) 206 { 207 return __snd_pcm_stream_lock_mode(substream, PCM_LOCK_IRQSAVE); 208 } 209 EXPORT_SYMBOL_GPL(_snd_pcm_stream_lock_irqsave); 210 211 /** 212 * snd_pcm_stream_unlock_irqrestore - Unlock the PCM stream 213 * @substream: PCM substream 214 * @flags: irq flags 215 * 216 * This is a counter-part of snd_pcm_stream_lock_irqsave(). 217 */ 218 void snd_pcm_stream_unlock_irqrestore(struct snd_pcm_substream *substream, 219 unsigned long flags) 220 { 221 __snd_pcm_stream_unlock_mode(substream, PCM_LOCK_IRQSAVE, flags); 222 } 223 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irqrestore); 224 225 int snd_pcm_info(struct snd_pcm_substream *substream, struct snd_pcm_info *info) 226 { 227 struct snd_pcm *pcm = substream->pcm; 228 struct snd_pcm_str *pstr = substream->pstr; 229 230 memset(info, 0, sizeof(*info)); 231 info->card = pcm->card->number; 232 info->device = pcm->device; 233 info->stream = substream->stream; 234 info->subdevice = substream->number; 235 strlcpy(info->id, pcm->id, sizeof(info->id)); 236 strlcpy(info->name, pcm->name, sizeof(info->name)); 237 info->dev_class = pcm->dev_class; 238 info->dev_subclass = pcm->dev_subclass; 239 info->subdevices_count = pstr->substream_count; 240 info->subdevices_avail = pstr->substream_count - pstr->substream_opened; 241 strlcpy(info->subname, substream->name, sizeof(info->subname)); 242 243 return 0; 244 } 245 246 int snd_pcm_info_user(struct snd_pcm_substream *substream, 247 struct snd_pcm_info __user * _info) 248 { 249 struct snd_pcm_info *info; 250 int err; 251 252 info = kmalloc(sizeof(*info), GFP_KERNEL); 253 if (! info) 254 return -ENOMEM; 255 err = snd_pcm_info(substream, info); 256 if (err >= 0) { 257 if (copy_to_user(_info, info, sizeof(*info))) 258 err = -EFAULT; 259 } 260 kfree(info); 261 return err; 262 } 263 264 static bool hw_support_mmap(struct snd_pcm_substream *substream) 265 { 266 if (!(substream->runtime->hw.info & SNDRV_PCM_INFO_MMAP)) 267 return false; 268 /* architecture supports dma_mmap_coherent()? */ 269 #if defined(CONFIG_ARCH_NO_COHERENT_DMA_MMAP) || !defined(CONFIG_HAS_DMA) 270 if (!substream->ops->mmap && 271 substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV) 272 return false; 273 #endif 274 return true; 275 } 276 277 static int constrain_mask_params(struct snd_pcm_substream *substream, 278 struct snd_pcm_hw_params *params) 279 { 280 struct snd_pcm_hw_constraints *constrs = 281 &substream->runtime->hw_constraints; 282 struct snd_mask *m; 283 unsigned int k; 284 struct snd_mask old_mask; 285 int changed; 286 287 for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) { 288 m = hw_param_mask(params, k); 289 if (snd_mask_empty(m)) 290 return -EINVAL; 291 292 /* This parameter is not requested to change by a caller. */ 293 if (!(params->rmask & (1 << k))) 294 continue; 295 296 if (trace_hw_mask_param_enabled()) 297 old_mask = *m; 298 299 changed = snd_mask_refine(m, constrs_mask(constrs, k)); 300 if (changed < 0) 301 return changed; 302 if (changed == 0) 303 continue; 304 305 /* Set corresponding flag so that the caller gets it. */ 306 trace_hw_mask_param(substream, k, 0, &old_mask, m); 307 params->cmask |= 1 << k; 308 } 309 310 return 0; 311 } 312 313 static int constrain_interval_params(struct snd_pcm_substream *substream, 314 struct snd_pcm_hw_params *params) 315 { 316 struct snd_pcm_hw_constraints *constrs = 317 &substream->runtime->hw_constraints; 318 struct snd_interval *i; 319 unsigned int k; 320 struct snd_interval old_interval; 321 int changed; 322 323 for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) { 324 i = hw_param_interval(params, k); 325 if (snd_interval_empty(i)) 326 return -EINVAL; 327 328 /* This parameter is not requested to change by a caller. */ 329 if (!(params->rmask & (1 << k))) 330 continue; 331 332 if (trace_hw_interval_param_enabled()) 333 old_interval = *i; 334 335 changed = snd_interval_refine(i, constrs_interval(constrs, k)); 336 if (changed < 0) 337 return changed; 338 if (changed == 0) 339 continue; 340 341 /* Set corresponding flag so that the caller gets it. */ 342 trace_hw_interval_param(substream, k, 0, &old_interval, i); 343 params->cmask |= 1 << k; 344 } 345 346 return 0; 347 } 348 349 static int constrain_params_by_rules(struct snd_pcm_substream *substream, 350 struct snd_pcm_hw_params *params) 351 { 352 struct snd_pcm_hw_constraints *constrs = 353 &substream->runtime->hw_constraints; 354 unsigned int k; 355 unsigned int *rstamps; 356 unsigned int vstamps[SNDRV_PCM_HW_PARAM_LAST_INTERVAL + 1]; 357 unsigned int stamp; 358 struct snd_pcm_hw_rule *r; 359 unsigned int d; 360 struct snd_mask old_mask; 361 struct snd_interval old_interval; 362 bool again; 363 int changed, err = 0; 364 365 /* 366 * Each application of rule has own sequence number. 367 * 368 * Each member of 'rstamps' array represents the sequence number of 369 * recent application of corresponding rule. 370 */ 371 rstamps = kcalloc(constrs->rules_num, sizeof(unsigned int), GFP_KERNEL); 372 if (!rstamps) 373 return -ENOMEM; 374 375 /* 376 * Each member of 'vstamps' array represents the sequence number of 377 * recent application of rule in which corresponding parameters were 378 * changed. 379 * 380 * In initial state, elements corresponding to parameters requested by 381 * a caller is 1. For unrequested parameters, corresponding members 382 * have 0 so that the parameters are never changed anymore. 383 */ 384 for (k = 0; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) 385 vstamps[k] = (params->rmask & (1 << k)) ? 1 : 0; 386 387 /* Due to the above design, actual sequence number starts at 2. */ 388 stamp = 2; 389 retry: 390 /* Apply all rules in order. */ 391 again = false; 392 for (k = 0; k < constrs->rules_num; k++) { 393 r = &constrs->rules[k]; 394 395 /* 396 * Check condition bits of this rule. When the rule has 397 * some condition bits, parameter without the bits is 398 * never processed. SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP 399 * is an example of the condition bits. 400 */ 401 if (r->cond && !(r->cond & params->flags)) 402 continue; 403 404 /* 405 * The 'deps' array includes maximum three dependencies 406 * to SNDRV_PCM_HW_PARAM_XXXs for this rule. The fourth 407 * member of this array is a sentinel and should be 408 * negative value. 409 * 410 * This rule should be processed in this time when dependent 411 * parameters were changed at former applications of the other 412 * rules. 413 */ 414 for (d = 0; r->deps[d] >= 0; d++) { 415 if (vstamps[r->deps[d]] > rstamps[k]) 416 break; 417 } 418 if (r->deps[d] < 0) 419 continue; 420 421 if (trace_hw_mask_param_enabled()) { 422 if (hw_is_mask(r->var)) 423 old_mask = *hw_param_mask(params, r->var); 424 } 425 if (trace_hw_interval_param_enabled()) { 426 if (hw_is_interval(r->var)) 427 old_interval = *hw_param_interval(params, r->var); 428 } 429 430 changed = r->func(params, r); 431 if (changed < 0) { 432 err = changed; 433 goto out; 434 } 435 436 /* 437 * When the parameter is changed, notify it to the caller 438 * by corresponding returned bit, then preparing for next 439 * iteration. 440 */ 441 if (changed && r->var >= 0) { 442 if (hw_is_mask(r->var)) { 443 trace_hw_mask_param(substream, r->var, 444 k + 1, &old_mask, 445 hw_param_mask(params, r->var)); 446 } 447 if (hw_is_interval(r->var)) { 448 trace_hw_interval_param(substream, r->var, 449 k + 1, &old_interval, 450 hw_param_interval(params, r->var)); 451 } 452 453 params->cmask |= (1 << r->var); 454 vstamps[r->var] = stamp; 455 again = true; 456 } 457 458 rstamps[k] = stamp++; 459 } 460 461 /* Iterate to evaluate all rules till no parameters are changed. */ 462 if (again) 463 goto retry; 464 465 out: 466 kfree(rstamps); 467 return err; 468 } 469 470 static int fixup_unreferenced_params(struct snd_pcm_substream *substream, 471 struct snd_pcm_hw_params *params) 472 { 473 const struct snd_interval *i; 474 const struct snd_mask *m; 475 int err; 476 477 if (!params->msbits) { 478 i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_SAMPLE_BITS); 479 if (snd_interval_single(i)) 480 params->msbits = snd_interval_value(i); 481 } 482 483 if (!params->rate_den) { 484 i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_RATE); 485 if (snd_interval_single(i)) { 486 params->rate_num = snd_interval_value(i); 487 params->rate_den = 1; 488 } 489 } 490 491 if (!params->fifo_size) { 492 m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT); 493 i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_CHANNELS); 494 if (snd_mask_single(m) && snd_interval_single(i)) { 495 err = substream->ops->ioctl(substream, 496 SNDRV_PCM_IOCTL1_FIFO_SIZE, params); 497 if (err < 0) 498 return err; 499 } 500 } 501 502 if (!params->info) { 503 params->info = substream->runtime->hw.info; 504 params->info &= ~(SNDRV_PCM_INFO_FIFO_IN_FRAMES | 505 SNDRV_PCM_INFO_DRAIN_TRIGGER); 506 if (!hw_support_mmap(substream)) 507 params->info &= ~(SNDRV_PCM_INFO_MMAP | 508 SNDRV_PCM_INFO_MMAP_VALID); 509 } 510 511 return 0; 512 } 513 514 int snd_pcm_hw_refine(struct snd_pcm_substream *substream, 515 struct snd_pcm_hw_params *params) 516 { 517 int err; 518 519 params->info = 0; 520 params->fifo_size = 0; 521 if (params->rmask & (1 << SNDRV_PCM_HW_PARAM_SAMPLE_BITS)) 522 params->msbits = 0; 523 if (params->rmask & (1 << SNDRV_PCM_HW_PARAM_RATE)) { 524 params->rate_num = 0; 525 params->rate_den = 0; 526 } 527 528 err = constrain_mask_params(substream, params); 529 if (err < 0) 530 return err; 531 532 err = constrain_interval_params(substream, params); 533 if (err < 0) 534 return err; 535 536 err = constrain_params_by_rules(substream, params); 537 if (err < 0) 538 return err; 539 540 params->rmask = 0; 541 542 return 0; 543 } 544 EXPORT_SYMBOL(snd_pcm_hw_refine); 545 546 static int snd_pcm_hw_refine_user(struct snd_pcm_substream *substream, 547 struct snd_pcm_hw_params __user * _params) 548 { 549 struct snd_pcm_hw_params *params; 550 int err; 551 552 params = memdup_user(_params, sizeof(*params)); 553 if (IS_ERR(params)) 554 return PTR_ERR(params); 555 556 err = snd_pcm_hw_refine(substream, params); 557 if (err < 0) 558 goto end; 559 560 err = fixup_unreferenced_params(substream, params); 561 if (err < 0) 562 goto end; 563 564 if (copy_to_user(_params, params, sizeof(*params))) 565 err = -EFAULT; 566 end: 567 kfree(params); 568 return err; 569 } 570 571 static int period_to_usecs(struct snd_pcm_runtime *runtime) 572 { 573 int usecs; 574 575 if (! runtime->rate) 576 return -1; /* invalid */ 577 578 /* take 75% of period time as the deadline */ 579 usecs = (750000 / runtime->rate) * runtime->period_size; 580 usecs += ((750000 % runtime->rate) * runtime->period_size) / 581 runtime->rate; 582 583 return usecs; 584 } 585 586 static void snd_pcm_set_state(struct snd_pcm_substream *substream, int state) 587 { 588 snd_pcm_stream_lock_irq(substream); 589 if (substream->runtime->status->state != SNDRV_PCM_STATE_DISCONNECTED) 590 substream->runtime->status->state = state; 591 snd_pcm_stream_unlock_irq(substream); 592 } 593 594 static inline void snd_pcm_timer_notify(struct snd_pcm_substream *substream, 595 int event) 596 { 597 #ifdef CONFIG_SND_PCM_TIMER 598 if (substream->timer) 599 snd_timer_notify(substream->timer, event, 600 &substream->runtime->trigger_tstamp); 601 #endif 602 } 603 604 /** 605 * snd_pcm_hw_param_choose - choose a configuration defined by @params 606 * @pcm: PCM instance 607 * @params: the hw_params instance 608 * 609 * Choose one configuration from configuration space defined by @params. 610 * The configuration chosen is that obtained fixing in this order: 611 * first access, first format, first subformat, min channels, 612 * min rate, min period time, max buffer size, min tick time 613 * 614 * Return: Zero if successful, or a negative error code on failure. 615 */ 616 static int snd_pcm_hw_params_choose(struct snd_pcm_substream *pcm, 617 struct snd_pcm_hw_params *params) 618 { 619 static const int vars[] = { 620 SNDRV_PCM_HW_PARAM_ACCESS, 621 SNDRV_PCM_HW_PARAM_FORMAT, 622 SNDRV_PCM_HW_PARAM_SUBFORMAT, 623 SNDRV_PCM_HW_PARAM_CHANNELS, 624 SNDRV_PCM_HW_PARAM_RATE, 625 SNDRV_PCM_HW_PARAM_PERIOD_TIME, 626 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 627 SNDRV_PCM_HW_PARAM_TICK_TIME, 628 -1 629 }; 630 const int *v; 631 struct snd_mask old_mask; 632 struct snd_interval old_interval; 633 int changed; 634 635 for (v = vars; *v != -1; v++) { 636 /* Keep old parameter to trace. */ 637 if (trace_hw_mask_param_enabled()) { 638 if (hw_is_mask(*v)) 639 old_mask = *hw_param_mask(params, *v); 640 } 641 if (trace_hw_interval_param_enabled()) { 642 if (hw_is_interval(*v)) 643 old_interval = *hw_param_interval(params, *v); 644 } 645 if (*v != SNDRV_PCM_HW_PARAM_BUFFER_SIZE) 646 changed = snd_pcm_hw_param_first(pcm, params, *v, NULL); 647 else 648 changed = snd_pcm_hw_param_last(pcm, params, *v, NULL); 649 if (changed < 0) 650 return changed; 651 if (changed == 0) 652 continue; 653 654 /* Trace the changed parameter. */ 655 if (hw_is_mask(*v)) { 656 trace_hw_mask_param(pcm, *v, 0, &old_mask, 657 hw_param_mask(params, *v)); 658 } 659 if (hw_is_interval(*v)) { 660 trace_hw_interval_param(pcm, *v, 0, &old_interval, 661 hw_param_interval(params, *v)); 662 } 663 } 664 665 return 0; 666 } 667 668 static int snd_pcm_hw_params(struct snd_pcm_substream *substream, 669 struct snd_pcm_hw_params *params) 670 { 671 struct snd_pcm_runtime *runtime; 672 int err, usecs; 673 unsigned int bits; 674 snd_pcm_uframes_t frames; 675 676 if (PCM_RUNTIME_CHECK(substream)) 677 return -ENXIO; 678 runtime = substream->runtime; 679 snd_pcm_stream_lock_irq(substream); 680 switch (runtime->status->state) { 681 case SNDRV_PCM_STATE_OPEN: 682 case SNDRV_PCM_STATE_SETUP: 683 case SNDRV_PCM_STATE_PREPARED: 684 break; 685 default: 686 snd_pcm_stream_unlock_irq(substream); 687 return -EBADFD; 688 } 689 snd_pcm_stream_unlock_irq(substream); 690 #if IS_ENABLED(CONFIG_SND_PCM_OSS) 691 if (!substream->oss.oss) 692 #endif 693 if (atomic_read(&substream->mmap_count)) 694 return -EBADFD; 695 696 params->rmask = ~0U; 697 err = snd_pcm_hw_refine(substream, params); 698 if (err < 0) 699 goto _error; 700 701 err = snd_pcm_hw_params_choose(substream, params); 702 if (err < 0) 703 goto _error; 704 705 err = fixup_unreferenced_params(substream, params); 706 if (err < 0) 707 goto _error; 708 709 if (substream->ops->hw_params != NULL) { 710 err = substream->ops->hw_params(substream, params); 711 if (err < 0) 712 goto _error; 713 } 714 715 runtime->access = params_access(params); 716 runtime->format = params_format(params); 717 runtime->subformat = params_subformat(params); 718 runtime->channels = params_channels(params); 719 runtime->rate = params_rate(params); 720 runtime->period_size = params_period_size(params); 721 runtime->periods = params_periods(params); 722 runtime->buffer_size = params_buffer_size(params); 723 runtime->info = params->info; 724 runtime->rate_num = params->rate_num; 725 runtime->rate_den = params->rate_den; 726 runtime->no_period_wakeup = 727 (params->info & SNDRV_PCM_INFO_NO_PERIOD_WAKEUP) && 728 (params->flags & SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP); 729 730 bits = snd_pcm_format_physical_width(runtime->format); 731 runtime->sample_bits = bits; 732 bits *= runtime->channels; 733 runtime->frame_bits = bits; 734 frames = 1; 735 while (bits % 8 != 0) { 736 bits *= 2; 737 frames *= 2; 738 } 739 runtime->byte_align = bits / 8; 740 runtime->min_align = frames; 741 742 /* Default sw params */ 743 runtime->tstamp_mode = SNDRV_PCM_TSTAMP_NONE; 744 runtime->period_step = 1; 745 runtime->control->avail_min = runtime->period_size; 746 runtime->start_threshold = 1; 747 runtime->stop_threshold = runtime->buffer_size; 748 runtime->silence_threshold = 0; 749 runtime->silence_size = 0; 750 runtime->boundary = runtime->buffer_size; 751 while (runtime->boundary * 2 <= LONG_MAX - runtime->buffer_size) 752 runtime->boundary *= 2; 753 754 snd_pcm_timer_resolution_change(substream); 755 snd_pcm_set_state(substream, SNDRV_PCM_STATE_SETUP); 756 757 if (pm_qos_request_active(&substream->latency_pm_qos_req)) 758 pm_qos_remove_request(&substream->latency_pm_qos_req); 759 if ((usecs = period_to_usecs(runtime)) >= 0) 760 pm_qos_add_request(&substream->latency_pm_qos_req, 761 PM_QOS_CPU_DMA_LATENCY, usecs); 762 return 0; 763 _error: 764 /* hardware might be unusable from this time, 765 so we force application to retry to set 766 the correct hardware parameter settings */ 767 snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN); 768 if (substream->ops->hw_free != NULL) 769 substream->ops->hw_free(substream); 770 return err; 771 } 772 773 static int snd_pcm_hw_params_user(struct snd_pcm_substream *substream, 774 struct snd_pcm_hw_params __user * _params) 775 { 776 struct snd_pcm_hw_params *params; 777 int err; 778 779 params = memdup_user(_params, sizeof(*params)); 780 if (IS_ERR(params)) 781 return PTR_ERR(params); 782 783 err = snd_pcm_hw_params(substream, params); 784 if (err < 0) 785 goto end; 786 787 if (copy_to_user(_params, params, sizeof(*params))) 788 err = -EFAULT; 789 end: 790 kfree(params); 791 return err; 792 } 793 794 static int snd_pcm_hw_free(struct snd_pcm_substream *substream) 795 { 796 struct snd_pcm_runtime *runtime; 797 int result = 0; 798 799 if (PCM_RUNTIME_CHECK(substream)) 800 return -ENXIO; 801 runtime = substream->runtime; 802 snd_pcm_stream_lock_irq(substream); 803 switch (runtime->status->state) { 804 case SNDRV_PCM_STATE_SETUP: 805 case SNDRV_PCM_STATE_PREPARED: 806 break; 807 default: 808 snd_pcm_stream_unlock_irq(substream); 809 return -EBADFD; 810 } 811 snd_pcm_stream_unlock_irq(substream); 812 if (atomic_read(&substream->mmap_count)) 813 return -EBADFD; 814 if (substream->ops->hw_free) 815 result = substream->ops->hw_free(substream); 816 snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN); 817 pm_qos_remove_request(&substream->latency_pm_qos_req); 818 return result; 819 } 820 821 static int snd_pcm_sw_params(struct snd_pcm_substream *substream, 822 struct snd_pcm_sw_params *params) 823 { 824 struct snd_pcm_runtime *runtime; 825 int err; 826 827 if (PCM_RUNTIME_CHECK(substream)) 828 return -ENXIO; 829 runtime = substream->runtime; 830 snd_pcm_stream_lock_irq(substream); 831 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) { 832 snd_pcm_stream_unlock_irq(substream); 833 return -EBADFD; 834 } 835 snd_pcm_stream_unlock_irq(substream); 836 837 if (params->tstamp_mode < 0 || 838 params->tstamp_mode > SNDRV_PCM_TSTAMP_LAST) 839 return -EINVAL; 840 if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12) && 841 params->tstamp_type > SNDRV_PCM_TSTAMP_TYPE_LAST) 842 return -EINVAL; 843 if (params->avail_min == 0) 844 return -EINVAL; 845 if (params->silence_size >= runtime->boundary) { 846 if (params->silence_threshold != 0) 847 return -EINVAL; 848 } else { 849 if (params->silence_size > params->silence_threshold) 850 return -EINVAL; 851 if (params->silence_threshold > runtime->buffer_size) 852 return -EINVAL; 853 } 854 err = 0; 855 snd_pcm_stream_lock_irq(substream); 856 runtime->tstamp_mode = params->tstamp_mode; 857 if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12)) 858 runtime->tstamp_type = params->tstamp_type; 859 runtime->period_step = params->period_step; 860 runtime->control->avail_min = params->avail_min; 861 runtime->start_threshold = params->start_threshold; 862 runtime->stop_threshold = params->stop_threshold; 863 runtime->silence_threshold = params->silence_threshold; 864 runtime->silence_size = params->silence_size; 865 params->boundary = runtime->boundary; 866 if (snd_pcm_running(substream)) { 867 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 868 runtime->silence_size > 0) 869 snd_pcm_playback_silence(substream, ULONG_MAX); 870 err = snd_pcm_update_state(substream, runtime); 871 } 872 snd_pcm_stream_unlock_irq(substream); 873 return err; 874 } 875 876 static int snd_pcm_sw_params_user(struct snd_pcm_substream *substream, 877 struct snd_pcm_sw_params __user * _params) 878 { 879 struct snd_pcm_sw_params params; 880 int err; 881 if (copy_from_user(¶ms, _params, sizeof(params))) 882 return -EFAULT; 883 err = snd_pcm_sw_params(substream, ¶ms); 884 if (copy_to_user(_params, ¶ms, sizeof(params))) 885 return -EFAULT; 886 return err; 887 } 888 889 static inline snd_pcm_uframes_t 890 snd_pcm_calc_delay(struct snd_pcm_substream *substream) 891 { 892 snd_pcm_uframes_t delay; 893 894 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 895 delay = snd_pcm_playback_hw_avail(substream->runtime); 896 else 897 delay = snd_pcm_capture_avail(substream->runtime); 898 return delay + substream->runtime->delay; 899 } 900 901 int snd_pcm_status(struct snd_pcm_substream *substream, 902 struct snd_pcm_status *status) 903 { 904 struct snd_pcm_runtime *runtime = substream->runtime; 905 906 snd_pcm_stream_lock_irq(substream); 907 908 snd_pcm_unpack_audio_tstamp_config(status->audio_tstamp_data, 909 &runtime->audio_tstamp_config); 910 911 /* backwards compatible behavior */ 912 if (runtime->audio_tstamp_config.type_requested == 913 SNDRV_PCM_AUDIO_TSTAMP_TYPE_COMPAT) { 914 if (runtime->hw.info & SNDRV_PCM_INFO_HAS_WALL_CLOCK) 915 runtime->audio_tstamp_config.type_requested = 916 SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK; 917 else 918 runtime->audio_tstamp_config.type_requested = 919 SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT; 920 runtime->audio_tstamp_report.valid = 0; 921 } else 922 runtime->audio_tstamp_report.valid = 1; 923 924 status->state = runtime->status->state; 925 status->suspended_state = runtime->status->suspended_state; 926 if (status->state == SNDRV_PCM_STATE_OPEN) 927 goto _end; 928 status->trigger_tstamp = runtime->trigger_tstamp; 929 if (snd_pcm_running(substream)) { 930 snd_pcm_update_hw_ptr(substream); 931 if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) { 932 status->tstamp = runtime->status->tstamp; 933 status->driver_tstamp = runtime->driver_tstamp; 934 status->audio_tstamp = 935 runtime->status->audio_tstamp; 936 if (runtime->audio_tstamp_report.valid == 1) 937 /* backwards compatibility, no report provided in COMPAT mode */ 938 snd_pcm_pack_audio_tstamp_report(&status->audio_tstamp_data, 939 &status->audio_tstamp_accuracy, 940 &runtime->audio_tstamp_report); 941 942 goto _tstamp_end; 943 } 944 } else { 945 /* get tstamp only in fallback mode and only if enabled */ 946 if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) 947 snd_pcm_gettime(runtime, &status->tstamp); 948 } 949 _tstamp_end: 950 status->appl_ptr = runtime->control->appl_ptr; 951 status->hw_ptr = runtime->status->hw_ptr; 952 status->avail = snd_pcm_avail(substream); 953 status->delay = snd_pcm_running(substream) ? 954 snd_pcm_calc_delay(substream) : 0; 955 status->avail_max = runtime->avail_max; 956 status->overrange = runtime->overrange; 957 runtime->avail_max = 0; 958 runtime->overrange = 0; 959 _end: 960 snd_pcm_stream_unlock_irq(substream); 961 return 0; 962 } 963 964 static int snd_pcm_status_user(struct snd_pcm_substream *substream, 965 struct snd_pcm_status __user * _status, 966 bool ext) 967 { 968 struct snd_pcm_status status; 969 int res; 970 971 memset(&status, 0, sizeof(status)); 972 /* 973 * with extension, parameters are read/write, 974 * get audio_tstamp_data from user, 975 * ignore rest of status structure 976 */ 977 if (ext && get_user(status.audio_tstamp_data, 978 (u32 __user *)(&_status->audio_tstamp_data))) 979 return -EFAULT; 980 res = snd_pcm_status(substream, &status); 981 if (res < 0) 982 return res; 983 if (copy_to_user(_status, &status, sizeof(status))) 984 return -EFAULT; 985 return 0; 986 } 987 988 static int snd_pcm_channel_info(struct snd_pcm_substream *substream, 989 struct snd_pcm_channel_info * info) 990 { 991 struct snd_pcm_runtime *runtime; 992 unsigned int channel; 993 994 channel = info->channel; 995 runtime = substream->runtime; 996 snd_pcm_stream_lock_irq(substream); 997 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) { 998 snd_pcm_stream_unlock_irq(substream); 999 return -EBADFD; 1000 } 1001 snd_pcm_stream_unlock_irq(substream); 1002 if (channel >= runtime->channels) 1003 return -EINVAL; 1004 memset(info, 0, sizeof(*info)); 1005 info->channel = channel; 1006 return substream->ops->ioctl(substream, SNDRV_PCM_IOCTL1_CHANNEL_INFO, info); 1007 } 1008 1009 static int snd_pcm_channel_info_user(struct snd_pcm_substream *substream, 1010 struct snd_pcm_channel_info __user * _info) 1011 { 1012 struct snd_pcm_channel_info info; 1013 int res; 1014 1015 if (copy_from_user(&info, _info, sizeof(info))) 1016 return -EFAULT; 1017 res = snd_pcm_channel_info(substream, &info); 1018 if (res < 0) 1019 return res; 1020 if (copy_to_user(_info, &info, sizeof(info))) 1021 return -EFAULT; 1022 return 0; 1023 } 1024 1025 static void snd_pcm_trigger_tstamp(struct snd_pcm_substream *substream) 1026 { 1027 struct snd_pcm_runtime *runtime = substream->runtime; 1028 if (runtime->trigger_master == NULL) 1029 return; 1030 if (runtime->trigger_master == substream) { 1031 if (!runtime->trigger_tstamp_latched) 1032 snd_pcm_gettime(runtime, &runtime->trigger_tstamp); 1033 } else { 1034 snd_pcm_trigger_tstamp(runtime->trigger_master); 1035 runtime->trigger_tstamp = runtime->trigger_master->runtime->trigger_tstamp; 1036 } 1037 runtime->trigger_master = NULL; 1038 } 1039 1040 struct action_ops { 1041 int (*pre_action)(struct snd_pcm_substream *substream, int state); 1042 int (*do_action)(struct snd_pcm_substream *substream, int state); 1043 void (*undo_action)(struct snd_pcm_substream *substream, int state); 1044 void (*post_action)(struct snd_pcm_substream *substream, int state); 1045 }; 1046 1047 /* 1048 * this functions is core for handling of linked stream 1049 * Note: the stream state might be changed also on failure 1050 * Note2: call with calling stream lock + link lock 1051 */ 1052 static int snd_pcm_action_group(const struct action_ops *ops, 1053 struct snd_pcm_substream *substream, 1054 int state, int do_lock) 1055 { 1056 struct snd_pcm_substream *s = NULL; 1057 struct snd_pcm_substream *s1; 1058 int res = 0, depth = 1; 1059 1060 snd_pcm_group_for_each_entry(s, substream) { 1061 if (do_lock && s != substream) { 1062 if (s->pcm->nonatomic) 1063 mutex_lock_nested(&s->self_group.mutex, depth); 1064 else 1065 spin_lock_nested(&s->self_group.lock, depth); 1066 depth++; 1067 } 1068 res = ops->pre_action(s, state); 1069 if (res < 0) 1070 goto _unlock; 1071 } 1072 snd_pcm_group_for_each_entry(s, substream) { 1073 res = ops->do_action(s, state); 1074 if (res < 0) { 1075 if (ops->undo_action) { 1076 snd_pcm_group_for_each_entry(s1, substream) { 1077 if (s1 == s) /* failed stream */ 1078 break; 1079 ops->undo_action(s1, state); 1080 } 1081 } 1082 s = NULL; /* unlock all */ 1083 goto _unlock; 1084 } 1085 } 1086 snd_pcm_group_for_each_entry(s, substream) { 1087 ops->post_action(s, state); 1088 } 1089 _unlock: 1090 if (do_lock) { 1091 /* unlock streams */ 1092 snd_pcm_group_for_each_entry(s1, substream) { 1093 if (s1 != substream) { 1094 if (s1->pcm->nonatomic) 1095 mutex_unlock(&s1->self_group.mutex); 1096 else 1097 spin_unlock(&s1->self_group.lock); 1098 } 1099 if (s1 == s) /* end */ 1100 break; 1101 } 1102 } 1103 return res; 1104 } 1105 1106 /* 1107 * Note: call with stream lock 1108 */ 1109 static int snd_pcm_action_single(const struct action_ops *ops, 1110 struct snd_pcm_substream *substream, 1111 int state) 1112 { 1113 int res; 1114 1115 res = ops->pre_action(substream, state); 1116 if (res < 0) 1117 return res; 1118 res = ops->do_action(substream, state); 1119 if (res == 0) 1120 ops->post_action(substream, state); 1121 else if (ops->undo_action) 1122 ops->undo_action(substream, state); 1123 return res; 1124 } 1125 1126 /* 1127 * Note: call with stream lock 1128 */ 1129 static int snd_pcm_action(const struct action_ops *ops, 1130 struct snd_pcm_substream *substream, 1131 int state) 1132 { 1133 int res; 1134 1135 if (!snd_pcm_stream_linked(substream)) 1136 return snd_pcm_action_single(ops, substream, state); 1137 1138 if (substream->pcm->nonatomic) { 1139 if (!mutex_trylock(&substream->group->mutex)) { 1140 mutex_unlock(&substream->self_group.mutex); 1141 mutex_lock(&substream->group->mutex); 1142 mutex_lock(&substream->self_group.mutex); 1143 } 1144 res = snd_pcm_action_group(ops, substream, state, 1); 1145 mutex_unlock(&substream->group->mutex); 1146 } else { 1147 if (!spin_trylock(&substream->group->lock)) { 1148 spin_unlock(&substream->self_group.lock); 1149 spin_lock(&substream->group->lock); 1150 spin_lock(&substream->self_group.lock); 1151 } 1152 res = snd_pcm_action_group(ops, substream, state, 1); 1153 spin_unlock(&substream->group->lock); 1154 } 1155 return res; 1156 } 1157 1158 /* 1159 * Note: don't use any locks before 1160 */ 1161 static int snd_pcm_action_lock_irq(const struct action_ops *ops, 1162 struct snd_pcm_substream *substream, 1163 int state) 1164 { 1165 int res; 1166 1167 snd_pcm_stream_lock_irq(substream); 1168 res = snd_pcm_action(ops, substream, state); 1169 snd_pcm_stream_unlock_irq(substream); 1170 return res; 1171 } 1172 1173 /* 1174 */ 1175 static int snd_pcm_action_nonatomic(const struct action_ops *ops, 1176 struct snd_pcm_substream *substream, 1177 int state) 1178 { 1179 int res; 1180 1181 down_read(&snd_pcm_link_rwsem); 1182 if (snd_pcm_stream_linked(substream)) 1183 res = snd_pcm_action_group(ops, substream, state, 0); 1184 else 1185 res = snd_pcm_action_single(ops, substream, state); 1186 up_read(&snd_pcm_link_rwsem); 1187 return res; 1188 } 1189 1190 /* 1191 * start callbacks 1192 */ 1193 static int snd_pcm_pre_start(struct snd_pcm_substream *substream, int state) 1194 { 1195 struct snd_pcm_runtime *runtime = substream->runtime; 1196 if (runtime->status->state != SNDRV_PCM_STATE_PREPARED) 1197 return -EBADFD; 1198 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 1199 !snd_pcm_playback_data(substream)) 1200 return -EPIPE; 1201 runtime->trigger_tstamp_latched = false; 1202 runtime->trigger_master = substream; 1203 return 0; 1204 } 1205 1206 static int snd_pcm_do_start(struct snd_pcm_substream *substream, int state) 1207 { 1208 if (substream->runtime->trigger_master != substream) 1209 return 0; 1210 return substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_START); 1211 } 1212 1213 static void snd_pcm_undo_start(struct snd_pcm_substream *substream, int state) 1214 { 1215 if (substream->runtime->trigger_master == substream) 1216 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP); 1217 } 1218 1219 static void snd_pcm_post_start(struct snd_pcm_substream *substream, int state) 1220 { 1221 struct snd_pcm_runtime *runtime = substream->runtime; 1222 snd_pcm_trigger_tstamp(substream); 1223 runtime->hw_ptr_jiffies = jiffies; 1224 runtime->hw_ptr_buffer_jiffies = (runtime->buffer_size * HZ) / 1225 runtime->rate; 1226 runtime->status->state = state; 1227 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 1228 runtime->silence_size > 0) 1229 snd_pcm_playback_silence(substream, ULONG_MAX); 1230 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSTART); 1231 } 1232 1233 static const struct action_ops snd_pcm_action_start = { 1234 .pre_action = snd_pcm_pre_start, 1235 .do_action = snd_pcm_do_start, 1236 .undo_action = snd_pcm_undo_start, 1237 .post_action = snd_pcm_post_start 1238 }; 1239 1240 /** 1241 * snd_pcm_start - start all linked streams 1242 * @substream: the PCM substream instance 1243 * 1244 * Return: Zero if successful, or a negative error code. 1245 * The stream lock must be acquired before calling this function. 1246 */ 1247 int snd_pcm_start(struct snd_pcm_substream *substream) 1248 { 1249 return snd_pcm_action(&snd_pcm_action_start, substream, 1250 SNDRV_PCM_STATE_RUNNING); 1251 } 1252 1253 /* take the stream lock and start the streams */ 1254 static int snd_pcm_start_lock_irq(struct snd_pcm_substream *substream) 1255 { 1256 return snd_pcm_action_lock_irq(&snd_pcm_action_start, substream, 1257 SNDRV_PCM_STATE_RUNNING); 1258 } 1259 1260 /* 1261 * stop callbacks 1262 */ 1263 static int snd_pcm_pre_stop(struct snd_pcm_substream *substream, int state) 1264 { 1265 struct snd_pcm_runtime *runtime = substream->runtime; 1266 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 1267 return -EBADFD; 1268 runtime->trigger_master = substream; 1269 return 0; 1270 } 1271 1272 static int snd_pcm_do_stop(struct snd_pcm_substream *substream, int state) 1273 { 1274 if (substream->runtime->trigger_master == substream && 1275 snd_pcm_running(substream)) 1276 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP); 1277 return 0; /* unconditonally stop all substreams */ 1278 } 1279 1280 static void snd_pcm_post_stop(struct snd_pcm_substream *substream, int state) 1281 { 1282 struct snd_pcm_runtime *runtime = substream->runtime; 1283 if (runtime->status->state != state) { 1284 snd_pcm_trigger_tstamp(substream); 1285 runtime->status->state = state; 1286 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSTOP); 1287 } 1288 wake_up(&runtime->sleep); 1289 wake_up(&runtime->tsleep); 1290 } 1291 1292 static const struct action_ops snd_pcm_action_stop = { 1293 .pre_action = snd_pcm_pre_stop, 1294 .do_action = snd_pcm_do_stop, 1295 .post_action = snd_pcm_post_stop 1296 }; 1297 1298 /** 1299 * snd_pcm_stop - try to stop all running streams in the substream group 1300 * @substream: the PCM substream instance 1301 * @state: PCM state after stopping the stream 1302 * 1303 * The state of each stream is then changed to the given state unconditionally. 1304 * 1305 * Return: Zero if successful, or a negative error code. 1306 */ 1307 int snd_pcm_stop(struct snd_pcm_substream *substream, snd_pcm_state_t state) 1308 { 1309 return snd_pcm_action(&snd_pcm_action_stop, substream, state); 1310 } 1311 EXPORT_SYMBOL(snd_pcm_stop); 1312 1313 /** 1314 * snd_pcm_drain_done - stop the DMA only when the given stream is playback 1315 * @substream: the PCM substream 1316 * 1317 * After stopping, the state is changed to SETUP. 1318 * Unlike snd_pcm_stop(), this affects only the given stream. 1319 * 1320 * Return: Zero if succesful, or a negative error code. 1321 */ 1322 int snd_pcm_drain_done(struct snd_pcm_substream *substream) 1323 { 1324 return snd_pcm_action_single(&snd_pcm_action_stop, substream, 1325 SNDRV_PCM_STATE_SETUP); 1326 } 1327 1328 /** 1329 * snd_pcm_stop_xrun - stop the running streams as XRUN 1330 * @substream: the PCM substream instance 1331 * 1332 * This stops the given running substream (and all linked substreams) as XRUN. 1333 * Unlike snd_pcm_stop(), this function takes the substream lock by itself. 1334 * 1335 * Return: Zero if successful, or a negative error code. 1336 */ 1337 int snd_pcm_stop_xrun(struct snd_pcm_substream *substream) 1338 { 1339 unsigned long flags; 1340 1341 snd_pcm_stream_lock_irqsave(substream, flags); 1342 if (substream->runtime && snd_pcm_running(substream)) 1343 __snd_pcm_xrun(substream); 1344 snd_pcm_stream_unlock_irqrestore(substream, flags); 1345 return 0; 1346 } 1347 EXPORT_SYMBOL_GPL(snd_pcm_stop_xrun); 1348 1349 /* 1350 * pause callbacks 1351 */ 1352 static int snd_pcm_pre_pause(struct snd_pcm_substream *substream, int push) 1353 { 1354 struct snd_pcm_runtime *runtime = substream->runtime; 1355 if (!(runtime->info & SNDRV_PCM_INFO_PAUSE)) 1356 return -ENOSYS; 1357 if (push) { 1358 if (runtime->status->state != SNDRV_PCM_STATE_RUNNING) 1359 return -EBADFD; 1360 } else if (runtime->status->state != SNDRV_PCM_STATE_PAUSED) 1361 return -EBADFD; 1362 runtime->trigger_master = substream; 1363 return 0; 1364 } 1365 1366 static int snd_pcm_do_pause(struct snd_pcm_substream *substream, int push) 1367 { 1368 if (substream->runtime->trigger_master != substream) 1369 return 0; 1370 /* some drivers might use hw_ptr to recover from the pause - 1371 update the hw_ptr now */ 1372 if (push) 1373 snd_pcm_update_hw_ptr(substream); 1374 /* The jiffies check in snd_pcm_update_hw_ptr*() is done by 1375 * a delta between the current jiffies, this gives a large enough 1376 * delta, effectively to skip the check once. 1377 */ 1378 substream->runtime->hw_ptr_jiffies = jiffies - HZ * 1000; 1379 return substream->ops->trigger(substream, 1380 push ? SNDRV_PCM_TRIGGER_PAUSE_PUSH : 1381 SNDRV_PCM_TRIGGER_PAUSE_RELEASE); 1382 } 1383 1384 static void snd_pcm_undo_pause(struct snd_pcm_substream *substream, int push) 1385 { 1386 if (substream->runtime->trigger_master == substream) 1387 substream->ops->trigger(substream, 1388 push ? SNDRV_PCM_TRIGGER_PAUSE_RELEASE : 1389 SNDRV_PCM_TRIGGER_PAUSE_PUSH); 1390 } 1391 1392 static void snd_pcm_post_pause(struct snd_pcm_substream *substream, int push) 1393 { 1394 struct snd_pcm_runtime *runtime = substream->runtime; 1395 snd_pcm_trigger_tstamp(substream); 1396 if (push) { 1397 runtime->status->state = SNDRV_PCM_STATE_PAUSED; 1398 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MPAUSE); 1399 wake_up(&runtime->sleep); 1400 wake_up(&runtime->tsleep); 1401 } else { 1402 runtime->status->state = SNDRV_PCM_STATE_RUNNING; 1403 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MCONTINUE); 1404 } 1405 } 1406 1407 static const struct action_ops snd_pcm_action_pause = { 1408 .pre_action = snd_pcm_pre_pause, 1409 .do_action = snd_pcm_do_pause, 1410 .undo_action = snd_pcm_undo_pause, 1411 .post_action = snd_pcm_post_pause 1412 }; 1413 1414 /* 1415 * Push/release the pause for all linked streams. 1416 */ 1417 static int snd_pcm_pause(struct snd_pcm_substream *substream, int push) 1418 { 1419 return snd_pcm_action(&snd_pcm_action_pause, substream, push); 1420 } 1421 1422 #ifdef CONFIG_PM 1423 /* suspend */ 1424 1425 static int snd_pcm_pre_suspend(struct snd_pcm_substream *substream, int state) 1426 { 1427 struct snd_pcm_runtime *runtime = substream->runtime; 1428 if (runtime->status->state == SNDRV_PCM_STATE_SUSPENDED) 1429 return -EBUSY; 1430 runtime->trigger_master = substream; 1431 return 0; 1432 } 1433 1434 static int snd_pcm_do_suspend(struct snd_pcm_substream *substream, int state) 1435 { 1436 struct snd_pcm_runtime *runtime = substream->runtime; 1437 if (runtime->trigger_master != substream) 1438 return 0; 1439 if (! snd_pcm_running(substream)) 1440 return 0; 1441 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND); 1442 return 0; /* suspend unconditionally */ 1443 } 1444 1445 static void snd_pcm_post_suspend(struct snd_pcm_substream *substream, int state) 1446 { 1447 struct snd_pcm_runtime *runtime = substream->runtime; 1448 snd_pcm_trigger_tstamp(substream); 1449 runtime->status->suspended_state = runtime->status->state; 1450 runtime->status->state = SNDRV_PCM_STATE_SUSPENDED; 1451 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSUSPEND); 1452 wake_up(&runtime->sleep); 1453 wake_up(&runtime->tsleep); 1454 } 1455 1456 static const struct action_ops snd_pcm_action_suspend = { 1457 .pre_action = snd_pcm_pre_suspend, 1458 .do_action = snd_pcm_do_suspend, 1459 .post_action = snd_pcm_post_suspend 1460 }; 1461 1462 /** 1463 * snd_pcm_suspend - trigger SUSPEND to all linked streams 1464 * @substream: the PCM substream 1465 * 1466 * After this call, all streams are changed to SUSPENDED state. 1467 * 1468 * Return: Zero if successful (or @substream is %NULL), or a negative error 1469 * code. 1470 */ 1471 int snd_pcm_suspend(struct snd_pcm_substream *substream) 1472 { 1473 int err; 1474 unsigned long flags; 1475 1476 if (! substream) 1477 return 0; 1478 1479 snd_pcm_stream_lock_irqsave(substream, flags); 1480 err = snd_pcm_action(&snd_pcm_action_suspend, substream, 0); 1481 snd_pcm_stream_unlock_irqrestore(substream, flags); 1482 return err; 1483 } 1484 EXPORT_SYMBOL(snd_pcm_suspend); 1485 1486 /** 1487 * snd_pcm_suspend_all - trigger SUSPEND to all substreams in the given pcm 1488 * @pcm: the PCM instance 1489 * 1490 * After this call, all streams are changed to SUSPENDED state. 1491 * 1492 * Return: Zero if successful (or @pcm is %NULL), or a negative error code. 1493 */ 1494 int snd_pcm_suspend_all(struct snd_pcm *pcm) 1495 { 1496 struct snd_pcm_substream *substream; 1497 int stream, err = 0; 1498 1499 if (! pcm) 1500 return 0; 1501 1502 for (stream = 0; stream < 2; stream++) { 1503 for (substream = pcm->streams[stream].substream; 1504 substream; substream = substream->next) { 1505 /* FIXME: the open/close code should lock this as well */ 1506 if (substream->runtime == NULL) 1507 continue; 1508 err = snd_pcm_suspend(substream); 1509 if (err < 0 && err != -EBUSY) 1510 return err; 1511 } 1512 } 1513 return 0; 1514 } 1515 EXPORT_SYMBOL(snd_pcm_suspend_all); 1516 1517 /* resume */ 1518 1519 static int snd_pcm_pre_resume(struct snd_pcm_substream *substream, int state) 1520 { 1521 struct snd_pcm_runtime *runtime = substream->runtime; 1522 if (!(runtime->info & SNDRV_PCM_INFO_RESUME)) 1523 return -ENOSYS; 1524 runtime->trigger_master = substream; 1525 return 0; 1526 } 1527 1528 static int snd_pcm_do_resume(struct snd_pcm_substream *substream, int state) 1529 { 1530 struct snd_pcm_runtime *runtime = substream->runtime; 1531 if (runtime->trigger_master != substream) 1532 return 0; 1533 /* DMA not running previously? */ 1534 if (runtime->status->suspended_state != SNDRV_PCM_STATE_RUNNING && 1535 (runtime->status->suspended_state != SNDRV_PCM_STATE_DRAINING || 1536 substream->stream != SNDRV_PCM_STREAM_PLAYBACK)) 1537 return 0; 1538 return substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_RESUME); 1539 } 1540 1541 static void snd_pcm_undo_resume(struct snd_pcm_substream *substream, int state) 1542 { 1543 if (substream->runtime->trigger_master == substream && 1544 snd_pcm_running(substream)) 1545 substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND); 1546 } 1547 1548 static void snd_pcm_post_resume(struct snd_pcm_substream *substream, int state) 1549 { 1550 struct snd_pcm_runtime *runtime = substream->runtime; 1551 snd_pcm_trigger_tstamp(substream); 1552 runtime->status->state = runtime->status->suspended_state; 1553 snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MRESUME); 1554 } 1555 1556 static const struct action_ops snd_pcm_action_resume = { 1557 .pre_action = snd_pcm_pre_resume, 1558 .do_action = snd_pcm_do_resume, 1559 .undo_action = snd_pcm_undo_resume, 1560 .post_action = snd_pcm_post_resume 1561 }; 1562 1563 static int snd_pcm_resume(struct snd_pcm_substream *substream) 1564 { 1565 return snd_pcm_action_lock_irq(&snd_pcm_action_resume, substream, 0); 1566 } 1567 1568 #else 1569 1570 static int snd_pcm_resume(struct snd_pcm_substream *substream) 1571 { 1572 return -ENOSYS; 1573 } 1574 1575 #endif /* CONFIG_PM */ 1576 1577 /* 1578 * xrun ioctl 1579 * 1580 * Change the RUNNING stream(s) to XRUN state. 1581 */ 1582 static int snd_pcm_xrun(struct snd_pcm_substream *substream) 1583 { 1584 struct snd_pcm_runtime *runtime = substream->runtime; 1585 int result; 1586 1587 snd_pcm_stream_lock_irq(substream); 1588 switch (runtime->status->state) { 1589 case SNDRV_PCM_STATE_XRUN: 1590 result = 0; /* already there */ 1591 break; 1592 case SNDRV_PCM_STATE_RUNNING: 1593 __snd_pcm_xrun(substream); 1594 result = 0; 1595 break; 1596 default: 1597 result = -EBADFD; 1598 } 1599 snd_pcm_stream_unlock_irq(substream); 1600 return result; 1601 } 1602 1603 /* 1604 * reset ioctl 1605 */ 1606 static int snd_pcm_pre_reset(struct snd_pcm_substream *substream, int state) 1607 { 1608 struct snd_pcm_runtime *runtime = substream->runtime; 1609 switch (runtime->status->state) { 1610 case SNDRV_PCM_STATE_RUNNING: 1611 case SNDRV_PCM_STATE_PREPARED: 1612 case SNDRV_PCM_STATE_PAUSED: 1613 case SNDRV_PCM_STATE_SUSPENDED: 1614 return 0; 1615 default: 1616 return -EBADFD; 1617 } 1618 } 1619 1620 static int snd_pcm_do_reset(struct snd_pcm_substream *substream, int state) 1621 { 1622 struct snd_pcm_runtime *runtime = substream->runtime; 1623 int err = substream->ops->ioctl(substream, SNDRV_PCM_IOCTL1_RESET, NULL); 1624 if (err < 0) 1625 return err; 1626 runtime->hw_ptr_base = 0; 1627 runtime->hw_ptr_interrupt = runtime->status->hw_ptr - 1628 runtime->status->hw_ptr % runtime->period_size; 1629 runtime->silence_start = runtime->status->hw_ptr; 1630 runtime->silence_filled = 0; 1631 return 0; 1632 } 1633 1634 static void snd_pcm_post_reset(struct snd_pcm_substream *substream, int state) 1635 { 1636 struct snd_pcm_runtime *runtime = substream->runtime; 1637 runtime->control->appl_ptr = runtime->status->hw_ptr; 1638 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK && 1639 runtime->silence_size > 0) 1640 snd_pcm_playback_silence(substream, ULONG_MAX); 1641 } 1642 1643 static const struct action_ops snd_pcm_action_reset = { 1644 .pre_action = snd_pcm_pre_reset, 1645 .do_action = snd_pcm_do_reset, 1646 .post_action = snd_pcm_post_reset 1647 }; 1648 1649 static int snd_pcm_reset(struct snd_pcm_substream *substream) 1650 { 1651 return snd_pcm_action_nonatomic(&snd_pcm_action_reset, substream, 0); 1652 } 1653 1654 /* 1655 * prepare ioctl 1656 */ 1657 /* we use the second argument for updating f_flags */ 1658 static int snd_pcm_pre_prepare(struct snd_pcm_substream *substream, 1659 int f_flags) 1660 { 1661 struct snd_pcm_runtime *runtime = substream->runtime; 1662 if (runtime->status->state == SNDRV_PCM_STATE_OPEN || 1663 runtime->status->state == SNDRV_PCM_STATE_DISCONNECTED) 1664 return -EBADFD; 1665 if (snd_pcm_running(substream)) 1666 return -EBUSY; 1667 substream->f_flags = f_flags; 1668 return 0; 1669 } 1670 1671 static int snd_pcm_do_prepare(struct snd_pcm_substream *substream, int state) 1672 { 1673 int err; 1674 err = substream->ops->prepare(substream); 1675 if (err < 0) 1676 return err; 1677 return snd_pcm_do_reset(substream, 0); 1678 } 1679 1680 static void snd_pcm_post_prepare(struct snd_pcm_substream *substream, int state) 1681 { 1682 struct snd_pcm_runtime *runtime = substream->runtime; 1683 runtime->control->appl_ptr = runtime->status->hw_ptr; 1684 snd_pcm_set_state(substream, SNDRV_PCM_STATE_PREPARED); 1685 } 1686 1687 static const struct action_ops snd_pcm_action_prepare = { 1688 .pre_action = snd_pcm_pre_prepare, 1689 .do_action = snd_pcm_do_prepare, 1690 .post_action = snd_pcm_post_prepare 1691 }; 1692 1693 /** 1694 * snd_pcm_prepare - prepare the PCM substream to be triggerable 1695 * @substream: the PCM substream instance 1696 * @file: file to refer f_flags 1697 * 1698 * Return: Zero if successful, or a negative error code. 1699 */ 1700 static int snd_pcm_prepare(struct snd_pcm_substream *substream, 1701 struct file *file) 1702 { 1703 int f_flags; 1704 1705 if (file) 1706 f_flags = file->f_flags; 1707 else 1708 f_flags = substream->f_flags; 1709 1710 snd_pcm_stream_lock_irq(substream); 1711 switch (substream->runtime->status->state) { 1712 case SNDRV_PCM_STATE_PAUSED: 1713 snd_pcm_pause(substream, 0); 1714 /* fallthru */ 1715 case SNDRV_PCM_STATE_SUSPENDED: 1716 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 1717 break; 1718 } 1719 snd_pcm_stream_unlock_irq(substream); 1720 1721 return snd_pcm_action_nonatomic(&snd_pcm_action_prepare, 1722 substream, f_flags); 1723 } 1724 1725 /* 1726 * drain ioctl 1727 */ 1728 1729 static int snd_pcm_pre_drain_init(struct snd_pcm_substream *substream, int state) 1730 { 1731 struct snd_pcm_runtime *runtime = substream->runtime; 1732 switch (runtime->status->state) { 1733 case SNDRV_PCM_STATE_OPEN: 1734 case SNDRV_PCM_STATE_DISCONNECTED: 1735 case SNDRV_PCM_STATE_SUSPENDED: 1736 return -EBADFD; 1737 } 1738 runtime->trigger_master = substream; 1739 return 0; 1740 } 1741 1742 static int snd_pcm_do_drain_init(struct snd_pcm_substream *substream, int state) 1743 { 1744 struct snd_pcm_runtime *runtime = substream->runtime; 1745 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { 1746 switch (runtime->status->state) { 1747 case SNDRV_PCM_STATE_PREPARED: 1748 /* start playback stream if possible */ 1749 if (! snd_pcm_playback_empty(substream)) { 1750 snd_pcm_do_start(substream, SNDRV_PCM_STATE_DRAINING); 1751 snd_pcm_post_start(substream, SNDRV_PCM_STATE_DRAINING); 1752 } else { 1753 runtime->status->state = SNDRV_PCM_STATE_SETUP; 1754 } 1755 break; 1756 case SNDRV_PCM_STATE_RUNNING: 1757 runtime->status->state = SNDRV_PCM_STATE_DRAINING; 1758 break; 1759 case SNDRV_PCM_STATE_XRUN: 1760 runtime->status->state = SNDRV_PCM_STATE_SETUP; 1761 break; 1762 default: 1763 break; 1764 } 1765 } else { 1766 /* stop running stream */ 1767 if (runtime->status->state == SNDRV_PCM_STATE_RUNNING) { 1768 int new_state = snd_pcm_capture_avail(runtime) > 0 ? 1769 SNDRV_PCM_STATE_DRAINING : SNDRV_PCM_STATE_SETUP; 1770 snd_pcm_do_stop(substream, new_state); 1771 snd_pcm_post_stop(substream, new_state); 1772 } 1773 } 1774 1775 if (runtime->status->state == SNDRV_PCM_STATE_DRAINING && 1776 runtime->trigger_master == substream && 1777 (runtime->hw.info & SNDRV_PCM_INFO_DRAIN_TRIGGER)) 1778 return substream->ops->trigger(substream, 1779 SNDRV_PCM_TRIGGER_DRAIN); 1780 1781 return 0; 1782 } 1783 1784 static void snd_pcm_post_drain_init(struct snd_pcm_substream *substream, int state) 1785 { 1786 } 1787 1788 static const struct action_ops snd_pcm_action_drain_init = { 1789 .pre_action = snd_pcm_pre_drain_init, 1790 .do_action = snd_pcm_do_drain_init, 1791 .post_action = snd_pcm_post_drain_init 1792 }; 1793 1794 static int snd_pcm_drop(struct snd_pcm_substream *substream); 1795 1796 /* 1797 * Drain the stream(s). 1798 * When the substream is linked, sync until the draining of all playback streams 1799 * is finished. 1800 * After this call, all streams are supposed to be either SETUP or DRAINING 1801 * (capture only) state. 1802 */ 1803 static int snd_pcm_drain(struct snd_pcm_substream *substream, 1804 struct file *file) 1805 { 1806 struct snd_card *card; 1807 struct snd_pcm_runtime *runtime; 1808 struct snd_pcm_substream *s; 1809 wait_queue_entry_t wait; 1810 int result = 0; 1811 int nonblock = 0; 1812 1813 card = substream->pcm->card; 1814 runtime = substream->runtime; 1815 1816 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 1817 return -EBADFD; 1818 1819 if (file) { 1820 if (file->f_flags & O_NONBLOCK) 1821 nonblock = 1; 1822 } else if (substream->f_flags & O_NONBLOCK) 1823 nonblock = 1; 1824 1825 down_read(&snd_pcm_link_rwsem); 1826 snd_pcm_stream_lock_irq(substream); 1827 /* resume pause */ 1828 if (runtime->status->state == SNDRV_PCM_STATE_PAUSED) 1829 snd_pcm_pause(substream, 0); 1830 1831 /* pre-start/stop - all running streams are changed to DRAINING state */ 1832 result = snd_pcm_action(&snd_pcm_action_drain_init, substream, 0); 1833 if (result < 0) 1834 goto unlock; 1835 /* in non-blocking, we don't wait in ioctl but let caller poll */ 1836 if (nonblock) { 1837 result = -EAGAIN; 1838 goto unlock; 1839 } 1840 1841 for (;;) { 1842 long tout; 1843 struct snd_pcm_runtime *to_check; 1844 if (signal_pending(current)) { 1845 result = -ERESTARTSYS; 1846 break; 1847 } 1848 /* find a substream to drain */ 1849 to_check = NULL; 1850 snd_pcm_group_for_each_entry(s, substream) { 1851 if (s->stream != SNDRV_PCM_STREAM_PLAYBACK) 1852 continue; 1853 runtime = s->runtime; 1854 if (runtime->status->state == SNDRV_PCM_STATE_DRAINING) { 1855 to_check = runtime; 1856 break; 1857 } 1858 } 1859 if (!to_check) 1860 break; /* all drained */ 1861 init_waitqueue_entry(&wait, current); 1862 add_wait_queue(&to_check->sleep, &wait); 1863 snd_pcm_stream_unlock_irq(substream); 1864 up_read(&snd_pcm_link_rwsem); 1865 if (runtime->no_period_wakeup) 1866 tout = MAX_SCHEDULE_TIMEOUT; 1867 else { 1868 tout = 10; 1869 if (runtime->rate) { 1870 long t = runtime->period_size * 2 / runtime->rate; 1871 tout = max(t, tout); 1872 } 1873 tout = msecs_to_jiffies(tout * 1000); 1874 } 1875 tout = schedule_timeout_interruptible(tout); 1876 down_read(&snd_pcm_link_rwsem); 1877 snd_pcm_stream_lock_irq(substream); 1878 remove_wait_queue(&to_check->sleep, &wait); 1879 if (card->shutdown) { 1880 result = -ENODEV; 1881 break; 1882 } 1883 if (tout == 0) { 1884 if (substream->runtime->status->state == SNDRV_PCM_STATE_SUSPENDED) 1885 result = -ESTRPIPE; 1886 else { 1887 dev_dbg(substream->pcm->card->dev, 1888 "playback drain error (DMA or IRQ trouble?)\n"); 1889 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 1890 result = -EIO; 1891 } 1892 break; 1893 } 1894 } 1895 1896 unlock: 1897 snd_pcm_stream_unlock_irq(substream); 1898 up_read(&snd_pcm_link_rwsem); 1899 1900 return result; 1901 } 1902 1903 /* 1904 * drop ioctl 1905 * 1906 * Immediately put all linked substreams into SETUP state. 1907 */ 1908 static int snd_pcm_drop(struct snd_pcm_substream *substream) 1909 { 1910 struct snd_pcm_runtime *runtime; 1911 int result = 0; 1912 1913 if (PCM_RUNTIME_CHECK(substream)) 1914 return -ENXIO; 1915 runtime = substream->runtime; 1916 1917 if (runtime->status->state == SNDRV_PCM_STATE_OPEN || 1918 runtime->status->state == SNDRV_PCM_STATE_DISCONNECTED) 1919 return -EBADFD; 1920 1921 snd_pcm_stream_lock_irq(substream); 1922 /* resume pause */ 1923 if (runtime->status->state == SNDRV_PCM_STATE_PAUSED) 1924 snd_pcm_pause(substream, 0); 1925 1926 snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP); 1927 /* runtime->control->appl_ptr = runtime->status->hw_ptr; */ 1928 snd_pcm_stream_unlock_irq(substream); 1929 1930 return result; 1931 } 1932 1933 1934 static bool is_pcm_file(struct file *file) 1935 { 1936 struct inode *inode = file_inode(file); 1937 unsigned int minor; 1938 1939 if (!S_ISCHR(inode->i_mode) || imajor(inode) != snd_major) 1940 return false; 1941 minor = iminor(inode); 1942 return snd_lookup_minor_data(minor, SNDRV_DEVICE_TYPE_PCM_PLAYBACK) || 1943 snd_lookup_minor_data(minor, SNDRV_DEVICE_TYPE_PCM_CAPTURE); 1944 } 1945 1946 /* 1947 * PCM link handling 1948 */ 1949 static int snd_pcm_link(struct snd_pcm_substream *substream, int fd) 1950 { 1951 int res = 0; 1952 struct snd_pcm_file *pcm_file; 1953 struct snd_pcm_substream *substream1; 1954 struct snd_pcm_group *group; 1955 struct fd f = fdget(fd); 1956 1957 if (!f.file) 1958 return -EBADFD; 1959 if (!is_pcm_file(f.file)) { 1960 res = -EBADFD; 1961 goto _badf; 1962 } 1963 pcm_file = f.file->private_data; 1964 substream1 = pcm_file->substream; 1965 group = kmalloc(sizeof(*group), GFP_KERNEL); 1966 if (!group) { 1967 res = -ENOMEM; 1968 goto _nolock; 1969 } 1970 down_write_nonblock(&snd_pcm_link_rwsem); 1971 write_lock_irq(&snd_pcm_link_rwlock); 1972 if (substream->runtime->status->state == SNDRV_PCM_STATE_OPEN || 1973 substream->runtime->status->state != substream1->runtime->status->state || 1974 substream->pcm->nonatomic != substream1->pcm->nonatomic) { 1975 res = -EBADFD; 1976 goto _end; 1977 } 1978 if (snd_pcm_stream_linked(substream1)) { 1979 res = -EALREADY; 1980 goto _end; 1981 } 1982 if (!snd_pcm_stream_linked(substream)) { 1983 substream->group = group; 1984 group = NULL; 1985 spin_lock_init(&substream->group->lock); 1986 mutex_init(&substream->group->mutex); 1987 INIT_LIST_HEAD(&substream->group->substreams); 1988 list_add_tail(&substream->link_list, &substream->group->substreams); 1989 substream->group->count = 1; 1990 } 1991 list_add_tail(&substream1->link_list, &substream->group->substreams); 1992 substream->group->count++; 1993 substream1->group = substream->group; 1994 _end: 1995 write_unlock_irq(&snd_pcm_link_rwlock); 1996 up_write(&snd_pcm_link_rwsem); 1997 _nolock: 1998 snd_card_unref(substream1->pcm->card); 1999 kfree(group); 2000 _badf: 2001 fdput(f); 2002 return res; 2003 } 2004 2005 static void relink_to_local(struct snd_pcm_substream *substream) 2006 { 2007 substream->group = &substream->self_group; 2008 INIT_LIST_HEAD(&substream->self_group.substreams); 2009 list_add_tail(&substream->link_list, &substream->self_group.substreams); 2010 } 2011 2012 static int snd_pcm_unlink(struct snd_pcm_substream *substream) 2013 { 2014 struct snd_pcm_substream *s; 2015 int res = 0; 2016 2017 down_write_nonblock(&snd_pcm_link_rwsem); 2018 write_lock_irq(&snd_pcm_link_rwlock); 2019 if (!snd_pcm_stream_linked(substream)) { 2020 res = -EALREADY; 2021 goto _end; 2022 } 2023 list_del(&substream->link_list); 2024 substream->group->count--; 2025 if (substream->group->count == 1) { /* detach the last stream, too */ 2026 snd_pcm_group_for_each_entry(s, substream) { 2027 relink_to_local(s); 2028 break; 2029 } 2030 kfree(substream->group); 2031 } 2032 relink_to_local(substream); 2033 _end: 2034 write_unlock_irq(&snd_pcm_link_rwlock); 2035 up_write(&snd_pcm_link_rwsem); 2036 return res; 2037 } 2038 2039 /* 2040 * hw configurator 2041 */ 2042 static int snd_pcm_hw_rule_mul(struct snd_pcm_hw_params *params, 2043 struct snd_pcm_hw_rule *rule) 2044 { 2045 struct snd_interval t; 2046 snd_interval_mul(hw_param_interval_c(params, rule->deps[0]), 2047 hw_param_interval_c(params, rule->deps[1]), &t); 2048 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2049 } 2050 2051 static int snd_pcm_hw_rule_div(struct snd_pcm_hw_params *params, 2052 struct snd_pcm_hw_rule *rule) 2053 { 2054 struct snd_interval t; 2055 snd_interval_div(hw_param_interval_c(params, rule->deps[0]), 2056 hw_param_interval_c(params, rule->deps[1]), &t); 2057 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2058 } 2059 2060 static int snd_pcm_hw_rule_muldivk(struct snd_pcm_hw_params *params, 2061 struct snd_pcm_hw_rule *rule) 2062 { 2063 struct snd_interval t; 2064 snd_interval_muldivk(hw_param_interval_c(params, rule->deps[0]), 2065 hw_param_interval_c(params, rule->deps[1]), 2066 (unsigned long) rule->private, &t); 2067 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2068 } 2069 2070 static int snd_pcm_hw_rule_mulkdiv(struct snd_pcm_hw_params *params, 2071 struct snd_pcm_hw_rule *rule) 2072 { 2073 struct snd_interval t; 2074 snd_interval_mulkdiv(hw_param_interval_c(params, rule->deps[0]), 2075 (unsigned long) rule->private, 2076 hw_param_interval_c(params, rule->deps[1]), &t); 2077 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2078 } 2079 2080 static int snd_pcm_hw_rule_format(struct snd_pcm_hw_params *params, 2081 struct snd_pcm_hw_rule *rule) 2082 { 2083 unsigned int k; 2084 const struct snd_interval *i = 2085 hw_param_interval_c(params, rule->deps[0]); 2086 struct snd_mask m; 2087 struct snd_mask *mask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT); 2088 snd_mask_any(&m); 2089 for (k = 0; k <= SNDRV_PCM_FORMAT_LAST; ++k) { 2090 int bits; 2091 if (! snd_mask_test(mask, k)) 2092 continue; 2093 bits = snd_pcm_format_physical_width(k); 2094 if (bits <= 0) 2095 continue; /* ignore invalid formats */ 2096 if ((unsigned)bits < i->min || (unsigned)bits > i->max) 2097 snd_mask_reset(&m, k); 2098 } 2099 return snd_mask_refine(mask, &m); 2100 } 2101 2102 static int snd_pcm_hw_rule_sample_bits(struct snd_pcm_hw_params *params, 2103 struct snd_pcm_hw_rule *rule) 2104 { 2105 struct snd_interval t; 2106 unsigned int k; 2107 t.min = UINT_MAX; 2108 t.max = 0; 2109 t.openmin = 0; 2110 t.openmax = 0; 2111 for (k = 0; k <= SNDRV_PCM_FORMAT_LAST; ++k) { 2112 int bits; 2113 if (! snd_mask_test(hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT), k)) 2114 continue; 2115 bits = snd_pcm_format_physical_width(k); 2116 if (bits <= 0) 2117 continue; /* ignore invalid formats */ 2118 if (t.min > (unsigned)bits) 2119 t.min = bits; 2120 if (t.max < (unsigned)bits) 2121 t.max = bits; 2122 } 2123 t.integer = 1; 2124 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2125 } 2126 2127 #if SNDRV_PCM_RATE_5512 != 1 << 0 || SNDRV_PCM_RATE_192000 != 1 << 12 2128 #error "Change this table" 2129 #endif 2130 2131 static const unsigned int rates[] = { 2132 5512, 8000, 11025, 16000, 22050, 32000, 44100, 2133 48000, 64000, 88200, 96000, 176400, 192000 2134 }; 2135 2136 const struct snd_pcm_hw_constraint_list snd_pcm_known_rates = { 2137 .count = ARRAY_SIZE(rates), 2138 .list = rates, 2139 }; 2140 2141 static int snd_pcm_hw_rule_rate(struct snd_pcm_hw_params *params, 2142 struct snd_pcm_hw_rule *rule) 2143 { 2144 struct snd_pcm_hardware *hw = rule->private; 2145 return snd_interval_list(hw_param_interval(params, rule->var), 2146 snd_pcm_known_rates.count, 2147 snd_pcm_known_rates.list, hw->rates); 2148 } 2149 2150 static int snd_pcm_hw_rule_buffer_bytes_max(struct snd_pcm_hw_params *params, 2151 struct snd_pcm_hw_rule *rule) 2152 { 2153 struct snd_interval t; 2154 struct snd_pcm_substream *substream = rule->private; 2155 t.min = 0; 2156 t.max = substream->buffer_bytes_max; 2157 t.openmin = 0; 2158 t.openmax = 0; 2159 t.integer = 1; 2160 return snd_interval_refine(hw_param_interval(params, rule->var), &t); 2161 } 2162 2163 int snd_pcm_hw_constraints_init(struct snd_pcm_substream *substream) 2164 { 2165 struct snd_pcm_runtime *runtime = substream->runtime; 2166 struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints; 2167 int k, err; 2168 2169 for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) { 2170 snd_mask_any(constrs_mask(constrs, k)); 2171 } 2172 2173 for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) { 2174 snd_interval_any(constrs_interval(constrs, k)); 2175 } 2176 2177 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_CHANNELS)); 2178 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_SIZE)); 2179 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_BYTES)); 2180 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_SAMPLE_BITS)); 2181 snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_FRAME_BITS)); 2182 2183 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FORMAT, 2184 snd_pcm_hw_rule_format, NULL, 2185 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 2186 if (err < 0) 2187 return err; 2188 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, 2189 snd_pcm_hw_rule_sample_bits, NULL, 2190 SNDRV_PCM_HW_PARAM_FORMAT, 2191 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 2192 if (err < 0) 2193 return err; 2194 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, 2195 snd_pcm_hw_rule_div, NULL, 2196 SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1); 2197 if (err < 0) 2198 return err; 2199 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, 2200 snd_pcm_hw_rule_mul, NULL, 2201 SNDRV_PCM_HW_PARAM_SAMPLE_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1); 2202 if (err < 0) 2203 return err; 2204 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, 2205 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2206 SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1); 2207 if (err < 0) 2208 return err; 2209 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS, 2210 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2211 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, -1); 2212 if (err < 0) 2213 return err; 2214 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_CHANNELS, 2215 snd_pcm_hw_rule_div, NULL, 2216 SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1); 2217 if (err < 0) 2218 return err; 2219 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE, 2220 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2221 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_TIME, -1); 2222 if (err < 0) 2223 return err; 2224 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE, 2225 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2226 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_BUFFER_TIME, -1); 2227 if (err < 0) 2228 return err; 2229 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIODS, 2230 snd_pcm_hw_rule_div, NULL, 2231 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1); 2232 if (err < 0) 2233 return err; 2234 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, 2235 snd_pcm_hw_rule_div, NULL, 2236 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1); 2237 if (err < 0) 2238 return err; 2239 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, 2240 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2241 SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2242 if (err < 0) 2243 return err; 2244 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, 2245 snd_pcm_hw_rule_muldivk, (void*) 1000000, 2246 SNDRV_PCM_HW_PARAM_PERIOD_TIME, SNDRV_PCM_HW_PARAM_RATE, -1); 2247 if (err < 0) 2248 return err; 2249 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 2250 snd_pcm_hw_rule_mul, NULL, 2251 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1); 2252 if (err < 0) 2253 return err; 2254 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 2255 snd_pcm_hw_rule_mulkdiv, (void*) 8, 2256 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2257 if (err < 0) 2258 return err; 2259 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, 2260 snd_pcm_hw_rule_muldivk, (void*) 1000000, 2261 SNDRV_PCM_HW_PARAM_BUFFER_TIME, SNDRV_PCM_HW_PARAM_RATE, -1); 2262 if (err < 0) 2263 return err; 2264 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, 2265 snd_pcm_hw_rule_muldivk, (void*) 8, 2266 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2267 if (err < 0) 2268 return err; 2269 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 2270 snd_pcm_hw_rule_muldivk, (void*) 8, 2271 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1); 2272 if (err < 0) 2273 return err; 2274 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_TIME, 2275 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2276 SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1); 2277 if (err < 0) 2278 return err; 2279 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_TIME, 2280 snd_pcm_hw_rule_mulkdiv, (void*) 1000000, 2281 SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1); 2282 if (err < 0) 2283 return err; 2284 return 0; 2285 } 2286 2287 int snd_pcm_hw_constraints_complete(struct snd_pcm_substream *substream) 2288 { 2289 struct snd_pcm_runtime *runtime = substream->runtime; 2290 struct snd_pcm_hardware *hw = &runtime->hw; 2291 int err; 2292 unsigned int mask = 0; 2293 2294 if (hw->info & SNDRV_PCM_INFO_INTERLEAVED) 2295 mask |= 1 << SNDRV_PCM_ACCESS_RW_INTERLEAVED; 2296 if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED) 2297 mask |= 1 << SNDRV_PCM_ACCESS_RW_NONINTERLEAVED; 2298 if (hw_support_mmap(substream)) { 2299 if (hw->info & SNDRV_PCM_INFO_INTERLEAVED) 2300 mask |= 1 << SNDRV_PCM_ACCESS_MMAP_INTERLEAVED; 2301 if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED) 2302 mask |= 1 << SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED; 2303 if (hw->info & SNDRV_PCM_INFO_COMPLEX) 2304 mask |= 1 << SNDRV_PCM_ACCESS_MMAP_COMPLEX; 2305 } 2306 err = snd_pcm_hw_constraint_mask(runtime, SNDRV_PCM_HW_PARAM_ACCESS, mask); 2307 if (err < 0) 2308 return err; 2309 2310 err = snd_pcm_hw_constraint_mask64(runtime, SNDRV_PCM_HW_PARAM_FORMAT, hw->formats); 2311 if (err < 0) 2312 return err; 2313 2314 err = snd_pcm_hw_constraint_mask(runtime, SNDRV_PCM_HW_PARAM_SUBFORMAT, 1 << SNDRV_PCM_SUBFORMAT_STD); 2315 if (err < 0) 2316 return err; 2317 2318 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_CHANNELS, 2319 hw->channels_min, hw->channels_max); 2320 if (err < 0) 2321 return err; 2322 2323 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_RATE, 2324 hw->rate_min, hw->rate_max); 2325 if (err < 0) 2326 return err; 2327 2328 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, 2329 hw->period_bytes_min, hw->period_bytes_max); 2330 if (err < 0) 2331 return err; 2332 2333 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIODS, 2334 hw->periods_min, hw->periods_max); 2335 if (err < 0) 2336 return err; 2337 2338 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 2339 hw->period_bytes_min, hw->buffer_bytes_max); 2340 if (err < 0) 2341 return err; 2342 2343 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 2344 snd_pcm_hw_rule_buffer_bytes_max, substream, 2345 SNDRV_PCM_HW_PARAM_BUFFER_BYTES, -1); 2346 if (err < 0) 2347 return err; 2348 2349 /* FIXME: remove */ 2350 if (runtime->dma_bytes) { 2351 err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 0, runtime->dma_bytes); 2352 if (err < 0) 2353 return err; 2354 } 2355 2356 if (!(hw->rates & (SNDRV_PCM_RATE_KNOT | SNDRV_PCM_RATE_CONTINUOUS))) { 2357 err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE, 2358 snd_pcm_hw_rule_rate, hw, 2359 SNDRV_PCM_HW_PARAM_RATE, -1); 2360 if (err < 0) 2361 return err; 2362 } 2363 2364 /* FIXME: this belong to lowlevel */ 2365 snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIOD_SIZE); 2366 2367 return 0; 2368 } 2369 2370 static void pcm_release_private(struct snd_pcm_substream *substream) 2371 { 2372 if (snd_pcm_stream_linked(substream)) 2373 snd_pcm_unlink(substream); 2374 } 2375 2376 void snd_pcm_release_substream(struct snd_pcm_substream *substream) 2377 { 2378 substream->ref_count--; 2379 if (substream->ref_count > 0) 2380 return; 2381 2382 snd_pcm_drop(substream); 2383 if (substream->hw_opened) { 2384 if (substream->ops->hw_free && 2385 substream->runtime->status->state != SNDRV_PCM_STATE_OPEN) 2386 substream->ops->hw_free(substream); 2387 substream->ops->close(substream); 2388 substream->hw_opened = 0; 2389 } 2390 if (pm_qos_request_active(&substream->latency_pm_qos_req)) 2391 pm_qos_remove_request(&substream->latency_pm_qos_req); 2392 if (substream->pcm_release) { 2393 substream->pcm_release(substream); 2394 substream->pcm_release = NULL; 2395 } 2396 snd_pcm_detach_substream(substream); 2397 } 2398 EXPORT_SYMBOL(snd_pcm_release_substream); 2399 2400 int snd_pcm_open_substream(struct snd_pcm *pcm, int stream, 2401 struct file *file, 2402 struct snd_pcm_substream **rsubstream) 2403 { 2404 struct snd_pcm_substream *substream; 2405 int err; 2406 2407 err = snd_pcm_attach_substream(pcm, stream, file, &substream); 2408 if (err < 0) 2409 return err; 2410 if (substream->ref_count > 1) { 2411 *rsubstream = substream; 2412 return 0; 2413 } 2414 2415 err = snd_pcm_hw_constraints_init(substream); 2416 if (err < 0) { 2417 pcm_dbg(pcm, "snd_pcm_hw_constraints_init failed\n"); 2418 goto error; 2419 } 2420 2421 if ((err = substream->ops->open(substream)) < 0) 2422 goto error; 2423 2424 substream->hw_opened = 1; 2425 2426 err = snd_pcm_hw_constraints_complete(substream); 2427 if (err < 0) { 2428 pcm_dbg(pcm, "snd_pcm_hw_constraints_complete failed\n"); 2429 goto error; 2430 } 2431 2432 *rsubstream = substream; 2433 return 0; 2434 2435 error: 2436 snd_pcm_release_substream(substream); 2437 return err; 2438 } 2439 EXPORT_SYMBOL(snd_pcm_open_substream); 2440 2441 static int snd_pcm_open_file(struct file *file, 2442 struct snd_pcm *pcm, 2443 int stream) 2444 { 2445 struct snd_pcm_file *pcm_file; 2446 struct snd_pcm_substream *substream; 2447 int err; 2448 2449 err = snd_pcm_open_substream(pcm, stream, file, &substream); 2450 if (err < 0) 2451 return err; 2452 2453 pcm_file = kzalloc(sizeof(*pcm_file), GFP_KERNEL); 2454 if (pcm_file == NULL) { 2455 snd_pcm_release_substream(substream); 2456 return -ENOMEM; 2457 } 2458 pcm_file->substream = substream; 2459 if (substream->ref_count == 1) { 2460 substream->file = pcm_file; 2461 substream->pcm_release = pcm_release_private; 2462 } 2463 file->private_data = pcm_file; 2464 2465 return 0; 2466 } 2467 2468 static int snd_pcm_playback_open(struct inode *inode, struct file *file) 2469 { 2470 struct snd_pcm *pcm; 2471 int err = nonseekable_open(inode, file); 2472 if (err < 0) 2473 return err; 2474 pcm = snd_lookup_minor_data(iminor(inode), 2475 SNDRV_DEVICE_TYPE_PCM_PLAYBACK); 2476 err = snd_pcm_open(file, pcm, SNDRV_PCM_STREAM_PLAYBACK); 2477 if (pcm) 2478 snd_card_unref(pcm->card); 2479 return err; 2480 } 2481 2482 static int snd_pcm_capture_open(struct inode *inode, struct file *file) 2483 { 2484 struct snd_pcm *pcm; 2485 int err = nonseekable_open(inode, file); 2486 if (err < 0) 2487 return err; 2488 pcm = snd_lookup_minor_data(iminor(inode), 2489 SNDRV_DEVICE_TYPE_PCM_CAPTURE); 2490 err = snd_pcm_open(file, pcm, SNDRV_PCM_STREAM_CAPTURE); 2491 if (pcm) 2492 snd_card_unref(pcm->card); 2493 return err; 2494 } 2495 2496 static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream) 2497 { 2498 int err; 2499 wait_queue_entry_t wait; 2500 2501 if (pcm == NULL) { 2502 err = -ENODEV; 2503 goto __error1; 2504 } 2505 err = snd_card_file_add(pcm->card, file); 2506 if (err < 0) 2507 goto __error1; 2508 if (!try_module_get(pcm->card->module)) { 2509 err = -EFAULT; 2510 goto __error2; 2511 } 2512 init_waitqueue_entry(&wait, current); 2513 add_wait_queue(&pcm->open_wait, &wait); 2514 mutex_lock(&pcm->open_mutex); 2515 while (1) { 2516 err = snd_pcm_open_file(file, pcm, stream); 2517 if (err >= 0) 2518 break; 2519 if (err == -EAGAIN) { 2520 if (file->f_flags & O_NONBLOCK) { 2521 err = -EBUSY; 2522 break; 2523 } 2524 } else 2525 break; 2526 set_current_state(TASK_INTERRUPTIBLE); 2527 mutex_unlock(&pcm->open_mutex); 2528 schedule(); 2529 mutex_lock(&pcm->open_mutex); 2530 if (pcm->card->shutdown) { 2531 err = -ENODEV; 2532 break; 2533 } 2534 if (signal_pending(current)) { 2535 err = -ERESTARTSYS; 2536 break; 2537 } 2538 } 2539 remove_wait_queue(&pcm->open_wait, &wait); 2540 mutex_unlock(&pcm->open_mutex); 2541 if (err < 0) 2542 goto __error; 2543 return err; 2544 2545 __error: 2546 module_put(pcm->card->module); 2547 __error2: 2548 snd_card_file_remove(pcm->card, file); 2549 __error1: 2550 return err; 2551 } 2552 2553 static int snd_pcm_release(struct inode *inode, struct file *file) 2554 { 2555 struct snd_pcm *pcm; 2556 struct snd_pcm_substream *substream; 2557 struct snd_pcm_file *pcm_file; 2558 2559 pcm_file = file->private_data; 2560 substream = pcm_file->substream; 2561 if (snd_BUG_ON(!substream)) 2562 return -ENXIO; 2563 pcm = substream->pcm; 2564 mutex_lock(&pcm->open_mutex); 2565 snd_pcm_release_substream(substream); 2566 kfree(pcm_file); 2567 mutex_unlock(&pcm->open_mutex); 2568 wake_up(&pcm->open_wait); 2569 module_put(pcm->card->module); 2570 snd_card_file_remove(pcm->card, file); 2571 return 0; 2572 } 2573 2574 /* check and update PCM state; return 0 or a negative error 2575 * call this inside PCM lock 2576 */ 2577 static int do_pcm_hwsync(struct snd_pcm_substream *substream) 2578 { 2579 switch (substream->runtime->status->state) { 2580 case SNDRV_PCM_STATE_DRAINING: 2581 if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) 2582 return -EBADFD; 2583 /* Fall through */ 2584 case SNDRV_PCM_STATE_RUNNING: 2585 return snd_pcm_update_hw_ptr(substream); 2586 case SNDRV_PCM_STATE_PREPARED: 2587 case SNDRV_PCM_STATE_PAUSED: 2588 return 0; 2589 case SNDRV_PCM_STATE_SUSPENDED: 2590 return -ESTRPIPE; 2591 case SNDRV_PCM_STATE_XRUN: 2592 return -EPIPE; 2593 default: 2594 return -EBADFD; 2595 } 2596 } 2597 2598 /* increase the appl_ptr; returns the processed frames or a negative error */ 2599 static snd_pcm_sframes_t forward_appl_ptr(struct snd_pcm_substream *substream, 2600 snd_pcm_uframes_t frames, 2601 snd_pcm_sframes_t avail) 2602 { 2603 struct snd_pcm_runtime *runtime = substream->runtime; 2604 snd_pcm_sframes_t appl_ptr; 2605 int ret; 2606 2607 if (avail <= 0) 2608 return 0; 2609 if (frames > (snd_pcm_uframes_t)avail) 2610 frames = avail; 2611 appl_ptr = runtime->control->appl_ptr + frames; 2612 if (appl_ptr >= (snd_pcm_sframes_t)runtime->boundary) 2613 appl_ptr -= runtime->boundary; 2614 ret = pcm_lib_apply_appl_ptr(substream, appl_ptr); 2615 return ret < 0 ? ret : frames; 2616 } 2617 2618 /* decrease the appl_ptr; returns the processed frames or zero for error */ 2619 static snd_pcm_sframes_t rewind_appl_ptr(struct snd_pcm_substream *substream, 2620 snd_pcm_uframes_t frames, 2621 snd_pcm_sframes_t avail) 2622 { 2623 struct snd_pcm_runtime *runtime = substream->runtime; 2624 snd_pcm_sframes_t appl_ptr; 2625 int ret; 2626 2627 if (avail <= 0) 2628 return 0; 2629 if (frames > (snd_pcm_uframes_t)avail) 2630 frames = avail; 2631 appl_ptr = runtime->control->appl_ptr - frames; 2632 if (appl_ptr < 0) 2633 appl_ptr += runtime->boundary; 2634 ret = pcm_lib_apply_appl_ptr(substream, appl_ptr); 2635 /* NOTE: we return zero for errors because PulseAudio gets depressed 2636 * upon receiving an error from rewind ioctl and stops processing 2637 * any longer. Returning zero means that no rewind is done, so 2638 * it's not absolutely wrong to answer like that. 2639 */ 2640 return ret < 0 ? 0 : frames; 2641 } 2642 2643 static snd_pcm_sframes_t snd_pcm_rewind(struct snd_pcm_substream *substream, 2644 snd_pcm_uframes_t frames) 2645 { 2646 snd_pcm_sframes_t ret; 2647 2648 if (frames == 0) 2649 return 0; 2650 2651 snd_pcm_stream_lock_irq(substream); 2652 ret = do_pcm_hwsync(substream); 2653 if (!ret) 2654 ret = rewind_appl_ptr(substream, frames, 2655 snd_pcm_hw_avail(substream)); 2656 snd_pcm_stream_unlock_irq(substream); 2657 return ret; 2658 } 2659 2660 static snd_pcm_sframes_t snd_pcm_forward(struct snd_pcm_substream *substream, 2661 snd_pcm_uframes_t frames) 2662 { 2663 snd_pcm_sframes_t ret; 2664 2665 if (frames == 0) 2666 return 0; 2667 2668 snd_pcm_stream_lock_irq(substream); 2669 ret = do_pcm_hwsync(substream); 2670 if (!ret) 2671 ret = forward_appl_ptr(substream, frames, 2672 snd_pcm_avail(substream)); 2673 snd_pcm_stream_unlock_irq(substream); 2674 return ret; 2675 } 2676 2677 static int snd_pcm_hwsync(struct snd_pcm_substream *substream) 2678 { 2679 int err; 2680 2681 snd_pcm_stream_lock_irq(substream); 2682 err = do_pcm_hwsync(substream); 2683 snd_pcm_stream_unlock_irq(substream); 2684 return err; 2685 } 2686 2687 static int snd_pcm_delay(struct snd_pcm_substream *substream, 2688 snd_pcm_sframes_t *delay) 2689 { 2690 int err; 2691 snd_pcm_sframes_t n = 0; 2692 2693 snd_pcm_stream_lock_irq(substream); 2694 err = do_pcm_hwsync(substream); 2695 if (!err) 2696 n = snd_pcm_calc_delay(substream); 2697 snd_pcm_stream_unlock_irq(substream); 2698 if (!err) 2699 *delay = n; 2700 return err; 2701 } 2702 2703 static int snd_pcm_sync_ptr(struct snd_pcm_substream *substream, 2704 struct snd_pcm_sync_ptr __user *_sync_ptr) 2705 { 2706 struct snd_pcm_runtime *runtime = substream->runtime; 2707 struct snd_pcm_sync_ptr sync_ptr; 2708 volatile struct snd_pcm_mmap_status *status; 2709 volatile struct snd_pcm_mmap_control *control; 2710 int err; 2711 2712 memset(&sync_ptr, 0, sizeof(sync_ptr)); 2713 if (get_user(sync_ptr.flags, (unsigned __user *)&(_sync_ptr->flags))) 2714 return -EFAULT; 2715 if (copy_from_user(&sync_ptr.c.control, &(_sync_ptr->c.control), sizeof(struct snd_pcm_mmap_control))) 2716 return -EFAULT; 2717 status = runtime->status; 2718 control = runtime->control; 2719 if (sync_ptr.flags & SNDRV_PCM_SYNC_PTR_HWSYNC) { 2720 err = snd_pcm_hwsync(substream); 2721 if (err < 0) 2722 return err; 2723 } 2724 snd_pcm_stream_lock_irq(substream); 2725 if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_APPL)) { 2726 err = pcm_lib_apply_appl_ptr(substream, 2727 sync_ptr.c.control.appl_ptr); 2728 if (err < 0) { 2729 snd_pcm_stream_unlock_irq(substream); 2730 return err; 2731 } 2732 } else { 2733 sync_ptr.c.control.appl_ptr = control->appl_ptr; 2734 } 2735 if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN)) 2736 control->avail_min = sync_ptr.c.control.avail_min; 2737 else 2738 sync_ptr.c.control.avail_min = control->avail_min; 2739 sync_ptr.s.status.state = status->state; 2740 sync_ptr.s.status.hw_ptr = status->hw_ptr; 2741 sync_ptr.s.status.tstamp = status->tstamp; 2742 sync_ptr.s.status.suspended_state = status->suspended_state; 2743 sync_ptr.s.status.audio_tstamp = status->audio_tstamp; 2744 snd_pcm_stream_unlock_irq(substream); 2745 if (copy_to_user(_sync_ptr, &sync_ptr, sizeof(sync_ptr))) 2746 return -EFAULT; 2747 return 0; 2748 } 2749 2750 static int snd_pcm_tstamp(struct snd_pcm_substream *substream, int __user *_arg) 2751 { 2752 struct snd_pcm_runtime *runtime = substream->runtime; 2753 int arg; 2754 2755 if (get_user(arg, _arg)) 2756 return -EFAULT; 2757 if (arg < 0 || arg > SNDRV_PCM_TSTAMP_TYPE_LAST) 2758 return -EINVAL; 2759 runtime->tstamp_type = arg; 2760 return 0; 2761 } 2762 2763 static int snd_pcm_xferi_frames_ioctl(struct snd_pcm_substream *substream, 2764 struct snd_xferi __user *_xferi) 2765 { 2766 struct snd_xferi xferi; 2767 struct snd_pcm_runtime *runtime = substream->runtime; 2768 snd_pcm_sframes_t result; 2769 2770 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 2771 return -EBADFD; 2772 if (put_user(0, &_xferi->result)) 2773 return -EFAULT; 2774 if (copy_from_user(&xferi, _xferi, sizeof(xferi))) 2775 return -EFAULT; 2776 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 2777 result = snd_pcm_lib_write(substream, xferi.buf, xferi.frames); 2778 else 2779 result = snd_pcm_lib_read(substream, xferi.buf, xferi.frames); 2780 __put_user(result, &_xferi->result); 2781 return result < 0 ? result : 0; 2782 } 2783 2784 static int snd_pcm_xfern_frames_ioctl(struct snd_pcm_substream *substream, 2785 struct snd_xfern __user *_xfern) 2786 { 2787 struct snd_xfern xfern; 2788 struct snd_pcm_runtime *runtime = substream->runtime; 2789 void *bufs; 2790 snd_pcm_sframes_t result; 2791 2792 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 2793 return -EBADFD; 2794 if (runtime->channels > 128) 2795 return -EINVAL; 2796 if (put_user(0, &_xfern->result)) 2797 return -EFAULT; 2798 if (copy_from_user(&xfern, _xfern, sizeof(xfern))) 2799 return -EFAULT; 2800 2801 bufs = memdup_user(xfern.bufs, sizeof(void *) * runtime->channels); 2802 if (IS_ERR(bufs)) 2803 return PTR_ERR(bufs); 2804 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 2805 result = snd_pcm_lib_writev(substream, bufs, xfern.frames); 2806 else 2807 result = snd_pcm_lib_readv(substream, bufs, xfern.frames); 2808 kfree(bufs); 2809 __put_user(result, &_xfern->result); 2810 return result < 0 ? result : 0; 2811 } 2812 2813 static int snd_pcm_rewind_ioctl(struct snd_pcm_substream *substream, 2814 snd_pcm_uframes_t __user *_frames) 2815 { 2816 snd_pcm_uframes_t frames; 2817 snd_pcm_sframes_t result; 2818 2819 if (get_user(frames, _frames)) 2820 return -EFAULT; 2821 if (put_user(0, _frames)) 2822 return -EFAULT; 2823 result = snd_pcm_rewind(substream, frames); 2824 __put_user(result, _frames); 2825 return result < 0 ? result : 0; 2826 } 2827 2828 static int snd_pcm_forward_ioctl(struct snd_pcm_substream *substream, 2829 snd_pcm_uframes_t __user *_frames) 2830 { 2831 snd_pcm_uframes_t frames; 2832 snd_pcm_sframes_t result; 2833 2834 if (get_user(frames, _frames)) 2835 return -EFAULT; 2836 if (put_user(0, _frames)) 2837 return -EFAULT; 2838 result = snd_pcm_forward(substream, frames); 2839 __put_user(result, _frames); 2840 return result < 0 ? result : 0; 2841 } 2842 2843 static int snd_pcm_common_ioctl(struct file *file, 2844 struct snd_pcm_substream *substream, 2845 unsigned int cmd, void __user *arg) 2846 { 2847 struct snd_pcm_file *pcm_file = file->private_data; 2848 int res; 2849 2850 if (PCM_RUNTIME_CHECK(substream)) 2851 return -ENXIO; 2852 2853 res = snd_power_wait(substream->pcm->card, SNDRV_CTL_POWER_D0); 2854 if (res < 0) 2855 return res; 2856 2857 switch (cmd) { 2858 case SNDRV_PCM_IOCTL_PVERSION: 2859 return put_user(SNDRV_PCM_VERSION, (int __user *)arg) ? -EFAULT : 0; 2860 case SNDRV_PCM_IOCTL_INFO: 2861 return snd_pcm_info_user(substream, arg); 2862 case SNDRV_PCM_IOCTL_TSTAMP: /* just for compatibility */ 2863 return 0; 2864 case SNDRV_PCM_IOCTL_TTSTAMP: 2865 return snd_pcm_tstamp(substream, arg); 2866 case SNDRV_PCM_IOCTL_USER_PVERSION: 2867 if (get_user(pcm_file->user_pversion, 2868 (unsigned int __user *)arg)) 2869 return -EFAULT; 2870 return 0; 2871 case SNDRV_PCM_IOCTL_HW_REFINE: 2872 return snd_pcm_hw_refine_user(substream, arg); 2873 case SNDRV_PCM_IOCTL_HW_PARAMS: 2874 return snd_pcm_hw_params_user(substream, arg); 2875 case SNDRV_PCM_IOCTL_HW_FREE: 2876 return snd_pcm_hw_free(substream); 2877 case SNDRV_PCM_IOCTL_SW_PARAMS: 2878 return snd_pcm_sw_params_user(substream, arg); 2879 case SNDRV_PCM_IOCTL_STATUS: 2880 return snd_pcm_status_user(substream, arg, false); 2881 case SNDRV_PCM_IOCTL_STATUS_EXT: 2882 return snd_pcm_status_user(substream, arg, true); 2883 case SNDRV_PCM_IOCTL_CHANNEL_INFO: 2884 return snd_pcm_channel_info_user(substream, arg); 2885 case SNDRV_PCM_IOCTL_PREPARE: 2886 return snd_pcm_prepare(substream, file); 2887 case SNDRV_PCM_IOCTL_RESET: 2888 return snd_pcm_reset(substream); 2889 case SNDRV_PCM_IOCTL_START: 2890 return snd_pcm_start_lock_irq(substream); 2891 case SNDRV_PCM_IOCTL_LINK: 2892 return snd_pcm_link(substream, (int)(unsigned long) arg); 2893 case SNDRV_PCM_IOCTL_UNLINK: 2894 return snd_pcm_unlink(substream); 2895 case SNDRV_PCM_IOCTL_RESUME: 2896 return snd_pcm_resume(substream); 2897 case SNDRV_PCM_IOCTL_XRUN: 2898 return snd_pcm_xrun(substream); 2899 case SNDRV_PCM_IOCTL_HWSYNC: 2900 return snd_pcm_hwsync(substream); 2901 case SNDRV_PCM_IOCTL_DELAY: 2902 { 2903 snd_pcm_sframes_t delay; 2904 snd_pcm_sframes_t __user *res = arg; 2905 int err; 2906 2907 err = snd_pcm_delay(substream, &delay); 2908 if (err) 2909 return err; 2910 if (put_user(delay, res)) 2911 return -EFAULT; 2912 return 0; 2913 } 2914 case SNDRV_PCM_IOCTL_SYNC_PTR: 2915 return snd_pcm_sync_ptr(substream, arg); 2916 #ifdef CONFIG_SND_SUPPORT_OLD_API 2917 case SNDRV_PCM_IOCTL_HW_REFINE_OLD: 2918 return snd_pcm_hw_refine_old_user(substream, arg); 2919 case SNDRV_PCM_IOCTL_HW_PARAMS_OLD: 2920 return snd_pcm_hw_params_old_user(substream, arg); 2921 #endif 2922 case SNDRV_PCM_IOCTL_DRAIN: 2923 return snd_pcm_drain(substream, file); 2924 case SNDRV_PCM_IOCTL_DROP: 2925 return snd_pcm_drop(substream); 2926 case SNDRV_PCM_IOCTL_PAUSE: 2927 return snd_pcm_action_lock_irq(&snd_pcm_action_pause, 2928 substream, 2929 (int)(unsigned long)arg); 2930 case SNDRV_PCM_IOCTL_WRITEI_FRAMES: 2931 case SNDRV_PCM_IOCTL_READI_FRAMES: 2932 return snd_pcm_xferi_frames_ioctl(substream, arg); 2933 case SNDRV_PCM_IOCTL_WRITEN_FRAMES: 2934 case SNDRV_PCM_IOCTL_READN_FRAMES: 2935 return snd_pcm_xfern_frames_ioctl(substream, arg); 2936 case SNDRV_PCM_IOCTL_REWIND: 2937 return snd_pcm_rewind_ioctl(substream, arg); 2938 case SNDRV_PCM_IOCTL_FORWARD: 2939 return snd_pcm_forward_ioctl(substream, arg); 2940 } 2941 pcm_dbg(substream->pcm, "unknown ioctl = 0x%x\n", cmd); 2942 return -ENOTTY; 2943 } 2944 2945 static long snd_pcm_ioctl(struct file *file, unsigned int cmd, 2946 unsigned long arg) 2947 { 2948 struct snd_pcm_file *pcm_file; 2949 2950 pcm_file = file->private_data; 2951 2952 if (((cmd >> 8) & 0xff) != 'A') 2953 return -ENOTTY; 2954 2955 return snd_pcm_common_ioctl(file, pcm_file->substream, cmd, 2956 (void __user *)arg); 2957 } 2958 2959 /** 2960 * snd_pcm_kernel_ioctl - Execute PCM ioctl in the kernel-space 2961 * @substream: PCM substream 2962 * @cmd: IOCTL cmd 2963 * @arg: IOCTL argument 2964 * 2965 * The function is provided primarily for OSS layer and USB gadget drivers, 2966 * and it allows only the limited set of ioctls (hw_params, sw_params, 2967 * prepare, start, drain, drop, forward). 2968 */ 2969 int snd_pcm_kernel_ioctl(struct snd_pcm_substream *substream, 2970 unsigned int cmd, void *arg) 2971 { 2972 snd_pcm_uframes_t *frames = arg; 2973 snd_pcm_sframes_t result; 2974 2975 switch (cmd) { 2976 case SNDRV_PCM_IOCTL_FORWARD: 2977 { 2978 /* provided only for OSS; capture-only and no value returned */ 2979 if (substream->stream != SNDRV_PCM_STREAM_CAPTURE) 2980 return -EINVAL; 2981 result = snd_pcm_forward(substream, *frames); 2982 return result < 0 ? result : 0; 2983 } 2984 case SNDRV_PCM_IOCTL_HW_PARAMS: 2985 return snd_pcm_hw_params(substream, arg); 2986 case SNDRV_PCM_IOCTL_SW_PARAMS: 2987 return snd_pcm_sw_params(substream, arg); 2988 case SNDRV_PCM_IOCTL_PREPARE: 2989 return snd_pcm_prepare(substream, NULL); 2990 case SNDRV_PCM_IOCTL_START: 2991 return snd_pcm_start_lock_irq(substream); 2992 case SNDRV_PCM_IOCTL_DRAIN: 2993 return snd_pcm_drain(substream, NULL); 2994 case SNDRV_PCM_IOCTL_DROP: 2995 return snd_pcm_drop(substream); 2996 case SNDRV_PCM_IOCTL_DELAY: 2997 return snd_pcm_delay(substream, frames); 2998 default: 2999 return -EINVAL; 3000 } 3001 } 3002 EXPORT_SYMBOL(snd_pcm_kernel_ioctl); 3003 3004 static ssize_t snd_pcm_read(struct file *file, char __user *buf, size_t count, 3005 loff_t * offset) 3006 { 3007 struct snd_pcm_file *pcm_file; 3008 struct snd_pcm_substream *substream; 3009 struct snd_pcm_runtime *runtime; 3010 snd_pcm_sframes_t result; 3011 3012 pcm_file = file->private_data; 3013 substream = pcm_file->substream; 3014 if (PCM_RUNTIME_CHECK(substream)) 3015 return -ENXIO; 3016 runtime = substream->runtime; 3017 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3018 return -EBADFD; 3019 if (!frame_aligned(runtime, count)) 3020 return -EINVAL; 3021 count = bytes_to_frames(runtime, count); 3022 result = snd_pcm_lib_read(substream, buf, count); 3023 if (result > 0) 3024 result = frames_to_bytes(runtime, result); 3025 return result; 3026 } 3027 3028 static ssize_t snd_pcm_write(struct file *file, const char __user *buf, 3029 size_t count, loff_t * offset) 3030 { 3031 struct snd_pcm_file *pcm_file; 3032 struct snd_pcm_substream *substream; 3033 struct snd_pcm_runtime *runtime; 3034 snd_pcm_sframes_t result; 3035 3036 pcm_file = file->private_data; 3037 substream = pcm_file->substream; 3038 if (PCM_RUNTIME_CHECK(substream)) 3039 return -ENXIO; 3040 runtime = substream->runtime; 3041 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3042 return -EBADFD; 3043 if (!frame_aligned(runtime, count)) 3044 return -EINVAL; 3045 count = bytes_to_frames(runtime, count); 3046 result = snd_pcm_lib_write(substream, buf, count); 3047 if (result > 0) 3048 result = frames_to_bytes(runtime, result); 3049 return result; 3050 } 3051 3052 static ssize_t snd_pcm_readv(struct kiocb *iocb, struct iov_iter *to) 3053 { 3054 struct snd_pcm_file *pcm_file; 3055 struct snd_pcm_substream *substream; 3056 struct snd_pcm_runtime *runtime; 3057 snd_pcm_sframes_t result; 3058 unsigned long i; 3059 void __user **bufs; 3060 snd_pcm_uframes_t frames; 3061 3062 pcm_file = iocb->ki_filp->private_data; 3063 substream = pcm_file->substream; 3064 if (PCM_RUNTIME_CHECK(substream)) 3065 return -ENXIO; 3066 runtime = substream->runtime; 3067 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3068 return -EBADFD; 3069 if (!iter_is_iovec(to)) 3070 return -EINVAL; 3071 if (to->nr_segs > 1024 || to->nr_segs != runtime->channels) 3072 return -EINVAL; 3073 if (!frame_aligned(runtime, to->iov->iov_len)) 3074 return -EINVAL; 3075 frames = bytes_to_samples(runtime, to->iov->iov_len); 3076 bufs = kmalloc_array(to->nr_segs, sizeof(void *), GFP_KERNEL); 3077 if (bufs == NULL) 3078 return -ENOMEM; 3079 for (i = 0; i < to->nr_segs; ++i) 3080 bufs[i] = to->iov[i].iov_base; 3081 result = snd_pcm_lib_readv(substream, bufs, frames); 3082 if (result > 0) 3083 result = frames_to_bytes(runtime, result); 3084 kfree(bufs); 3085 return result; 3086 } 3087 3088 static ssize_t snd_pcm_writev(struct kiocb *iocb, struct iov_iter *from) 3089 { 3090 struct snd_pcm_file *pcm_file; 3091 struct snd_pcm_substream *substream; 3092 struct snd_pcm_runtime *runtime; 3093 snd_pcm_sframes_t result; 3094 unsigned long i; 3095 void __user **bufs; 3096 snd_pcm_uframes_t frames; 3097 3098 pcm_file = iocb->ki_filp->private_data; 3099 substream = pcm_file->substream; 3100 if (PCM_RUNTIME_CHECK(substream)) 3101 return -ENXIO; 3102 runtime = substream->runtime; 3103 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3104 return -EBADFD; 3105 if (!iter_is_iovec(from)) 3106 return -EINVAL; 3107 if (from->nr_segs > 128 || from->nr_segs != runtime->channels || 3108 !frame_aligned(runtime, from->iov->iov_len)) 3109 return -EINVAL; 3110 frames = bytes_to_samples(runtime, from->iov->iov_len); 3111 bufs = kmalloc_array(from->nr_segs, sizeof(void *), GFP_KERNEL); 3112 if (bufs == NULL) 3113 return -ENOMEM; 3114 for (i = 0; i < from->nr_segs; ++i) 3115 bufs[i] = from->iov[i].iov_base; 3116 result = snd_pcm_lib_writev(substream, bufs, frames); 3117 if (result > 0) 3118 result = frames_to_bytes(runtime, result); 3119 kfree(bufs); 3120 return result; 3121 } 3122 3123 static __poll_t snd_pcm_poll(struct file *file, poll_table *wait) 3124 { 3125 struct snd_pcm_file *pcm_file; 3126 struct snd_pcm_substream *substream; 3127 struct snd_pcm_runtime *runtime; 3128 __poll_t mask, ok; 3129 snd_pcm_uframes_t avail; 3130 3131 pcm_file = file->private_data; 3132 3133 substream = pcm_file->substream; 3134 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 3135 ok = EPOLLOUT | EPOLLWRNORM; 3136 else 3137 ok = EPOLLIN | EPOLLRDNORM; 3138 if (PCM_RUNTIME_CHECK(substream)) 3139 return ok | EPOLLERR; 3140 3141 runtime = substream->runtime; 3142 poll_wait(file, &runtime->sleep, wait); 3143 3144 mask = 0; 3145 snd_pcm_stream_lock_irq(substream); 3146 avail = snd_pcm_avail(substream); 3147 switch (runtime->status->state) { 3148 case SNDRV_PCM_STATE_RUNNING: 3149 case SNDRV_PCM_STATE_PREPARED: 3150 case SNDRV_PCM_STATE_PAUSED: 3151 if (avail >= runtime->control->avail_min) 3152 mask = ok; 3153 break; 3154 case SNDRV_PCM_STATE_DRAINING: 3155 if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) { 3156 mask = ok; 3157 if (!avail) 3158 mask |= EPOLLERR; 3159 } 3160 break; 3161 default: 3162 mask = ok | EPOLLERR; 3163 break; 3164 } 3165 snd_pcm_stream_unlock_irq(substream); 3166 return mask; 3167 } 3168 3169 /* 3170 * mmap support 3171 */ 3172 3173 /* 3174 * Only on coherent architectures, we can mmap the status and the control records 3175 * for effcient data transfer. On others, we have to use HWSYNC ioctl... 3176 */ 3177 #if defined(CONFIG_X86) || defined(CONFIG_PPC) || defined(CONFIG_ALPHA) 3178 /* 3179 * mmap status record 3180 */ 3181 static vm_fault_t snd_pcm_mmap_status_fault(struct vm_fault *vmf) 3182 { 3183 struct snd_pcm_substream *substream = vmf->vma->vm_private_data; 3184 struct snd_pcm_runtime *runtime; 3185 3186 if (substream == NULL) 3187 return VM_FAULT_SIGBUS; 3188 runtime = substream->runtime; 3189 vmf->page = virt_to_page(runtime->status); 3190 get_page(vmf->page); 3191 return 0; 3192 } 3193 3194 static const struct vm_operations_struct snd_pcm_vm_ops_status = 3195 { 3196 .fault = snd_pcm_mmap_status_fault, 3197 }; 3198 3199 static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file, 3200 struct vm_area_struct *area) 3201 { 3202 long size; 3203 if (!(area->vm_flags & VM_READ)) 3204 return -EINVAL; 3205 size = area->vm_end - area->vm_start; 3206 if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_status))) 3207 return -EINVAL; 3208 area->vm_ops = &snd_pcm_vm_ops_status; 3209 area->vm_private_data = substream; 3210 area->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; 3211 return 0; 3212 } 3213 3214 /* 3215 * mmap control record 3216 */ 3217 static vm_fault_t snd_pcm_mmap_control_fault(struct vm_fault *vmf) 3218 { 3219 struct snd_pcm_substream *substream = vmf->vma->vm_private_data; 3220 struct snd_pcm_runtime *runtime; 3221 3222 if (substream == NULL) 3223 return VM_FAULT_SIGBUS; 3224 runtime = substream->runtime; 3225 vmf->page = virt_to_page(runtime->control); 3226 get_page(vmf->page); 3227 return 0; 3228 } 3229 3230 static const struct vm_operations_struct snd_pcm_vm_ops_control = 3231 { 3232 .fault = snd_pcm_mmap_control_fault, 3233 }; 3234 3235 static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file, 3236 struct vm_area_struct *area) 3237 { 3238 long size; 3239 if (!(area->vm_flags & VM_READ)) 3240 return -EINVAL; 3241 size = area->vm_end - area->vm_start; 3242 if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_control))) 3243 return -EINVAL; 3244 area->vm_ops = &snd_pcm_vm_ops_control; 3245 area->vm_private_data = substream; 3246 area->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; 3247 return 0; 3248 } 3249 3250 static bool pcm_status_mmap_allowed(struct snd_pcm_file *pcm_file) 3251 { 3252 if (pcm_file->no_compat_mmap) 3253 return false; 3254 /* See pcm_control_mmap_allowed() below. 3255 * Since older alsa-lib requires both status and control mmaps to be 3256 * coupled, we have to disable the status mmap for old alsa-lib, too. 3257 */ 3258 if (pcm_file->user_pversion < SNDRV_PROTOCOL_VERSION(2, 0, 14) && 3259 (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR)) 3260 return false; 3261 return true; 3262 } 3263 3264 static bool pcm_control_mmap_allowed(struct snd_pcm_file *pcm_file) 3265 { 3266 if (pcm_file->no_compat_mmap) 3267 return false; 3268 /* Disallow the control mmap when SYNC_APPLPTR flag is set; 3269 * it enforces the user-space to fall back to snd_pcm_sync_ptr(), 3270 * thus it effectively assures the manual update of appl_ptr. 3271 */ 3272 if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR) 3273 return false; 3274 return true; 3275 } 3276 3277 #else /* ! coherent mmap */ 3278 /* 3279 * don't support mmap for status and control records. 3280 */ 3281 #define pcm_status_mmap_allowed(pcm_file) false 3282 #define pcm_control_mmap_allowed(pcm_file) false 3283 3284 static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file, 3285 struct vm_area_struct *area) 3286 { 3287 return -ENXIO; 3288 } 3289 static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file, 3290 struct vm_area_struct *area) 3291 { 3292 return -ENXIO; 3293 } 3294 #endif /* coherent mmap */ 3295 3296 static inline struct page * 3297 snd_pcm_default_page_ops(struct snd_pcm_substream *substream, unsigned long ofs) 3298 { 3299 void *vaddr = substream->runtime->dma_area + ofs; 3300 return virt_to_page(vaddr); 3301 } 3302 3303 /* 3304 * fault callback for mmapping a RAM page 3305 */ 3306 static vm_fault_t snd_pcm_mmap_data_fault(struct vm_fault *vmf) 3307 { 3308 struct snd_pcm_substream *substream = vmf->vma->vm_private_data; 3309 struct snd_pcm_runtime *runtime; 3310 unsigned long offset; 3311 struct page * page; 3312 size_t dma_bytes; 3313 3314 if (substream == NULL) 3315 return VM_FAULT_SIGBUS; 3316 runtime = substream->runtime; 3317 offset = vmf->pgoff << PAGE_SHIFT; 3318 dma_bytes = PAGE_ALIGN(runtime->dma_bytes); 3319 if (offset > dma_bytes - PAGE_SIZE) 3320 return VM_FAULT_SIGBUS; 3321 if (substream->ops->page) 3322 page = substream->ops->page(substream, offset); 3323 else 3324 page = snd_pcm_default_page_ops(substream, offset); 3325 if (!page) 3326 return VM_FAULT_SIGBUS; 3327 get_page(page); 3328 vmf->page = page; 3329 return 0; 3330 } 3331 3332 static const struct vm_operations_struct snd_pcm_vm_ops_data = { 3333 .open = snd_pcm_mmap_data_open, 3334 .close = snd_pcm_mmap_data_close, 3335 }; 3336 3337 static const struct vm_operations_struct snd_pcm_vm_ops_data_fault = { 3338 .open = snd_pcm_mmap_data_open, 3339 .close = snd_pcm_mmap_data_close, 3340 .fault = snd_pcm_mmap_data_fault, 3341 }; 3342 3343 /* 3344 * mmap the DMA buffer on RAM 3345 */ 3346 3347 /** 3348 * snd_pcm_lib_default_mmap - Default PCM data mmap function 3349 * @substream: PCM substream 3350 * @area: VMA 3351 * 3352 * This is the default mmap handler for PCM data. When mmap pcm_ops is NULL, 3353 * this function is invoked implicitly. 3354 */ 3355 int snd_pcm_lib_default_mmap(struct snd_pcm_substream *substream, 3356 struct vm_area_struct *area) 3357 { 3358 area->vm_flags |= VM_DONTEXPAND | VM_DONTDUMP; 3359 #ifdef CONFIG_GENERIC_ALLOCATOR 3360 if (substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV_IRAM) { 3361 area->vm_page_prot = pgprot_writecombine(area->vm_page_prot); 3362 return remap_pfn_range(area, area->vm_start, 3363 substream->dma_buffer.addr >> PAGE_SHIFT, 3364 area->vm_end - area->vm_start, area->vm_page_prot); 3365 } 3366 #endif /* CONFIG_GENERIC_ALLOCATOR */ 3367 #ifndef CONFIG_X86 /* for avoiding warnings arch/x86/mm/pat.c */ 3368 if (IS_ENABLED(CONFIG_HAS_DMA) && !substream->ops->page && 3369 substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV) 3370 return dma_mmap_coherent(substream->dma_buffer.dev.dev, 3371 area, 3372 substream->runtime->dma_area, 3373 substream->runtime->dma_addr, 3374 substream->runtime->dma_bytes); 3375 #endif /* CONFIG_X86 */ 3376 /* mmap with fault handler */ 3377 area->vm_ops = &snd_pcm_vm_ops_data_fault; 3378 return 0; 3379 } 3380 EXPORT_SYMBOL_GPL(snd_pcm_lib_default_mmap); 3381 3382 /* 3383 * mmap the DMA buffer on I/O memory area 3384 */ 3385 #if SNDRV_PCM_INFO_MMAP_IOMEM 3386 /** 3387 * snd_pcm_lib_mmap_iomem - Default PCM data mmap function for I/O mem 3388 * @substream: PCM substream 3389 * @area: VMA 3390 * 3391 * When your hardware uses the iomapped pages as the hardware buffer and 3392 * wants to mmap it, pass this function as mmap pcm_ops. Note that this 3393 * is supposed to work only on limited architectures. 3394 */ 3395 int snd_pcm_lib_mmap_iomem(struct snd_pcm_substream *substream, 3396 struct vm_area_struct *area) 3397 { 3398 struct snd_pcm_runtime *runtime = substream->runtime; 3399 3400 area->vm_page_prot = pgprot_noncached(area->vm_page_prot); 3401 return vm_iomap_memory(area, runtime->dma_addr, runtime->dma_bytes); 3402 } 3403 EXPORT_SYMBOL(snd_pcm_lib_mmap_iomem); 3404 #endif /* SNDRV_PCM_INFO_MMAP */ 3405 3406 /* 3407 * mmap DMA buffer 3408 */ 3409 int snd_pcm_mmap_data(struct snd_pcm_substream *substream, struct file *file, 3410 struct vm_area_struct *area) 3411 { 3412 struct snd_pcm_runtime *runtime; 3413 long size; 3414 unsigned long offset; 3415 size_t dma_bytes; 3416 int err; 3417 3418 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) { 3419 if (!(area->vm_flags & (VM_WRITE|VM_READ))) 3420 return -EINVAL; 3421 } else { 3422 if (!(area->vm_flags & VM_READ)) 3423 return -EINVAL; 3424 } 3425 runtime = substream->runtime; 3426 if (runtime->status->state == SNDRV_PCM_STATE_OPEN) 3427 return -EBADFD; 3428 if (!(runtime->info & SNDRV_PCM_INFO_MMAP)) 3429 return -ENXIO; 3430 if (runtime->access == SNDRV_PCM_ACCESS_RW_INTERLEAVED || 3431 runtime->access == SNDRV_PCM_ACCESS_RW_NONINTERLEAVED) 3432 return -EINVAL; 3433 size = area->vm_end - area->vm_start; 3434 offset = area->vm_pgoff << PAGE_SHIFT; 3435 dma_bytes = PAGE_ALIGN(runtime->dma_bytes); 3436 if ((size_t)size > dma_bytes) 3437 return -EINVAL; 3438 if (offset > dma_bytes - size) 3439 return -EINVAL; 3440 3441 area->vm_ops = &snd_pcm_vm_ops_data; 3442 area->vm_private_data = substream; 3443 if (substream->ops->mmap) 3444 err = substream->ops->mmap(substream, area); 3445 else 3446 err = snd_pcm_lib_default_mmap(substream, area); 3447 if (!err) 3448 atomic_inc(&substream->mmap_count); 3449 return err; 3450 } 3451 EXPORT_SYMBOL(snd_pcm_mmap_data); 3452 3453 static int snd_pcm_mmap(struct file *file, struct vm_area_struct *area) 3454 { 3455 struct snd_pcm_file * pcm_file; 3456 struct snd_pcm_substream *substream; 3457 unsigned long offset; 3458 3459 pcm_file = file->private_data; 3460 substream = pcm_file->substream; 3461 if (PCM_RUNTIME_CHECK(substream)) 3462 return -ENXIO; 3463 3464 offset = area->vm_pgoff << PAGE_SHIFT; 3465 switch (offset) { 3466 case SNDRV_PCM_MMAP_OFFSET_STATUS: 3467 if (!pcm_status_mmap_allowed(pcm_file)) 3468 return -ENXIO; 3469 return snd_pcm_mmap_status(substream, file, area); 3470 case SNDRV_PCM_MMAP_OFFSET_CONTROL: 3471 if (!pcm_control_mmap_allowed(pcm_file)) 3472 return -ENXIO; 3473 return snd_pcm_mmap_control(substream, file, area); 3474 default: 3475 return snd_pcm_mmap_data(substream, file, area); 3476 } 3477 return 0; 3478 } 3479 3480 static int snd_pcm_fasync(int fd, struct file * file, int on) 3481 { 3482 struct snd_pcm_file * pcm_file; 3483 struct snd_pcm_substream *substream; 3484 struct snd_pcm_runtime *runtime; 3485 3486 pcm_file = file->private_data; 3487 substream = pcm_file->substream; 3488 if (PCM_RUNTIME_CHECK(substream)) 3489 return -ENXIO; 3490 runtime = substream->runtime; 3491 return fasync_helper(fd, file, on, &runtime->fasync); 3492 } 3493 3494 /* 3495 * ioctl32 compat 3496 */ 3497 #ifdef CONFIG_COMPAT 3498 #include "pcm_compat.c" 3499 #else 3500 #define snd_pcm_ioctl_compat NULL 3501 #endif 3502 3503 /* 3504 * To be removed helpers to keep binary compatibility 3505 */ 3506 3507 #ifdef CONFIG_SND_SUPPORT_OLD_API 3508 #define __OLD_TO_NEW_MASK(x) ((x&7)|((x&0x07fffff8)<<5)) 3509 #define __NEW_TO_OLD_MASK(x) ((x&7)|((x&0xffffff00)>>5)) 3510 3511 static void snd_pcm_hw_convert_from_old_params(struct snd_pcm_hw_params *params, 3512 struct snd_pcm_hw_params_old *oparams) 3513 { 3514 unsigned int i; 3515 3516 memset(params, 0, sizeof(*params)); 3517 params->flags = oparams->flags; 3518 for (i = 0; i < ARRAY_SIZE(oparams->masks); i++) 3519 params->masks[i].bits[0] = oparams->masks[i]; 3520 memcpy(params->intervals, oparams->intervals, sizeof(oparams->intervals)); 3521 params->rmask = __OLD_TO_NEW_MASK(oparams->rmask); 3522 params->cmask = __OLD_TO_NEW_MASK(oparams->cmask); 3523 params->info = oparams->info; 3524 params->msbits = oparams->msbits; 3525 params->rate_num = oparams->rate_num; 3526 params->rate_den = oparams->rate_den; 3527 params->fifo_size = oparams->fifo_size; 3528 } 3529 3530 static void snd_pcm_hw_convert_to_old_params(struct snd_pcm_hw_params_old *oparams, 3531 struct snd_pcm_hw_params *params) 3532 { 3533 unsigned int i; 3534 3535 memset(oparams, 0, sizeof(*oparams)); 3536 oparams->flags = params->flags; 3537 for (i = 0; i < ARRAY_SIZE(oparams->masks); i++) 3538 oparams->masks[i] = params->masks[i].bits[0]; 3539 memcpy(oparams->intervals, params->intervals, sizeof(oparams->intervals)); 3540 oparams->rmask = __NEW_TO_OLD_MASK(params->rmask); 3541 oparams->cmask = __NEW_TO_OLD_MASK(params->cmask); 3542 oparams->info = params->info; 3543 oparams->msbits = params->msbits; 3544 oparams->rate_num = params->rate_num; 3545 oparams->rate_den = params->rate_den; 3546 oparams->fifo_size = params->fifo_size; 3547 } 3548 3549 static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream, 3550 struct snd_pcm_hw_params_old __user * _oparams) 3551 { 3552 struct snd_pcm_hw_params *params; 3553 struct snd_pcm_hw_params_old *oparams = NULL; 3554 int err; 3555 3556 params = kmalloc(sizeof(*params), GFP_KERNEL); 3557 if (!params) 3558 return -ENOMEM; 3559 3560 oparams = memdup_user(_oparams, sizeof(*oparams)); 3561 if (IS_ERR(oparams)) { 3562 err = PTR_ERR(oparams); 3563 goto out; 3564 } 3565 snd_pcm_hw_convert_from_old_params(params, oparams); 3566 err = snd_pcm_hw_refine(substream, params); 3567 if (err < 0) 3568 goto out_old; 3569 3570 err = fixup_unreferenced_params(substream, params); 3571 if (err < 0) 3572 goto out_old; 3573 3574 snd_pcm_hw_convert_to_old_params(oparams, params); 3575 if (copy_to_user(_oparams, oparams, sizeof(*oparams))) 3576 err = -EFAULT; 3577 out_old: 3578 kfree(oparams); 3579 out: 3580 kfree(params); 3581 return err; 3582 } 3583 3584 static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream, 3585 struct snd_pcm_hw_params_old __user * _oparams) 3586 { 3587 struct snd_pcm_hw_params *params; 3588 struct snd_pcm_hw_params_old *oparams = NULL; 3589 int err; 3590 3591 params = kmalloc(sizeof(*params), GFP_KERNEL); 3592 if (!params) 3593 return -ENOMEM; 3594 3595 oparams = memdup_user(_oparams, sizeof(*oparams)); 3596 if (IS_ERR(oparams)) { 3597 err = PTR_ERR(oparams); 3598 goto out; 3599 } 3600 3601 snd_pcm_hw_convert_from_old_params(params, oparams); 3602 err = snd_pcm_hw_params(substream, params); 3603 if (err < 0) 3604 goto out_old; 3605 3606 snd_pcm_hw_convert_to_old_params(oparams, params); 3607 if (copy_to_user(_oparams, oparams, sizeof(*oparams))) 3608 err = -EFAULT; 3609 out_old: 3610 kfree(oparams); 3611 out: 3612 kfree(params); 3613 return err; 3614 } 3615 #endif /* CONFIG_SND_SUPPORT_OLD_API */ 3616 3617 #ifndef CONFIG_MMU 3618 static unsigned long snd_pcm_get_unmapped_area(struct file *file, 3619 unsigned long addr, 3620 unsigned long len, 3621 unsigned long pgoff, 3622 unsigned long flags) 3623 { 3624 struct snd_pcm_file *pcm_file = file->private_data; 3625 struct snd_pcm_substream *substream = pcm_file->substream; 3626 struct snd_pcm_runtime *runtime = substream->runtime; 3627 unsigned long offset = pgoff << PAGE_SHIFT; 3628 3629 switch (offset) { 3630 case SNDRV_PCM_MMAP_OFFSET_STATUS: 3631 return (unsigned long)runtime->status; 3632 case SNDRV_PCM_MMAP_OFFSET_CONTROL: 3633 return (unsigned long)runtime->control; 3634 default: 3635 return (unsigned long)runtime->dma_area + offset; 3636 } 3637 } 3638 #else 3639 # define snd_pcm_get_unmapped_area NULL 3640 #endif 3641 3642 /* 3643 * Register section 3644 */ 3645 3646 const struct file_operations snd_pcm_f_ops[2] = { 3647 { 3648 .owner = THIS_MODULE, 3649 .write = snd_pcm_write, 3650 .write_iter = snd_pcm_writev, 3651 .open = snd_pcm_playback_open, 3652 .release = snd_pcm_release, 3653 .llseek = no_llseek, 3654 .poll = snd_pcm_poll, 3655 .unlocked_ioctl = snd_pcm_ioctl, 3656 .compat_ioctl = snd_pcm_ioctl_compat, 3657 .mmap = snd_pcm_mmap, 3658 .fasync = snd_pcm_fasync, 3659 .get_unmapped_area = snd_pcm_get_unmapped_area, 3660 }, 3661 { 3662 .owner = THIS_MODULE, 3663 .read = snd_pcm_read, 3664 .read_iter = snd_pcm_readv, 3665 .open = snd_pcm_capture_open, 3666 .release = snd_pcm_release, 3667 .llseek = no_llseek, 3668 .poll = snd_pcm_poll, 3669 .unlocked_ioctl = snd_pcm_ioctl, 3670 .compat_ioctl = snd_pcm_ioctl_compat, 3671 .mmap = snd_pcm_mmap, 3672 .fasync = snd_pcm_fasync, 3673 .get_unmapped_area = snd_pcm_get_unmapped_area, 3674 } 3675 }; 3676