1 /* 2 * 3 * Implementation of primary alsa driver code base for Intel HD Audio. 4 * 5 * Copyright(c) 2004 Intel Corporation. All rights reserved. 6 * 7 * Copyright (c) 2004 Takashi Iwai <tiwai@suse.de> 8 * PeiSen Hou <pshou@realtek.com.tw> 9 * 10 * This program is free software; you can redistribute it and/or modify it 11 * under the terms of the GNU General Public License as published by the Free 12 * Software Foundation; either version 2 of the License, or (at your option) 13 * any later version. 14 * 15 * This program is distributed in the hope that it will be useful, but WITHOUT 16 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 17 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 18 * more details. 19 * 20 * 21 */ 22 23 #include <linux/clocksource.h> 24 #include <linux/delay.h> 25 #include <linux/interrupt.h> 26 #include <linux/kernel.h> 27 #include <linux/module.h> 28 #include <linux/pm_runtime.h> 29 #include <linux/slab.h> 30 31 #ifdef CONFIG_X86 32 /* for art-tsc conversion */ 33 #include <asm/tsc.h> 34 #endif 35 36 #include <sound/core.h> 37 #include <sound/initval.h> 38 #include "hda_controller.h" 39 40 #define CREATE_TRACE_POINTS 41 #include "hda_controller_trace.h" 42 43 /* DSP lock helpers */ 44 #define dsp_lock(dev) snd_hdac_dsp_lock(azx_stream(dev)) 45 #define dsp_unlock(dev) snd_hdac_dsp_unlock(azx_stream(dev)) 46 #define dsp_is_locked(dev) snd_hdac_stream_is_locked(azx_stream(dev)) 47 48 /* assign a stream for the PCM */ 49 static inline struct azx_dev * 50 azx_assign_device(struct azx *chip, struct snd_pcm_substream *substream) 51 { 52 struct hdac_stream *s; 53 54 s = snd_hdac_stream_assign(azx_bus(chip), substream); 55 if (!s) 56 return NULL; 57 return stream_to_azx_dev(s); 58 } 59 60 /* release the assigned stream */ 61 static inline void azx_release_device(struct azx_dev *azx_dev) 62 { 63 snd_hdac_stream_release(azx_stream(azx_dev)); 64 } 65 66 static inline struct hda_pcm_stream * 67 to_hda_pcm_stream(struct snd_pcm_substream *substream) 68 { 69 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 70 return &apcm->info->stream[substream->stream]; 71 } 72 73 static u64 azx_adjust_codec_delay(struct snd_pcm_substream *substream, 74 u64 nsec) 75 { 76 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 77 struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream); 78 u64 codec_frames, codec_nsecs; 79 80 if (!hinfo->ops.get_delay) 81 return nsec; 82 83 codec_frames = hinfo->ops.get_delay(hinfo, apcm->codec, substream); 84 codec_nsecs = div_u64(codec_frames * 1000000000LL, 85 substream->runtime->rate); 86 87 if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) 88 return nsec + codec_nsecs; 89 90 return (nsec > codec_nsecs) ? nsec - codec_nsecs : 0; 91 } 92 93 /* 94 * PCM ops 95 */ 96 97 static int azx_pcm_close(struct snd_pcm_substream *substream) 98 { 99 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 100 struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream); 101 struct azx *chip = apcm->chip; 102 struct azx_dev *azx_dev = get_azx_dev(substream); 103 104 trace_azx_pcm_close(chip, azx_dev); 105 mutex_lock(&chip->open_mutex); 106 azx_release_device(azx_dev); 107 if (hinfo->ops.close) 108 hinfo->ops.close(hinfo, apcm->codec, substream); 109 snd_hda_power_down(apcm->codec); 110 mutex_unlock(&chip->open_mutex); 111 snd_hda_codec_pcm_put(apcm->info); 112 return 0; 113 } 114 115 static int azx_pcm_hw_params(struct snd_pcm_substream *substream, 116 struct snd_pcm_hw_params *hw_params) 117 { 118 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 119 struct azx *chip = apcm->chip; 120 struct azx_dev *azx_dev = get_azx_dev(substream); 121 int ret; 122 123 trace_azx_pcm_hw_params(chip, azx_dev); 124 dsp_lock(azx_dev); 125 if (dsp_is_locked(azx_dev)) { 126 ret = -EBUSY; 127 goto unlock; 128 } 129 130 azx_dev->core.bufsize = 0; 131 azx_dev->core.period_bytes = 0; 132 azx_dev->core.format_val = 0; 133 ret = snd_pcm_lib_malloc_pages(substream, 134 params_buffer_bytes(hw_params)); 135 136 unlock: 137 dsp_unlock(azx_dev); 138 return ret; 139 } 140 141 static int azx_pcm_hw_free(struct snd_pcm_substream *substream) 142 { 143 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 144 struct azx_dev *azx_dev = get_azx_dev(substream); 145 struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream); 146 int err; 147 148 /* reset BDL address */ 149 dsp_lock(azx_dev); 150 if (!dsp_is_locked(azx_dev)) 151 snd_hdac_stream_cleanup(azx_stream(azx_dev)); 152 153 snd_hda_codec_cleanup(apcm->codec, hinfo, substream); 154 155 err = snd_pcm_lib_free_pages(substream); 156 azx_stream(azx_dev)->prepared = 0; 157 dsp_unlock(azx_dev); 158 return err; 159 } 160 161 static int azx_pcm_prepare(struct snd_pcm_substream *substream) 162 { 163 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 164 struct azx *chip = apcm->chip; 165 struct azx_dev *azx_dev = get_azx_dev(substream); 166 struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream); 167 struct snd_pcm_runtime *runtime = substream->runtime; 168 unsigned int format_val, stream_tag; 169 int err; 170 struct hda_spdif_out *spdif = 171 snd_hda_spdif_out_of_nid(apcm->codec, hinfo->nid); 172 unsigned short ctls = spdif ? spdif->ctls : 0; 173 174 trace_azx_pcm_prepare(chip, azx_dev); 175 dsp_lock(azx_dev); 176 if (dsp_is_locked(azx_dev)) { 177 err = -EBUSY; 178 goto unlock; 179 } 180 181 snd_hdac_stream_reset(azx_stream(azx_dev)); 182 format_val = snd_hdac_calc_stream_format(runtime->rate, 183 runtime->channels, 184 runtime->format, 185 hinfo->maxbps, 186 ctls); 187 if (!format_val) { 188 dev_err(chip->card->dev, 189 "invalid format_val, rate=%d, ch=%d, format=%d\n", 190 runtime->rate, runtime->channels, runtime->format); 191 err = -EINVAL; 192 goto unlock; 193 } 194 195 err = snd_hdac_stream_set_params(azx_stream(azx_dev), format_val); 196 if (err < 0) 197 goto unlock; 198 199 snd_hdac_stream_setup(azx_stream(azx_dev)); 200 201 stream_tag = azx_dev->core.stream_tag; 202 /* CA-IBG chips need the playback stream starting from 1 */ 203 if ((chip->driver_caps & AZX_DCAPS_CTX_WORKAROUND) && 204 stream_tag > chip->capture_streams) 205 stream_tag -= chip->capture_streams; 206 err = snd_hda_codec_prepare(apcm->codec, hinfo, stream_tag, 207 azx_dev->core.format_val, substream); 208 209 unlock: 210 if (!err) 211 azx_stream(azx_dev)->prepared = 1; 212 dsp_unlock(azx_dev); 213 return err; 214 } 215 216 static int azx_pcm_trigger(struct snd_pcm_substream *substream, int cmd) 217 { 218 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 219 struct azx *chip = apcm->chip; 220 struct hdac_bus *bus = azx_bus(chip); 221 struct azx_dev *azx_dev; 222 struct snd_pcm_substream *s; 223 struct hdac_stream *hstr; 224 bool start; 225 int sbits = 0; 226 int sync_reg; 227 228 azx_dev = get_azx_dev(substream); 229 trace_azx_pcm_trigger(chip, azx_dev, cmd); 230 231 hstr = azx_stream(azx_dev); 232 if (chip->driver_caps & AZX_DCAPS_OLD_SSYNC) 233 sync_reg = AZX_REG_OLD_SSYNC; 234 else 235 sync_reg = AZX_REG_SSYNC; 236 237 if (dsp_is_locked(azx_dev) || !hstr->prepared) 238 return -EPIPE; 239 240 switch (cmd) { 241 case SNDRV_PCM_TRIGGER_START: 242 case SNDRV_PCM_TRIGGER_PAUSE_RELEASE: 243 case SNDRV_PCM_TRIGGER_RESUME: 244 start = true; 245 break; 246 case SNDRV_PCM_TRIGGER_PAUSE_PUSH: 247 case SNDRV_PCM_TRIGGER_SUSPEND: 248 case SNDRV_PCM_TRIGGER_STOP: 249 start = false; 250 break; 251 default: 252 return -EINVAL; 253 } 254 255 snd_pcm_group_for_each_entry(s, substream) { 256 if (s->pcm->card != substream->pcm->card) 257 continue; 258 azx_dev = get_azx_dev(s); 259 sbits |= 1 << azx_dev->core.index; 260 snd_pcm_trigger_done(s, substream); 261 } 262 263 spin_lock(&bus->reg_lock); 264 265 /* first, set SYNC bits of corresponding streams */ 266 snd_hdac_stream_sync_trigger(hstr, true, sbits, sync_reg); 267 268 snd_pcm_group_for_each_entry(s, substream) { 269 if (s->pcm->card != substream->pcm->card) 270 continue; 271 azx_dev = get_azx_dev(s); 272 if (start) { 273 azx_dev->insufficient = 1; 274 snd_hdac_stream_start(azx_stream(azx_dev), true); 275 } else { 276 snd_hdac_stream_stop(azx_stream(azx_dev)); 277 } 278 } 279 spin_unlock(&bus->reg_lock); 280 281 snd_hdac_stream_sync(hstr, start, sbits); 282 283 spin_lock(&bus->reg_lock); 284 /* reset SYNC bits */ 285 snd_hdac_stream_sync_trigger(hstr, false, sbits, sync_reg); 286 if (start) 287 snd_hdac_stream_timecounter_init(hstr, sbits); 288 spin_unlock(&bus->reg_lock); 289 return 0; 290 } 291 292 unsigned int azx_get_pos_lpib(struct azx *chip, struct azx_dev *azx_dev) 293 { 294 return snd_hdac_stream_get_pos_lpib(azx_stream(azx_dev)); 295 } 296 EXPORT_SYMBOL_GPL(azx_get_pos_lpib); 297 298 unsigned int azx_get_pos_posbuf(struct azx *chip, struct azx_dev *azx_dev) 299 { 300 return snd_hdac_stream_get_pos_posbuf(azx_stream(azx_dev)); 301 } 302 EXPORT_SYMBOL_GPL(azx_get_pos_posbuf); 303 304 unsigned int azx_get_position(struct azx *chip, 305 struct azx_dev *azx_dev) 306 { 307 struct snd_pcm_substream *substream = azx_dev->core.substream; 308 unsigned int pos; 309 int stream = substream->stream; 310 int delay = 0; 311 312 if (chip->get_position[stream]) 313 pos = chip->get_position[stream](chip, azx_dev); 314 else /* use the position buffer as default */ 315 pos = azx_get_pos_posbuf(chip, azx_dev); 316 317 if (pos >= azx_dev->core.bufsize) 318 pos = 0; 319 320 if (substream->runtime) { 321 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 322 struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream); 323 324 if (chip->get_delay[stream]) 325 delay += chip->get_delay[stream](chip, azx_dev, pos); 326 if (hinfo->ops.get_delay) 327 delay += hinfo->ops.get_delay(hinfo, apcm->codec, 328 substream); 329 substream->runtime->delay = delay; 330 } 331 332 trace_azx_get_position(chip, azx_dev, pos, delay); 333 return pos; 334 } 335 EXPORT_SYMBOL_GPL(azx_get_position); 336 337 static snd_pcm_uframes_t azx_pcm_pointer(struct snd_pcm_substream *substream) 338 { 339 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 340 struct azx *chip = apcm->chip; 341 struct azx_dev *azx_dev = get_azx_dev(substream); 342 return bytes_to_frames(substream->runtime, 343 azx_get_position(chip, azx_dev)); 344 } 345 346 /* 347 * azx_scale64: Scale base by mult/div while not overflowing sanely 348 * 349 * Derived from scale64_check_overflow in kernel/time/timekeeping.c 350 * 351 * The tmestamps for a 48Khz stream can overflow after (2^64/10^9)/48K which 352 * is about 384307 ie ~4.5 days. 353 * 354 * This scales the calculation so that overflow will happen but after 2^64 / 355 * 48000 secs, which is pretty large! 356 * 357 * In caln below: 358 * base may overflow, but since there isn’t any additional division 359 * performed on base it’s OK 360 * rem can’t overflow because both are 32-bit values 361 */ 362 363 #ifdef CONFIG_X86 364 static u64 azx_scale64(u64 base, u32 num, u32 den) 365 { 366 u64 rem; 367 368 rem = do_div(base, den); 369 370 base *= num; 371 rem *= num; 372 373 do_div(rem, den); 374 375 return base + rem; 376 } 377 378 static int azx_get_sync_time(ktime_t *device, 379 struct system_counterval_t *system, void *ctx) 380 { 381 struct snd_pcm_substream *substream = ctx; 382 struct azx_dev *azx_dev = get_azx_dev(substream); 383 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 384 struct azx *chip = apcm->chip; 385 struct snd_pcm_runtime *runtime; 386 u64 ll_counter, ll_counter_l, ll_counter_h; 387 u64 tsc_counter, tsc_counter_l, tsc_counter_h; 388 u32 wallclk_ctr, wallclk_cycles; 389 bool direction; 390 u32 dma_select; 391 u32 timeout = 200; 392 u32 retry_count = 0; 393 394 runtime = substream->runtime; 395 396 if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) 397 direction = 1; 398 else 399 direction = 0; 400 401 /* 0th stream tag is not used, so DMA ch 0 is for 1st stream tag */ 402 do { 403 timeout = 100; 404 dma_select = (direction << GTSCC_CDMAS_DMA_DIR_SHIFT) | 405 (azx_dev->core.stream_tag - 1); 406 snd_hdac_chip_writel(azx_bus(chip), GTSCC, dma_select); 407 408 /* Enable the capture */ 409 snd_hdac_chip_updatel(azx_bus(chip), GTSCC, 0, GTSCC_TSCCI_MASK); 410 411 while (timeout) { 412 if (snd_hdac_chip_readl(azx_bus(chip), GTSCC) & 413 GTSCC_TSCCD_MASK) 414 break; 415 416 timeout--; 417 } 418 419 if (!timeout) { 420 dev_err(chip->card->dev, "GTSCC capture Timedout!\n"); 421 return -EIO; 422 } 423 424 /* Read wall clock counter */ 425 wallclk_ctr = snd_hdac_chip_readl(azx_bus(chip), WALFCC); 426 427 /* Read TSC counter */ 428 tsc_counter_l = snd_hdac_chip_readl(azx_bus(chip), TSCCL); 429 tsc_counter_h = snd_hdac_chip_readl(azx_bus(chip), TSCCU); 430 431 /* Read Link counter */ 432 ll_counter_l = snd_hdac_chip_readl(azx_bus(chip), LLPCL); 433 ll_counter_h = snd_hdac_chip_readl(azx_bus(chip), LLPCU); 434 435 /* Ack: registers read done */ 436 snd_hdac_chip_writel(azx_bus(chip), GTSCC, GTSCC_TSCCD_SHIFT); 437 438 tsc_counter = (tsc_counter_h << TSCCU_CCU_SHIFT) | 439 tsc_counter_l; 440 441 ll_counter = (ll_counter_h << LLPC_CCU_SHIFT) | ll_counter_l; 442 wallclk_cycles = wallclk_ctr & WALFCC_CIF_MASK; 443 444 /* 445 * An error occurs near frame "rollover". The clocks in 446 * frame value indicates whether this error may have 447 * occurred. Here we use the value of 10 i.e., 448 * HDA_MAX_CYCLE_OFFSET 449 */ 450 if (wallclk_cycles < HDA_MAX_CYCLE_VALUE - HDA_MAX_CYCLE_OFFSET 451 && wallclk_cycles > HDA_MAX_CYCLE_OFFSET) 452 break; 453 454 /* 455 * Sleep before we read again, else we may again get 456 * value near to MAX_CYCLE. Try to sleep for different 457 * amount of time so we dont hit the same number again 458 */ 459 udelay(retry_count++); 460 461 } while (retry_count != HDA_MAX_CYCLE_READ_RETRY); 462 463 if (retry_count == HDA_MAX_CYCLE_READ_RETRY) { 464 dev_err_ratelimited(chip->card->dev, 465 "Error in WALFCC cycle count\n"); 466 return -EIO; 467 } 468 469 *device = ns_to_ktime(azx_scale64(ll_counter, 470 NSEC_PER_SEC, runtime->rate)); 471 *device = ktime_add_ns(*device, (wallclk_cycles * NSEC_PER_SEC) / 472 ((HDA_MAX_CYCLE_VALUE + 1) * runtime->rate)); 473 474 *system = convert_art_to_tsc(tsc_counter); 475 476 return 0; 477 } 478 479 #else 480 static int azx_get_sync_time(ktime_t *device, 481 struct system_counterval_t *system, void *ctx) 482 { 483 return -ENXIO; 484 } 485 #endif 486 487 static int azx_get_crosststamp(struct snd_pcm_substream *substream, 488 struct system_device_crosststamp *xtstamp) 489 { 490 return get_device_system_crosststamp(azx_get_sync_time, 491 substream, NULL, xtstamp); 492 } 493 494 static inline bool is_link_time_supported(struct snd_pcm_runtime *runtime, 495 struct snd_pcm_audio_tstamp_config *ts) 496 { 497 if (runtime->hw.info & SNDRV_PCM_INFO_HAS_LINK_SYNCHRONIZED_ATIME) 498 if (ts->type_requested == SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK_SYNCHRONIZED) 499 return true; 500 501 return false; 502 } 503 504 static int azx_get_time_info(struct snd_pcm_substream *substream, 505 struct timespec *system_ts, struct timespec *audio_ts, 506 struct snd_pcm_audio_tstamp_config *audio_tstamp_config, 507 struct snd_pcm_audio_tstamp_report *audio_tstamp_report) 508 { 509 struct azx_dev *azx_dev = get_azx_dev(substream); 510 struct snd_pcm_runtime *runtime = substream->runtime; 511 struct system_device_crosststamp xtstamp; 512 int ret; 513 u64 nsec; 514 515 if ((substream->runtime->hw.info & SNDRV_PCM_INFO_HAS_LINK_ATIME) && 516 (audio_tstamp_config->type_requested == SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK)) { 517 518 snd_pcm_gettime(substream->runtime, system_ts); 519 520 nsec = timecounter_read(&azx_dev->core.tc); 521 nsec = div_u64(nsec, 3); /* can be optimized */ 522 if (audio_tstamp_config->report_delay) 523 nsec = azx_adjust_codec_delay(substream, nsec); 524 525 *audio_ts = ns_to_timespec(nsec); 526 527 audio_tstamp_report->actual_type = SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK; 528 audio_tstamp_report->accuracy_report = 1; /* rest of structure is valid */ 529 audio_tstamp_report->accuracy = 42; /* 24 MHz WallClock == 42ns resolution */ 530 531 } else if (is_link_time_supported(runtime, audio_tstamp_config)) { 532 533 ret = azx_get_crosststamp(substream, &xtstamp); 534 if (ret) 535 return ret; 536 537 switch (runtime->tstamp_type) { 538 case SNDRV_PCM_TSTAMP_TYPE_MONOTONIC: 539 return -EINVAL; 540 541 case SNDRV_PCM_TSTAMP_TYPE_MONOTONIC_RAW: 542 *system_ts = ktime_to_timespec(xtstamp.sys_monoraw); 543 break; 544 545 default: 546 *system_ts = ktime_to_timespec(xtstamp.sys_realtime); 547 break; 548 549 } 550 551 *audio_ts = ktime_to_timespec(xtstamp.device); 552 553 audio_tstamp_report->actual_type = 554 SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK_SYNCHRONIZED; 555 audio_tstamp_report->accuracy_report = 1; 556 /* 24 MHz WallClock == 42ns resolution */ 557 audio_tstamp_report->accuracy = 42; 558 559 } else { 560 audio_tstamp_report->actual_type = SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT; 561 } 562 563 return 0; 564 } 565 566 static struct snd_pcm_hardware azx_pcm_hw = { 567 .info = (SNDRV_PCM_INFO_MMAP | 568 SNDRV_PCM_INFO_INTERLEAVED | 569 SNDRV_PCM_INFO_BLOCK_TRANSFER | 570 SNDRV_PCM_INFO_MMAP_VALID | 571 /* No full-resume yet implemented */ 572 /* SNDRV_PCM_INFO_RESUME |*/ 573 SNDRV_PCM_INFO_PAUSE | 574 SNDRV_PCM_INFO_SYNC_START | 575 SNDRV_PCM_INFO_HAS_WALL_CLOCK | /* legacy */ 576 SNDRV_PCM_INFO_HAS_LINK_ATIME | 577 SNDRV_PCM_INFO_NO_PERIOD_WAKEUP), 578 .formats = SNDRV_PCM_FMTBIT_S16_LE, 579 .rates = SNDRV_PCM_RATE_48000, 580 .rate_min = 48000, 581 .rate_max = 48000, 582 .channels_min = 2, 583 .channels_max = 2, 584 .buffer_bytes_max = AZX_MAX_BUF_SIZE, 585 .period_bytes_min = 128, 586 .period_bytes_max = AZX_MAX_BUF_SIZE / 2, 587 .periods_min = 2, 588 .periods_max = AZX_MAX_FRAG, 589 .fifo_size = 0, 590 }; 591 592 static int azx_pcm_open(struct snd_pcm_substream *substream) 593 { 594 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 595 struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream); 596 struct azx *chip = apcm->chip; 597 struct azx_dev *azx_dev; 598 struct snd_pcm_runtime *runtime = substream->runtime; 599 int err; 600 int buff_step; 601 602 snd_hda_codec_pcm_get(apcm->info); 603 mutex_lock(&chip->open_mutex); 604 azx_dev = azx_assign_device(chip, substream); 605 trace_azx_pcm_open(chip, azx_dev); 606 if (azx_dev == NULL) { 607 err = -EBUSY; 608 goto unlock; 609 } 610 runtime->private_data = azx_dev; 611 612 if (chip->gts_present) 613 azx_pcm_hw.info = azx_pcm_hw.info | 614 SNDRV_PCM_INFO_HAS_LINK_SYNCHRONIZED_ATIME; 615 616 runtime->hw = azx_pcm_hw; 617 runtime->hw.channels_min = hinfo->channels_min; 618 runtime->hw.channels_max = hinfo->channels_max; 619 runtime->hw.formats = hinfo->formats; 620 runtime->hw.rates = hinfo->rates; 621 snd_pcm_limit_hw_rates(runtime); 622 snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIODS); 623 624 /* avoid wrap-around with wall-clock */ 625 snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_TIME, 626 20, 627 178000000); 628 629 if (chip->align_buffer_size) 630 /* constrain buffer sizes to be multiple of 128 631 bytes. This is more efficient in terms of memory 632 access but isn't required by the HDA spec and 633 prevents users from specifying exact period/buffer 634 sizes. For example for 44.1kHz, a period size set 635 to 20ms will be rounded to 19.59ms. */ 636 buff_step = 128; 637 else 638 /* Don't enforce steps on buffer sizes, still need to 639 be multiple of 4 bytes (HDA spec). Tested on Intel 640 HDA controllers, may not work on all devices where 641 option needs to be disabled */ 642 buff_step = 4; 643 644 snd_pcm_hw_constraint_step(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 645 buff_step); 646 snd_pcm_hw_constraint_step(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, 647 buff_step); 648 snd_hda_power_up(apcm->codec); 649 if (hinfo->ops.open) 650 err = hinfo->ops.open(hinfo, apcm->codec, substream); 651 else 652 err = -ENODEV; 653 if (err < 0) { 654 azx_release_device(azx_dev); 655 goto powerdown; 656 } 657 snd_pcm_limit_hw_rates(runtime); 658 /* sanity check */ 659 if (snd_BUG_ON(!runtime->hw.channels_min) || 660 snd_BUG_ON(!runtime->hw.channels_max) || 661 snd_BUG_ON(!runtime->hw.formats) || 662 snd_BUG_ON(!runtime->hw.rates)) { 663 azx_release_device(azx_dev); 664 if (hinfo->ops.close) 665 hinfo->ops.close(hinfo, apcm->codec, substream); 666 err = -EINVAL; 667 goto powerdown; 668 } 669 670 /* disable LINK_ATIME timestamps for capture streams 671 until we figure out how to handle digital inputs */ 672 if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) { 673 runtime->hw.info &= ~SNDRV_PCM_INFO_HAS_WALL_CLOCK; /* legacy */ 674 runtime->hw.info &= ~SNDRV_PCM_INFO_HAS_LINK_ATIME; 675 } 676 677 snd_pcm_set_sync(substream); 678 mutex_unlock(&chip->open_mutex); 679 return 0; 680 681 powerdown: 682 snd_hda_power_down(apcm->codec); 683 unlock: 684 mutex_unlock(&chip->open_mutex); 685 snd_hda_codec_pcm_put(apcm->info); 686 return err; 687 } 688 689 static int azx_pcm_mmap(struct snd_pcm_substream *substream, 690 struct vm_area_struct *area) 691 { 692 struct azx_pcm *apcm = snd_pcm_substream_chip(substream); 693 struct azx *chip = apcm->chip; 694 if (chip->ops->pcm_mmap_prepare) 695 chip->ops->pcm_mmap_prepare(substream, area); 696 return snd_pcm_lib_default_mmap(substream, area); 697 } 698 699 static const struct snd_pcm_ops azx_pcm_ops = { 700 .open = azx_pcm_open, 701 .close = azx_pcm_close, 702 .ioctl = snd_pcm_lib_ioctl, 703 .hw_params = azx_pcm_hw_params, 704 .hw_free = azx_pcm_hw_free, 705 .prepare = azx_pcm_prepare, 706 .trigger = azx_pcm_trigger, 707 .pointer = azx_pcm_pointer, 708 .get_time_info = azx_get_time_info, 709 .mmap = azx_pcm_mmap, 710 .page = snd_pcm_sgbuf_ops_page, 711 }; 712 713 static void azx_pcm_free(struct snd_pcm *pcm) 714 { 715 struct azx_pcm *apcm = pcm->private_data; 716 if (apcm) { 717 list_del(&apcm->list); 718 apcm->info->pcm = NULL; 719 kfree(apcm); 720 } 721 } 722 723 #define MAX_PREALLOC_SIZE (32 * 1024 * 1024) 724 725 int snd_hda_attach_pcm_stream(struct hda_bus *_bus, struct hda_codec *codec, 726 struct hda_pcm *cpcm) 727 { 728 struct hdac_bus *bus = &_bus->core; 729 struct azx *chip = bus_to_azx(bus); 730 struct snd_pcm *pcm; 731 struct azx_pcm *apcm; 732 int pcm_dev = cpcm->device; 733 unsigned int size; 734 int s, err; 735 int type = SNDRV_DMA_TYPE_DEV_SG; 736 737 list_for_each_entry(apcm, &chip->pcm_list, list) { 738 if (apcm->pcm->device == pcm_dev) { 739 dev_err(chip->card->dev, "PCM %d already exists\n", 740 pcm_dev); 741 return -EBUSY; 742 } 743 } 744 err = snd_pcm_new(chip->card, cpcm->name, pcm_dev, 745 cpcm->stream[SNDRV_PCM_STREAM_PLAYBACK].substreams, 746 cpcm->stream[SNDRV_PCM_STREAM_CAPTURE].substreams, 747 &pcm); 748 if (err < 0) 749 return err; 750 strlcpy(pcm->name, cpcm->name, sizeof(pcm->name)); 751 apcm = kzalloc(sizeof(*apcm), GFP_KERNEL); 752 if (apcm == NULL) { 753 snd_device_free(chip->card, pcm); 754 return -ENOMEM; 755 } 756 apcm->chip = chip; 757 apcm->pcm = pcm; 758 apcm->codec = codec; 759 apcm->info = cpcm; 760 pcm->private_data = apcm; 761 pcm->private_free = azx_pcm_free; 762 if (cpcm->pcm_type == HDA_PCM_TYPE_MODEM) 763 pcm->dev_class = SNDRV_PCM_CLASS_MODEM; 764 list_add_tail(&apcm->list, &chip->pcm_list); 765 cpcm->pcm = pcm; 766 for (s = 0; s < 2; s++) { 767 if (cpcm->stream[s].substreams) 768 snd_pcm_set_ops(pcm, s, &azx_pcm_ops); 769 } 770 /* buffer pre-allocation */ 771 size = CONFIG_SND_HDA_PREALLOC_SIZE * 1024; 772 if (size > MAX_PREALLOC_SIZE) 773 size = MAX_PREALLOC_SIZE; 774 if (chip->uc_buffer) 775 type = SNDRV_DMA_TYPE_DEV_UC_SG; 776 snd_pcm_lib_preallocate_pages_for_all(pcm, type, 777 chip->card->dev, 778 size, MAX_PREALLOC_SIZE); 779 return 0; 780 } 781 782 static unsigned int azx_command_addr(u32 cmd) 783 { 784 unsigned int addr = cmd >> 28; 785 786 if (addr >= AZX_MAX_CODECS) { 787 snd_BUG(); 788 addr = 0; 789 } 790 791 return addr; 792 } 793 794 /* receive a response */ 795 static int azx_rirb_get_response(struct hdac_bus *bus, unsigned int addr, 796 unsigned int *res) 797 { 798 struct azx *chip = bus_to_azx(bus); 799 struct hda_bus *hbus = &chip->bus; 800 unsigned long timeout; 801 unsigned long loopcounter; 802 int do_poll = 0; 803 804 again: 805 timeout = jiffies + msecs_to_jiffies(1000); 806 807 for (loopcounter = 0;; loopcounter++) { 808 spin_lock_irq(&bus->reg_lock); 809 if (chip->polling_mode || do_poll) 810 snd_hdac_bus_update_rirb(bus); 811 if (!bus->rirb.cmds[addr]) { 812 if (!do_poll) 813 chip->poll_count = 0; 814 if (res) 815 *res = bus->rirb.res[addr]; /* the last value */ 816 spin_unlock_irq(&bus->reg_lock); 817 return 0; 818 } 819 spin_unlock_irq(&bus->reg_lock); 820 if (time_after(jiffies, timeout)) 821 break; 822 if (hbus->needs_damn_long_delay || loopcounter > 3000) 823 msleep(2); /* temporary workaround */ 824 else { 825 udelay(10); 826 cond_resched(); 827 } 828 } 829 830 if (hbus->no_response_fallback) 831 return -EIO; 832 833 if (!chip->polling_mode && chip->poll_count < 2) { 834 dev_dbg(chip->card->dev, 835 "azx_get_response timeout, polling the codec once: last cmd=0x%08x\n", 836 bus->last_cmd[addr]); 837 do_poll = 1; 838 chip->poll_count++; 839 goto again; 840 } 841 842 843 if (!chip->polling_mode) { 844 dev_warn(chip->card->dev, 845 "azx_get_response timeout, switching to polling mode: last cmd=0x%08x\n", 846 bus->last_cmd[addr]); 847 chip->polling_mode = 1; 848 goto again; 849 } 850 851 if (chip->msi) { 852 dev_warn(chip->card->dev, 853 "No response from codec, disabling MSI: last cmd=0x%08x\n", 854 bus->last_cmd[addr]); 855 if (chip->ops->disable_msi_reset_irq && 856 chip->ops->disable_msi_reset_irq(chip) < 0) 857 return -EIO; 858 goto again; 859 } 860 861 if (chip->probing) { 862 /* If this critical timeout happens during the codec probing 863 * phase, this is likely an access to a non-existing codec 864 * slot. Better to return an error and reset the system. 865 */ 866 return -EIO; 867 } 868 869 /* no fallback mechanism? */ 870 if (!chip->fallback_to_single_cmd) 871 return -EIO; 872 873 /* a fatal communication error; need either to reset or to fallback 874 * to the single_cmd mode 875 */ 876 if (hbus->allow_bus_reset && !hbus->response_reset && !hbus->in_reset) { 877 hbus->response_reset = 1; 878 return -EAGAIN; /* give a chance to retry */ 879 } 880 881 dev_err(chip->card->dev, 882 "azx_get_response timeout, switching to single_cmd mode: last cmd=0x%08x\n", 883 bus->last_cmd[addr]); 884 chip->single_cmd = 1; 885 hbus->response_reset = 0; 886 snd_hdac_bus_stop_cmd_io(bus); 887 return -EIO; 888 } 889 890 /* 891 * Use the single immediate command instead of CORB/RIRB for simplicity 892 * 893 * Note: according to Intel, this is not preferred use. The command was 894 * intended for the BIOS only, and may get confused with unsolicited 895 * responses. So, we shouldn't use it for normal operation from the 896 * driver. 897 * I left the codes, however, for debugging/testing purposes. 898 */ 899 900 /* receive a response */ 901 static int azx_single_wait_for_response(struct azx *chip, unsigned int addr) 902 { 903 int timeout = 50; 904 905 while (timeout--) { 906 /* check IRV busy bit */ 907 if (azx_readw(chip, IRS) & AZX_IRS_VALID) { 908 /* reuse rirb.res as the response return value */ 909 azx_bus(chip)->rirb.res[addr] = azx_readl(chip, IR); 910 return 0; 911 } 912 udelay(1); 913 } 914 if (printk_ratelimit()) 915 dev_dbg(chip->card->dev, "get_response timeout: IRS=0x%x\n", 916 azx_readw(chip, IRS)); 917 azx_bus(chip)->rirb.res[addr] = -1; 918 return -EIO; 919 } 920 921 /* send a command */ 922 static int azx_single_send_cmd(struct hdac_bus *bus, u32 val) 923 { 924 struct azx *chip = bus_to_azx(bus); 925 unsigned int addr = azx_command_addr(val); 926 int timeout = 50; 927 928 bus->last_cmd[azx_command_addr(val)] = val; 929 while (timeout--) { 930 /* check ICB busy bit */ 931 if (!((azx_readw(chip, IRS) & AZX_IRS_BUSY))) { 932 /* Clear IRV valid bit */ 933 azx_writew(chip, IRS, azx_readw(chip, IRS) | 934 AZX_IRS_VALID); 935 azx_writel(chip, IC, val); 936 azx_writew(chip, IRS, azx_readw(chip, IRS) | 937 AZX_IRS_BUSY); 938 return azx_single_wait_for_response(chip, addr); 939 } 940 udelay(1); 941 } 942 if (printk_ratelimit()) 943 dev_dbg(chip->card->dev, 944 "send_cmd timeout: IRS=0x%x, val=0x%x\n", 945 azx_readw(chip, IRS), val); 946 return -EIO; 947 } 948 949 /* receive a response */ 950 static int azx_single_get_response(struct hdac_bus *bus, unsigned int addr, 951 unsigned int *res) 952 { 953 if (res) 954 *res = bus->rirb.res[addr]; 955 return 0; 956 } 957 958 /* 959 * The below are the main callbacks from hda_codec. 960 * 961 * They are just the skeleton to call sub-callbacks according to the 962 * current setting of chip->single_cmd. 963 */ 964 965 /* send a command */ 966 static int azx_send_cmd(struct hdac_bus *bus, unsigned int val) 967 { 968 struct azx *chip = bus_to_azx(bus); 969 970 if (chip->disabled) 971 return 0; 972 if (chip->single_cmd) 973 return azx_single_send_cmd(bus, val); 974 else 975 return snd_hdac_bus_send_cmd(bus, val); 976 } 977 978 /* get a response */ 979 static int azx_get_response(struct hdac_bus *bus, unsigned int addr, 980 unsigned int *res) 981 { 982 struct azx *chip = bus_to_azx(bus); 983 984 if (chip->disabled) 985 return 0; 986 if (chip->single_cmd) 987 return azx_single_get_response(bus, addr, res); 988 else 989 return azx_rirb_get_response(bus, addr, res); 990 } 991 992 static int azx_link_power(struct hdac_bus *bus, bool enable) 993 { 994 struct azx *chip = bus_to_azx(bus); 995 996 if (chip->ops->link_power) 997 return chip->ops->link_power(chip, enable); 998 else 999 return -EINVAL; 1000 } 1001 1002 static const struct hdac_bus_ops bus_core_ops = { 1003 .command = azx_send_cmd, 1004 .get_response = azx_get_response, 1005 .link_power = azx_link_power, 1006 }; 1007 1008 #ifdef CONFIG_SND_HDA_DSP_LOADER 1009 /* 1010 * DSP loading code (e.g. for CA0132) 1011 */ 1012 1013 /* use the first stream for loading DSP */ 1014 static struct azx_dev * 1015 azx_get_dsp_loader_dev(struct azx *chip) 1016 { 1017 struct hdac_bus *bus = azx_bus(chip); 1018 struct hdac_stream *s; 1019 1020 list_for_each_entry(s, &bus->stream_list, list) 1021 if (s->index == chip->playback_index_offset) 1022 return stream_to_azx_dev(s); 1023 1024 return NULL; 1025 } 1026 1027 int snd_hda_codec_load_dsp_prepare(struct hda_codec *codec, unsigned int format, 1028 unsigned int byte_size, 1029 struct snd_dma_buffer *bufp) 1030 { 1031 struct hdac_bus *bus = &codec->bus->core; 1032 struct azx *chip = bus_to_azx(bus); 1033 struct azx_dev *azx_dev; 1034 struct hdac_stream *hstr; 1035 bool saved = false; 1036 int err; 1037 1038 azx_dev = azx_get_dsp_loader_dev(chip); 1039 hstr = azx_stream(azx_dev); 1040 spin_lock_irq(&bus->reg_lock); 1041 if (hstr->opened) { 1042 chip->saved_azx_dev = *azx_dev; 1043 saved = true; 1044 } 1045 spin_unlock_irq(&bus->reg_lock); 1046 1047 err = snd_hdac_dsp_prepare(hstr, format, byte_size, bufp); 1048 if (err < 0) { 1049 spin_lock_irq(&bus->reg_lock); 1050 if (saved) 1051 *azx_dev = chip->saved_azx_dev; 1052 spin_unlock_irq(&bus->reg_lock); 1053 return err; 1054 } 1055 1056 hstr->prepared = 0; 1057 return err; 1058 } 1059 EXPORT_SYMBOL_GPL(snd_hda_codec_load_dsp_prepare); 1060 1061 void snd_hda_codec_load_dsp_trigger(struct hda_codec *codec, bool start) 1062 { 1063 struct hdac_bus *bus = &codec->bus->core; 1064 struct azx *chip = bus_to_azx(bus); 1065 struct azx_dev *azx_dev = azx_get_dsp_loader_dev(chip); 1066 1067 snd_hdac_dsp_trigger(azx_stream(azx_dev), start); 1068 } 1069 EXPORT_SYMBOL_GPL(snd_hda_codec_load_dsp_trigger); 1070 1071 void snd_hda_codec_load_dsp_cleanup(struct hda_codec *codec, 1072 struct snd_dma_buffer *dmab) 1073 { 1074 struct hdac_bus *bus = &codec->bus->core; 1075 struct azx *chip = bus_to_azx(bus); 1076 struct azx_dev *azx_dev = azx_get_dsp_loader_dev(chip); 1077 struct hdac_stream *hstr = azx_stream(azx_dev); 1078 1079 if (!dmab->area || !hstr->locked) 1080 return; 1081 1082 snd_hdac_dsp_cleanup(hstr, dmab); 1083 spin_lock_irq(&bus->reg_lock); 1084 if (hstr->opened) 1085 *azx_dev = chip->saved_azx_dev; 1086 hstr->locked = false; 1087 spin_unlock_irq(&bus->reg_lock); 1088 } 1089 EXPORT_SYMBOL_GPL(snd_hda_codec_load_dsp_cleanup); 1090 #endif /* CONFIG_SND_HDA_DSP_LOADER */ 1091 1092 /* 1093 * reset and start the controller registers 1094 */ 1095 void azx_init_chip(struct azx *chip, bool full_reset) 1096 { 1097 if (snd_hdac_bus_init_chip(azx_bus(chip), full_reset)) { 1098 /* correct RINTCNT for CXT */ 1099 if (chip->driver_caps & AZX_DCAPS_CTX_WORKAROUND) 1100 azx_writew(chip, RINTCNT, 0xc0); 1101 } 1102 } 1103 EXPORT_SYMBOL_GPL(azx_init_chip); 1104 1105 void azx_stop_all_streams(struct azx *chip) 1106 { 1107 struct hdac_bus *bus = azx_bus(chip); 1108 struct hdac_stream *s; 1109 1110 list_for_each_entry(s, &bus->stream_list, list) 1111 snd_hdac_stream_stop(s); 1112 } 1113 EXPORT_SYMBOL_GPL(azx_stop_all_streams); 1114 1115 void azx_stop_chip(struct azx *chip) 1116 { 1117 snd_hdac_bus_stop_chip(azx_bus(chip)); 1118 } 1119 EXPORT_SYMBOL_GPL(azx_stop_chip); 1120 1121 /* 1122 * interrupt handler 1123 */ 1124 static void stream_update(struct hdac_bus *bus, struct hdac_stream *s) 1125 { 1126 struct azx *chip = bus_to_azx(bus); 1127 struct azx_dev *azx_dev = stream_to_azx_dev(s); 1128 1129 /* check whether this IRQ is really acceptable */ 1130 if (!chip->ops->position_check || 1131 chip->ops->position_check(chip, azx_dev)) { 1132 spin_unlock(&bus->reg_lock); 1133 snd_pcm_period_elapsed(azx_stream(azx_dev)->substream); 1134 spin_lock(&bus->reg_lock); 1135 } 1136 } 1137 1138 irqreturn_t azx_interrupt(int irq, void *dev_id) 1139 { 1140 struct azx *chip = dev_id; 1141 struct hdac_bus *bus = azx_bus(chip); 1142 u32 status; 1143 bool active, handled = false; 1144 int repeat = 0; /* count for avoiding endless loop */ 1145 1146 #ifdef CONFIG_PM 1147 if (azx_has_pm_runtime(chip)) 1148 if (!pm_runtime_active(chip->card->dev)) 1149 return IRQ_NONE; 1150 #endif 1151 1152 spin_lock(&bus->reg_lock); 1153 1154 if (chip->disabled) 1155 goto unlock; 1156 1157 do { 1158 status = azx_readl(chip, INTSTS); 1159 if (status == 0 || status == 0xffffffff) 1160 break; 1161 1162 handled = true; 1163 active = false; 1164 if (snd_hdac_bus_handle_stream_irq(bus, status, stream_update)) 1165 active = true; 1166 1167 /* clear rirb int */ 1168 status = azx_readb(chip, RIRBSTS); 1169 if (status & RIRB_INT_MASK) { 1170 active = true; 1171 if (status & RIRB_INT_RESPONSE) { 1172 if (chip->driver_caps & AZX_DCAPS_CTX_WORKAROUND) 1173 udelay(80); 1174 snd_hdac_bus_update_rirb(bus); 1175 } 1176 azx_writeb(chip, RIRBSTS, RIRB_INT_MASK); 1177 } 1178 } while (active && ++repeat < 10); 1179 1180 unlock: 1181 spin_unlock(&bus->reg_lock); 1182 1183 return IRQ_RETVAL(handled); 1184 } 1185 EXPORT_SYMBOL_GPL(azx_interrupt); 1186 1187 /* 1188 * Codec initerface 1189 */ 1190 1191 /* 1192 * Probe the given codec address 1193 */ 1194 static int probe_codec(struct azx *chip, int addr) 1195 { 1196 unsigned int cmd = (addr << 28) | (AC_NODE_ROOT << 20) | 1197 (AC_VERB_PARAMETERS << 8) | AC_PAR_VENDOR_ID; 1198 struct hdac_bus *bus = azx_bus(chip); 1199 int err; 1200 unsigned int res = -1; 1201 1202 mutex_lock(&bus->cmd_mutex); 1203 chip->probing = 1; 1204 azx_send_cmd(bus, cmd); 1205 err = azx_get_response(bus, addr, &res); 1206 chip->probing = 0; 1207 mutex_unlock(&bus->cmd_mutex); 1208 if (err < 0 || res == -1) 1209 return -EIO; 1210 dev_dbg(chip->card->dev, "codec #%d probed OK\n", addr); 1211 return 0; 1212 } 1213 1214 void snd_hda_bus_reset(struct hda_bus *bus) 1215 { 1216 struct azx *chip = bus_to_azx(&bus->core); 1217 1218 bus->in_reset = 1; 1219 azx_stop_chip(chip); 1220 azx_init_chip(chip, true); 1221 if (bus->core.chip_init) 1222 snd_hda_bus_reset_codecs(bus); 1223 bus->in_reset = 0; 1224 } 1225 1226 /* HD-audio bus initialization */ 1227 int azx_bus_init(struct azx *chip, const char *model, 1228 const struct hdac_io_ops *io_ops) 1229 { 1230 struct hda_bus *bus = &chip->bus; 1231 int err; 1232 1233 err = snd_hdac_bus_init(&bus->core, chip->card->dev, &bus_core_ops, 1234 io_ops); 1235 if (err < 0) 1236 return err; 1237 1238 bus->card = chip->card; 1239 mutex_init(&bus->prepare_mutex); 1240 bus->pci = chip->pci; 1241 bus->modelname = model; 1242 bus->mixer_assigned = -1; 1243 bus->core.snoop = azx_snoop(chip); 1244 if (chip->get_position[0] != azx_get_pos_lpib || 1245 chip->get_position[1] != azx_get_pos_lpib) 1246 bus->core.use_posbuf = true; 1247 bus->core.bdl_pos_adj = chip->bdl_pos_adj; 1248 if (chip->driver_caps & AZX_DCAPS_CORBRP_SELF_CLEAR) 1249 bus->core.corbrp_self_clear = true; 1250 1251 if (chip->driver_caps & AZX_DCAPS_4K_BDLE_BOUNDARY) 1252 bus->core.align_bdle_4k = true; 1253 1254 /* AMD chipsets often cause the communication stalls upon certain 1255 * sequence like the pin-detection. It seems that forcing the synced 1256 * access works around the stall. Grrr... 1257 */ 1258 if (chip->driver_caps & AZX_DCAPS_SYNC_WRITE) { 1259 dev_dbg(chip->card->dev, "Enable sync_write for stable communication\n"); 1260 bus->core.sync_write = 1; 1261 bus->allow_bus_reset = 1; 1262 } 1263 1264 return 0; 1265 } 1266 EXPORT_SYMBOL_GPL(azx_bus_init); 1267 1268 /* Probe codecs */ 1269 int azx_probe_codecs(struct azx *chip, unsigned int max_slots) 1270 { 1271 struct hdac_bus *bus = azx_bus(chip); 1272 int c, codecs, err; 1273 1274 codecs = 0; 1275 if (!max_slots) 1276 max_slots = AZX_DEFAULT_CODECS; 1277 1278 /* First try to probe all given codec slots */ 1279 for (c = 0; c < max_slots; c++) { 1280 if ((bus->codec_mask & (1 << c)) & chip->codec_probe_mask) { 1281 if (probe_codec(chip, c) < 0) { 1282 /* Some BIOSen give you wrong codec addresses 1283 * that don't exist 1284 */ 1285 dev_warn(chip->card->dev, 1286 "Codec #%d probe error; disabling it...\n", c); 1287 bus->codec_mask &= ~(1 << c); 1288 /* More badly, accessing to a non-existing 1289 * codec often screws up the controller chip, 1290 * and disturbs the further communications. 1291 * Thus if an error occurs during probing, 1292 * better to reset the controller chip to 1293 * get back to the sanity state. 1294 */ 1295 azx_stop_chip(chip); 1296 azx_init_chip(chip, true); 1297 } 1298 } 1299 } 1300 1301 /* Then create codec instances */ 1302 for (c = 0; c < max_slots; c++) { 1303 if ((bus->codec_mask & (1 << c)) & chip->codec_probe_mask) { 1304 struct hda_codec *codec; 1305 err = snd_hda_codec_new(&chip->bus, chip->card, c, &codec); 1306 if (err < 0) 1307 continue; 1308 codec->jackpoll_interval = chip->jackpoll_interval; 1309 codec->beep_mode = chip->beep_mode; 1310 codecs++; 1311 } 1312 } 1313 if (!codecs) { 1314 dev_err(chip->card->dev, "no codecs initialized\n"); 1315 return -ENXIO; 1316 } 1317 return 0; 1318 } 1319 EXPORT_SYMBOL_GPL(azx_probe_codecs); 1320 1321 /* configure each codec instance */ 1322 int azx_codec_configure(struct azx *chip) 1323 { 1324 struct hda_codec *codec, *next; 1325 1326 /* use _safe version here since snd_hda_codec_configure() deregisters 1327 * the device upon error and deletes itself from the bus list. 1328 */ 1329 list_for_each_codec_safe(codec, next, &chip->bus) { 1330 snd_hda_codec_configure(codec); 1331 } 1332 1333 if (!azx_bus(chip)->num_codecs) 1334 return -ENODEV; 1335 return 0; 1336 } 1337 EXPORT_SYMBOL_GPL(azx_codec_configure); 1338 1339 static int stream_direction(struct azx *chip, unsigned char index) 1340 { 1341 if (index >= chip->capture_index_offset && 1342 index < chip->capture_index_offset + chip->capture_streams) 1343 return SNDRV_PCM_STREAM_CAPTURE; 1344 return SNDRV_PCM_STREAM_PLAYBACK; 1345 } 1346 1347 /* initialize SD streams */ 1348 int azx_init_streams(struct azx *chip) 1349 { 1350 int i; 1351 int stream_tags[2] = { 0, 0 }; 1352 1353 /* initialize each stream (aka device) 1354 * assign the starting bdl address to each stream (device) 1355 * and initialize 1356 */ 1357 for (i = 0; i < chip->num_streams; i++) { 1358 struct azx_dev *azx_dev = kzalloc(sizeof(*azx_dev), GFP_KERNEL); 1359 int dir, tag; 1360 1361 if (!azx_dev) 1362 return -ENOMEM; 1363 1364 dir = stream_direction(chip, i); 1365 /* stream tag must be unique throughout 1366 * the stream direction group, 1367 * valid values 1...15 1368 * use separate stream tag if the flag 1369 * AZX_DCAPS_SEPARATE_STREAM_TAG is used 1370 */ 1371 if (chip->driver_caps & AZX_DCAPS_SEPARATE_STREAM_TAG) 1372 tag = ++stream_tags[dir]; 1373 else 1374 tag = i + 1; 1375 snd_hdac_stream_init(azx_bus(chip), azx_stream(azx_dev), 1376 i, dir, tag); 1377 } 1378 1379 return 0; 1380 } 1381 EXPORT_SYMBOL_GPL(azx_init_streams); 1382 1383 void azx_free_streams(struct azx *chip) 1384 { 1385 struct hdac_bus *bus = azx_bus(chip); 1386 struct hdac_stream *s; 1387 1388 while (!list_empty(&bus->stream_list)) { 1389 s = list_first_entry(&bus->stream_list, struct hdac_stream, list); 1390 list_del(&s->list); 1391 kfree(stream_to_azx_dev(s)); 1392 } 1393 } 1394 EXPORT_SYMBOL_GPL(azx_free_streams); 1395