xref: /openbmc/linux/sound/pci/hda/hda_controller.c (revision c4c3c32d)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *
4  *  Implementation of primary alsa driver code base for Intel HD Audio.
5  *
6  *  Copyright(c) 2004 Intel Corporation. All rights reserved.
7  *
8  *  Copyright (c) 2004 Takashi Iwai <tiwai@suse.de>
9  *                     PeiSen Hou <pshou@realtek.com.tw>
10  */
11 
12 #include <linux/clocksource.h>
13 #include <linux/delay.h>
14 #include <linux/interrupt.h>
15 #include <linux/kernel.h>
16 #include <linux/module.h>
17 #include <linux/pm_runtime.h>
18 #include <linux/slab.h>
19 
20 #ifdef CONFIG_X86
21 /* for art-tsc conversion */
22 #include <asm/tsc.h>
23 #endif
24 
25 #include <sound/core.h>
26 #include <sound/initval.h>
27 #include "hda_controller.h"
28 #include "hda_local.h"
29 
30 #define CREATE_TRACE_POINTS
31 #include "hda_controller_trace.h"
32 
33 /* DSP lock helpers */
34 #define dsp_lock(dev)		snd_hdac_dsp_lock(azx_stream(dev))
35 #define dsp_unlock(dev)		snd_hdac_dsp_unlock(azx_stream(dev))
36 #define dsp_is_locked(dev)	snd_hdac_stream_is_locked(azx_stream(dev))
37 
38 /* assign a stream for the PCM */
39 static inline struct azx_dev *
40 azx_assign_device(struct azx *chip, struct snd_pcm_substream *substream)
41 {
42 	struct hdac_stream *s;
43 
44 	s = snd_hdac_stream_assign(azx_bus(chip), substream);
45 	if (!s)
46 		return NULL;
47 	return stream_to_azx_dev(s);
48 }
49 
50 /* release the assigned stream */
51 static inline void azx_release_device(struct azx_dev *azx_dev)
52 {
53 	snd_hdac_stream_release(azx_stream(azx_dev));
54 }
55 
56 static inline struct hda_pcm_stream *
57 to_hda_pcm_stream(struct snd_pcm_substream *substream)
58 {
59 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
60 	return &apcm->info->stream[substream->stream];
61 }
62 
63 static u64 azx_adjust_codec_delay(struct snd_pcm_substream *substream,
64 				u64 nsec)
65 {
66 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
67 	struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream);
68 	u64 codec_frames, codec_nsecs;
69 
70 	if (!hinfo->ops.get_delay)
71 		return nsec;
72 
73 	codec_frames = hinfo->ops.get_delay(hinfo, apcm->codec, substream);
74 	codec_nsecs = div_u64(codec_frames * 1000000000LL,
75 			      substream->runtime->rate);
76 
77 	if (substream->stream == SNDRV_PCM_STREAM_CAPTURE)
78 		return nsec + codec_nsecs;
79 
80 	return (nsec > codec_nsecs) ? nsec - codec_nsecs : 0;
81 }
82 
83 /*
84  * PCM ops
85  */
86 
87 static int azx_pcm_close(struct snd_pcm_substream *substream)
88 {
89 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
90 	struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream);
91 	struct azx *chip = apcm->chip;
92 	struct azx_dev *azx_dev = get_azx_dev(substream);
93 
94 	trace_azx_pcm_close(chip, azx_dev);
95 	mutex_lock(&chip->open_mutex);
96 	azx_release_device(azx_dev);
97 	if (hinfo->ops.close)
98 		hinfo->ops.close(hinfo, apcm->codec, substream);
99 	snd_hda_power_down(apcm->codec);
100 	mutex_unlock(&chip->open_mutex);
101 	snd_hda_codec_pcm_put(apcm->info);
102 	return 0;
103 }
104 
105 static int azx_pcm_hw_params(struct snd_pcm_substream *substream,
106 			     struct snd_pcm_hw_params *hw_params)
107 {
108 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
109 	struct azx *chip = apcm->chip;
110 	struct azx_dev *azx_dev = get_azx_dev(substream);
111 	int ret = 0;
112 
113 	trace_azx_pcm_hw_params(chip, azx_dev);
114 	dsp_lock(azx_dev);
115 	if (dsp_is_locked(azx_dev)) {
116 		ret = -EBUSY;
117 		goto unlock;
118 	}
119 
120 	azx_dev->core.bufsize = 0;
121 	azx_dev->core.period_bytes = 0;
122 	azx_dev->core.format_val = 0;
123 
124 unlock:
125 	dsp_unlock(azx_dev);
126 	return ret;
127 }
128 
129 static int azx_pcm_hw_free(struct snd_pcm_substream *substream)
130 {
131 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
132 	struct azx_dev *azx_dev = get_azx_dev(substream);
133 	struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream);
134 
135 	/* reset BDL address */
136 	dsp_lock(azx_dev);
137 	if (!dsp_is_locked(azx_dev))
138 		snd_hdac_stream_cleanup(azx_stream(azx_dev));
139 
140 	snd_hda_codec_cleanup(apcm->codec, hinfo, substream);
141 
142 	azx_stream(azx_dev)->prepared = 0;
143 	dsp_unlock(azx_dev);
144 	return 0;
145 }
146 
147 static int azx_pcm_prepare(struct snd_pcm_substream *substream)
148 {
149 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
150 	struct azx *chip = apcm->chip;
151 	struct azx_dev *azx_dev = get_azx_dev(substream);
152 	struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream);
153 	struct snd_pcm_runtime *runtime = substream->runtime;
154 	unsigned int format_val, stream_tag;
155 	int err;
156 	struct hda_spdif_out *spdif =
157 		snd_hda_spdif_out_of_nid(apcm->codec, hinfo->nid);
158 	unsigned short ctls = spdif ? spdif->ctls : 0;
159 
160 	trace_azx_pcm_prepare(chip, azx_dev);
161 	dsp_lock(azx_dev);
162 	if (dsp_is_locked(azx_dev)) {
163 		err = -EBUSY;
164 		goto unlock;
165 	}
166 
167 	snd_hdac_stream_reset(azx_stream(azx_dev));
168 	format_val = snd_hdac_calc_stream_format(runtime->rate,
169 						runtime->channels,
170 						runtime->format,
171 						hinfo->maxbps,
172 						ctls);
173 	if (!format_val) {
174 		dev_err(chip->card->dev,
175 			"invalid format_val, rate=%d, ch=%d, format=%d\n",
176 			runtime->rate, runtime->channels, runtime->format);
177 		err = -EINVAL;
178 		goto unlock;
179 	}
180 
181 	err = snd_hdac_stream_set_params(azx_stream(azx_dev), format_val);
182 	if (err < 0)
183 		goto unlock;
184 
185 	snd_hdac_stream_setup(azx_stream(azx_dev));
186 
187 	stream_tag = azx_dev->core.stream_tag;
188 	/* CA-IBG chips need the playback stream starting from 1 */
189 	if ((chip->driver_caps & AZX_DCAPS_CTX_WORKAROUND) &&
190 	    stream_tag > chip->capture_streams)
191 		stream_tag -= chip->capture_streams;
192 	err = snd_hda_codec_prepare(apcm->codec, hinfo, stream_tag,
193 				     azx_dev->core.format_val, substream);
194 
195  unlock:
196 	if (!err)
197 		azx_stream(azx_dev)->prepared = 1;
198 	dsp_unlock(azx_dev);
199 	return err;
200 }
201 
202 static int azx_pcm_trigger(struct snd_pcm_substream *substream, int cmd)
203 {
204 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
205 	struct azx *chip = apcm->chip;
206 	struct hdac_bus *bus = azx_bus(chip);
207 	struct azx_dev *azx_dev;
208 	struct snd_pcm_substream *s;
209 	struct hdac_stream *hstr;
210 	bool start;
211 	int sbits = 0;
212 	int sync_reg;
213 
214 	azx_dev = get_azx_dev(substream);
215 	trace_azx_pcm_trigger(chip, azx_dev, cmd);
216 
217 	hstr = azx_stream(azx_dev);
218 	if (chip->driver_caps & AZX_DCAPS_OLD_SSYNC)
219 		sync_reg = AZX_REG_OLD_SSYNC;
220 	else
221 		sync_reg = AZX_REG_SSYNC;
222 
223 	if (dsp_is_locked(azx_dev) || !hstr->prepared)
224 		return -EPIPE;
225 
226 	switch (cmd) {
227 	case SNDRV_PCM_TRIGGER_START:
228 	case SNDRV_PCM_TRIGGER_PAUSE_RELEASE:
229 	case SNDRV_PCM_TRIGGER_RESUME:
230 		start = true;
231 		break;
232 	case SNDRV_PCM_TRIGGER_PAUSE_PUSH:
233 	case SNDRV_PCM_TRIGGER_SUSPEND:
234 	case SNDRV_PCM_TRIGGER_STOP:
235 		start = false;
236 		break;
237 	default:
238 		return -EINVAL;
239 	}
240 
241 	snd_pcm_group_for_each_entry(s, substream) {
242 		if (s->pcm->card != substream->pcm->card)
243 			continue;
244 		azx_dev = get_azx_dev(s);
245 		sbits |= 1 << azx_dev->core.index;
246 		snd_pcm_trigger_done(s, substream);
247 	}
248 
249 	spin_lock(&bus->reg_lock);
250 
251 	/* first, set SYNC bits of corresponding streams */
252 	snd_hdac_stream_sync_trigger(hstr, true, sbits, sync_reg);
253 
254 	snd_pcm_group_for_each_entry(s, substream) {
255 		if (s->pcm->card != substream->pcm->card)
256 			continue;
257 		azx_dev = get_azx_dev(s);
258 		if (start) {
259 			azx_dev->insufficient = 1;
260 			snd_hdac_stream_start(azx_stream(azx_dev));
261 		} else {
262 			snd_hdac_stream_stop(azx_stream(azx_dev));
263 		}
264 	}
265 	spin_unlock(&bus->reg_lock);
266 
267 	snd_hdac_stream_sync(hstr, start, sbits);
268 
269 	spin_lock(&bus->reg_lock);
270 	/* reset SYNC bits */
271 	snd_hdac_stream_sync_trigger(hstr, false, sbits, sync_reg);
272 	if (start)
273 		snd_hdac_stream_timecounter_init(hstr, sbits);
274 	spin_unlock(&bus->reg_lock);
275 	return 0;
276 }
277 
278 unsigned int azx_get_pos_lpib(struct azx *chip, struct azx_dev *azx_dev)
279 {
280 	return snd_hdac_stream_get_pos_lpib(azx_stream(azx_dev));
281 }
282 EXPORT_SYMBOL_GPL(azx_get_pos_lpib);
283 
284 unsigned int azx_get_pos_posbuf(struct azx *chip, struct azx_dev *azx_dev)
285 {
286 	return snd_hdac_stream_get_pos_posbuf(azx_stream(azx_dev));
287 }
288 EXPORT_SYMBOL_GPL(azx_get_pos_posbuf);
289 
290 unsigned int azx_get_position(struct azx *chip,
291 			      struct azx_dev *azx_dev)
292 {
293 	struct snd_pcm_substream *substream = azx_dev->core.substream;
294 	unsigned int pos;
295 	int stream = substream->stream;
296 	int delay = 0;
297 
298 	if (chip->get_position[stream])
299 		pos = chip->get_position[stream](chip, azx_dev);
300 	else /* use the position buffer as default */
301 		pos = azx_get_pos_posbuf(chip, azx_dev);
302 
303 	if (pos >= azx_dev->core.bufsize)
304 		pos = 0;
305 
306 	if (substream->runtime) {
307 		struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
308 		struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream);
309 
310 		if (chip->get_delay[stream])
311 			delay += chip->get_delay[stream](chip, azx_dev, pos);
312 		if (hinfo->ops.get_delay)
313 			delay += hinfo->ops.get_delay(hinfo, apcm->codec,
314 						      substream);
315 		substream->runtime->delay = delay;
316 	}
317 
318 	trace_azx_get_position(chip, azx_dev, pos, delay);
319 	return pos;
320 }
321 EXPORT_SYMBOL_GPL(azx_get_position);
322 
323 static snd_pcm_uframes_t azx_pcm_pointer(struct snd_pcm_substream *substream)
324 {
325 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
326 	struct azx *chip = apcm->chip;
327 	struct azx_dev *azx_dev = get_azx_dev(substream);
328 	return bytes_to_frames(substream->runtime,
329 			       azx_get_position(chip, azx_dev));
330 }
331 
332 /*
333  * azx_scale64: Scale base by mult/div while not overflowing sanely
334  *
335  * Derived from scale64_check_overflow in kernel/time/timekeeping.c
336  *
337  * The tmestamps for a 48Khz stream can overflow after (2^64/10^9)/48K which
338  * is about 384307 ie ~4.5 days.
339  *
340  * This scales the calculation so that overflow will happen but after 2^64 /
341  * 48000 secs, which is pretty large!
342  *
343  * In caln below:
344  *	base may overflow, but since there isn’t any additional division
345  *	performed on base it’s OK
346  *	rem can’t overflow because both are 32-bit values
347  */
348 
349 #ifdef CONFIG_X86
350 static u64 azx_scale64(u64 base, u32 num, u32 den)
351 {
352 	u64 rem;
353 
354 	rem = do_div(base, den);
355 
356 	base *= num;
357 	rem *= num;
358 
359 	do_div(rem, den);
360 
361 	return base + rem;
362 }
363 
364 static int azx_get_sync_time(ktime_t *device,
365 		struct system_counterval_t *system, void *ctx)
366 {
367 	struct snd_pcm_substream *substream = ctx;
368 	struct azx_dev *azx_dev = get_azx_dev(substream);
369 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
370 	struct azx *chip = apcm->chip;
371 	struct snd_pcm_runtime *runtime;
372 	u64 ll_counter, ll_counter_l, ll_counter_h;
373 	u64 tsc_counter, tsc_counter_l, tsc_counter_h;
374 	u32 wallclk_ctr, wallclk_cycles;
375 	bool direction;
376 	u32 dma_select;
377 	u32 timeout;
378 	u32 retry_count = 0;
379 
380 	runtime = substream->runtime;
381 
382 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
383 		direction = 1;
384 	else
385 		direction = 0;
386 
387 	/* 0th stream tag is not used, so DMA ch 0 is for 1st stream tag */
388 	do {
389 		timeout = 100;
390 		dma_select = (direction << GTSCC_CDMAS_DMA_DIR_SHIFT) |
391 					(azx_dev->core.stream_tag - 1);
392 		snd_hdac_chip_writel(azx_bus(chip), GTSCC, dma_select);
393 
394 		/* Enable the capture */
395 		snd_hdac_chip_updatel(azx_bus(chip), GTSCC, 0, GTSCC_TSCCI_MASK);
396 
397 		while (timeout) {
398 			if (snd_hdac_chip_readl(azx_bus(chip), GTSCC) &
399 						GTSCC_TSCCD_MASK)
400 				break;
401 
402 			timeout--;
403 		}
404 
405 		if (!timeout) {
406 			dev_err(chip->card->dev, "GTSCC capture Timedout!\n");
407 			return -EIO;
408 		}
409 
410 		/* Read wall clock counter */
411 		wallclk_ctr = snd_hdac_chip_readl(azx_bus(chip), WALFCC);
412 
413 		/* Read TSC counter */
414 		tsc_counter_l = snd_hdac_chip_readl(azx_bus(chip), TSCCL);
415 		tsc_counter_h = snd_hdac_chip_readl(azx_bus(chip), TSCCU);
416 
417 		/* Read Link counter */
418 		ll_counter_l = snd_hdac_chip_readl(azx_bus(chip), LLPCL);
419 		ll_counter_h = snd_hdac_chip_readl(azx_bus(chip), LLPCU);
420 
421 		/* Ack: registers read done */
422 		snd_hdac_chip_writel(azx_bus(chip), GTSCC, GTSCC_TSCCD_SHIFT);
423 
424 		tsc_counter = (tsc_counter_h << TSCCU_CCU_SHIFT) |
425 						tsc_counter_l;
426 
427 		ll_counter = (ll_counter_h << LLPC_CCU_SHIFT) |	ll_counter_l;
428 		wallclk_cycles = wallclk_ctr & WALFCC_CIF_MASK;
429 
430 		/*
431 		 * An error occurs near frame "rollover". The clocks in
432 		 * frame value indicates whether this error may have
433 		 * occurred. Here we use the value of 10 i.e.,
434 		 * HDA_MAX_CYCLE_OFFSET
435 		 */
436 		if (wallclk_cycles < HDA_MAX_CYCLE_VALUE - HDA_MAX_CYCLE_OFFSET
437 					&& wallclk_cycles > HDA_MAX_CYCLE_OFFSET)
438 			break;
439 
440 		/*
441 		 * Sleep before we read again, else we may again get
442 		 * value near to MAX_CYCLE. Try to sleep for different
443 		 * amount of time so we dont hit the same number again
444 		 */
445 		udelay(retry_count++);
446 
447 	} while (retry_count != HDA_MAX_CYCLE_READ_RETRY);
448 
449 	if (retry_count == HDA_MAX_CYCLE_READ_RETRY) {
450 		dev_err_ratelimited(chip->card->dev,
451 			"Error in WALFCC cycle count\n");
452 		return -EIO;
453 	}
454 
455 	*device = ns_to_ktime(azx_scale64(ll_counter,
456 				NSEC_PER_SEC, runtime->rate));
457 	*device = ktime_add_ns(*device, (wallclk_cycles * NSEC_PER_SEC) /
458 			       ((HDA_MAX_CYCLE_VALUE + 1) * runtime->rate));
459 
460 	*system = convert_art_to_tsc(tsc_counter);
461 
462 	return 0;
463 }
464 
465 #else
466 static int azx_get_sync_time(ktime_t *device,
467 		struct system_counterval_t *system, void *ctx)
468 {
469 	return -ENXIO;
470 }
471 #endif
472 
473 static int azx_get_crosststamp(struct snd_pcm_substream *substream,
474 			      struct system_device_crosststamp *xtstamp)
475 {
476 	return get_device_system_crosststamp(azx_get_sync_time,
477 					substream, NULL, xtstamp);
478 }
479 
480 static inline bool is_link_time_supported(struct snd_pcm_runtime *runtime,
481 				struct snd_pcm_audio_tstamp_config *ts)
482 {
483 	if (runtime->hw.info & SNDRV_PCM_INFO_HAS_LINK_SYNCHRONIZED_ATIME)
484 		if (ts->type_requested == SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK_SYNCHRONIZED)
485 			return true;
486 
487 	return false;
488 }
489 
490 static int azx_get_time_info(struct snd_pcm_substream *substream,
491 			struct timespec64 *system_ts, struct timespec64 *audio_ts,
492 			struct snd_pcm_audio_tstamp_config *audio_tstamp_config,
493 			struct snd_pcm_audio_tstamp_report *audio_tstamp_report)
494 {
495 	struct azx_dev *azx_dev = get_azx_dev(substream);
496 	struct snd_pcm_runtime *runtime = substream->runtime;
497 	struct system_device_crosststamp xtstamp;
498 	int ret;
499 	u64 nsec;
500 
501 	if ((substream->runtime->hw.info & SNDRV_PCM_INFO_HAS_LINK_ATIME) &&
502 		(audio_tstamp_config->type_requested == SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK)) {
503 
504 		snd_pcm_gettime(substream->runtime, system_ts);
505 
506 		nsec = timecounter_read(&azx_dev->core.tc);
507 		if (audio_tstamp_config->report_delay)
508 			nsec = azx_adjust_codec_delay(substream, nsec);
509 
510 		*audio_ts = ns_to_timespec64(nsec);
511 
512 		audio_tstamp_report->actual_type = SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK;
513 		audio_tstamp_report->accuracy_report = 1; /* rest of structure is valid */
514 		audio_tstamp_report->accuracy = 42; /* 24 MHz WallClock == 42ns resolution */
515 
516 	} else if (is_link_time_supported(runtime, audio_tstamp_config)) {
517 
518 		ret = azx_get_crosststamp(substream, &xtstamp);
519 		if (ret)
520 			return ret;
521 
522 		switch (runtime->tstamp_type) {
523 		case SNDRV_PCM_TSTAMP_TYPE_MONOTONIC:
524 			return -EINVAL;
525 
526 		case SNDRV_PCM_TSTAMP_TYPE_MONOTONIC_RAW:
527 			*system_ts = ktime_to_timespec64(xtstamp.sys_monoraw);
528 			break;
529 
530 		default:
531 			*system_ts = ktime_to_timespec64(xtstamp.sys_realtime);
532 			break;
533 
534 		}
535 
536 		*audio_ts = ktime_to_timespec64(xtstamp.device);
537 
538 		audio_tstamp_report->actual_type =
539 			SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK_SYNCHRONIZED;
540 		audio_tstamp_report->accuracy_report = 1;
541 		/* 24 MHz WallClock == 42ns resolution */
542 		audio_tstamp_report->accuracy = 42;
543 
544 	} else {
545 		audio_tstamp_report->actual_type = SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT;
546 	}
547 
548 	return 0;
549 }
550 
551 static const struct snd_pcm_hardware azx_pcm_hw = {
552 	.info =			(SNDRV_PCM_INFO_MMAP |
553 				 SNDRV_PCM_INFO_INTERLEAVED |
554 				 SNDRV_PCM_INFO_BLOCK_TRANSFER |
555 				 SNDRV_PCM_INFO_MMAP_VALID |
556 				 /* No full-resume yet implemented */
557 				 /* SNDRV_PCM_INFO_RESUME |*/
558 				 SNDRV_PCM_INFO_PAUSE |
559 				 SNDRV_PCM_INFO_SYNC_START |
560 				 SNDRV_PCM_INFO_HAS_WALL_CLOCK | /* legacy */
561 				 SNDRV_PCM_INFO_HAS_LINK_ATIME |
562 				 SNDRV_PCM_INFO_NO_PERIOD_WAKEUP),
563 	.formats =		SNDRV_PCM_FMTBIT_S16_LE,
564 	.rates =		SNDRV_PCM_RATE_48000,
565 	.rate_min =		48000,
566 	.rate_max =		48000,
567 	.channels_min =		2,
568 	.channels_max =		2,
569 	.buffer_bytes_max =	AZX_MAX_BUF_SIZE,
570 	.period_bytes_min =	128,
571 	.period_bytes_max =	AZX_MAX_BUF_SIZE / 2,
572 	.periods_min =		2,
573 	.periods_max =		AZX_MAX_FRAG,
574 	.fifo_size =		0,
575 };
576 
577 static int azx_pcm_open(struct snd_pcm_substream *substream)
578 {
579 	struct azx_pcm *apcm = snd_pcm_substream_chip(substream);
580 	struct hda_pcm_stream *hinfo = to_hda_pcm_stream(substream);
581 	struct azx *chip = apcm->chip;
582 	struct azx_dev *azx_dev;
583 	struct snd_pcm_runtime *runtime = substream->runtime;
584 	int err;
585 	int buff_step;
586 
587 	snd_hda_codec_pcm_get(apcm->info);
588 	mutex_lock(&chip->open_mutex);
589 	azx_dev = azx_assign_device(chip, substream);
590 	trace_azx_pcm_open(chip, azx_dev);
591 	if (azx_dev == NULL) {
592 		err = -EBUSY;
593 		goto unlock;
594 	}
595 	runtime->private_data = azx_dev;
596 
597 	runtime->hw = azx_pcm_hw;
598 	if (chip->gts_present)
599 		runtime->hw.info |= SNDRV_PCM_INFO_HAS_LINK_SYNCHRONIZED_ATIME;
600 	runtime->hw.channels_min = hinfo->channels_min;
601 	runtime->hw.channels_max = hinfo->channels_max;
602 	runtime->hw.formats = hinfo->formats;
603 	runtime->hw.rates = hinfo->rates;
604 	snd_pcm_limit_hw_rates(runtime);
605 	snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIODS);
606 
607 	/* avoid wrap-around with wall-clock */
608 	snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_TIME,
609 				     20,
610 				     178000000);
611 
612 	if (chip->align_buffer_size)
613 		/* constrain buffer sizes to be multiple of 128
614 		   bytes. This is more efficient in terms of memory
615 		   access but isn't required by the HDA spec and
616 		   prevents users from specifying exact period/buffer
617 		   sizes. For example for 44.1kHz, a period size set
618 		   to 20ms will be rounded to 19.59ms. */
619 		buff_step = 128;
620 	else
621 		/* Don't enforce steps on buffer sizes, still need to
622 		   be multiple of 4 bytes (HDA spec). Tested on Intel
623 		   HDA controllers, may not work on all devices where
624 		   option needs to be disabled */
625 		buff_step = 4;
626 
627 	snd_pcm_hw_constraint_step(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES,
628 				   buff_step);
629 	snd_pcm_hw_constraint_step(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES,
630 				   buff_step);
631 	snd_hda_power_up(apcm->codec);
632 	if (hinfo->ops.open)
633 		err = hinfo->ops.open(hinfo, apcm->codec, substream);
634 	else
635 		err = -ENODEV;
636 	if (err < 0) {
637 		azx_release_device(azx_dev);
638 		goto powerdown;
639 	}
640 	snd_pcm_limit_hw_rates(runtime);
641 	/* sanity check */
642 	if (snd_BUG_ON(!runtime->hw.channels_min) ||
643 	    snd_BUG_ON(!runtime->hw.channels_max) ||
644 	    snd_BUG_ON(!runtime->hw.formats) ||
645 	    snd_BUG_ON(!runtime->hw.rates)) {
646 		azx_release_device(azx_dev);
647 		if (hinfo->ops.close)
648 			hinfo->ops.close(hinfo, apcm->codec, substream);
649 		err = -EINVAL;
650 		goto powerdown;
651 	}
652 
653 	/* disable LINK_ATIME timestamps for capture streams
654 	   until we figure out how to handle digital inputs */
655 	if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) {
656 		runtime->hw.info &= ~SNDRV_PCM_INFO_HAS_WALL_CLOCK; /* legacy */
657 		runtime->hw.info &= ~SNDRV_PCM_INFO_HAS_LINK_ATIME;
658 	}
659 
660 	snd_pcm_set_sync(substream);
661 	mutex_unlock(&chip->open_mutex);
662 	return 0;
663 
664  powerdown:
665 	snd_hda_power_down(apcm->codec);
666  unlock:
667 	mutex_unlock(&chip->open_mutex);
668 	snd_hda_codec_pcm_put(apcm->info);
669 	return err;
670 }
671 
672 static const struct snd_pcm_ops azx_pcm_ops = {
673 	.open = azx_pcm_open,
674 	.close = azx_pcm_close,
675 	.hw_params = azx_pcm_hw_params,
676 	.hw_free = azx_pcm_hw_free,
677 	.prepare = azx_pcm_prepare,
678 	.trigger = azx_pcm_trigger,
679 	.pointer = azx_pcm_pointer,
680 	.get_time_info =  azx_get_time_info,
681 };
682 
683 static void azx_pcm_free(struct snd_pcm *pcm)
684 {
685 	struct azx_pcm *apcm = pcm->private_data;
686 	if (apcm) {
687 		list_del(&apcm->list);
688 		apcm->info->pcm = NULL;
689 		kfree(apcm);
690 	}
691 }
692 
693 #define MAX_PREALLOC_SIZE	(32 * 1024 * 1024)
694 
695 int snd_hda_attach_pcm_stream(struct hda_bus *_bus, struct hda_codec *codec,
696 			      struct hda_pcm *cpcm)
697 {
698 	struct hdac_bus *bus = &_bus->core;
699 	struct azx *chip = bus_to_azx(bus);
700 	struct snd_pcm *pcm;
701 	struct azx_pcm *apcm;
702 	int pcm_dev = cpcm->device;
703 	unsigned int size;
704 	int s, err;
705 	int type = SNDRV_DMA_TYPE_DEV_SG;
706 
707 	list_for_each_entry(apcm, &chip->pcm_list, list) {
708 		if (apcm->pcm->device == pcm_dev) {
709 			dev_err(chip->card->dev, "PCM %d already exists\n",
710 				pcm_dev);
711 			return -EBUSY;
712 		}
713 	}
714 	err = snd_pcm_new(chip->card, cpcm->name, pcm_dev,
715 			  cpcm->stream[SNDRV_PCM_STREAM_PLAYBACK].substreams,
716 			  cpcm->stream[SNDRV_PCM_STREAM_CAPTURE].substreams,
717 			  &pcm);
718 	if (err < 0)
719 		return err;
720 	strscpy(pcm->name, cpcm->name, sizeof(pcm->name));
721 	apcm = kzalloc(sizeof(*apcm), GFP_KERNEL);
722 	if (apcm == NULL) {
723 		snd_device_free(chip->card, pcm);
724 		return -ENOMEM;
725 	}
726 	apcm->chip = chip;
727 	apcm->pcm = pcm;
728 	apcm->codec = codec;
729 	apcm->info = cpcm;
730 	pcm->private_data = apcm;
731 	pcm->private_free = azx_pcm_free;
732 	if (cpcm->pcm_type == HDA_PCM_TYPE_MODEM)
733 		pcm->dev_class = SNDRV_PCM_CLASS_MODEM;
734 	list_add_tail(&apcm->list, &chip->pcm_list);
735 	cpcm->pcm = pcm;
736 	for (s = 0; s < 2; s++) {
737 		if (cpcm->stream[s].substreams)
738 			snd_pcm_set_ops(pcm, s, &azx_pcm_ops);
739 	}
740 	/* buffer pre-allocation */
741 	size = CONFIG_SND_HDA_PREALLOC_SIZE * 1024;
742 	if (size > MAX_PREALLOC_SIZE)
743 		size = MAX_PREALLOC_SIZE;
744 	if (chip->uc_buffer)
745 		type = SNDRV_DMA_TYPE_DEV_WC_SG;
746 	snd_pcm_set_managed_buffer_all(pcm, type, chip->card->dev,
747 				       size, MAX_PREALLOC_SIZE);
748 	return 0;
749 }
750 
751 static unsigned int azx_command_addr(u32 cmd)
752 {
753 	unsigned int addr = cmd >> 28;
754 
755 	if (addr >= AZX_MAX_CODECS) {
756 		snd_BUG();
757 		addr = 0;
758 	}
759 
760 	return addr;
761 }
762 
763 /* receive a response */
764 static int azx_rirb_get_response(struct hdac_bus *bus, unsigned int addr,
765 				 unsigned int *res)
766 {
767 	struct azx *chip = bus_to_azx(bus);
768 	struct hda_bus *hbus = &chip->bus;
769 	int err;
770 
771  again:
772 	err = snd_hdac_bus_get_response(bus, addr, res);
773 	if (!err)
774 		return 0;
775 
776 	if (hbus->no_response_fallback)
777 		return -EIO;
778 
779 	if (!bus->polling_mode) {
780 		dev_warn(chip->card->dev,
781 			 "azx_get_response timeout, switching to polling mode: last cmd=0x%08x\n",
782 			 bus->last_cmd[addr]);
783 		bus->polling_mode = 1;
784 		goto again;
785 	}
786 
787 	if (chip->msi) {
788 		dev_warn(chip->card->dev,
789 			 "No response from codec, disabling MSI: last cmd=0x%08x\n",
790 			 bus->last_cmd[addr]);
791 		if (chip->ops->disable_msi_reset_irq &&
792 		    chip->ops->disable_msi_reset_irq(chip) < 0)
793 			return -EIO;
794 		goto again;
795 	}
796 
797 	if (chip->probing) {
798 		/* If this critical timeout happens during the codec probing
799 		 * phase, this is likely an access to a non-existing codec
800 		 * slot.  Better to return an error and reset the system.
801 		 */
802 		return -EIO;
803 	}
804 
805 	/* no fallback mechanism? */
806 	if (!chip->fallback_to_single_cmd)
807 		return -EIO;
808 
809 	/* a fatal communication error; need either to reset or to fallback
810 	 * to the single_cmd mode
811 	 */
812 	if (hbus->allow_bus_reset && !hbus->response_reset && !hbus->in_reset) {
813 		hbus->response_reset = 1;
814 		dev_err(chip->card->dev,
815 			"No response from codec, resetting bus: last cmd=0x%08x\n",
816 			bus->last_cmd[addr]);
817 		return -EAGAIN; /* give a chance to retry */
818 	}
819 
820 	dev_err(chip->card->dev,
821 		"azx_get_response timeout, switching to single_cmd mode: last cmd=0x%08x\n",
822 		bus->last_cmd[addr]);
823 	chip->single_cmd = 1;
824 	hbus->response_reset = 0;
825 	snd_hdac_bus_stop_cmd_io(bus);
826 	return -EIO;
827 }
828 
829 /*
830  * Use the single immediate command instead of CORB/RIRB for simplicity
831  *
832  * Note: according to Intel, this is not preferred use.  The command was
833  *       intended for the BIOS only, and may get confused with unsolicited
834  *       responses.  So, we shouldn't use it for normal operation from the
835  *       driver.
836  *       I left the codes, however, for debugging/testing purposes.
837  */
838 
839 /* receive a response */
840 static int azx_single_wait_for_response(struct azx *chip, unsigned int addr)
841 {
842 	int timeout = 50;
843 
844 	while (timeout--) {
845 		/* check IRV busy bit */
846 		if (azx_readw(chip, IRS) & AZX_IRS_VALID) {
847 			/* reuse rirb.res as the response return value */
848 			azx_bus(chip)->rirb.res[addr] = azx_readl(chip, IR);
849 			return 0;
850 		}
851 		udelay(1);
852 	}
853 	if (printk_ratelimit())
854 		dev_dbg(chip->card->dev, "get_response timeout: IRS=0x%x\n",
855 			azx_readw(chip, IRS));
856 	azx_bus(chip)->rirb.res[addr] = -1;
857 	return -EIO;
858 }
859 
860 /* send a command */
861 static int azx_single_send_cmd(struct hdac_bus *bus, u32 val)
862 {
863 	struct azx *chip = bus_to_azx(bus);
864 	unsigned int addr = azx_command_addr(val);
865 	int timeout = 50;
866 
867 	bus->last_cmd[azx_command_addr(val)] = val;
868 	while (timeout--) {
869 		/* check ICB busy bit */
870 		if (!((azx_readw(chip, IRS) & AZX_IRS_BUSY))) {
871 			/* Clear IRV valid bit */
872 			azx_writew(chip, IRS, azx_readw(chip, IRS) |
873 				   AZX_IRS_VALID);
874 			azx_writel(chip, IC, val);
875 			azx_writew(chip, IRS, azx_readw(chip, IRS) |
876 				   AZX_IRS_BUSY);
877 			return azx_single_wait_for_response(chip, addr);
878 		}
879 		udelay(1);
880 	}
881 	if (printk_ratelimit())
882 		dev_dbg(chip->card->dev,
883 			"send_cmd timeout: IRS=0x%x, val=0x%x\n",
884 			azx_readw(chip, IRS), val);
885 	return -EIO;
886 }
887 
888 /* receive a response */
889 static int azx_single_get_response(struct hdac_bus *bus, unsigned int addr,
890 				   unsigned int *res)
891 {
892 	if (res)
893 		*res = bus->rirb.res[addr];
894 	return 0;
895 }
896 
897 /*
898  * The below are the main callbacks from hda_codec.
899  *
900  * They are just the skeleton to call sub-callbacks according to the
901  * current setting of chip->single_cmd.
902  */
903 
904 /* send a command */
905 static int azx_send_cmd(struct hdac_bus *bus, unsigned int val)
906 {
907 	struct azx *chip = bus_to_azx(bus);
908 
909 	if (chip->disabled)
910 		return 0;
911 	if (chip->single_cmd)
912 		return azx_single_send_cmd(bus, val);
913 	else
914 		return snd_hdac_bus_send_cmd(bus, val);
915 }
916 
917 /* get a response */
918 static int azx_get_response(struct hdac_bus *bus, unsigned int addr,
919 			    unsigned int *res)
920 {
921 	struct azx *chip = bus_to_azx(bus);
922 
923 	if (chip->disabled)
924 		return 0;
925 	if (chip->single_cmd)
926 		return azx_single_get_response(bus, addr, res);
927 	else
928 		return azx_rirb_get_response(bus, addr, res);
929 }
930 
931 static const struct hdac_bus_ops bus_core_ops = {
932 	.command = azx_send_cmd,
933 	.get_response = azx_get_response,
934 };
935 
936 #ifdef CONFIG_SND_HDA_DSP_LOADER
937 /*
938  * DSP loading code (e.g. for CA0132)
939  */
940 
941 /* use the first stream for loading DSP */
942 static struct azx_dev *
943 azx_get_dsp_loader_dev(struct azx *chip)
944 {
945 	struct hdac_bus *bus = azx_bus(chip);
946 	struct hdac_stream *s;
947 
948 	list_for_each_entry(s, &bus->stream_list, list)
949 		if (s->index == chip->playback_index_offset)
950 			return stream_to_azx_dev(s);
951 
952 	return NULL;
953 }
954 
955 int snd_hda_codec_load_dsp_prepare(struct hda_codec *codec, unsigned int format,
956 				   unsigned int byte_size,
957 				   struct snd_dma_buffer *bufp)
958 {
959 	struct hdac_bus *bus = &codec->bus->core;
960 	struct azx *chip = bus_to_azx(bus);
961 	struct azx_dev *azx_dev;
962 	struct hdac_stream *hstr;
963 	bool saved = false;
964 	int err;
965 
966 	azx_dev = azx_get_dsp_loader_dev(chip);
967 	hstr = azx_stream(azx_dev);
968 	spin_lock_irq(&bus->reg_lock);
969 	if (hstr->opened) {
970 		chip->saved_azx_dev = *azx_dev;
971 		saved = true;
972 	}
973 	spin_unlock_irq(&bus->reg_lock);
974 
975 	err = snd_hdac_dsp_prepare(hstr, format, byte_size, bufp);
976 	if (err < 0) {
977 		spin_lock_irq(&bus->reg_lock);
978 		if (saved)
979 			*azx_dev = chip->saved_azx_dev;
980 		spin_unlock_irq(&bus->reg_lock);
981 		return err;
982 	}
983 
984 	hstr->prepared = 0;
985 	return err;
986 }
987 EXPORT_SYMBOL_GPL(snd_hda_codec_load_dsp_prepare);
988 
989 void snd_hda_codec_load_dsp_trigger(struct hda_codec *codec, bool start)
990 {
991 	struct hdac_bus *bus = &codec->bus->core;
992 	struct azx *chip = bus_to_azx(bus);
993 	struct azx_dev *azx_dev = azx_get_dsp_loader_dev(chip);
994 
995 	snd_hdac_dsp_trigger(azx_stream(azx_dev), start);
996 }
997 EXPORT_SYMBOL_GPL(snd_hda_codec_load_dsp_trigger);
998 
999 void snd_hda_codec_load_dsp_cleanup(struct hda_codec *codec,
1000 				    struct snd_dma_buffer *dmab)
1001 {
1002 	struct hdac_bus *bus = &codec->bus->core;
1003 	struct azx *chip = bus_to_azx(bus);
1004 	struct azx_dev *azx_dev = azx_get_dsp_loader_dev(chip);
1005 	struct hdac_stream *hstr = azx_stream(azx_dev);
1006 
1007 	if (!dmab->area || !hstr->locked)
1008 		return;
1009 
1010 	snd_hdac_dsp_cleanup(hstr, dmab);
1011 	spin_lock_irq(&bus->reg_lock);
1012 	if (hstr->opened)
1013 		*azx_dev = chip->saved_azx_dev;
1014 	hstr->locked = false;
1015 	spin_unlock_irq(&bus->reg_lock);
1016 }
1017 EXPORT_SYMBOL_GPL(snd_hda_codec_load_dsp_cleanup);
1018 #endif /* CONFIG_SND_HDA_DSP_LOADER */
1019 
1020 /*
1021  * reset and start the controller registers
1022  */
1023 void azx_init_chip(struct azx *chip, bool full_reset)
1024 {
1025 	if (snd_hdac_bus_init_chip(azx_bus(chip), full_reset)) {
1026 		/* correct RINTCNT for CXT */
1027 		if (chip->driver_caps & AZX_DCAPS_CTX_WORKAROUND)
1028 			azx_writew(chip, RINTCNT, 0xc0);
1029 	}
1030 }
1031 EXPORT_SYMBOL_GPL(azx_init_chip);
1032 
1033 void azx_stop_all_streams(struct azx *chip)
1034 {
1035 	struct hdac_bus *bus = azx_bus(chip);
1036 
1037 	snd_hdac_stop_streams(bus);
1038 }
1039 EXPORT_SYMBOL_GPL(azx_stop_all_streams);
1040 
1041 void azx_stop_chip(struct azx *chip)
1042 {
1043 	snd_hdac_bus_stop_chip(azx_bus(chip));
1044 }
1045 EXPORT_SYMBOL_GPL(azx_stop_chip);
1046 
1047 /*
1048  * interrupt handler
1049  */
1050 static void stream_update(struct hdac_bus *bus, struct hdac_stream *s)
1051 {
1052 	struct azx *chip = bus_to_azx(bus);
1053 	struct azx_dev *azx_dev = stream_to_azx_dev(s);
1054 
1055 	/* check whether this IRQ is really acceptable */
1056 	if (!chip->ops->position_check ||
1057 	    chip->ops->position_check(chip, azx_dev)) {
1058 		spin_unlock(&bus->reg_lock);
1059 		snd_pcm_period_elapsed(azx_stream(azx_dev)->substream);
1060 		spin_lock(&bus->reg_lock);
1061 	}
1062 }
1063 
1064 irqreturn_t azx_interrupt(int irq, void *dev_id)
1065 {
1066 	struct azx *chip = dev_id;
1067 	struct hdac_bus *bus = azx_bus(chip);
1068 	u32 status;
1069 	bool active, handled = false;
1070 	int repeat = 0; /* count for avoiding endless loop */
1071 
1072 #ifdef CONFIG_PM
1073 	if (azx_has_pm_runtime(chip))
1074 		if (!pm_runtime_active(chip->card->dev))
1075 			return IRQ_NONE;
1076 #endif
1077 
1078 	spin_lock(&bus->reg_lock);
1079 
1080 	if (chip->disabled)
1081 		goto unlock;
1082 
1083 	do {
1084 		status = azx_readl(chip, INTSTS);
1085 		if (status == 0 || status == 0xffffffff)
1086 			break;
1087 
1088 		handled = true;
1089 		active = false;
1090 		if (snd_hdac_bus_handle_stream_irq(bus, status, stream_update))
1091 			active = true;
1092 
1093 		status = azx_readb(chip, RIRBSTS);
1094 		if (status & RIRB_INT_MASK) {
1095 			/*
1096 			 * Clearing the interrupt status here ensures that no
1097 			 * interrupt gets masked after the RIRB wp is read in
1098 			 * snd_hdac_bus_update_rirb. This avoids a possible
1099 			 * race condition where codec response in RIRB may
1100 			 * remain unserviced by IRQ, eventually falling back
1101 			 * to polling mode in azx_rirb_get_response.
1102 			 */
1103 			azx_writeb(chip, RIRBSTS, RIRB_INT_MASK);
1104 			active = true;
1105 			if (status & RIRB_INT_RESPONSE) {
1106 				if (chip->driver_caps & AZX_DCAPS_CTX_WORKAROUND)
1107 					udelay(80);
1108 				snd_hdac_bus_update_rirb(bus);
1109 			}
1110 		}
1111 	} while (active && ++repeat < 10);
1112 
1113  unlock:
1114 	spin_unlock(&bus->reg_lock);
1115 
1116 	return IRQ_RETVAL(handled);
1117 }
1118 EXPORT_SYMBOL_GPL(azx_interrupt);
1119 
1120 /*
1121  * Codec initerface
1122  */
1123 
1124 /*
1125  * Probe the given codec address
1126  */
1127 static int probe_codec(struct azx *chip, int addr)
1128 {
1129 	unsigned int cmd = (addr << 28) | (AC_NODE_ROOT << 20) |
1130 		(AC_VERB_PARAMETERS << 8) | AC_PAR_VENDOR_ID;
1131 	struct hdac_bus *bus = azx_bus(chip);
1132 	int err;
1133 	unsigned int res = -1;
1134 
1135 	mutex_lock(&bus->cmd_mutex);
1136 	chip->probing = 1;
1137 	azx_send_cmd(bus, cmd);
1138 	err = azx_get_response(bus, addr, &res);
1139 	chip->probing = 0;
1140 	mutex_unlock(&bus->cmd_mutex);
1141 	if (err < 0 || res == -1)
1142 		return -EIO;
1143 	dev_dbg(chip->card->dev, "codec #%d probed OK\n", addr);
1144 	return 0;
1145 }
1146 
1147 void snd_hda_bus_reset(struct hda_bus *bus)
1148 {
1149 	struct azx *chip = bus_to_azx(&bus->core);
1150 
1151 	bus->in_reset = 1;
1152 	azx_stop_chip(chip);
1153 	azx_init_chip(chip, true);
1154 	if (bus->core.chip_init)
1155 		snd_hda_bus_reset_codecs(bus);
1156 	bus->in_reset = 0;
1157 }
1158 
1159 /* HD-audio bus initialization */
1160 int azx_bus_init(struct azx *chip, const char *model)
1161 {
1162 	struct hda_bus *bus = &chip->bus;
1163 	int err;
1164 
1165 	err = snd_hdac_bus_init(&bus->core, chip->card->dev, &bus_core_ops);
1166 	if (err < 0)
1167 		return err;
1168 
1169 	bus->card = chip->card;
1170 	mutex_init(&bus->prepare_mutex);
1171 	bus->pci = chip->pci;
1172 	bus->modelname = model;
1173 	bus->mixer_assigned = -1;
1174 	bus->core.snoop = azx_snoop(chip);
1175 	if (chip->get_position[0] != azx_get_pos_lpib ||
1176 	    chip->get_position[1] != azx_get_pos_lpib)
1177 		bus->core.use_posbuf = true;
1178 	bus->core.bdl_pos_adj = chip->bdl_pos_adj;
1179 	if (chip->driver_caps & AZX_DCAPS_CORBRP_SELF_CLEAR)
1180 		bus->core.corbrp_self_clear = true;
1181 
1182 	if (chip->driver_caps & AZX_DCAPS_4K_BDLE_BOUNDARY)
1183 		bus->core.align_bdle_4k = true;
1184 
1185 	/* enable sync_write flag for stable communication as default */
1186 	bus->core.sync_write = 1;
1187 
1188 	return 0;
1189 }
1190 EXPORT_SYMBOL_GPL(azx_bus_init);
1191 
1192 /* Probe codecs */
1193 int azx_probe_codecs(struct azx *chip, unsigned int max_slots)
1194 {
1195 	struct hdac_bus *bus = azx_bus(chip);
1196 	int c, codecs, err;
1197 
1198 	codecs = 0;
1199 	if (!max_slots)
1200 		max_slots = AZX_DEFAULT_CODECS;
1201 
1202 	/* First try to probe all given codec slots */
1203 	for (c = 0; c < max_slots; c++) {
1204 		if ((bus->codec_mask & (1 << c)) & chip->codec_probe_mask) {
1205 			if (probe_codec(chip, c) < 0) {
1206 				/* Some BIOSen give you wrong codec addresses
1207 				 * that don't exist
1208 				 */
1209 				dev_warn(chip->card->dev,
1210 					 "Codec #%d probe error; disabling it...\n", c);
1211 				bus->codec_mask &= ~(1 << c);
1212 				/* More badly, accessing to a non-existing
1213 				 * codec often screws up the controller chip,
1214 				 * and disturbs the further communications.
1215 				 * Thus if an error occurs during probing,
1216 				 * better to reset the controller chip to
1217 				 * get back to the sanity state.
1218 				 */
1219 				azx_stop_chip(chip);
1220 				azx_init_chip(chip, true);
1221 			}
1222 		}
1223 	}
1224 
1225 	/* Then create codec instances */
1226 	for (c = 0; c < max_slots; c++) {
1227 		if ((bus->codec_mask & (1 << c)) & chip->codec_probe_mask) {
1228 			struct hda_codec *codec;
1229 			err = snd_hda_codec_new(&chip->bus, chip->card, c, &codec);
1230 			if (err < 0)
1231 				continue;
1232 			codec->jackpoll_interval = chip->jackpoll_interval;
1233 			codec->beep_mode = chip->beep_mode;
1234 			codec->ctl_dev_id = chip->ctl_dev_id;
1235 			codecs++;
1236 		}
1237 	}
1238 	if (!codecs) {
1239 		dev_err(chip->card->dev, "no codecs initialized\n");
1240 		return -ENXIO;
1241 	}
1242 	return 0;
1243 }
1244 EXPORT_SYMBOL_GPL(azx_probe_codecs);
1245 
1246 /* configure each codec instance */
1247 int azx_codec_configure(struct azx *chip)
1248 {
1249 	struct hda_codec *codec, *next;
1250 	int success = 0;
1251 
1252 	list_for_each_codec(codec, &chip->bus) {
1253 		if (!snd_hda_codec_configure(codec))
1254 			success++;
1255 	}
1256 
1257 	if (success) {
1258 		/* unregister failed codecs if any codec has been probed */
1259 		list_for_each_codec_safe(codec, next, &chip->bus) {
1260 			if (!codec->configured) {
1261 				codec_err(codec, "Unable to configure, disabling\n");
1262 				snd_hdac_device_unregister(&codec->core);
1263 			}
1264 		}
1265 	}
1266 
1267 	return success ? 0 : -ENODEV;
1268 }
1269 EXPORT_SYMBOL_GPL(azx_codec_configure);
1270 
1271 static int stream_direction(struct azx *chip, unsigned char index)
1272 {
1273 	if (index >= chip->capture_index_offset &&
1274 	    index < chip->capture_index_offset + chip->capture_streams)
1275 		return SNDRV_PCM_STREAM_CAPTURE;
1276 	return SNDRV_PCM_STREAM_PLAYBACK;
1277 }
1278 
1279 /* initialize SD streams */
1280 int azx_init_streams(struct azx *chip)
1281 {
1282 	int i;
1283 	int stream_tags[2] = { 0, 0 };
1284 
1285 	/* initialize each stream (aka device)
1286 	 * assign the starting bdl address to each stream (device)
1287 	 * and initialize
1288 	 */
1289 	for (i = 0; i < chip->num_streams; i++) {
1290 		struct azx_dev *azx_dev = kzalloc(sizeof(*azx_dev), GFP_KERNEL);
1291 		int dir, tag;
1292 
1293 		if (!azx_dev)
1294 			return -ENOMEM;
1295 
1296 		dir = stream_direction(chip, i);
1297 		/* stream tag must be unique throughout
1298 		 * the stream direction group,
1299 		 * valid values 1...15
1300 		 * use separate stream tag if the flag
1301 		 * AZX_DCAPS_SEPARATE_STREAM_TAG is used
1302 		 */
1303 		if (chip->driver_caps & AZX_DCAPS_SEPARATE_STREAM_TAG)
1304 			tag = ++stream_tags[dir];
1305 		else
1306 			tag = i + 1;
1307 		snd_hdac_stream_init(azx_bus(chip), azx_stream(azx_dev),
1308 				     i, dir, tag);
1309 	}
1310 
1311 	return 0;
1312 }
1313 EXPORT_SYMBOL_GPL(azx_init_streams);
1314 
1315 void azx_free_streams(struct azx *chip)
1316 {
1317 	struct hdac_bus *bus = azx_bus(chip);
1318 	struct hdac_stream *s;
1319 
1320 	while (!list_empty(&bus->stream_list)) {
1321 		s = list_first_entry(&bus->stream_list, struct hdac_stream, list);
1322 		list_del(&s->list);
1323 		kfree(stream_to_azx_dev(s));
1324 	}
1325 }
1326 EXPORT_SYMBOL_GPL(azx_free_streams);
1327