1 // SPDX-License-Identifier: GPL-2.0+
2 //
3 //  Copyright (C) 2013, Analog Devices Inc.
4 //	Author: Lars-Peter Clausen <lars@metafoo.de>
5 
6 #include <linux/module.h>
7 #include <linux/init.h>
8 #include <linux/dmaengine.h>
9 #include <linux/slab.h>
10 #include <sound/pcm.h>
11 #include <sound/pcm_params.h>
12 #include <sound/soc.h>
13 #include <linux/dma-mapping.h>
14 #include <linux/of.h>
15 
16 #include <sound/dmaengine_pcm.h>
17 
18 static unsigned int prealloc_buffer_size_kbytes = 512;
19 module_param(prealloc_buffer_size_kbytes, uint, 0444);
20 MODULE_PARM_DESC(prealloc_buffer_size_kbytes, "Preallocate DMA buffer size (KB).");
21 
22 /*
23  * The platforms dmaengine driver does not support reporting the amount of
24  * bytes that are still left to transfer.
25  */
26 #define SND_DMAENGINE_PCM_FLAG_NO_RESIDUE BIT(31)
27 
28 static struct device *dmaengine_dma_dev(struct dmaengine_pcm *pcm,
29 	struct snd_pcm_substream *substream)
30 {
31 	if (!pcm->chan[substream->stream])
32 		return NULL;
33 
34 	return pcm->chan[substream->stream]->device->dev;
35 }
36 
37 /**
38  * snd_dmaengine_pcm_prepare_slave_config() - Generic prepare_slave_config callback
39  * @substream: PCM substream
40  * @params: hw_params
41  * @slave_config: DMA slave config to prepare
42  *
43  * This function can be used as a generic prepare_slave_config callback for
44  * platforms which make use of the snd_dmaengine_dai_dma_data struct for their
45  * DAI DMA data. Internally the function will first call
46  * snd_hwparams_to_dma_slave_config to fill in the slave config based on the
47  * hw_params, followed by snd_dmaengine_pcm_set_config_from_dai_data to fill in
48  * the remaining fields based on the DAI DMA data.
49  */
50 int snd_dmaengine_pcm_prepare_slave_config(struct snd_pcm_substream *substream,
51 	struct snd_pcm_hw_params *params, struct dma_slave_config *slave_config)
52 {
53 	struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream);
54 	struct snd_dmaengine_dai_dma_data *dma_data;
55 	int ret;
56 
57 	if (rtd->dai_link->num_cpus > 1) {
58 		dev_err(rtd->dev,
59 			"%s doesn't support Multi CPU yet\n", __func__);
60 		return -EINVAL;
61 	}
62 
63 	dma_data = snd_soc_dai_get_dma_data(asoc_rtd_to_cpu(rtd, 0), substream);
64 
65 	ret = snd_hwparams_to_dma_slave_config(substream, params, slave_config);
66 	if (ret)
67 		return ret;
68 
69 	snd_dmaengine_pcm_set_config_from_dai_data(substream, dma_data,
70 		slave_config);
71 
72 	return 0;
73 }
74 EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_prepare_slave_config);
75 
76 static int dmaengine_pcm_hw_params(struct snd_soc_component *component,
77 				   struct snd_pcm_substream *substream,
78 				   struct snd_pcm_hw_params *params)
79 {
80 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
81 	struct dma_chan *chan = snd_dmaengine_pcm_get_chan(substream);
82 	struct dma_slave_config slave_config;
83 	int ret;
84 
85 	if (!pcm->config->prepare_slave_config)
86 		return 0;
87 
88 	memset(&slave_config, 0, sizeof(slave_config));
89 
90 	ret = pcm->config->prepare_slave_config(substream, params, &slave_config);
91 	if (ret)
92 		return ret;
93 
94 	return dmaengine_slave_config(chan, &slave_config);
95 }
96 
97 static int
98 dmaengine_pcm_set_runtime_hwparams(struct snd_soc_component *component,
99 				   struct snd_pcm_substream *substream)
100 {
101 	struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream);
102 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
103 	struct device *dma_dev = dmaengine_dma_dev(pcm, substream);
104 	struct dma_chan *chan = pcm->chan[substream->stream];
105 	struct snd_dmaengine_dai_dma_data *dma_data;
106 	struct snd_pcm_hardware hw;
107 
108 	if (rtd->dai_link->num_cpus > 1) {
109 		dev_err(rtd->dev,
110 			"%s doesn't support Multi CPU yet\n", __func__);
111 		return -EINVAL;
112 	}
113 
114 	if (pcm->config->pcm_hardware)
115 		return snd_soc_set_runtime_hwparams(substream,
116 				pcm->config->pcm_hardware);
117 
118 	dma_data = snd_soc_dai_get_dma_data(asoc_rtd_to_cpu(rtd, 0), substream);
119 
120 	memset(&hw, 0, sizeof(hw));
121 	hw.info = SNDRV_PCM_INFO_MMAP | SNDRV_PCM_INFO_MMAP_VALID |
122 			SNDRV_PCM_INFO_INTERLEAVED;
123 	hw.periods_min = 2;
124 	hw.periods_max = UINT_MAX;
125 	hw.period_bytes_min = dma_data->maxburst * DMA_SLAVE_BUSWIDTH_8_BYTES;
126 	if (!hw.period_bytes_min)
127 		hw.period_bytes_min = 256;
128 	hw.period_bytes_max = dma_get_max_seg_size(dma_dev);
129 	hw.buffer_bytes_max = SIZE_MAX;
130 	hw.fifo_size = dma_data->fifo_size;
131 
132 	if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE)
133 		hw.info |= SNDRV_PCM_INFO_BATCH;
134 
135 	/**
136 	 * FIXME: Remove the return value check to align with the code
137 	 * before adding snd_dmaengine_pcm_refine_runtime_hwparams
138 	 * function.
139 	 */
140 	snd_dmaengine_pcm_refine_runtime_hwparams(substream,
141 						  dma_data,
142 						  &hw,
143 						  chan);
144 
145 	return snd_soc_set_runtime_hwparams(substream, &hw);
146 }
147 
148 static int dmaengine_pcm_open(struct snd_soc_component *component,
149 			      struct snd_pcm_substream *substream)
150 {
151 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
152 	struct dma_chan *chan = pcm->chan[substream->stream];
153 	int ret;
154 
155 	ret = dmaengine_pcm_set_runtime_hwparams(component, substream);
156 	if (ret)
157 		return ret;
158 
159 	return snd_dmaengine_pcm_open(substream, chan);
160 }
161 
162 static int dmaengine_pcm_close(struct snd_soc_component *component,
163 			       struct snd_pcm_substream *substream)
164 {
165 	return snd_dmaengine_pcm_close(substream);
166 }
167 
168 static int dmaengine_pcm_trigger(struct snd_soc_component *component,
169 				 struct snd_pcm_substream *substream, int cmd)
170 {
171 	return snd_dmaengine_pcm_trigger(substream, cmd);
172 }
173 
174 static struct dma_chan *dmaengine_pcm_compat_request_channel(
175 	struct snd_soc_component *component,
176 	struct snd_soc_pcm_runtime *rtd,
177 	struct snd_pcm_substream *substream)
178 {
179 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
180 	struct snd_dmaengine_dai_dma_data *dma_data;
181 
182 	if (rtd->dai_link->num_cpus > 1) {
183 		dev_err(rtd->dev,
184 			"%s doesn't support Multi CPU yet\n", __func__);
185 		return NULL;
186 	}
187 
188 	dma_data = snd_soc_dai_get_dma_data(asoc_rtd_to_cpu(rtd, 0), substream);
189 
190 	if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) && pcm->chan[0])
191 		return pcm->chan[0];
192 
193 	if (pcm->config->compat_request_channel)
194 		return pcm->config->compat_request_channel(rtd, substream);
195 
196 	return snd_dmaengine_pcm_request_channel(pcm->config->compat_filter_fn,
197 						 dma_data->filter_data);
198 }
199 
200 static bool dmaengine_pcm_can_report_residue(struct device *dev,
201 	struct dma_chan *chan)
202 {
203 	struct dma_slave_caps dma_caps;
204 	int ret;
205 
206 	ret = dma_get_slave_caps(chan, &dma_caps);
207 	if (ret != 0) {
208 		dev_warn(dev, "Failed to get DMA channel capabilities, falling back to period counting: %d\n",
209 			 ret);
210 		return false;
211 	}
212 
213 	if (dma_caps.residue_granularity == DMA_RESIDUE_GRANULARITY_DESCRIPTOR)
214 		return false;
215 
216 	return true;
217 }
218 
219 static int dmaengine_pcm_new(struct snd_soc_component *component,
220 			     struct snd_soc_pcm_runtime *rtd)
221 {
222 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
223 	const struct snd_dmaengine_pcm_config *config = pcm->config;
224 	struct device *dev = component->dev;
225 	size_t prealloc_buffer_size;
226 	size_t max_buffer_size;
227 	unsigned int i;
228 
229 	if (config->prealloc_buffer_size)
230 		prealloc_buffer_size = config->prealloc_buffer_size;
231 	else
232 		prealloc_buffer_size = prealloc_buffer_size_kbytes * 1024;
233 
234 	if (config->pcm_hardware && config->pcm_hardware->buffer_bytes_max)
235 		max_buffer_size = config->pcm_hardware->buffer_bytes_max;
236 	else
237 		max_buffer_size = SIZE_MAX;
238 
239 	for_each_pcm_streams(i) {
240 		struct snd_pcm_substream *substream = rtd->pcm->streams[i].substream;
241 		if (!substream)
242 			continue;
243 
244 		if (!pcm->chan[i] && config->chan_names[i])
245 			pcm->chan[i] = dma_request_slave_channel(dev,
246 				config->chan_names[i]);
247 
248 		if (!pcm->chan[i] && (pcm->flags & SND_DMAENGINE_PCM_FLAG_COMPAT)) {
249 			pcm->chan[i] = dmaengine_pcm_compat_request_channel(
250 				component, rtd, substream);
251 		}
252 
253 		if (!pcm->chan[i]) {
254 			dev_err(component->dev,
255 				"Missing dma channel for stream: %d\n", i);
256 			return -EINVAL;
257 		}
258 
259 		snd_pcm_set_managed_buffer(substream,
260 				SNDRV_DMA_TYPE_DEV_IRAM,
261 				dmaengine_dma_dev(pcm, substream),
262 				prealloc_buffer_size,
263 				max_buffer_size);
264 
265 		if (!dmaengine_pcm_can_report_residue(dev, pcm->chan[i]))
266 			pcm->flags |= SND_DMAENGINE_PCM_FLAG_NO_RESIDUE;
267 
268 		if (rtd->pcm->streams[i].pcm->name[0] == '\0') {
269 			strscpy_pad(rtd->pcm->streams[i].pcm->name,
270 				    rtd->pcm->streams[i].pcm->id,
271 				    sizeof(rtd->pcm->streams[i].pcm->name));
272 		}
273 	}
274 
275 	return 0;
276 }
277 
278 static snd_pcm_uframes_t dmaengine_pcm_pointer(
279 	struct snd_soc_component *component,
280 	struct snd_pcm_substream *substream)
281 {
282 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
283 
284 	if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE)
285 		return snd_dmaengine_pcm_pointer_no_residue(substream);
286 	else
287 		return snd_dmaengine_pcm_pointer(substream);
288 }
289 
290 static int dmaengine_copy(struct snd_soc_component *component,
291 			  struct snd_pcm_substream *substream,
292 			  int channel, unsigned long hwoff,
293 			  struct iov_iter *iter, unsigned long bytes)
294 {
295 	struct snd_pcm_runtime *runtime = substream->runtime;
296 	struct dmaengine_pcm *pcm = soc_component_to_pcm(component);
297 	int (*process)(struct snd_pcm_substream *substream,
298 		       int channel, unsigned long hwoff,
299 		       unsigned long bytes) = pcm->config->process;
300 	bool is_playback = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
301 	void *dma_ptr = runtime->dma_area + hwoff +
302 			channel * (runtime->dma_bytes / runtime->channels);
303 
304 	if (is_playback)
305 		if (copy_from_iter(dma_ptr, bytes, iter) != bytes)
306 			return -EFAULT;
307 
308 	if (process) {
309 		int ret = process(substream, channel, hwoff, bytes);
310 		if (ret < 0)
311 			return ret;
312 	}
313 
314 	if (!is_playback)
315 		if (copy_to_iter(dma_ptr, bytes, iter) != bytes)
316 			return -EFAULT;
317 
318 	return 0;
319 }
320 
321 static int dmaengine_pcm_sync_stop(struct snd_soc_component *component,
322 				   struct snd_pcm_substream *substream)
323 {
324 	return snd_dmaengine_pcm_sync_stop(substream);
325 }
326 
327 static const struct snd_soc_component_driver dmaengine_pcm_component = {
328 	.name		= SND_DMAENGINE_PCM_DRV_NAME,
329 	.probe_order	= SND_SOC_COMP_ORDER_LATE,
330 	.open		= dmaengine_pcm_open,
331 	.close		= dmaengine_pcm_close,
332 	.hw_params	= dmaengine_pcm_hw_params,
333 	.trigger	= dmaengine_pcm_trigger,
334 	.pointer	= dmaengine_pcm_pointer,
335 	.pcm_construct	= dmaengine_pcm_new,
336 	.sync_stop	= dmaengine_pcm_sync_stop,
337 };
338 
339 static const struct snd_soc_component_driver dmaengine_pcm_component_process = {
340 	.name		= SND_DMAENGINE_PCM_DRV_NAME,
341 	.probe_order	= SND_SOC_COMP_ORDER_LATE,
342 	.open		= dmaengine_pcm_open,
343 	.close		= dmaengine_pcm_close,
344 	.hw_params	= dmaengine_pcm_hw_params,
345 	.trigger	= dmaengine_pcm_trigger,
346 	.pointer	= dmaengine_pcm_pointer,
347 	.copy		= dmaengine_copy,
348 	.pcm_construct	= dmaengine_pcm_new,
349 	.sync_stop	= dmaengine_pcm_sync_stop,
350 };
351 
352 static const char * const dmaengine_pcm_dma_channel_names[] = {
353 	[SNDRV_PCM_STREAM_PLAYBACK] = "tx",
354 	[SNDRV_PCM_STREAM_CAPTURE] = "rx",
355 };
356 
357 static int dmaengine_pcm_request_chan_of(struct dmaengine_pcm *pcm,
358 	struct device *dev, const struct snd_dmaengine_pcm_config *config)
359 {
360 	unsigned int i;
361 	const char *name;
362 	struct dma_chan *chan;
363 
364 	if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_DT) || (!dev->of_node &&
365 	    !(config->dma_dev && config->dma_dev->of_node)))
366 		return 0;
367 
368 	if (config->dma_dev) {
369 		/*
370 		 * If this warning is seen, it probably means that your Linux
371 		 * device structure does not match your HW device structure.
372 		 * It would be best to refactor the Linux device structure to
373 		 * correctly match the HW structure.
374 		 */
375 		dev_warn(dev, "DMA channels sourced from device %s",
376 			 dev_name(config->dma_dev));
377 		dev = config->dma_dev;
378 	}
379 
380 	for_each_pcm_streams(i) {
381 		if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
382 			name = "rx-tx";
383 		else
384 			name = dmaengine_pcm_dma_channel_names[i];
385 		if (config->chan_names[i])
386 			name = config->chan_names[i];
387 		chan = dma_request_chan(dev, name);
388 		if (IS_ERR(chan)) {
389 			/*
390 			 * Only report probe deferral errors, channels
391 			 * might not be present for devices that
392 			 * support only TX or only RX.
393 			 */
394 			if (PTR_ERR(chan) == -EPROBE_DEFER)
395 				return -EPROBE_DEFER;
396 			pcm->chan[i] = NULL;
397 		} else {
398 			pcm->chan[i] = chan;
399 		}
400 		if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
401 			break;
402 	}
403 
404 	if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
405 		pcm->chan[1] = pcm->chan[0];
406 
407 	return 0;
408 }
409 
410 static void dmaengine_pcm_release_chan(struct dmaengine_pcm *pcm)
411 {
412 	unsigned int i;
413 
414 	for_each_pcm_streams(i) {
415 		if (!pcm->chan[i])
416 			continue;
417 		dma_release_channel(pcm->chan[i]);
418 		if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX)
419 			break;
420 	}
421 }
422 
423 static const struct snd_dmaengine_pcm_config snd_dmaengine_pcm_default_config = {
424 	.prepare_slave_config = snd_dmaengine_pcm_prepare_slave_config,
425 };
426 
427 /**
428  * snd_dmaengine_pcm_register - Register a dmaengine based PCM device
429  * @dev: The parent device for the PCM device
430  * @config: Platform specific PCM configuration
431  * @flags: Platform specific quirks
432  */
433 int snd_dmaengine_pcm_register(struct device *dev,
434 	const struct snd_dmaengine_pcm_config *config, unsigned int flags)
435 {
436 	const struct snd_soc_component_driver *driver;
437 	struct dmaengine_pcm *pcm;
438 	int ret;
439 
440 	pcm = kzalloc(sizeof(*pcm), GFP_KERNEL);
441 	if (!pcm)
442 		return -ENOMEM;
443 
444 #ifdef CONFIG_DEBUG_FS
445 	pcm->component.debugfs_prefix = "dma";
446 #endif
447 	if (!config)
448 		config = &snd_dmaengine_pcm_default_config;
449 	pcm->config = config;
450 	pcm->flags = flags;
451 
452 	ret = dmaengine_pcm_request_chan_of(pcm, dev, config);
453 	if (ret)
454 		goto err_free_dma;
455 
456 	if (config->process)
457 		driver = &dmaengine_pcm_component_process;
458 	else
459 		driver = &dmaengine_pcm_component;
460 
461 	ret = snd_soc_component_initialize(&pcm->component, driver, dev);
462 	if (ret)
463 		goto err_free_dma;
464 
465 	ret = snd_soc_add_component(&pcm->component, NULL, 0);
466 	if (ret)
467 		goto err_free_dma;
468 
469 	return 0;
470 
471 err_free_dma:
472 	dmaengine_pcm_release_chan(pcm);
473 	kfree(pcm);
474 	return ret;
475 }
476 EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_register);
477 
478 /**
479  * snd_dmaengine_pcm_unregister - Removes a dmaengine based PCM device
480  * @dev: Parent device the PCM was register with
481  *
482  * Removes a dmaengine based PCM device previously registered with
483  * snd_dmaengine_pcm_register.
484  */
485 void snd_dmaengine_pcm_unregister(struct device *dev)
486 {
487 	struct snd_soc_component *component;
488 	struct dmaengine_pcm *pcm;
489 
490 	component = snd_soc_lookup_component(dev, SND_DMAENGINE_PCM_DRV_NAME);
491 	if (!component)
492 		return;
493 
494 	pcm = soc_component_to_pcm(component);
495 
496 	snd_soc_unregister_component_by_driver(dev, component->driver);
497 	dmaengine_pcm_release_chan(pcm);
498 	kfree(pcm);
499 }
500 EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_unregister);
501 
502 MODULE_LICENSE("GPL");
503