1 // SPDX-License-Identifier: GPL-2.0+ 2 // 3 // Copyright (C) 2013, Analog Devices Inc. 4 // Author: Lars-Peter Clausen <lars@metafoo.de> 5 6 #include <linux/module.h> 7 #include <linux/init.h> 8 #include <linux/dmaengine.h> 9 #include <linux/slab.h> 10 #include <sound/pcm.h> 11 #include <sound/pcm_params.h> 12 #include <sound/soc.h> 13 #include <linux/dma-mapping.h> 14 #include <linux/of.h> 15 16 #include <sound/dmaengine_pcm.h> 17 18 /* 19 * The platforms dmaengine driver does not support reporting the amount of 20 * bytes that are still left to transfer. 21 */ 22 #define SND_DMAENGINE_PCM_FLAG_NO_RESIDUE BIT(31) 23 24 static struct device *dmaengine_dma_dev(struct dmaengine_pcm *pcm, 25 struct snd_pcm_substream *substream) 26 { 27 if (!pcm->chan[substream->stream]) 28 return NULL; 29 30 return pcm->chan[substream->stream]->device->dev; 31 } 32 33 /** 34 * snd_dmaengine_pcm_prepare_slave_config() - Generic prepare_slave_config callback 35 * @substream: PCM substream 36 * @params: hw_params 37 * @slave_config: DMA slave config to prepare 38 * 39 * This function can be used as a generic prepare_slave_config callback for 40 * platforms which make use of the snd_dmaengine_dai_dma_data struct for their 41 * DAI DMA data. Internally the function will first call 42 * snd_hwparams_to_dma_slave_config to fill in the slave config based on the 43 * hw_params, followed by snd_dmaengine_set_config_from_dai_data to fill in the 44 * remaining fields based on the DAI DMA data. 45 */ 46 int snd_dmaengine_pcm_prepare_slave_config(struct snd_pcm_substream *substream, 47 struct snd_pcm_hw_params *params, struct dma_slave_config *slave_config) 48 { 49 struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); 50 struct snd_dmaengine_dai_dma_data *dma_data; 51 int ret; 52 53 if (rtd->num_cpus > 1) { 54 dev_err(rtd->dev, 55 "%s doesn't support Multi CPU yet\n", __func__); 56 return -EINVAL; 57 } 58 59 dma_data = snd_soc_dai_get_dma_data(asoc_rtd_to_cpu(rtd, 0), substream); 60 61 ret = snd_hwparams_to_dma_slave_config(substream, params, slave_config); 62 if (ret) 63 return ret; 64 65 snd_dmaengine_pcm_set_config_from_dai_data(substream, dma_data, 66 slave_config); 67 68 return 0; 69 } 70 EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_prepare_slave_config); 71 72 static int dmaengine_pcm_hw_params(struct snd_soc_component *component, 73 struct snd_pcm_substream *substream, 74 struct snd_pcm_hw_params *params) 75 { 76 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 77 struct dma_chan *chan = snd_dmaengine_pcm_get_chan(substream); 78 int (*prepare_slave_config)(struct snd_pcm_substream *substream, 79 struct snd_pcm_hw_params *params, 80 struct dma_slave_config *slave_config); 81 struct dma_slave_config slave_config; 82 int ret; 83 84 memset(&slave_config, 0, sizeof(slave_config)); 85 86 if (!pcm->config) 87 prepare_slave_config = snd_dmaengine_pcm_prepare_slave_config; 88 else 89 prepare_slave_config = pcm->config->prepare_slave_config; 90 91 if (prepare_slave_config) { 92 ret = prepare_slave_config(substream, params, &slave_config); 93 if (ret) 94 return ret; 95 96 ret = dmaengine_slave_config(chan, &slave_config); 97 if (ret) 98 return ret; 99 } 100 101 return 0; 102 } 103 104 static int 105 dmaengine_pcm_set_runtime_hwparams(struct snd_soc_component *component, 106 struct snd_pcm_substream *substream) 107 { 108 struct snd_soc_pcm_runtime *rtd = asoc_substream_to_rtd(substream); 109 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 110 struct device *dma_dev = dmaengine_dma_dev(pcm, substream); 111 struct dma_chan *chan = pcm->chan[substream->stream]; 112 struct snd_dmaengine_dai_dma_data *dma_data; 113 struct snd_pcm_hardware hw; 114 115 if (rtd->num_cpus > 1) { 116 dev_err(rtd->dev, 117 "%s doesn't support Multi CPU yet\n", __func__); 118 return -EINVAL; 119 } 120 121 if (pcm->config && pcm->config->pcm_hardware) 122 return snd_soc_set_runtime_hwparams(substream, 123 pcm->config->pcm_hardware); 124 125 dma_data = snd_soc_dai_get_dma_data(asoc_rtd_to_cpu(rtd, 0), substream); 126 127 memset(&hw, 0, sizeof(hw)); 128 hw.info = SNDRV_PCM_INFO_MMAP | SNDRV_PCM_INFO_MMAP_VALID | 129 SNDRV_PCM_INFO_INTERLEAVED; 130 hw.periods_min = 2; 131 hw.periods_max = UINT_MAX; 132 hw.period_bytes_min = 256; 133 hw.period_bytes_max = dma_get_max_seg_size(dma_dev); 134 hw.buffer_bytes_max = SIZE_MAX; 135 hw.fifo_size = dma_data->fifo_size; 136 137 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE) 138 hw.info |= SNDRV_PCM_INFO_BATCH; 139 140 /** 141 * FIXME: Remove the return value check to align with the code 142 * before adding snd_dmaengine_pcm_refine_runtime_hwparams 143 * function. 144 */ 145 snd_dmaengine_pcm_refine_runtime_hwparams(substream, 146 dma_data, 147 &hw, 148 chan); 149 150 return snd_soc_set_runtime_hwparams(substream, &hw); 151 } 152 153 static int dmaengine_pcm_open(struct snd_soc_component *component, 154 struct snd_pcm_substream *substream) 155 { 156 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 157 struct dma_chan *chan = pcm->chan[substream->stream]; 158 int ret; 159 160 ret = dmaengine_pcm_set_runtime_hwparams(component, substream); 161 if (ret) 162 return ret; 163 164 return snd_dmaengine_pcm_open(substream, chan); 165 } 166 167 static int dmaengine_pcm_close(struct snd_soc_component *component, 168 struct snd_pcm_substream *substream) 169 { 170 return snd_dmaengine_pcm_close(substream); 171 } 172 173 static int dmaengine_pcm_trigger(struct snd_soc_component *component, 174 struct snd_pcm_substream *substream, int cmd) 175 { 176 return snd_dmaengine_pcm_trigger(substream, cmd); 177 } 178 179 static struct dma_chan *dmaengine_pcm_compat_request_channel( 180 struct snd_soc_component *component, 181 struct snd_soc_pcm_runtime *rtd, 182 struct snd_pcm_substream *substream) 183 { 184 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 185 struct snd_dmaengine_dai_dma_data *dma_data; 186 dma_filter_fn fn = NULL; 187 188 if (rtd->num_cpus > 1) { 189 dev_err(rtd->dev, 190 "%s doesn't support Multi CPU yet\n", __func__); 191 return NULL; 192 } 193 194 dma_data = snd_soc_dai_get_dma_data(asoc_rtd_to_cpu(rtd, 0), substream); 195 196 if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) && pcm->chan[0]) 197 return pcm->chan[0]; 198 199 if (pcm->config && pcm->config->compat_request_channel) 200 return pcm->config->compat_request_channel(rtd, substream); 201 202 if (pcm->config) 203 fn = pcm->config->compat_filter_fn; 204 205 return snd_dmaengine_pcm_request_channel(fn, dma_data->filter_data); 206 } 207 208 static bool dmaengine_pcm_can_report_residue(struct device *dev, 209 struct dma_chan *chan) 210 { 211 struct dma_slave_caps dma_caps; 212 int ret; 213 214 ret = dma_get_slave_caps(chan, &dma_caps); 215 if (ret != 0) { 216 dev_warn(dev, "Failed to get DMA channel capabilities, falling back to period counting: %d\n", 217 ret); 218 return false; 219 } 220 221 if (dma_caps.residue_granularity == DMA_RESIDUE_GRANULARITY_DESCRIPTOR) 222 return false; 223 224 return true; 225 } 226 227 static int dmaengine_pcm_new(struct snd_soc_component *component, 228 struct snd_soc_pcm_runtime *rtd) 229 { 230 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 231 const struct snd_dmaengine_pcm_config *config = pcm->config; 232 struct device *dev = component->dev; 233 struct snd_pcm_substream *substream; 234 size_t prealloc_buffer_size; 235 size_t max_buffer_size; 236 unsigned int i; 237 238 if (config && config->prealloc_buffer_size) { 239 prealloc_buffer_size = config->prealloc_buffer_size; 240 max_buffer_size = config->pcm_hardware->buffer_bytes_max; 241 } else { 242 prealloc_buffer_size = 512 * 1024; 243 max_buffer_size = SIZE_MAX; 244 } 245 246 for_each_pcm_streams(i) { 247 substream = rtd->pcm->streams[i].substream; 248 if (!substream) 249 continue; 250 251 if (!pcm->chan[i] && config && config->chan_names[i]) 252 pcm->chan[i] = dma_request_slave_channel(dev, 253 config->chan_names[i]); 254 255 if (!pcm->chan[i] && (pcm->flags & SND_DMAENGINE_PCM_FLAG_COMPAT)) { 256 pcm->chan[i] = dmaengine_pcm_compat_request_channel( 257 component, rtd, substream); 258 } 259 260 if (!pcm->chan[i]) { 261 dev_err(component->dev, 262 "Missing dma channel for stream: %d\n", i); 263 return -EINVAL; 264 } 265 266 snd_pcm_set_managed_buffer(substream, 267 SNDRV_DMA_TYPE_DEV_IRAM, 268 dmaengine_dma_dev(pcm, substream), 269 prealloc_buffer_size, 270 max_buffer_size); 271 272 if (!dmaengine_pcm_can_report_residue(dev, pcm->chan[i])) 273 pcm->flags |= SND_DMAENGINE_PCM_FLAG_NO_RESIDUE; 274 275 if (rtd->pcm->streams[i].pcm->name[0] == '\0') { 276 strscpy_pad(rtd->pcm->streams[i].pcm->name, 277 rtd->pcm->streams[i].pcm->id, 278 sizeof(rtd->pcm->streams[i].pcm->name)); 279 } 280 } 281 282 return 0; 283 } 284 285 static snd_pcm_uframes_t dmaengine_pcm_pointer( 286 struct snd_soc_component *component, 287 struct snd_pcm_substream *substream) 288 { 289 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 290 291 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_RESIDUE) 292 return snd_dmaengine_pcm_pointer_no_residue(substream); 293 else 294 return snd_dmaengine_pcm_pointer(substream); 295 } 296 297 static int dmaengine_copy_user(struct snd_soc_component *component, 298 struct snd_pcm_substream *substream, 299 int channel, unsigned long hwoff, 300 void __user *buf, unsigned long bytes) 301 { 302 struct snd_pcm_runtime *runtime = substream->runtime; 303 struct dmaengine_pcm *pcm = soc_component_to_pcm(component); 304 int (*process)(struct snd_pcm_substream *substream, 305 int channel, unsigned long hwoff, 306 void *buf, unsigned long bytes) = pcm->config->process; 307 bool is_playback = substream->stream == SNDRV_PCM_STREAM_PLAYBACK; 308 void *dma_ptr = runtime->dma_area + hwoff + 309 channel * (runtime->dma_bytes / runtime->channels); 310 int ret; 311 312 if (is_playback) 313 if (copy_from_user(dma_ptr, buf, bytes)) 314 return -EFAULT; 315 316 if (process) { 317 ret = process(substream, channel, hwoff, (__force void *)buf, bytes); 318 if (ret < 0) 319 return ret; 320 } 321 322 if (!is_playback) 323 if (copy_to_user(buf, dma_ptr, bytes)) 324 return -EFAULT; 325 326 return 0; 327 } 328 329 static const struct snd_soc_component_driver dmaengine_pcm_component = { 330 .name = SND_DMAENGINE_PCM_DRV_NAME, 331 .probe_order = SND_SOC_COMP_ORDER_LATE, 332 .open = dmaengine_pcm_open, 333 .close = dmaengine_pcm_close, 334 .hw_params = dmaengine_pcm_hw_params, 335 .trigger = dmaengine_pcm_trigger, 336 .pointer = dmaengine_pcm_pointer, 337 .pcm_construct = dmaengine_pcm_new, 338 }; 339 340 static const struct snd_soc_component_driver dmaengine_pcm_component_process = { 341 .name = SND_DMAENGINE_PCM_DRV_NAME, 342 .probe_order = SND_SOC_COMP_ORDER_LATE, 343 .open = dmaengine_pcm_open, 344 .close = dmaengine_pcm_close, 345 .hw_params = dmaengine_pcm_hw_params, 346 .trigger = dmaengine_pcm_trigger, 347 .pointer = dmaengine_pcm_pointer, 348 .copy_user = dmaengine_copy_user, 349 .pcm_construct = dmaengine_pcm_new, 350 }; 351 352 static const char * const dmaengine_pcm_dma_channel_names[] = { 353 [SNDRV_PCM_STREAM_PLAYBACK] = "tx", 354 [SNDRV_PCM_STREAM_CAPTURE] = "rx", 355 }; 356 357 static int dmaengine_pcm_request_chan_of(struct dmaengine_pcm *pcm, 358 struct device *dev, const struct snd_dmaengine_pcm_config *config) 359 { 360 unsigned int i; 361 const char *name; 362 struct dma_chan *chan; 363 364 if ((pcm->flags & SND_DMAENGINE_PCM_FLAG_NO_DT) || (!dev->of_node && 365 !(config && config->dma_dev && config->dma_dev->of_node))) 366 return 0; 367 368 if (config && config->dma_dev) { 369 /* 370 * If this warning is seen, it probably means that your Linux 371 * device structure does not match your HW device structure. 372 * It would be best to refactor the Linux device structure to 373 * correctly match the HW structure. 374 */ 375 dev_warn(dev, "DMA channels sourced from device %s", 376 dev_name(config->dma_dev)); 377 dev = config->dma_dev; 378 } 379 380 for_each_pcm_streams(i) { 381 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) 382 name = "rx-tx"; 383 else 384 name = dmaengine_pcm_dma_channel_names[i]; 385 if (config && config->chan_names[i]) 386 name = config->chan_names[i]; 387 chan = dma_request_chan(dev, name); 388 if (IS_ERR(chan)) { 389 if (PTR_ERR(chan) == -EPROBE_DEFER) 390 return -EPROBE_DEFER; 391 pcm->chan[i] = NULL; 392 } else { 393 pcm->chan[i] = chan; 394 } 395 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) 396 break; 397 } 398 399 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) 400 pcm->chan[1] = pcm->chan[0]; 401 402 return 0; 403 } 404 405 static void dmaengine_pcm_release_chan(struct dmaengine_pcm *pcm) 406 { 407 unsigned int i; 408 409 for_each_pcm_streams(i) { 410 if (!pcm->chan[i]) 411 continue; 412 dma_release_channel(pcm->chan[i]); 413 if (pcm->flags & SND_DMAENGINE_PCM_FLAG_HALF_DUPLEX) 414 break; 415 } 416 } 417 418 /** 419 * snd_dmaengine_pcm_register - Register a dmaengine based PCM device 420 * @dev: The parent device for the PCM device 421 * @config: Platform specific PCM configuration 422 * @flags: Platform specific quirks 423 */ 424 int snd_dmaengine_pcm_register(struct device *dev, 425 const struct snd_dmaengine_pcm_config *config, unsigned int flags) 426 { 427 const struct snd_soc_component_driver *driver; 428 struct dmaengine_pcm *pcm; 429 int ret; 430 431 pcm = kzalloc(sizeof(*pcm), GFP_KERNEL); 432 if (!pcm) 433 return -ENOMEM; 434 435 #ifdef CONFIG_DEBUG_FS 436 pcm->component.debugfs_prefix = "dma"; 437 #endif 438 pcm->config = config; 439 pcm->flags = flags; 440 441 ret = dmaengine_pcm_request_chan_of(pcm, dev, config); 442 if (ret) 443 goto err_free_dma; 444 445 if (config && config->process) 446 driver = &dmaengine_pcm_component_process; 447 else 448 driver = &dmaengine_pcm_component; 449 450 ret = snd_soc_component_initialize(&pcm->component, driver, dev); 451 if (ret) 452 goto err_free_dma; 453 454 ret = snd_soc_add_component(&pcm->component, NULL, 0); 455 if (ret) 456 goto err_free_dma; 457 458 return 0; 459 460 err_free_dma: 461 dmaengine_pcm_release_chan(pcm); 462 kfree(pcm); 463 return ret; 464 } 465 EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_register); 466 467 /** 468 * snd_dmaengine_pcm_unregister - Removes a dmaengine based PCM device 469 * @dev: Parent device the PCM was register with 470 * 471 * Removes a dmaengine based PCM device previously registered with 472 * snd_dmaengine_pcm_register. 473 */ 474 void snd_dmaengine_pcm_unregister(struct device *dev) 475 { 476 struct snd_soc_component *component; 477 struct dmaengine_pcm *pcm; 478 479 component = snd_soc_lookup_component(dev, SND_DMAENGINE_PCM_DRV_NAME); 480 if (!component) 481 return; 482 483 pcm = soc_component_to_pcm(component); 484 485 snd_soc_unregister_component_by_driver(dev, component->driver); 486 dmaengine_pcm_release_chan(pcm); 487 kfree(pcm); 488 } 489 EXPORT_SYMBOL_GPL(snd_dmaengine_pcm_unregister); 490 491 MODULE_LICENSE("GPL"); 492