1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * sst.c - Intel SST Driver for audio engine 4 * 5 * Copyright (C) 2008-14 Intel Corp 6 * Authors: Vinod Koul <vinod.koul@intel.com> 7 * Harsha Priya <priya.harsha@intel.com> 8 * Dharageswari R <dharageswari.r@intel.com> 9 * KP Jeeja <jeeja.kp@intel.com> 10 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 11 * 12 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 13 */ 14 #include <linux/module.h> 15 #include <linux/fs.h> 16 #include <linux/interrupt.h> 17 #include <linux/io.h> 18 #include <linux/firmware.h> 19 #include <linux/pm_runtime.h> 20 #include <linux/pm_qos.h> 21 #include <linux/async.h> 22 #include <linux/acpi.h> 23 #include <linux/sysfs.h> 24 #include <sound/core.h> 25 #include <sound/soc.h> 26 #include <asm/platform_sst_audio.h> 27 #include "../sst-mfld-platform.h" 28 #include "sst.h" 29 30 MODULE_AUTHOR("Vinod Koul <vinod.koul@intel.com>"); 31 MODULE_AUTHOR("Harsha Priya <priya.harsha@intel.com>"); 32 MODULE_DESCRIPTION("Intel (R) SST(R) Audio Engine Driver"); 33 MODULE_LICENSE("GPL v2"); 34 35 static inline bool sst_is_process_reply(u32 msg_id) 36 { 37 return ((msg_id & PROCESS_MSG) ? true : false); 38 } 39 40 static inline bool sst_validate_mailbox_size(unsigned int size) 41 { 42 return ((size <= SST_MAILBOX_SIZE) ? true : false); 43 } 44 45 static irqreturn_t intel_sst_interrupt_mrfld(int irq, void *context) 46 { 47 union interrupt_reg_mrfld isr; 48 union ipc_header_mrfld header; 49 union sst_imr_reg_mrfld imr; 50 struct ipc_post *msg = NULL; 51 unsigned int size; 52 struct intel_sst_drv *drv = (struct intel_sst_drv *) context; 53 irqreturn_t retval = IRQ_HANDLED; 54 55 /* Interrupt arrived, check src */ 56 isr.full = sst_shim_read64(drv->shim, SST_ISRX); 57 58 if (isr.part.done_interrupt) { 59 /* Clear done bit */ 60 spin_lock(&drv->ipc_spin_lock); 61 header.full = sst_shim_read64(drv->shim, 62 drv->ipc_reg.ipcx); 63 header.p.header_high.part.done = 0; 64 sst_shim_write64(drv->shim, drv->ipc_reg.ipcx, header.full); 65 66 /* write 1 to clear status register */; 67 isr.part.done_interrupt = 1; 68 sst_shim_write64(drv->shim, SST_ISRX, isr.full); 69 spin_unlock(&drv->ipc_spin_lock); 70 71 /* we can send more messages to DSP so trigger work */ 72 queue_work(drv->post_msg_wq, &drv->ipc_post_msg_wq); 73 retval = IRQ_HANDLED; 74 } 75 76 if (isr.part.busy_interrupt) { 77 /* message from dsp so copy that */ 78 spin_lock(&drv->ipc_spin_lock); 79 imr.full = sst_shim_read64(drv->shim, SST_IMRX); 80 imr.part.busy_interrupt = 1; 81 sst_shim_write64(drv->shim, SST_IMRX, imr.full); 82 spin_unlock(&drv->ipc_spin_lock); 83 header.full = sst_shim_read64(drv->shim, drv->ipc_reg.ipcd); 84 85 if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) { 86 drv->ops->clear_interrupt(drv); 87 return IRQ_HANDLED; 88 } 89 90 if (header.p.header_high.part.large) { 91 size = header.p.header_low_payload; 92 if (sst_validate_mailbox_size(size)) { 93 memcpy_fromio(msg->mailbox_data, 94 drv->mailbox + drv->mailbox_recv_offset, size); 95 } else { 96 dev_err(drv->dev, 97 "Mailbox not copied, payload size is: %u\n", size); 98 header.p.header_low_payload = 0; 99 } 100 } 101 102 msg->mrfld_header = header; 103 msg->is_process_reply = 104 sst_is_process_reply(header.p.header_high.part.msg_id); 105 spin_lock(&drv->rx_msg_lock); 106 list_add_tail(&msg->node, &drv->rx_list); 107 spin_unlock(&drv->rx_msg_lock); 108 drv->ops->clear_interrupt(drv); 109 retval = IRQ_WAKE_THREAD; 110 } 111 return retval; 112 } 113 114 static irqreturn_t intel_sst_irq_thread_mrfld(int irq, void *context) 115 { 116 struct intel_sst_drv *drv = (struct intel_sst_drv *) context; 117 struct ipc_post *__msg, *msg = NULL; 118 unsigned long irq_flags; 119 120 spin_lock_irqsave(&drv->rx_msg_lock, irq_flags); 121 if (list_empty(&drv->rx_list)) { 122 spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags); 123 return IRQ_HANDLED; 124 } 125 126 list_for_each_entry_safe(msg, __msg, &drv->rx_list, node) { 127 list_del(&msg->node); 128 spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags); 129 if (msg->is_process_reply) 130 drv->ops->process_message(msg); 131 else 132 drv->ops->process_reply(drv, msg); 133 134 if (msg->is_large) 135 kfree(msg->mailbox_data); 136 kfree(msg); 137 spin_lock_irqsave(&drv->rx_msg_lock, irq_flags); 138 } 139 spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags); 140 return IRQ_HANDLED; 141 } 142 143 static int sst_save_dsp_context_v2(struct intel_sst_drv *sst) 144 { 145 int ret = 0; 146 147 ret = sst_prepare_and_post_msg(sst, SST_TASK_ID_MEDIA, IPC_CMD, 148 IPC_PREP_D3, PIPE_RSVD, 0, NULL, NULL, 149 true, true, false, true); 150 151 if (ret < 0) { 152 dev_err(sst->dev, "not suspending FW!!, Err: %d\n", ret); 153 return -EIO; 154 } 155 156 return 0; 157 } 158 159 160 static struct intel_sst_ops mrfld_ops = { 161 .interrupt = intel_sst_interrupt_mrfld, 162 .irq_thread = intel_sst_irq_thread_mrfld, 163 .clear_interrupt = intel_sst_clear_intr_mrfld, 164 .start = sst_start_mrfld, 165 .reset = intel_sst_reset_dsp_mrfld, 166 .post_message = sst_post_message_mrfld, 167 .process_reply = sst_process_reply_mrfld, 168 .save_dsp_context = sst_save_dsp_context_v2, 169 .alloc_stream = sst_alloc_stream_mrfld, 170 .post_download = sst_post_download_mrfld, 171 }; 172 173 int sst_driver_ops(struct intel_sst_drv *sst) 174 { 175 176 switch (sst->dev_id) { 177 case SST_MRFLD_PCI_ID: 178 case SST_BYT_ACPI_ID: 179 case SST_CHV_ACPI_ID: 180 sst->tstamp = SST_TIME_STAMP_MRFLD; 181 sst->ops = &mrfld_ops; 182 return 0; 183 184 default: 185 dev_err(sst->dev, 186 "SST Driver capabilities missing for dev_id: %x", 187 sst->dev_id); 188 return -EINVAL; 189 } 190 } 191 192 void sst_process_pending_msg(struct work_struct *work) 193 { 194 struct intel_sst_drv *ctx = container_of(work, 195 struct intel_sst_drv, ipc_post_msg_wq); 196 197 ctx->ops->post_message(ctx, NULL, false); 198 } 199 200 static int sst_workqueue_init(struct intel_sst_drv *ctx) 201 { 202 INIT_LIST_HEAD(&ctx->memcpy_list); 203 INIT_LIST_HEAD(&ctx->rx_list); 204 INIT_LIST_HEAD(&ctx->ipc_dispatch_list); 205 INIT_LIST_HEAD(&ctx->block_list); 206 INIT_WORK(&ctx->ipc_post_msg_wq, sst_process_pending_msg); 207 init_waitqueue_head(&ctx->wait_queue); 208 209 ctx->post_msg_wq = 210 create_singlethread_workqueue("sst_post_msg_wq"); 211 if (!ctx->post_msg_wq) 212 return -EBUSY; 213 return 0; 214 } 215 216 static void sst_init_locks(struct intel_sst_drv *ctx) 217 { 218 mutex_init(&ctx->sst_lock); 219 spin_lock_init(&ctx->rx_msg_lock); 220 spin_lock_init(&ctx->ipc_spin_lock); 221 spin_lock_init(&ctx->block_lock); 222 } 223 224 int sst_alloc_drv_context(struct intel_sst_drv **ctx, 225 struct device *dev, unsigned int dev_id) 226 { 227 *ctx = devm_kzalloc(dev, sizeof(struct intel_sst_drv), GFP_KERNEL); 228 if (!(*ctx)) 229 return -ENOMEM; 230 231 (*ctx)->dev = dev; 232 (*ctx)->dev_id = dev_id; 233 234 return 0; 235 } 236 EXPORT_SYMBOL_GPL(sst_alloc_drv_context); 237 238 static ssize_t firmware_version_show(struct device *dev, 239 struct device_attribute *attr, char *buf) 240 { 241 struct intel_sst_drv *ctx = dev_get_drvdata(dev); 242 243 if (ctx->fw_version.type == 0 && ctx->fw_version.major == 0 && 244 ctx->fw_version.minor == 0 && ctx->fw_version.build == 0) 245 return sprintf(buf, "FW not yet loaded\n"); 246 else 247 return sprintf(buf, "v%02x.%02x.%02x.%02x\n", 248 ctx->fw_version.type, ctx->fw_version.major, 249 ctx->fw_version.minor, ctx->fw_version.build); 250 251 } 252 253 static DEVICE_ATTR_RO(firmware_version); 254 255 static const struct attribute *sst_fw_version_attrs[] = { 256 &dev_attr_firmware_version.attr, 257 NULL, 258 }; 259 260 static const struct attribute_group sst_fw_version_attr_group = { 261 .attrs = (struct attribute **)sst_fw_version_attrs, 262 }; 263 264 int sst_context_init(struct intel_sst_drv *ctx) 265 { 266 int ret = 0, i; 267 268 if (!ctx->pdata) 269 return -EINVAL; 270 271 if (!ctx->pdata->probe_data) 272 return -EINVAL; 273 274 memcpy(&ctx->info, ctx->pdata->probe_data, sizeof(ctx->info)); 275 276 ret = sst_driver_ops(ctx); 277 if (ret != 0) 278 return -EINVAL; 279 280 sst_init_locks(ctx); 281 sst_set_fw_state_locked(ctx, SST_RESET); 282 283 /* pvt_id 0 reserved for async messages */ 284 ctx->pvt_id = 1; 285 ctx->stream_cnt = 0; 286 ctx->fw_in_mem = NULL; 287 /* we use memcpy, so set to 0 */ 288 ctx->use_dma = 0; 289 ctx->use_lli = 0; 290 291 if (sst_workqueue_init(ctx)) 292 return -EINVAL; 293 294 ctx->mailbox_recv_offset = ctx->pdata->ipc_info->mbox_recv_off; 295 ctx->ipc_reg.ipcx = SST_IPCX + ctx->pdata->ipc_info->ipc_offset; 296 ctx->ipc_reg.ipcd = SST_IPCD + ctx->pdata->ipc_info->ipc_offset; 297 298 dev_info(ctx->dev, "Got drv data max stream %d\n", 299 ctx->info.max_streams); 300 301 for (i = 1; i <= ctx->info.max_streams; i++) { 302 struct stream_info *stream = &ctx->streams[i]; 303 304 memset(stream, 0, sizeof(*stream)); 305 stream->pipe_id = PIPE_RSVD; 306 mutex_init(&stream->lock); 307 } 308 309 /* Register the ISR */ 310 ret = devm_request_threaded_irq(ctx->dev, ctx->irq_num, ctx->ops->interrupt, 311 ctx->ops->irq_thread, 0, SST_DRV_NAME, 312 ctx); 313 if (ret) 314 goto do_free_mem; 315 316 dev_dbg(ctx->dev, "Registered IRQ %#x\n", ctx->irq_num); 317 318 /* default intr are unmasked so set this as masked */ 319 sst_shim_write64(ctx->shim, SST_IMRX, 0xFFFF0038); 320 321 ctx->qos = devm_kzalloc(ctx->dev, 322 sizeof(struct pm_qos_request), GFP_KERNEL); 323 if (!ctx->qos) { 324 ret = -ENOMEM; 325 goto do_free_mem; 326 } 327 cpu_latency_qos_add_request(ctx->qos, PM_QOS_DEFAULT_VALUE); 328 329 dev_dbg(ctx->dev, "Requesting FW %s now...\n", ctx->firmware_name); 330 ret = request_firmware_nowait(THIS_MODULE, true, ctx->firmware_name, 331 ctx->dev, GFP_KERNEL, ctx, sst_firmware_load_cb); 332 if (ret) { 333 dev_err(ctx->dev, "Firmware download failed:%d\n", ret); 334 goto do_free_mem; 335 } 336 337 ret = sysfs_create_group(&ctx->dev->kobj, 338 &sst_fw_version_attr_group); 339 if (ret) { 340 dev_err(ctx->dev, 341 "Unable to create sysfs\n"); 342 goto err_sysfs; 343 } 344 345 sst_register(ctx->dev); 346 return 0; 347 err_sysfs: 348 sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group); 349 350 do_free_mem: 351 destroy_workqueue(ctx->post_msg_wq); 352 return ret; 353 } 354 EXPORT_SYMBOL_GPL(sst_context_init); 355 356 void sst_context_cleanup(struct intel_sst_drv *ctx) 357 { 358 pm_runtime_get_noresume(ctx->dev); 359 pm_runtime_disable(ctx->dev); 360 sst_unregister(ctx->dev); 361 sst_set_fw_state_locked(ctx, SST_SHUTDOWN); 362 sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group); 363 flush_scheduled_work(); 364 destroy_workqueue(ctx->post_msg_wq); 365 cpu_latency_qos_remove_request(ctx->qos); 366 kfree(ctx->fw_sg_list.src); 367 kfree(ctx->fw_sg_list.dst); 368 ctx->fw_sg_list.list_len = 0; 369 kfree(ctx->fw_in_mem); 370 ctx->fw_in_mem = NULL; 371 sst_memcpy_free_resources(ctx); 372 } 373 EXPORT_SYMBOL_GPL(sst_context_cleanup); 374 375 void sst_configure_runtime_pm(struct intel_sst_drv *ctx) 376 { 377 pm_runtime_set_autosuspend_delay(ctx->dev, SST_SUSPEND_DELAY); 378 pm_runtime_use_autosuspend(ctx->dev); 379 /* 380 * For acpi devices, the actual physical device state is 381 * initially active. So change the state to active before 382 * enabling the pm 383 */ 384 385 if (!acpi_disabled) 386 pm_runtime_set_active(ctx->dev); 387 388 pm_runtime_enable(ctx->dev); 389 390 if (acpi_disabled) 391 pm_runtime_set_active(ctx->dev); 392 else 393 pm_runtime_put_noidle(ctx->dev); 394 } 395 EXPORT_SYMBOL_GPL(sst_configure_runtime_pm); 396 397 static int intel_sst_runtime_suspend(struct device *dev) 398 { 399 int ret = 0; 400 struct intel_sst_drv *ctx = dev_get_drvdata(dev); 401 402 if (ctx->sst_state == SST_RESET) { 403 dev_dbg(dev, "LPE is already in RESET state, No action\n"); 404 return 0; 405 } 406 /* save fw context */ 407 if (ctx->ops->save_dsp_context(ctx)) 408 return -EBUSY; 409 410 /* Move the SST state to Reset */ 411 sst_set_fw_state_locked(ctx, SST_RESET); 412 413 synchronize_irq(ctx->irq_num); 414 flush_workqueue(ctx->post_msg_wq); 415 416 ctx->ops->reset(ctx); 417 418 return ret; 419 } 420 421 static int intel_sst_suspend(struct device *dev) 422 { 423 struct intel_sst_drv *ctx = dev_get_drvdata(dev); 424 struct sst_fw_save *fw_save; 425 int i, ret; 426 427 /* check first if we are already in SW reset */ 428 if (ctx->sst_state == SST_RESET) 429 return 0; 430 431 /* 432 * check if any stream is active and running 433 * they should already by suspend by soc_suspend 434 */ 435 for (i = 1; i <= ctx->info.max_streams; i++) { 436 struct stream_info *stream = &ctx->streams[i]; 437 438 if (stream->status == STREAM_RUNNING) { 439 dev_err(dev, "stream %d is running, can't suspend, abort\n", i); 440 return -EBUSY; 441 } 442 443 if (ctx->pdata->streams_lost_on_suspend) { 444 stream->resume_status = stream->status; 445 stream->resume_prev = stream->prev; 446 if (stream->status != STREAM_UN_INIT) 447 sst_free_stream(ctx, i); 448 } 449 } 450 synchronize_irq(ctx->irq_num); 451 flush_workqueue(ctx->post_msg_wq); 452 453 /* Move the SST state to Reset */ 454 sst_set_fw_state_locked(ctx, SST_RESET); 455 456 /* tell DSP we are suspending */ 457 if (ctx->ops->save_dsp_context(ctx)) 458 return -EBUSY; 459 460 /* save the memories */ 461 fw_save = kzalloc(sizeof(*fw_save), GFP_KERNEL); 462 if (!fw_save) 463 return -ENOMEM; 464 fw_save->iram = kvzalloc(ctx->iram_end - ctx->iram_base, GFP_KERNEL); 465 if (!fw_save->iram) { 466 ret = -ENOMEM; 467 goto iram; 468 } 469 fw_save->dram = kvzalloc(ctx->dram_end - ctx->dram_base, GFP_KERNEL); 470 if (!fw_save->dram) { 471 ret = -ENOMEM; 472 goto dram; 473 } 474 fw_save->sram = kvzalloc(SST_MAILBOX_SIZE, GFP_KERNEL); 475 if (!fw_save->sram) { 476 ret = -ENOMEM; 477 goto sram; 478 } 479 480 fw_save->ddr = kvzalloc(ctx->ddr_end - ctx->ddr_base, GFP_KERNEL); 481 if (!fw_save->ddr) { 482 ret = -ENOMEM; 483 goto ddr; 484 } 485 486 memcpy32_fromio(fw_save->iram, ctx->iram, ctx->iram_end - ctx->iram_base); 487 memcpy32_fromio(fw_save->dram, ctx->dram, ctx->dram_end - ctx->dram_base); 488 memcpy32_fromio(fw_save->sram, ctx->mailbox, SST_MAILBOX_SIZE); 489 memcpy32_fromio(fw_save->ddr, ctx->ddr, ctx->ddr_end - ctx->ddr_base); 490 491 ctx->fw_save = fw_save; 492 ctx->ops->reset(ctx); 493 return 0; 494 ddr: 495 kvfree(fw_save->sram); 496 sram: 497 kvfree(fw_save->dram); 498 dram: 499 kvfree(fw_save->iram); 500 iram: 501 kfree(fw_save); 502 return ret; 503 } 504 505 static int intel_sst_resume(struct device *dev) 506 { 507 struct intel_sst_drv *ctx = dev_get_drvdata(dev); 508 struct sst_fw_save *fw_save = ctx->fw_save; 509 struct sst_block *block; 510 int i, ret = 0; 511 512 if (!fw_save) 513 return 0; 514 515 sst_set_fw_state_locked(ctx, SST_FW_LOADING); 516 517 /* we have to restore the memory saved */ 518 ctx->ops->reset(ctx); 519 520 ctx->fw_save = NULL; 521 522 memcpy32_toio(ctx->iram, fw_save->iram, ctx->iram_end - ctx->iram_base); 523 memcpy32_toio(ctx->dram, fw_save->dram, ctx->dram_end - ctx->dram_base); 524 memcpy32_toio(ctx->mailbox, fw_save->sram, SST_MAILBOX_SIZE); 525 memcpy32_toio(ctx->ddr, fw_save->ddr, ctx->ddr_end - ctx->ddr_base); 526 527 kvfree(fw_save->sram); 528 kvfree(fw_save->dram); 529 kvfree(fw_save->iram); 530 kvfree(fw_save->ddr); 531 kfree(fw_save); 532 533 block = sst_create_block(ctx, 0, FW_DWNL_ID); 534 if (block == NULL) 535 return -ENOMEM; 536 537 538 /* start and wait for ack */ 539 ctx->ops->start(ctx); 540 ret = sst_wait_timeout(ctx, block); 541 if (ret) { 542 dev_err(ctx->dev, "fw download failed %d\n", ret); 543 /* FW download failed due to timeout */ 544 ret = -EBUSY; 545 546 } else { 547 sst_set_fw_state_locked(ctx, SST_FW_RUNNING); 548 } 549 550 if (ctx->pdata->streams_lost_on_suspend) { 551 for (i = 1; i <= ctx->info.max_streams; i++) { 552 struct stream_info *stream = &ctx->streams[i]; 553 554 if (stream->resume_status != STREAM_UN_INIT) { 555 dev_dbg(ctx->dev, "Re-allocing stream %d status %d prev %d\n", 556 i, stream->resume_status, 557 stream->resume_prev); 558 sst_realloc_stream(ctx, i); 559 stream->status = stream->resume_status; 560 stream->prev = stream->resume_prev; 561 } 562 } 563 } 564 565 sst_free_block(ctx, block); 566 return ret; 567 } 568 569 const struct dev_pm_ops intel_sst_pm = { 570 .suspend = intel_sst_suspend, 571 .resume = intel_sst_resume, 572 .runtime_suspend = intel_sst_runtime_suspend, 573 }; 574 EXPORT_SYMBOL_GPL(intel_sst_pm); 575