xref: /openbmc/linux/sound/soc/intel/atom/sst/sst.c (revision aa311596)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *  sst.c - Intel SST Driver for audio engine
4  *
5  *  Copyright (C) 2008-14	Intel Corp
6  *  Authors:	Vinod Koul <vinod.koul@intel.com>
7  *		Harsha Priya <priya.harsha@intel.com>
8  *		Dharageswari R <dharageswari.r@intel.com>
9  *		KP Jeeja <jeeja.kp@intel.com>
10  *  ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
11  *
12  * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
13  */
14 #include <linux/module.h>
15 #include <linux/fs.h>
16 #include <linux/interrupt.h>
17 #include <linux/io.h>
18 #include <linux/firmware.h>
19 #include <linux/pci.h>
20 #include <linux/pm_runtime.h>
21 #include <linux/pm_qos.h>
22 #include <linux/async.h>
23 #include <linux/acpi.h>
24 #include <linux/sysfs.h>
25 #include <sound/core.h>
26 #include <sound/soc.h>
27 #include <asm/platform_sst_audio.h>
28 #include "../sst-mfld-platform.h"
29 #include "sst.h"
30 
31 MODULE_AUTHOR("Vinod Koul <vinod.koul@intel.com>");
32 MODULE_AUTHOR("Harsha Priya <priya.harsha@intel.com>");
33 MODULE_DESCRIPTION("Intel (R) SST(R) Audio Engine Driver");
34 MODULE_LICENSE("GPL v2");
35 
36 static inline bool sst_is_process_reply(u32 msg_id)
37 {
38 	return ((msg_id & PROCESS_MSG) ? true : false);
39 }
40 
41 static inline bool sst_validate_mailbox_size(unsigned int size)
42 {
43 	return ((size <= SST_MAILBOX_SIZE) ? true : false);
44 }
45 
46 static irqreturn_t intel_sst_interrupt_mrfld(int irq, void *context)
47 {
48 	union interrupt_reg_mrfld isr;
49 	union ipc_header_mrfld header;
50 	union sst_imr_reg_mrfld imr;
51 	struct ipc_post *msg = NULL;
52 	unsigned int size;
53 	struct intel_sst_drv *drv = (struct intel_sst_drv *) context;
54 	irqreturn_t retval = IRQ_HANDLED;
55 
56 	/* Interrupt arrived, check src */
57 	isr.full = sst_shim_read64(drv->shim, SST_ISRX);
58 
59 	if (isr.part.done_interrupt) {
60 		/* Clear done bit */
61 		spin_lock(&drv->ipc_spin_lock);
62 		header.full = sst_shim_read64(drv->shim,
63 					drv->ipc_reg.ipcx);
64 		header.p.header_high.part.done = 0;
65 		sst_shim_write64(drv->shim, drv->ipc_reg.ipcx, header.full);
66 
67 		/* write 1 to clear status register */;
68 		isr.part.done_interrupt = 1;
69 		sst_shim_write64(drv->shim, SST_ISRX, isr.full);
70 		spin_unlock(&drv->ipc_spin_lock);
71 
72 		/* we can send more messages to DSP so trigger work */
73 		queue_work(drv->post_msg_wq, &drv->ipc_post_msg_wq);
74 		retval = IRQ_HANDLED;
75 	}
76 
77 	if (isr.part.busy_interrupt) {
78 		/* message from dsp so copy that */
79 		spin_lock(&drv->ipc_spin_lock);
80 		imr.full = sst_shim_read64(drv->shim, SST_IMRX);
81 		imr.part.busy_interrupt = 1;
82 		sst_shim_write64(drv->shim, SST_IMRX, imr.full);
83 		spin_unlock(&drv->ipc_spin_lock);
84 		header.full =  sst_shim_read64(drv->shim, drv->ipc_reg.ipcd);
85 
86 		if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) {
87 			drv->ops->clear_interrupt(drv);
88 			return IRQ_HANDLED;
89 		}
90 
91 		if (header.p.header_high.part.large) {
92 			size = header.p.header_low_payload;
93 			if (sst_validate_mailbox_size(size)) {
94 				memcpy_fromio(msg->mailbox_data,
95 					drv->mailbox + drv->mailbox_recv_offset, size);
96 			} else {
97 				dev_err(drv->dev,
98 					"Mailbox not copied, payload size is: %u\n", size);
99 				header.p.header_low_payload = 0;
100 			}
101 		}
102 
103 		msg->mrfld_header = header;
104 		msg->is_process_reply =
105 			sst_is_process_reply(header.p.header_high.part.msg_id);
106 		spin_lock(&drv->rx_msg_lock);
107 		list_add_tail(&msg->node, &drv->rx_list);
108 		spin_unlock(&drv->rx_msg_lock);
109 		drv->ops->clear_interrupt(drv);
110 		retval = IRQ_WAKE_THREAD;
111 	}
112 	return retval;
113 }
114 
115 static irqreturn_t intel_sst_irq_thread_mrfld(int irq, void *context)
116 {
117 	struct intel_sst_drv *drv = (struct intel_sst_drv *) context;
118 	struct ipc_post *__msg, *msg;
119 	unsigned long irq_flags;
120 
121 	spin_lock_irqsave(&drv->rx_msg_lock, irq_flags);
122 	if (list_empty(&drv->rx_list)) {
123 		spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
124 		return IRQ_HANDLED;
125 	}
126 
127 	list_for_each_entry_safe(msg, __msg, &drv->rx_list, node) {
128 		list_del(&msg->node);
129 		spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
130 		if (msg->is_process_reply)
131 			drv->ops->process_message(msg);
132 		else
133 			drv->ops->process_reply(drv, msg);
134 
135 		if (msg->is_large)
136 			kfree(msg->mailbox_data);
137 		kfree(msg);
138 		spin_lock_irqsave(&drv->rx_msg_lock, irq_flags);
139 	}
140 	spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
141 	return IRQ_HANDLED;
142 }
143 
144 static int sst_save_dsp_context_v2(struct intel_sst_drv *sst)
145 {
146 	int ret = 0;
147 
148 	ret = sst_prepare_and_post_msg(sst, SST_TASK_ID_MEDIA, IPC_CMD,
149 			IPC_PREP_D3, PIPE_RSVD, 0, NULL, NULL,
150 			true, true, false, true);
151 
152 	if (ret < 0) {
153 		dev_err(sst->dev, "not suspending FW!!, Err: %d\n", ret);
154 		return -EIO;
155 	}
156 
157 	return 0;
158 }
159 
160 
161 static struct intel_sst_ops mrfld_ops = {
162 	.interrupt = intel_sst_interrupt_mrfld,
163 	.irq_thread = intel_sst_irq_thread_mrfld,
164 	.clear_interrupt = intel_sst_clear_intr_mrfld,
165 	.start = sst_start_mrfld,
166 	.reset = intel_sst_reset_dsp_mrfld,
167 	.post_message = sst_post_message_mrfld,
168 	.process_reply = sst_process_reply_mrfld,
169 	.save_dsp_context =  sst_save_dsp_context_v2,
170 	.alloc_stream = sst_alloc_stream_mrfld,
171 	.post_download = sst_post_download_mrfld,
172 };
173 
174 int sst_driver_ops(struct intel_sst_drv *sst)
175 {
176 
177 	switch (sst->dev_id) {
178 	case PCI_DEVICE_ID_INTEL_SST_TNG:
179 	case PCI_DEVICE_ID_INTEL_SST_BYT:
180 	case PCI_DEVICE_ID_INTEL_SST_BSW:
181 		sst->tstamp = SST_TIME_STAMP_MRFLD;
182 		sst->ops = &mrfld_ops;
183 		return 0;
184 
185 	default:
186 		dev_err(sst->dev,
187 			"SST Driver capabilities missing for dev_id: %x",
188 			sst->dev_id);
189 		return -EINVAL;
190 	}
191 }
192 
193 void sst_process_pending_msg(struct work_struct *work)
194 {
195 	struct intel_sst_drv *ctx = container_of(work,
196 			struct intel_sst_drv, ipc_post_msg_wq);
197 
198 	ctx->ops->post_message(ctx, NULL, false);
199 }
200 
201 static int sst_workqueue_init(struct intel_sst_drv *ctx)
202 {
203 	INIT_LIST_HEAD(&ctx->memcpy_list);
204 	INIT_LIST_HEAD(&ctx->rx_list);
205 	INIT_LIST_HEAD(&ctx->ipc_dispatch_list);
206 	INIT_LIST_HEAD(&ctx->block_list);
207 	INIT_WORK(&ctx->ipc_post_msg_wq, sst_process_pending_msg);
208 	init_waitqueue_head(&ctx->wait_queue);
209 
210 	ctx->post_msg_wq =
211 		create_singlethread_workqueue("sst_post_msg_wq");
212 	if (!ctx->post_msg_wq)
213 		return -EBUSY;
214 	return 0;
215 }
216 
217 static void sst_init_locks(struct intel_sst_drv *ctx)
218 {
219 	mutex_init(&ctx->sst_lock);
220 	spin_lock_init(&ctx->rx_msg_lock);
221 	spin_lock_init(&ctx->ipc_spin_lock);
222 	spin_lock_init(&ctx->block_lock);
223 }
224 
225 /*
226  * Driver handles PCI IDs in ACPI - sst_acpi_probe() - and we are using only
227  * device ID part. If real ACPI ID appears, the kstrtouint() returns error, so
228  * we are fine with using unsigned short as dev_id type.
229  */
230 int sst_alloc_drv_context(struct intel_sst_drv **ctx,
231 		struct device *dev, unsigned short dev_id)
232 {
233 	*ctx = devm_kzalloc(dev, sizeof(struct intel_sst_drv), GFP_KERNEL);
234 	if (!(*ctx))
235 		return -ENOMEM;
236 
237 	(*ctx)->dev = dev;
238 	(*ctx)->dev_id = dev_id;
239 
240 	return 0;
241 }
242 EXPORT_SYMBOL_GPL(sst_alloc_drv_context);
243 
244 static ssize_t firmware_version_show(struct device *dev,
245 			    struct device_attribute *attr, char *buf)
246 {
247 	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
248 
249 	if (ctx->fw_version.type == 0 && ctx->fw_version.major == 0 &&
250 	    ctx->fw_version.minor == 0 && ctx->fw_version.build == 0)
251 		return sysfs_emit(buf, "FW not yet loaded\n");
252 	else
253 		return sysfs_emit(buf, "v%02x.%02x.%02x.%02x\n",
254 				  ctx->fw_version.type, ctx->fw_version.major,
255 				  ctx->fw_version.minor, ctx->fw_version.build);
256 
257 }
258 
259 static DEVICE_ATTR_RO(firmware_version);
260 
261 static const struct attribute *sst_fw_version_attrs[] = {
262 	&dev_attr_firmware_version.attr,
263 	NULL,
264 };
265 
266 static const struct attribute_group sst_fw_version_attr_group = {
267 	.attrs = (struct attribute **)sst_fw_version_attrs,
268 };
269 
270 int sst_context_init(struct intel_sst_drv *ctx)
271 {
272 	int ret = 0, i;
273 
274 	if (!ctx->pdata)
275 		return -EINVAL;
276 
277 	if (!ctx->pdata->probe_data)
278 		return -EINVAL;
279 
280 	memcpy(&ctx->info, ctx->pdata->probe_data, sizeof(ctx->info));
281 
282 	ret = sst_driver_ops(ctx);
283 	if (ret != 0)
284 		return -EINVAL;
285 
286 	sst_init_locks(ctx);
287 	sst_set_fw_state_locked(ctx, SST_RESET);
288 
289 	/* pvt_id 0 reserved for async messages */
290 	ctx->pvt_id = 1;
291 	ctx->stream_cnt = 0;
292 	ctx->fw_in_mem = NULL;
293 	/* we use memcpy, so set to 0 */
294 	ctx->use_dma = 0;
295 	ctx->use_lli = 0;
296 
297 	if (sst_workqueue_init(ctx))
298 		return -EINVAL;
299 
300 	ctx->mailbox_recv_offset = ctx->pdata->ipc_info->mbox_recv_off;
301 	ctx->ipc_reg.ipcx = SST_IPCX + ctx->pdata->ipc_info->ipc_offset;
302 	ctx->ipc_reg.ipcd = SST_IPCD + ctx->pdata->ipc_info->ipc_offset;
303 
304 	dev_info(ctx->dev, "Got drv data max stream %d\n",
305 				ctx->info.max_streams);
306 
307 	for (i = 1; i <= ctx->info.max_streams; i++) {
308 		struct stream_info *stream = &ctx->streams[i];
309 
310 		memset(stream, 0, sizeof(*stream));
311 		stream->pipe_id = PIPE_RSVD;
312 		mutex_init(&stream->lock);
313 	}
314 
315 	/* Register the ISR */
316 	ret = devm_request_threaded_irq(ctx->dev, ctx->irq_num, ctx->ops->interrupt,
317 					ctx->ops->irq_thread, 0, SST_DRV_NAME,
318 					ctx);
319 	if (ret)
320 		goto do_free_mem;
321 
322 	dev_dbg(ctx->dev, "Registered IRQ %#x\n", ctx->irq_num);
323 
324 	/* default intr are unmasked so set this as masked */
325 	sst_shim_write64(ctx->shim, SST_IMRX, 0xFFFF0038);
326 
327 	ctx->qos = devm_kzalloc(ctx->dev,
328 		sizeof(struct pm_qos_request), GFP_KERNEL);
329 	if (!ctx->qos) {
330 		ret = -ENOMEM;
331 		goto do_free_mem;
332 	}
333 	cpu_latency_qos_add_request(ctx->qos, PM_QOS_DEFAULT_VALUE);
334 
335 	dev_dbg(ctx->dev, "Requesting FW %s now...\n", ctx->firmware_name);
336 	ret = request_firmware_nowait(THIS_MODULE, true, ctx->firmware_name,
337 				      ctx->dev, GFP_KERNEL, ctx, sst_firmware_load_cb);
338 	if (ret) {
339 		dev_err(ctx->dev, "Firmware download failed:%d\n", ret);
340 		goto do_free_mem;
341 	}
342 
343 	ret = sysfs_create_group(&ctx->dev->kobj,
344 				 &sst_fw_version_attr_group);
345 	if (ret) {
346 		dev_err(ctx->dev,
347 			"Unable to create sysfs\n");
348 		goto err_sysfs;
349 	}
350 
351 	sst_register(ctx->dev);
352 	return 0;
353 err_sysfs:
354 	sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group);
355 
356 do_free_mem:
357 	destroy_workqueue(ctx->post_msg_wq);
358 	return ret;
359 }
360 EXPORT_SYMBOL_GPL(sst_context_init);
361 
362 void sst_context_cleanup(struct intel_sst_drv *ctx)
363 {
364 	pm_runtime_get_noresume(ctx->dev);
365 	pm_runtime_disable(ctx->dev);
366 	sst_unregister(ctx->dev);
367 	sst_set_fw_state_locked(ctx, SST_SHUTDOWN);
368 	sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group);
369 	destroy_workqueue(ctx->post_msg_wq);
370 	cpu_latency_qos_remove_request(ctx->qos);
371 	kfree(ctx->fw_sg_list.src);
372 	kfree(ctx->fw_sg_list.dst);
373 	ctx->fw_sg_list.list_len = 0;
374 	kfree(ctx->fw_in_mem);
375 	ctx->fw_in_mem = NULL;
376 	sst_memcpy_free_resources(ctx);
377 }
378 EXPORT_SYMBOL_GPL(sst_context_cleanup);
379 
380 void sst_configure_runtime_pm(struct intel_sst_drv *ctx)
381 {
382 	pm_runtime_set_autosuspend_delay(ctx->dev, SST_SUSPEND_DELAY);
383 	pm_runtime_use_autosuspend(ctx->dev);
384 	/*
385 	 * For acpi devices, the actual physical device state is
386 	 * initially active. So change the state to active before
387 	 * enabling the pm
388 	 */
389 
390 	if (!acpi_disabled)
391 		pm_runtime_set_active(ctx->dev);
392 
393 	pm_runtime_enable(ctx->dev);
394 
395 	if (acpi_disabled)
396 		pm_runtime_set_active(ctx->dev);
397 	else
398 		pm_runtime_put_noidle(ctx->dev);
399 }
400 EXPORT_SYMBOL_GPL(sst_configure_runtime_pm);
401 
402 static int intel_sst_runtime_suspend(struct device *dev)
403 {
404 	int ret = 0;
405 	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
406 
407 	if (ctx->sst_state == SST_RESET) {
408 		dev_dbg(dev, "LPE is already in RESET state, No action\n");
409 		return 0;
410 	}
411 	/* save fw context */
412 	if (ctx->ops->save_dsp_context(ctx))
413 		return -EBUSY;
414 
415 	/* Move the SST state to Reset */
416 	sst_set_fw_state_locked(ctx, SST_RESET);
417 
418 	synchronize_irq(ctx->irq_num);
419 	flush_workqueue(ctx->post_msg_wq);
420 
421 	ctx->ops->reset(ctx);
422 
423 	return ret;
424 }
425 
426 static int intel_sst_suspend(struct device *dev)
427 {
428 	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
429 	struct sst_fw_save *fw_save;
430 	int i, ret;
431 
432 	/* check first if we are already in SW reset */
433 	if (ctx->sst_state == SST_RESET)
434 		return 0;
435 
436 	/*
437 	 * check if any stream is active and running
438 	 * they should already by suspend by soc_suspend
439 	 */
440 	for (i = 1; i <= ctx->info.max_streams; i++) {
441 		struct stream_info *stream = &ctx->streams[i];
442 
443 		if (stream->status == STREAM_RUNNING) {
444 			dev_err(dev, "stream %d is running, can't suspend, abort\n", i);
445 			return -EBUSY;
446 		}
447 
448 		if (ctx->pdata->streams_lost_on_suspend) {
449 			stream->resume_status = stream->status;
450 			stream->resume_prev = stream->prev;
451 			if (stream->status != STREAM_UN_INIT)
452 				sst_free_stream(ctx, i);
453 		}
454 	}
455 	synchronize_irq(ctx->irq_num);
456 	flush_workqueue(ctx->post_msg_wq);
457 
458 	/* Move the SST state to Reset */
459 	sst_set_fw_state_locked(ctx, SST_RESET);
460 
461 	/* tell DSP we are suspending */
462 	if (ctx->ops->save_dsp_context(ctx))
463 		return -EBUSY;
464 
465 	/* save the memories */
466 	fw_save = kzalloc(sizeof(*fw_save), GFP_KERNEL);
467 	if (!fw_save)
468 		return -ENOMEM;
469 	fw_save->iram = kvzalloc(ctx->iram_end - ctx->iram_base, GFP_KERNEL);
470 	if (!fw_save->iram) {
471 		ret = -ENOMEM;
472 		goto iram;
473 	}
474 	fw_save->dram = kvzalloc(ctx->dram_end - ctx->dram_base, GFP_KERNEL);
475 	if (!fw_save->dram) {
476 		ret = -ENOMEM;
477 		goto dram;
478 	}
479 	fw_save->sram = kvzalloc(SST_MAILBOX_SIZE, GFP_KERNEL);
480 	if (!fw_save->sram) {
481 		ret = -ENOMEM;
482 		goto sram;
483 	}
484 
485 	fw_save->ddr = kvzalloc(ctx->ddr_end - ctx->ddr_base, GFP_KERNEL);
486 	if (!fw_save->ddr) {
487 		ret = -ENOMEM;
488 		goto ddr;
489 	}
490 
491 	memcpy32_fromio(fw_save->iram, ctx->iram, ctx->iram_end - ctx->iram_base);
492 	memcpy32_fromio(fw_save->dram, ctx->dram, ctx->dram_end - ctx->dram_base);
493 	memcpy32_fromio(fw_save->sram, ctx->mailbox, SST_MAILBOX_SIZE);
494 	memcpy32_fromio(fw_save->ddr, ctx->ddr, ctx->ddr_end - ctx->ddr_base);
495 
496 	ctx->fw_save = fw_save;
497 	ctx->ops->reset(ctx);
498 	return 0;
499 ddr:
500 	kvfree(fw_save->sram);
501 sram:
502 	kvfree(fw_save->dram);
503 dram:
504 	kvfree(fw_save->iram);
505 iram:
506 	kfree(fw_save);
507 	return ret;
508 }
509 
510 static int intel_sst_resume(struct device *dev)
511 {
512 	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
513 	struct sst_fw_save *fw_save = ctx->fw_save;
514 	struct sst_block *block;
515 	int i, ret = 0;
516 
517 	if (!fw_save)
518 		return 0;
519 
520 	sst_set_fw_state_locked(ctx, SST_FW_LOADING);
521 
522 	/* we have to restore the memory saved */
523 	ctx->ops->reset(ctx);
524 
525 	ctx->fw_save = NULL;
526 
527 	memcpy32_toio(ctx->iram, fw_save->iram, ctx->iram_end - ctx->iram_base);
528 	memcpy32_toio(ctx->dram, fw_save->dram, ctx->dram_end - ctx->dram_base);
529 	memcpy32_toio(ctx->mailbox, fw_save->sram, SST_MAILBOX_SIZE);
530 	memcpy32_toio(ctx->ddr, fw_save->ddr, ctx->ddr_end - ctx->ddr_base);
531 
532 	kvfree(fw_save->sram);
533 	kvfree(fw_save->dram);
534 	kvfree(fw_save->iram);
535 	kvfree(fw_save->ddr);
536 	kfree(fw_save);
537 
538 	block = sst_create_block(ctx, 0, FW_DWNL_ID);
539 	if (block == NULL)
540 		return -ENOMEM;
541 
542 
543 	/* start and wait for ack */
544 	ctx->ops->start(ctx);
545 	ret = sst_wait_timeout(ctx, block);
546 	if (ret) {
547 		dev_err(ctx->dev, "fw download failed %d\n", ret);
548 		/* FW download failed due to timeout */
549 		ret = -EBUSY;
550 
551 	} else {
552 		sst_set_fw_state_locked(ctx, SST_FW_RUNNING);
553 	}
554 
555 	if (ctx->pdata->streams_lost_on_suspend) {
556 		for (i = 1; i <= ctx->info.max_streams; i++) {
557 			struct stream_info *stream = &ctx->streams[i];
558 
559 			if (stream->resume_status != STREAM_UN_INIT) {
560 				dev_dbg(ctx->dev, "Re-allocing stream %d status %d prev %d\n",
561 					i, stream->resume_status,
562 					stream->resume_prev);
563 				sst_realloc_stream(ctx, i);
564 				stream->status = stream->resume_status;
565 				stream->prev = stream->resume_prev;
566 			}
567 		}
568 	}
569 
570 	sst_free_block(ctx, block);
571 	return ret;
572 }
573 
574 const struct dev_pm_ops intel_sst_pm = {
575 	.suspend = intel_sst_suspend,
576 	.resume = intel_sst_resume,
577 	.runtime_suspend = intel_sst_runtime_suspend,
578 };
579 EXPORT_SYMBOL_GPL(intel_sst_pm);
580