// SPDX-License-Identifier: GPL-2.0+ /* * SoundWire AMD Manager driver * * Copyright 2023 Advanced Micro Devices, Inc. */ #include #include #include #include #include #include #include #include #include #include #include #include #include "bus.h" #include "amd_manager.h" #define DRV_NAME "amd_sdw_manager" #define to_amd_sdw(b) container_of(b, struct amd_sdw_manager, bus) static void amd_enable_sdw_pads(struct amd_sdw_manager *amd_manager) { u32 sw_pad_pulldown_val; u32 val; mutex_lock(amd_manager->acp_sdw_lock); val = readl(amd_manager->acp_mmio + ACP_SW_PAD_KEEPER_EN); val |= amd_manager->reg_mask->sw_pad_enable_mask; writel(val, amd_manager->acp_mmio + ACP_SW_PAD_KEEPER_EN); usleep_range(1000, 1500); sw_pad_pulldown_val = readl(amd_manager->acp_mmio + ACP_PAD_PULLDOWN_CTRL); sw_pad_pulldown_val &= amd_manager->reg_mask->sw_pad_pulldown_mask; writel(sw_pad_pulldown_val, amd_manager->acp_mmio + ACP_PAD_PULLDOWN_CTRL); mutex_unlock(amd_manager->acp_sdw_lock); } static int amd_init_sdw_manager(struct amd_sdw_manager *amd_manager) { u32 val; int ret; writel(AMD_SDW_ENABLE, amd_manager->mmio + ACP_SW_EN); ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, val, ACP_DELAY_US, AMD_SDW_TIMEOUT); if (ret) return ret; /* SoundWire manager bus reset */ writel(AMD_SDW_BUS_RESET_REQ, amd_manager->mmio + ACP_SW_BUS_RESET_CTRL); ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_BUS_RESET_CTRL, val, (val & AMD_SDW_BUS_RESET_DONE), ACP_DELAY_US, AMD_SDW_TIMEOUT); if (ret) return ret; writel(AMD_SDW_BUS_RESET_CLEAR_REQ, amd_manager->mmio + ACP_SW_BUS_RESET_CTRL); ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_BUS_RESET_CTRL, val, !val, ACP_DELAY_US, AMD_SDW_TIMEOUT); if (ret) { dev_err(amd_manager->dev, "Failed to reset SoundWire manager instance%d\n", amd_manager->instance); return ret; } writel(AMD_SDW_DISABLE, amd_manager->mmio + ACP_SW_EN); return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, !val, ACP_DELAY_US, AMD_SDW_TIMEOUT); } static int amd_enable_sdw_manager(struct amd_sdw_manager *amd_manager) { u32 val; writel(AMD_SDW_ENABLE, amd_manager->mmio + ACP_SW_EN); return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, val, ACP_DELAY_US, AMD_SDW_TIMEOUT); } static int amd_disable_sdw_manager(struct amd_sdw_manager *amd_manager) { u32 val; writel(AMD_SDW_DISABLE, amd_manager->mmio + ACP_SW_EN); /* * After invoking manager disable sequence, check whether * manager has executed clock stop sequence. In this case, * manager should ignore checking enable status register. */ val = readl(amd_manager->mmio + ACP_SW_CLK_RESUME_CTRL); if (val) return 0; return readl_poll_timeout(amd_manager->mmio + ACP_SW_EN_STATUS, val, !val, ACP_DELAY_US, AMD_SDW_TIMEOUT); } static void amd_enable_sdw_interrupts(struct amd_sdw_manager *amd_manager) { struct sdw_manager_reg_mask *reg_mask = amd_manager->reg_mask; u32 val; mutex_lock(amd_manager->acp_sdw_lock); val = readl(amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance)); val |= reg_mask->acp_sdw_intr_mask; writel(val, amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance)); mutex_unlock(amd_manager->acp_sdw_lock); writel(AMD_SDW_IRQ_MASK_0TO7, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7); writel(AMD_SDW_IRQ_MASK_8TO11, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11); writel(AMD_SDW_IRQ_ERROR_MASK, amd_manager->mmio + ACP_SW_ERROR_INTR_MASK); } static void amd_disable_sdw_interrupts(struct amd_sdw_manager *amd_manager) { struct sdw_manager_reg_mask *reg_mask = amd_manager->reg_mask; u32 val; mutex_lock(amd_manager->acp_sdw_lock); val = readl(amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance)); val &= ~reg_mask->acp_sdw_intr_mask; writel(val, amd_manager->acp_mmio + ACP_EXTERNAL_INTR_CNTL(amd_manager->instance)); mutex_unlock(amd_manager->acp_sdw_lock); writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_0TO7); writel(0x00, amd_manager->mmio + ACP_SW_STATE_CHANGE_STATUS_MASK_8TO11); writel(0x00, amd_manager->mmio + ACP_SW_ERROR_INTR_MASK); } static void amd_sdw_set_frameshape(struct amd_sdw_manager *amd_manager) { u32 frame_size; frame_size = (amd_manager->rows_index << 3) | amd_manager->cols_index; writel(frame_size, amd_manager->mmio + ACP_SW_FRAMESIZE); } static void amd_sdw_ctl_word_prep(u32 *lower_word, u32 *upper_word, struct sdw_msg *msg, int cmd_offset) { u32 upper_data; u32 lower_data = 0; u16 addr; u8 upper_addr, lower_addr; u8 data = 0; addr = msg->addr + cmd_offset; upper_addr = (addr & 0xFF00) >> 8; lower_addr = addr & 0xFF; if (msg->flags == SDW_MSG_FLAG_WRITE) data = msg->buf[cmd_offset]; upper_data = FIELD_PREP(AMD_SDW_MCP_CMD_DEV_ADDR, msg->dev_num); upper_data |= FIELD_PREP(AMD_SDW_MCP_CMD_COMMAND, msg->flags + 2); upper_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_ADDR_HIGH, upper_addr); lower_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_ADDR_LOW, lower_addr); lower_data |= FIELD_PREP(AMD_SDW_MCP_CMD_REG_DATA, data); *upper_word = upper_data; *lower_word = lower_data; } static u64 amd_sdw_send_cmd_get_resp(struct amd_sdw_manager *amd_manager, u32 lower_data, u32 upper_data) { u64 resp; u32 lower_resp, upper_resp; u32 sts; int ret; ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts, !(sts & AMD_SDW_IMM_CMD_BUSY), ACP_DELAY_US, AMD_SDW_TIMEOUT); if (ret) { dev_err(amd_manager->dev, "SDW%x previous cmd status clear failed\n", amd_manager->instance); return ret; } if (sts & AMD_SDW_IMM_RES_VALID) { dev_err(amd_manager->dev, "SDW%x manager is in bad state\n", amd_manager->instance); writel(0x00, amd_manager->mmio + ACP_SW_IMM_CMD_STS); } writel(upper_data, amd_manager->mmio + ACP_SW_IMM_CMD_UPPER_WORD); writel(lower_data, amd_manager->mmio + ACP_SW_IMM_CMD_LOWER_QWORD); ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts, (sts & AMD_SDW_IMM_RES_VALID), ACP_DELAY_US, AMD_SDW_TIMEOUT); if (ret) { dev_err(amd_manager->dev, "SDW%x cmd response timeout occurred\n", amd_manager->instance); return ret; } upper_resp = readl(amd_manager->mmio + ACP_SW_IMM_RESP_UPPER_WORD); lower_resp = readl(amd_manager->mmio + ACP_SW_IMM_RESP_LOWER_QWORD); writel(AMD_SDW_IMM_RES_VALID, amd_manager->mmio + ACP_SW_IMM_CMD_STS); ret = readl_poll_timeout(amd_manager->mmio + ACP_SW_IMM_CMD_STS, sts, !(sts & AMD_SDW_IMM_RES_VALID), ACP_DELAY_US, AMD_SDW_TIMEOUT); if (ret) { dev_err(amd_manager->dev, "SDW%x cmd status retry failed\n", amd_manager->instance); return ret; } resp = upper_resp; resp = (resp << 32) | lower_resp; return resp; } static enum sdw_command_response amd_program_scp_addr(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg) { struct sdw_msg scp_msg = {0}; u64 response_buf[2] = {0}; u32 upper_data = 0, lower_data = 0; int index; scp_msg.dev_num = msg->dev_num; scp_msg.addr = SDW_SCP_ADDRPAGE1; scp_msg.buf = &msg->addr_page1; scp_msg.flags = SDW_MSG_FLAG_WRITE; amd_sdw_ctl_word_prep(&lower_data, &upper_data, &scp_msg, 0); response_buf[0] = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data); scp_msg.addr = SDW_SCP_ADDRPAGE2; scp_msg.buf = &msg->addr_page2; amd_sdw_ctl_word_prep(&lower_data, &upper_data, &scp_msg, 0); response_buf[1] = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data); for (index = 0; index < 2; index++) { if (response_buf[index] == -ETIMEDOUT) { dev_err_ratelimited(amd_manager->dev, "SCP_addrpage command timeout for Slave %d\n", msg->dev_num); return SDW_CMD_TIMEOUT; } else if (!(response_buf[index] & AMD_SDW_MCP_RESP_ACK)) { if (response_buf[index] & AMD_SDW_MCP_RESP_NACK) { dev_err_ratelimited(amd_manager->dev, "SCP_addrpage NACKed for Slave %d\n", msg->dev_num); return SDW_CMD_FAIL; } dev_dbg_ratelimited(amd_manager->dev, "SCP_addrpage ignored for Slave %d\n", msg->dev_num); return SDW_CMD_IGNORED; } } return SDW_CMD_OK; } static int amd_prep_msg(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg) { int ret; if (msg->page) { ret = amd_program_scp_addr(amd_manager, msg); if (ret) { msg->len = 0; return ret; } } switch (msg->flags) { case SDW_MSG_FLAG_READ: case SDW_MSG_FLAG_WRITE: break; default: dev_err(amd_manager->dev, "Invalid msg cmd: %d\n", msg->flags); return -EINVAL; } return 0; } static enum sdw_command_response amd_sdw_fill_msg_resp(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg, u64 response, int offset) { if (response & AMD_SDW_MCP_RESP_ACK) { if (msg->flags == SDW_MSG_FLAG_READ) msg->buf[offset] = FIELD_GET(AMD_SDW_MCP_RESP_RDATA, response); } else { if (response == -ETIMEDOUT) { dev_err_ratelimited(amd_manager->dev, "command timeout for Slave %d\n", msg->dev_num); return SDW_CMD_TIMEOUT; } else if (response & AMD_SDW_MCP_RESP_NACK) { dev_err_ratelimited(amd_manager->dev, "command response NACK received for Slave %d\n", msg->dev_num); return SDW_CMD_FAIL; } dev_err_ratelimited(amd_manager->dev, "command is ignored for Slave %d\n", msg->dev_num); return SDW_CMD_IGNORED; } return SDW_CMD_OK; } static unsigned int _amd_sdw_xfer_msg(struct amd_sdw_manager *amd_manager, struct sdw_msg *msg, int cmd_offset) { u64 response; u32 upper_data = 0, lower_data = 0; amd_sdw_ctl_word_prep(&lower_data, &upper_data, msg, cmd_offset); response = amd_sdw_send_cmd_get_resp(amd_manager, lower_data, upper_data); return amd_sdw_fill_msg_resp(amd_manager, msg, response, cmd_offset); } static enum sdw_command_response amd_sdw_xfer_msg(struct sdw_bus *bus, struct sdw_msg *msg) { struct amd_sdw_manager *amd_manager = to_amd_sdw(bus); int ret, i; ret = amd_prep_msg(amd_manager, msg); if (ret) return SDW_CMD_FAIL_OTHER; for (i = 0; i < msg->len; i++) { ret = _amd_sdw_xfer_msg(amd_manager, msg, i); if (ret) return ret; } return SDW_CMD_OK; } static u32 amd_sdw_read_ping_status(struct sdw_bus *bus) { struct amd_sdw_manager *amd_manager = to_amd_sdw(bus); u64 response; u32 slave_stat; response = amd_sdw_send_cmd_get_resp(amd_manager, 0, 0); /* slave status from ping response */ slave_stat = FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_0_3, response); slave_stat |= FIELD_GET(AMD_SDW_MCP_SLAVE_STAT_4_11, response) << 8; dev_dbg(amd_manager->dev, "slave_stat:0x%x\n", slave_stat); return slave_stat; } static int amd_sdw_compute_params(struct sdw_bus *bus) { struct sdw_transport_data t_data = {0}; struct sdw_master_runtime *m_rt; struct sdw_port_runtime *p_rt; struct sdw_bus_params *b_params = &bus->params; int port_bo, hstart, hstop, sample_int; unsigned int rate, bps; port_bo = 0; hstart = 1; hstop = bus->params.col - 1; t_data.hstop = hstop; t_data.hstart = hstart; list_for_each_entry(m_rt, &bus->m_rt_list, bus_node) { rate = m_rt->stream->params.rate; bps = m_rt->stream->params.bps; sample_int = (bus->params.curr_dr_freq / rate); list_for_each_entry(p_rt, &m_rt->port_list, port_node) { port_bo = (p_rt->num * 64) + 1; dev_dbg(bus->dev, "p_rt->num=%d hstart=%d hstop=%d port_bo=%d\n", p_rt->num, hstart, hstop, port_bo); sdw_fill_xport_params(&p_rt->transport_params, p_rt->num, false, SDW_BLK_GRP_CNT_1, sample_int, port_bo, port_bo >> 8, hstart, hstop, SDW_BLK_PKG_PER_PORT, 0x0); sdw_fill_port_params(&p_rt->port_params, p_rt->num, bps, SDW_PORT_FLOW_MODE_ISOCH, b_params->m_data_mode); t_data.hstart = hstart; t_data.hstop = hstop; t_data.block_offset = port_bo; t_data.sub_block_offset = 0; } sdw_compute_slave_ports(m_rt, &t_data); } return 0; } static int amd_sdw_port_params(struct sdw_bus *bus, struct sdw_port_params *p_params, unsigned int bank) { struct amd_sdw_manager *amd_manager = to_amd_sdw(bus); u32 frame_fmt_reg, dpn_frame_fmt; dev_dbg(amd_manager->dev, "p_params->num:0x%x\n", p_params->num); switch (amd_manager->instance) { case ACP_SDW0: frame_fmt_reg = sdw0_manager_dp_reg[p_params->num].frame_fmt_reg; break; case ACP_SDW1: frame_fmt_reg = sdw1_manager_dp_reg[p_params->num].frame_fmt_reg; break; default: return -EINVAL; } dpn_frame_fmt = readl(amd_manager->mmio + frame_fmt_reg); u32p_replace_bits(&dpn_frame_fmt, p_params->flow_mode, AMD_DPN_FRAME_FMT_PFM); u32p_replace_bits(&dpn_frame_fmt, p_params->data_mode, AMD_DPN_FRAME_FMT_PDM); u32p_replace_bits(&dpn_frame_fmt, p_params->bps - 1, AMD_DPN_FRAME_FMT_WORD_LEN); writel(dpn_frame_fmt, amd_manager->mmio + frame_fmt_reg); return 0; } static int amd_sdw_transport_params(struct sdw_bus *bus, struct sdw_transport_params *params, enum sdw_reg_bank bank) { struct amd_sdw_manager *amd_manager = to_amd_sdw(bus); u32 dpn_frame_fmt; u32 dpn_sampleinterval; u32 dpn_hctrl; u32 dpn_offsetctrl; u32 dpn_lanectrl; u32 frame_fmt_reg, sample_int_reg, hctrl_dp0_reg; u32 offset_reg, lane_ctrl_ch_en_reg; switch (amd_manager->instance) { case ACP_SDW0: frame_fmt_reg = sdw0_manager_dp_reg[params->port_num].frame_fmt_reg; sample_int_reg = sdw0_manager_dp_reg[params->port_num].sample_int_reg; hctrl_dp0_reg = sdw0_manager_dp_reg[params->port_num].hctrl_dp0_reg; offset_reg = sdw0_manager_dp_reg[params->port_num].offset_reg; lane_ctrl_ch_en_reg = sdw0_manager_dp_reg[params->port_num].lane_ctrl_ch_en_reg; break; case ACP_SDW1: frame_fmt_reg = sdw1_manager_dp_reg[params->port_num].frame_fmt_reg; sample_int_reg = sdw1_manager_dp_reg[params->port_num].sample_int_reg; hctrl_dp0_reg = sdw1_manager_dp_reg[params->port_num].hctrl_dp0_reg; offset_reg = sdw1_manager_dp_reg[params->port_num].offset_reg; lane_ctrl_ch_en_reg = sdw1_manager_dp_reg[params->port_num].lane_ctrl_ch_en_reg; break; default: return -EINVAL; } writel(AMD_SDW_SSP_COUNTER_VAL, amd_manager->mmio + ACP_SW_SSP_COUNTER); dpn_frame_fmt = readl(amd_manager->mmio + frame_fmt_reg); u32p_replace_bits(&dpn_frame_fmt, params->blk_pkg_mode, AMD_DPN_FRAME_FMT_BLK_PKG_MODE); u32p_replace_bits(&dpn_frame_fmt, params->blk_grp_ctrl, AMD_DPN_FRAME_FMT_BLK_GRP_CTRL); u32p_replace_bits(&dpn_frame_fmt, SDW_STREAM_PCM, AMD_DPN_FRAME_FMT_PCM_OR_PDM); writel(dpn_frame_fmt, amd_manager->mmio + frame_fmt_reg); dpn_sampleinterval = params->sample_interval - 1; writel(dpn_sampleinterval, amd_manager->mmio + sample_int_reg); dpn_hctrl = FIELD_PREP(AMD_DPN_HCTRL_HSTOP, params->hstop); dpn_hctrl |= FIELD_PREP(AMD_DPN_HCTRL_HSTART, params->hstart); writel(dpn_hctrl, amd_manager->mmio + hctrl_dp0_reg); dpn_offsetctrl = FIELD_PREP(AMD_DPN_OFFSET_CTRL_1, params->offset1); dpn_offsetctrl |= FIELD_PREP(AMD_DPN_OFFSET_CTRL_2, params->offset2); writel(dpn_offsetctrl, amd_manager->mmio + offset_reg); /* * lane_ctrl_ch_en_reg will be used to program lane_ctrl and ch_mask * parameters. */ dpn_lanectrl = readl(amd_manager->mmio + lane_ctrl_ch_en_reg); u32p_replace_bits(&dpn_lanectrl, params->lane_ctrl, AMD_DPN_CH_EN_LCTRL); writel(dpn_lanectrl, amd_manager->mmio + lane_ctrl_ch_en_reg); return 0; } static int amd_sdw_port_enable(struct sdw_bus *bus, struct sdw_enable_ch *enable_ch, unsigned int bank) { struct amd_sdw_manager *amd_manager = to_amd_sdw(bus); u32 dpn_ch_enable; u32 lane_ctrl_ch_en_reg; switch (amd_manager->instance) { case ACP_SDW0: lane_ctrl_ch_en_reg = sdw0_manager_dp_reg[enable_ch->port_num].lane_ctrl_ch_en_reg; break; case ACP_SDW1: lane_ctrl_ch_en_reg = sdw1_manager_dp_reg[enable_ch->port_num].lane_ctrl_ch_en_reg; break; default: return -EINVAL; } /* * lane_ctrl_ch_en_reg will be used to program lane_ctrl and ch_mask * parameters. */ dpn_ch_enable = readl(amd_manager->mmio + lane_ctrl_ch_en_reg); u32p_replace_bits(&dpn_ch_enable, enable_ch->ch_mask, AMD_DPN_CH_EN_CHMASK); if (enable_ch->enable) writel(dpn_ch_enable, amd_manager->mmio + lane_ctrl_ch_en_reg); else writel(0, amd_manager->mmio + lane_ctrl_ch_en_reg); return 0; } static int sdw_master_read_amd_prop(struct sdw_bus *bus) { struct amd_sdw_manager *amd_manager = to_amd_sdw(bus); struct fwnode_handle *link; struct sdw_master_prop *prop; u32 quirk_mask = 0; u32 wake_en_mask = 0; u32 power_mode_mask = 0; char name[32]; prop = &bus->prop; /* Find manager handle */ snprintf(name, sizeof(name), "mipi-sdw-link-%d-subproperties", bus->link_id); link = device_get_named_child_node(bus->dev, name); if (!link) { dev_err(bus->dev, "Manager node %s not found\n", name); return -EIO; } fwnode_property_read_u32(link, "amd-sdw-enable", &quirk_mask); if (!(quirk_mask & AMD_SDW_QUIRK_MASK_BUS_ENABLE)) prop->hw_disabled = true; prop->quirks = SDW_MASTER_QUIRKS_CLEAR_INITIAL_CLASH | SDW_MASTER_QUIRKS_CLEAR_INITIAL_PARITY; fwnode_property_read_u32(link, "amd-sdw-wakeup-enable", &wake_en_mask); amd_manager->wake_en_mask = wake_en_mask; fwnode_property_read_u32(link, "amd-sdw-power-mode", &power_mode_mask); amd_manager->power_mode_mask = power_mode_mask; return 0; } static int amd_prop_read(struct sdw_bus *bus) { sdw_master_read_prop(bus); sdw_master_read_amd_prop(bus); return 0; } static const struct sdw_master_port_ops amd_sdw_port_ops = { .dpn_set_port_params = amd_sdw_port_params, .dpn_set_port_transport_params = amd_sdw_transport_params, .dpn_port_enable_ch = amd_sdw_port_enable, }; static const struct sdw_master_ops amd_sdw_ops = { .read_prop = amd_prop_read, .xfer_msg = amd_sdw_xfer_msg, .read_ping_status = amd_sdw_read_ping_status, }; static void amd_sdw_probe_work(struct work_struct *work) { struct amd_sdw_manager *amd_manager = container_of(work, struct amd_sdw_manager, probe_work); struct sdw_master_prop *prop; int ret; prop = &amd_manager->bus.prop; if (!prop->hw_disabled) { amd_enable_sdw_pads(amd_manager); ret = amd_init_sdw_manager(amd_manager); if (ret) return; amd_enable_sdw_interrupts(amd_manager); ret = amd_enable_sdw_manager(amd_manager); if (ret) return; amd_sdw_set_frameshape(amd_manager); } } static int amd_sdw_manager_probe(struct platform_device *pdev) { const struct acp_sdw_pdata *pdata = pdev->dev.platform_data; struct resource *res; struct device *dev = &pdev->dev; struct sdw_master_prop *prop; struct sdw_bus_params *params; struct amd_sdw_manager *amd_manager; int ret; amd_manager = devm_kzalloc(dev, sizeof(struct amd_sdw_manager), GFP_KERNEL); if (!amd_manager) return -ENOMEM; res = platform_get_resource(pdev, IORESOURCE_MEM, 0); if (!res) return -ENOMEM; amd_manager->acp_mmio = devm_ioremap(dev, res->start, resource_size(res)); if (IS_ERR(amd_manager->mmio)) { dev_err(dev, "mmio not found\n"); return PTR_ERR(amd_manager->mmio); } amd_manager->instance = pdata->instance; amd_manager->mmio = amd_manager->acp_mmio + (amd_manager->instance * SDW_MANAGER_REG_OFFSET); amd_manager->acp_sdw_lock = pdata->acp_sdw_lock; amd_manager->cols_index = sdw_find_col_index(AMD_SDW_DEFAULT_COLUMNS); amd_manager->rows_index = sdw_find_row_index(AMD_SDW_DEFAULT_ROWS); amd_manager->dev = dev; amd_manager->bus.ops = &amd_sdw_ops; amd_manager->bus.port_ops = &amd_sdw_port_ops; amd_manager->bus.compute_params = &amd_sdw_compute_params; amd_manager->bus.clk_stop_timeout = 200; amd_manager->bus.link_id = amd_manager->instance; switch (amd_manager->instance) { case ACP_SDW0: amd_manager->num_dout_ports = AMD_SDW0_MAX_TX_PORTS; amd_manager->num_din_ports = AMD_SDW0_MAX_RX_PORTS; break; case ACP_SDW1: amd_manager->num_dout_ports = AMD_SDW1_MAX_TX_PORTS; amd_manager->num_din_ports = AMD_SDW1_MAX_RX_PORTS; break; default: return -EINVAL; } amd_manager->reg_mask = &sdw_manager_reg_mask_array[amd_manager->instance]; params = &amd_manager->bus.params; params->max_dr_freq = AMD_SDW_DEFAULT_CLK_FREQ * 2; params->curr_dr_freq = AMD_SDW_DEFAULT_CLK_FREQ * 2; params->col = AMD_SDW_DEFAULT_COLUMNS; params->row = AMD_SDW_DEFAULT_ROWS; prop = &amd_manager->bus.prop; prop->clk_freq = &amd_sdw_freq_tbl[0]; prop->mclk_freq = AMD_SDW_BUS_BASE_FREQ; ret = sdw_bus_master_add(&amd_manager->bus, dev, dev->fwnode); if (ret) { dev_err(dev, "Failed to register SoundWire manager(%d)\n", ret); return ret; } dev_set_drvdata(dev, amd_manager); INIT_WORK(&amd_manager->probe_work, amd_sdw_probe_work); /* * Instead of having lengthy probe sequence, use deferred probe. */ schedule_work(&amd_manager->probe_work); return 0; } static int amd_sdw_manager_remove(struct platform_device *pdev) { struct amd_sdw_manager *amd_manager = dev_get_drvdata(&pdev->dev); cancel_work_sync(&amd_manager->probe_work); amd_disable_sdw_interrupts(amd_manager); sdw_bus_master_delete(&amd_manager->bus); return amd_disable_sdw_manager(amd_manager); } static struct platform_driver amd_sdw_driver = { .probe = &amd_sdw_manager_probe, .remove = &amd_sdw_manager_remove, .driver = { .name = "amd_sdw_manager", } }; module_platform_driver(amd_sdw_driver); MODULE_AUTHOR("Vijendar.Mukunda@amd.com"); MODULE_DESCRIPTION("AMD SoundWire driver"); MODULE_LICENSE("GPL"); MODULE_ALIAS("platform:" DRV_NAME);