1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause 2 /* 3 * Copyright (C) 2015 Intel Mobile Communications GmbH 4 * Copyright (C) 2016-2017 Intel Deutschland GmbH 5 * Copyright (C) 2019-2021 Intel Corporation 6 */ 7 #include <linux/kernel.h> 8 #include <linux/bsearch.h> 9 10 #include "fw/api/tx.h" 11 #include "iwl-trans.h" 12 #include "iwl-drv.h" 13 #include "iwl-fh.h" 14 #include "queue/tx.h" 15 #include <linux/dmapool.h> 16 #include "fw/api/commands.h" 17 18 struct iwl_trans *iwl_trans_alloc(unsigned int priv_size, 19 struct device *dev, 20 const struct iwl_trans_ops *ops, 21 const struct iwl_cfg_trans_params *cfg_trans) 22 { 23 struct iwl_trans *trans; 24 #ifdef CONFIG_LOCKDEP 25 static struct lock_class_key __key; 26 #endif 27 28 trans = devm_kzalloc(dev, sizeof(*trans) + priv_size, GFP_KERNEL); 29 if (!trans) 30 return NULL; 31 32 trans->trans_cfg = cfg_trans; 33 34 #ifdef CONFIG_LOCKDEP 35 lockdep_init_map(&trans->sync_cmd_lockdep_map, "sync_cmd_lockdep_map", 36 &__key, 0); 37 #endif 38 39 trans->dev = dev; 40 trans->ops = ops; 41 trans->num_rx_queues = 1; 42 43 WARN_ON(!ops->wait_txq_empty && !ops->wait_tx_queues_empty); 44 45 if (trans->trans_cfg->use_tfh) { 46 trans->txqs.tfd.addr_size = 64; 47 trans->txqs.tfd.max_tbs = IWL_TFH_NUM_TBS; 48 trans->txqs.tfd.size = sizeof(struct iwl_tfh_tfd); 49 } else { 50 trans->txqs.tfd.addr_size = 36; 51 trans->txqs.tfd.max_tbs = IWL_NUM_OF_TBS; 52 trans->txqs.tfd.size = sizeof(struct iwl_tfd); 53 } 54 trans->max_skb_frags = IWL_TRANS_MAX_FRAGS(trans); 55 56 return trans; 57 } 58 59 int iwl_trans_init(struct iwl_trans *trans) 60 { 61 int txcmd_size, txcmd_align; 62 63 if (!trans->trans_cfg->gen2) { 64 txcmd_size = sizeof(struct iwl_tx_cmd); 65 txcmd_align = sizeof(void *); 66 } else if (trans->trans_cfg->device_family < IWL_DEVICE_FAMILY_AX210) { 67 txcmd_size = sizeof(struct iwl_tx_cmd_gen2); 68 txcmd_align = 64; 69 } else { 70 txcmd_size = sizeof(struct iwl_tx_cmd_gen3); 71 txcmd_align = 128; 72 } 73 74 txcmd_size += sizeof(struct iwl_cmd_header); 75 txcmd_size += 36; /* biggest possible 802.11 header */ 76 77 /* Ensure device TX cmd cannot reach/cross a page boundary in gen2 */ 78 if (WARN_ON(trans->trans_cfg->gen2 && txcmd_size >= txcmd_align)) 79 return -EINVAL; 80 81 if (trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_BZ) 82 trans->txqs.bc_tbl_size = 83 sizeof(struct iwl_gen3_bc_tbl_entry) * TFD_QUEUE_BC_SIZE_GEN3_BZ; 84 else if (trans->trans_cfg->device_family >= IWL_DEVICE_FAMILY_AX210) 85 trans->txqs.bc_tbl_size = 86 sizeof(struct iwl_gen3_bc_tbl_entry) * TFD_QUEUE_BC_SIZE_GEN3_AX210; 87 else 88 trans->txqs.bc_tbl_size = sizeof(struct iwlagn_scd_bc_tbl); 89 /* 90 * For gen2 devices, we use a single allocation for each byte-count 91 * table, but they're pretty small (1k) so use a DMA pool that we 92 * allocate here. 93 */ 94 if (trans->trans_cfg->gen2) { 95 trans->txqs.bc_pool = dmam_pool_create("iwlwifi:bc", trans->dev, 96 trans->txqs.bc_tbl_size, 97 256, 0); 98 if (!trans->txqs.bc_pool) 99 return -ENOMEM; 100 } 101 102 /* Some things must not change even if the config does */ 103 WARN_ON(trans->txqs.tfd.addr_size != 104 (trans->trans_cfg->use_tfh ? 64 : 36)); 105 106 snprintf(trans->dev_cmd_pool_name, sizeof(trans->dev_cmd_pool_name), 107 "iwl_cmd_pool:%s", dev_name(trans->dev)); 108 trans->dev_cmd_pool = 109 kmem_cache_create(trans->dev_cmd_pool_name, 110 txcmd_size, txcmd_align, 111 SLAB_HWCACHE_ALIGN, NULL); 112 if (!trans->dev_cmd_pool) 113 return -ENOMEM; 114 115 trans->txqs.tso_hdr_page = alloc_percpu(struct iwl_tso_hdr_page); 116 if (!trans->txqs.tso_hdr_page) { 117 kmem_cache_destroy(trans->dev_cmd_pool); 118 return -ENOMEM; 119 } 120 121 /* Initialize the wait queue for commands */ 122 init_waitqueue_head(&trans->wait_command_queue); 123 124 return 0; 125 } 126 127 void iwl_trans_free(struct iwl_trans *trans) 128 { 129 int i; 130 131 if (trans->txqs.tso_hdr_page) { 132 for_each_possible_cpu(i) { 133 struct iwl_tso_hdr_page *p = 134 per_cpu_ptr(trans->txqs.tso_hdr_page, i); 135 136 if (p && p->page) 137 __free_page(p->page); 138 } 139 140 free_percpu(trans->txqs.tso_hdr_page); 141 } 142 143 kmem_cache_destroy(trans->dev_cmd_pool); 144 } 145 146 int iwl_trans_send_cmd(struct iwl_trans *trans, struct iwl_host_cmd *cmd) 147 { 148 int ret; 149 150 if (unlikely(!(cmd->flags & CMD_SEND_IN_RFKILL) && 151 test_bit(STATUS_RFKILL_OPMODE, &trans->status))) 152 return -ERFKILL; 153 154 /* 155 * We can't test IWL_MVM_STATUS_IN_D3 in mvm->status because this 156 * bit is set early in the D3 flow, before we send all the commands 157 * that configure the firmware for D3 operation (power, patterns, ...) 158 * and we don't want to flag all those with CMD_SEND_IN_D3. 159 * So use the system_pm_mode instead. The only command sent after 160 * we set system_pm_mode is D3_CONFIG_CMD, which we now flag with 161 * CMD_SEND_IN_D3. 162 */ 163 if (unlikely(trans->system_pm_mode == IWL_PLAT_PM_MODE_D3 && 164 !(cmd->flags & CMD_SEND_IN_D3))) 165 return -EHOSTDOWN; 166 167 if (unlikely(test_bit(STATUS_FW_ERROR, &trans->status))) 168 return -EIO; 169 170 if (unlikely(trans->state != IWL_TRANS_FW_ALIVE)) { 171 IWL_ERR(trans, "%s bad state = %d\n", __func__, trans->state); 172 return -EIO; 173 } 174 175 if (WARN_ON((cmd->flags & CMD_WANT_ASYNC_CALLBACK) && 176 !(cmd->flags & CMD_ASYNC))) 177 return -EINVAL; 178 179 if (!(cmd->flags & CMD_ASYNC)) 180 lock_map_acquire_read(&trans->sync_cmd_lockdep_map); 181 182 if (trans->wide_cmd_header && !iwl_cmd_groupid(cmd->id)) { 183 if (cmd->id != REPLY_ERROR) 184 cmd->id = DEF_ID(cmd->id); 185 } 186 187 ret = iwl_trans_txq_send_hcmd(trans, cmd); 188 189 if (!(cmd->flags & CMD_ASYNC)) 190 lock_map_release(&trans->sync_cmd_lockdep_map); 191 192 if (WARN_ON((cmd->flags & CMD_WANT_SKB) && !ret && !cmd->resp_pkt)) 193 return -EIO; 194 195 return ret; 196 } 197 IWL_EXPORT_SYMBOL(iwl_trans_send_cmd); 198 199 /* Comparator for struct iwl_hcmd_names. 200 * Used in the binary search over a list of host commands. 201 * 202 * @key: command_id that we're looking for. 203 * @elt: struct iwl_hcmd_names candidate for match. 204 * 205 * @return 0 iff equal. 206 */ 207 static int iwl_hcmd_names_cmp(const void *key, const void *elt) 208 { 209 const struct iwl_hcmd_names *name = elt; 210 const u8 *cmd1 = key; 211 u8 cmd2 = name->cmd_id; 212 213 return (*cmd1 - cmd2); 214 } 215 216 const char *iwl_get_cmd_string(struct iwl_trans *trans, u32 id) 217 { 218 u8 grp, cmd; 219 struct iwl_hcmd_names *ret; 220 const struct iwl_hcmd_arr *arr; 221 size_t size = sizeof(struct iwl_hcmd_names); 222 223 grp = iwl_cmd_groupid(id); 224 cmd = iwl_cmd_opcode(id); 225 226 if (!trans->command_groups || grp >= trans->command_groups_size || 227 !trans->command_groups[grp].arr) 228 return "UNKNOWN"; 229 230 arr = &trans->command_groups[grp]; 231 ret = bsearch(&cmd, arr->arr, arr->size, size, iwl_hcmd_names_cmp); 232 if (!ret) 233 return "UNKNOWN"; 234 return ret->cmd_name; 235 } 236 IWL_EXPORT_SYMBOL(iwl_get_cmd_string); 237 238 int iwl_cmd_groups_verify_sorted(const struct iwl_trans_config *trans) 239 { 240 int i, j; 241 const struct iwl_hcmd_arr *arr; 242 243 for (i = 0; i < trans->command_groups_size; i++) { 244 arr = &trans->command_groups[i]; 245 if (!arr->arr) 246 continue; 247 for (j = 0; j < arr->size - 1; j++) 248 if (arr->arr[j].cmd_id > arr->arr[j + 1].cmd_id) 249 return -1; 250 } 251 return 0; 252 } 253 IWL_EXPORT_SYMBOL(iwl_cmd_groups_verify_sorted); 254