1 // SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only) 2 /* Copyright(c) 2014 - 2020 Intel Corporation */ 3 #include <linux/module.h> 4 #include <linux/slab.h> 5 #include "adf_accel_devices.h" 6 #include "adf_common_drv.h" 7 #include "adf_transport.h" 8 #include "adf_cfg.h" 9 #include "adf_cfg_strings.h" 10 #include "adf_gen2_hw_data.h" 11 #include "qat_crypto.h" 12 #include "icp_qat_fw.h" 13 14 #define SEC ADF_KERNEL_SEC 15 16 static struct service_hndl qat_crypto; 17 18 void qat_crypto_put_instance(struct qat_crypto_instance *inst) 19 { 20 atomic_dec(&inst->refctr); 21 adf_dev_put(inst->accel_dev); 22 } 23 24 static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) 25 { 26 struct qat_crypto_instance *inst, *tmp; 27 int i; 28 29 list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { 30 for (i = 0; i < atomic_read(&inst->refctr); i++) 31 qat_crypto_put_instance(inst); 32 33 if (inst->sym_tx) 34 adf_remove_ring(inst->sym_tx); 35 36 if (inst->sym_rx) 37 adf_remove_ring(inst->sym_rx); 38 39 if (inst->pke_tx) 40 adf_remove_ring(inst->pke_tx); 41 42 if (inst->pke_rx) 43 adf_remove_ring(inst->pke_rx); 44 45 list_del(&inst->list); 46 kfree(inst); 47 } 48 return 0; 49 } 50 51 struct qat_crypto_instance *qat_crypto_get_instance_node(int node) 52 { 53 struct adf_accel_dev *accel_dev = NULL, *tmp_dev; 54 struct qat_crypto_instance *inst = NULL, *tmp_inst; 55 unsigned long best = ~0; 56 57 list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { 58 unsigned long ctr; 59 60 if ((node == dev_to_node(&GET_DEV(tmp_dev)) || 61 dev_to_node(&GET_DEV(tmp_dev)) < 0) && 62 adf_dev_started(tmp_dev) && 63 !list_empty(&tmp_dev->crypto_list)) { 64 ctr = atomic_read(&tmp_dev->ref_count); 65 if (best > ctr) { 66 accel_dev = tmp_dev; 67 best = ctr; 68 } 69 } 70 } 71 72 if (!accel_dev) { 73 pr_debug_ratelimited("QAT: Could not find a device on node %d\n", node); 74 /* Get any started device */ 75 list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { 76 if (adf_dev_started(tmp_dev) && 77 !list_empty(&tmp_dev->crypto_list)) { 78 accel_dev = tmp_dev; 79 break; 80 } 81 } 82 } 83 84 if (!accel_dev) 85 return NULL; 86 87 best = ~0; 88 list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { 89 unsigned long ctr; 90 91 ctr = atomic_read(&tmp_inst->refctr); 92 if (best > ctr) { 93 inst = tmp_inst; 94 best = ctr; 95 } 96 } 97 if (inst) { 98 if (adf_dev_get(accel_dev)) { 99 dev_err(&GET_DEV(accel_dev), "Could not increment dev refctr\n"); 100 return NULL; 101 } 102 atomic_inc(&inst->refctr); 103 } 104 return inst; 105 } 106 107 /** 108 * qat_crypto_vf_dev_config() 109 * create dev config required to create crypto inst. 110 * 111 * @accel_dev: Pointer to acceleration device. 112 * 113 * Function creates device configuration required to create 114 * asym, sym or, crypto instances 115 * 116 * Return: 0 on success, error code otherwise. 117 */ 118 int qat_crypto_vf_dev_config(struct adf_accel_dev *accel_dev) 119 { 120 u16 ring_to_svc_map = GET_HW_DATA(accel_dev)->ring_to_svc_map; 121 122 if (ring_to_svc_map != ADF_GEN2_DEFAULT_RING_TO_SRV_MAP) { 123 dev_err(&GET_DEV(accel_dev), 124 "Unsupported ring/service mapping present on PF"); 125 return -EFAULT; 126 } 127 128 return GET_HW_DATA(accel_dev)->dev_config(accel_dev); 129 } 130 131 static int qat_crypto_create_instances(struct adf_accel_dev *accel_dev) 132 { 133 unsigned long num_inst, num_msg_sym, num_msg_asym; 134 char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; 135 char val[ADF_CFG_MAX_VAL_LEN_IN_BYTES]; 136 unsigned long sym_bank, asym_bank; 137 struct qat_crypto_instance *inst; 138 int msg_size; 139 int ret; 140 int i; 141 142 INIT_LIST_HEAD(&accel_dev->crypto_list); 143 ret = adf_cfg_get_param_value(accel_dev, SEC, ADF_NUM_CY, val); 144 if (ret) 145 return ret; 146 147 ret = kstrtoul(val, 0, &num_inst); 148 if (ret) 149 return ret; 150 151 for (i = 0; i < num_inst; i++) { 152 inst = kzalloc_node(sizeof(*inst), GFP_KERNEL, 153 dev_to_node(&GET_DEV(accel_dev))); 154 if (!inst) { 155 ret = -ENOMEM; 156 goto err; 157 } 158 159 list_add_tail(&inst->list, &accel_dev->crypto_list); 160 inst->id = i; 161 atomic_set(&inst->refctr, 0); 162 inst->accel_dev = accel_dev; 163 164 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_BANK_NUM, i); 165 ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); 166 if (ret) 167 goto err; 168 169 ret = kstrtoul(val, 10, &sym_bank); 170 if (ret) 171 goto err; 172 173 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_BANK_NUM, i); 174 ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); 175 if (ret) 176 goto err; 177 178 ret = kstrtoul(val, 10, &asym_bank); 179 if (ret) 180 goto err; 181 182 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_SIZE, i); 183 ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); 184 if (ret) 185 goto err; 186 187 ret = kstrtoul(val, 10, &num_msg_sym); 188 if (ret) 189 goto err; 190 191 num_msg_sym = num_msg_sym >> 1; 192 193 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_SIZE, i); 194 ret = adf_cfg_get_param_value(accel_dev, SEC, key, val); 195 if (ret) 196 goto err; 197 198 ret = kstrtoul(val, 10, &num_msg_asym); 199 if (ret) 200 goto err; 201 num_msg_asym = num_msg_asym >> 1; 202 203 msg_size = ICP_QAT_FW_REQ_DEFAULT_SZ; 204 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_TX, i); 205 ret = adf_create_ring(accel_dev, SEC, sym_bank, num_msg_sym, 206 msg_size, key, NULL, 0, &inst->sym_tx); 207 if (ret) 208 goto err; 209 210 msg_size = msg_size >> 1; 211 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_TX, i); 212 ret = adf_create_ring(accel_dev, SEC, asym_bank, num_msg_asym, 213 msg_size, key, NULL, 0, &inst->pke_tx); 214 if (ret) 215 goto err; 216 217 msg_size = ICP_QAT_FW_RESP_DEFAULT_SZ; 218 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_RX, i); 219 ret = adf_create_ring(accel_dev, SEC, sym_bank, num_msg_sym, 220 msg_size, key, qat_alg_callback, 0, 221 &inst->sym_rx); 222 if (ret) 223 goto err; 224 225 snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_RX, i); 226 ret = adf_create_ring(accel_dev, SEC, asym_bank, num_msg_asym, 227 msg_size, key, qat_alg_asym_callback, 0, 228 &inst->pke_rx); 229 if (ret) 230 goto err; 231 232 INIT_LIST_HEAD(&inst->backlog.list); 233 spin_lock_init(&inst->backlog.lock); 234 } 235 return 0; 236 err: 237 qat_crypto_free_instances(accel_dev); 238 return ret; 239 } 240 241 static int qat_crypto_init(struct adf_accel_dev *accel_dev) 242 { 243 if (qat_crypto_create_instances(accel_dev)) 244 return -EFAULT; 245 246 return 0; 247 } 248 249 static int qat_crypto_shutdown(struct adf_accel_dev *accel_dev) 250 { 251 return qat_crypto_free_instances(accel_dev); 252 } 253 254 static int qat_crypto_event_handler(struct adf_accel_dev *accel_dev, 255 enum adf_event event) 256 { 257 int ret; 258 259 switch (event) { 260 case ADF_EVENT_INIT: 261 ret = qat_crypto_init(accel_dev); 262 break; 263 case ADF_EVENT_SHUTDOWN: 264 ret = qat_crypto_shutdown(accel_dev); 265 break; 266 case ADF_EVENT_RESTARTING: 267 case ADF_EVENT_RESTARTED: 268 case ADF_EVENT_START: 269 case ADF_EVENT_STOP: 270 default: 271 ret = 0; 272 } 273 return ret; 274 } 275 276 int qat_crypto_register(void) 277 { 278 memset(&qat_crypto, 0, sizeof(qat_crypto)); 279 qat_crypto.event_hld = qat_crypto_event_handler; 280 qat_crypto.name = "qat_crypto"; 281 return adf_service_register(&qat_crypto); 282 } 283 284 int qat_crypto_unregister(void) 285 { 286 return adf_service_unregister(&qat_crypto); 287 } 288