xref: /openbmc/linux/drivers/nvme/host/auth.c (revision f4356947)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (c) 2020 Hannes Reinecke, SUSE Linux
4  */
5 
6 #include <linux/crc32.h>
7 #include <linux/base64.h>
8 #include <linux/prandom.h>
9 #include <asm/unaligned.h>
10 #include <crypto/hash.h>
11 #include <crypto/dh.h>
12 #include "nvme.h"
13 #include "fabrics.h"
14 #include <linux/nvme-auth.h>
15 
16 #define CHAP_BUF_SIZE 4096
17 static struct kmem_cache *nvme_chap_buf_cache;
18 static mempool_t *nvme_chap_buf_pool;
19 
20 struct nvme_dhchap_queue_context {
21 	struct list_head entry;
22 	struct work_struct auth_work;
23 	struct nvme_ctrl *ctrl;
24 	struct crypto_shash *shash_tfm;
25 	struct crypto_kpp *dh_tfm;
26 	void *buf;
27 	int qid;
28 	int error;
29 	u32 s1;
30 	u32 s2;
31 	u16 transaction;
32 	u8 status;
33 	u8 hash_id;
34 	size_t hash_len;
35 	u8 dhgroup_id;
36 	u8 c1[64];
37 	u8 c2[64];
38 	u8 response[64];
39 	u8 *host_response;
40 	u8 *ctrl_key;
41 	int ctrl_key_len;
42 	u8 *host_key;
43 	int host_key_len;
44 	u8 *sess_key;
45 	int sess_key_len;
46 };
47 
48 static struct workqueue_struct *nvme_auth_wq;
49 
50 #define nvme_auth_flags_from_qid(qid) \
51 	(qid == 0) ? 0 : BLK_MQ_REQ_NOWAIT | BLK_MQ_REQ_RESERVED
52 #define nvme_auth_queue_from_qid(ctrl, qid) \
53 	(qid == 0) ? (ctrl)->fabrics_q : (ctrl)->connect_q
54 
55 static inline int ctrl_max_dhchaps(struct nvme_ctrl *ctrl)
56 {
57 	return ctrl->opts->nr_io_queues + ctrl->opts->nr_write_queues +
58 			ctrl->opts->nr_poll_queues + 1;
59 }
60 
61 static int nvme_auth_submit(struct nvme_ctrl *ctrl, int qid,
62 			    void *data, size_t data_len, bool auth_send)
63 {
64 	struct nvme_command cmd = {};
65 	blk_mq_req_flags_t flags = nvme_auth_flags_from_qid(qid);
66 	struct request_queue *q = nvme_auth_queue_from_qid(ctrl, qid);
67 	int ret;
68 
69 	cmd.auth_common.opcode = nvme_fabrics_command;
70 	cmd.auth_common.secp = NVME_AUTH_DHCHAP_PROTOCOL_IDENTIFIER;
71 	cmd.auth_common.spsp0 = 0x01;
72 	cmd.auth_common.spsp1 = 0x01;
73 	if (auth_send) {
74 		cmd.auth_send.fctype = nvme_fabrics_type_auth_send;
75 		cmd.auth_send.tl = cpu_to_le32(data_len);
76 	} else {
77 		cmd.auth_receive.fctype = nvme_fabrics_type_auth_receive;
78 		cmd.auth_receive.al = cpu_to_le32(data_len);
79 	}
80 
81 	ret = __nvme_submit_sync_cmd(q, &cmd, NULL, data, data_len,
82 				     qid == 0 ? NVME_QID_ANY : qid,
83 				     0, flags);
84 	if (ret > 0)
85 		dev_warn(ctrl->device,
86 			"qid %d auth_send failed with status %d\n", qid, ret);
87 	else if (ret < 0)
88 		dev_err(ctrl->device,
89 			"qid %d auth_send failed with error %d\n", qid, ret);
90 	return ret;
91 }
92 
93 static int nvme_auth_receive_validate(struct nvme_ctrl *ctrl, int qid,
94 		struct nvmf_auth_dhchap_failure_data *data,
95 		u16 transaction, u8 expected_msg)
96 {
97 	dev_dbg(ctrl->device, "%s: qid %d auth_type %d auth_id %x\n",
98 		__func__, qid, data->auth_type, data->auth_id);
99 
100 	if (data->auth_type == NVME_AUTH_COMMON_MESSAGES &&
101 	    data->auth_id == NVME_AUTH_DHCHAP_MESSAGE_FAILURE1) {
102 		return data->rescode_exp;
103 	}
104 	if (data->auth_type != NVME_AUTH_DHCHAP_MESSAGES ||
105 	    data->auth_id != expected_msg) {
106 		dev_warn(ctrl->device,
107 			 "qid %d invalid message %02x/%02x\n",
108 			 qid, data->auth_type, data->auth_id);
109 		return NVME_AUTH_DHCHAP_FAILURE_INCORRECT_MESSAGE;
110 	}
111 	if (le16_to_cpu(data->t_id) != transaction) {
112 		dev_warn(ctrl->device,
113 			 "qid %d invalid transaction ID %d\n",
114 			 qid, le16_to_cpu(data->t_id));
115 		return NVME_AUTH_DHCHAP_FAILURE_INCORRECT_MESSAGE;
116 	}
117 	return 0;
118 }
119 
120 static int nvme_auth_set_dhchap_negotiate_data(struct nvme_ctrl *ctrl,
121 		struct nvme_dhchap_queue_context *chap)
122 {
123 	struct nvmf_auth_dhchap_negotiate_data *data = chap->buf;
124 	size_t size = sizeof(*data) + sizeof(union nvmf_auth_protocol);
125 
126 	if (size > CHAP_BUF_SIZE) {
127 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
128 		return -EINVAL;
129 	}
130 	memset((u8 *)chap->buf, 0, size);
131 	data->auth_type = NVME_AUTH_COMMON_MESSAGES;
132 	data->auth_id = NVME_AUTH_DHCHAP_MESSAGE_NEGOTIATE;
133 	data->t_id = cpu_to_le16(chap->transaction);
134 	data->sc_c = 0; /* No secure channel concatenation */
135 	data->napd = 1;
136 	data->auth_protocol[0].dhchap.authid = NVME_AUTH_DHCHAP_AUTH_ID;
137 	data->auth_protocol[0].dhchap.halen = 3;
138 	data->auth_protocol[0].dhchap.dhlen = 6;
139 	data->auth_protocol[0].dhchap.idlist[0] = NVME_AUTH_HASH_SHA256;
140 	data->auth_protocol[0].dhchap.idlist[1] = NVME_AUTH_HASH_SHA384;
141 	data->auth_protocol[0].dhchap.idlist[2] = NVME_AUTH_HASH_SHA512;
142 	data->auth_protocol[0].dhchap.idlist[30] = NVME_AUTH_DHGROUP_NULL;
143 	data->auth_protocol[0].dhchap.idlist[31] = NVME_AUTH_DHGROUP_2048;
144 	data->auth_protocol[0].dhchap.idlist[32] = NVME_AUTH_DHGROUP_3072;
145 	data->auth_protocol[0].dhchap.idlist[33] = NVME_AUTH_DHGROUP_4096;
146 	data->auth_protocol[0].dhchap.idlist[34] = NVME_AUTH_DHGROUP_6144;
147 	data->auth_protocol[0].dhchap.idlist[35] = NVME_AUTH_DHGROUP_8192;
148 
149 	return size;
150 }
151 
152 static int nvme_auth_process_dhchap_challenge(struct nvme_ctrl *ctrl,
153 		struct nvme_dhchap_queue_context *chap)
154 {
155 	struct nvmf_auth_dhchap_challenge_data *data = chap->buf;
156 	u16 dhvlen = le16_to_cpu(data->dhvlen);
157 	size_t size = sizeof(*data) + data->hl + dhvlen;
158 	const char *gid_name = nvme_auth_dhgroup_name(data->dhgid);
159 	const char *hmac_name, *kpp_name;
160 
161 	if (size > CHAP_BUF_SIZE) {
162 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
163 		return -EINVAL;
164 	}
165 
166 	hmac_name = nvme_auth_hmac_name(data->hashid);
167 	if (!hmac_name) {
168 		dev_warn(ctrl->device,
169 			 "qid %d: invalid HASH ID %d\n",
170 			 chap->qid, data->hashid);
171 		chap->status = NVME_AUTH_DHCHAP_FAILURE_HASH_UNUSABLE;
172 		return -EPROTO;
173 	}
174 
175 	if (chap->hash_id == data->hashid && chap->shash_tfm &&
176 	    !strcmp(crypto_shash_alg_name(chap->shash_tfm), hmac_name) &&
177 	    crypto_shash_digestsize(chap->shash_tfm) == data->hl) {
178 		dev_dbg(ctrl->device,
179 			"qid %d: reuse existing hash %s\n",
180 			chap->qid, hmac_name);
181 		goto select_kpp;
182 	}
183 
184 	/* Reset if hash cannot be reused */
185 	if (chap->shash_tfm) {
186 		crypto_free_shash(chap->shash_tfm);
187 		chap->hash_id = 0;
188 		chap->hash_len = 0;
189 	}
190 	chap->shash_tfm = crypto_alloc_shash(hmac_name, 0,
191 					     CRYPTO_ALG_ALLOCATES_MEMORY);
192 	if (IS_ERR(chap->shash_tfm)) {
193 		dev_warn(ctrl->device,
194 			 "qid %d: failed to allocate hash %s, error %ld\n",
195 			 chap->qid, hmac_name, PTR_ERR(chap->shash_tfm));
196 		chap->shash_tfm = NULL;
197 		chap->status = NVME_AUTH_DHCHAP_FAILURE_FAILED;
198 		return -ENOMEM;
199 	}
200 
201 	if (crypto_shash_digestsize(chap->shash_tfm) != data->hl) {
202 		dev_warn(ctrl->device,
203 			 "qid %d: invalid hash length %d\n",
204 			 chap->qid, data->hl);
205 		crypto_free_shash(chap->shash_tfm);
206 		chap->shash_tfm = NULL;
207 		chap->status = NVME_AUTH_DHCHAP_FAILURE_HASH_UNUSABLE;
208 		return -EPROTO;
209 	}
210 
211 	chap->hash_id = data->hashid;
212 	chap->hash_len = data->hl;
213 	dev_dbg(ctrl->device, "qid %d: selected hash %s\n",
214 		chap->qid, hmac_name);
215 
216 select_kpp:
217 	kpp_name = nvme_auth_dhgroup_kpp(data->dhgid);
218 	if (!kpp_name) {
219 		dev_warn(ctrl->device,
220 			 "qid %d: invalid DH group id %d\n",
221 			 chap->qid, data->dhgid);
222 		chap->status = NVME_AUTH_DHCHAP_FAILURE_DHGROUP_UNUSABLE;
223 		/* Leave previous dh_tfm intact */
224 		return -EPROTO;
225 	}
226 
227 	if (chap->dhgroup_id == data->dhgid &&
228 	    (data->dhgid == NVME_AUTH_DHGROUP_NULL || chap->dh_tfm)) {
229 		dev_dbg(ctrl->device,
230 			"qid %d: reuse existing DH group %s\n",
231 			chap->qid, gid_name);
232 		goto skip_kpp;
233 	}
234 
235 	/* Reset dh_tfm if it can't be reused */
236 	if (chap->dh_tfm) {
237 		crypto_free_kpp(chap->dh_tfm);
238 		chap->dh_tfm = NULL;
239 	}
240 
241 	if (data->dhgid != NVME_AUTH_DHGROUP_NULL) {
242 		if (dhvlen == 0) {
243 			dev_warn(ctrl->device,
244 				 "qid %d: empty DH value\n",
245 				 chap->qid);
246 			chap->status = NVME_AUTH_DHCHAP_FAILURE_DHGROUP_UNUSABLE;
247 			return -EPROTO;
248 		}
249 
250 		chap->dh_tfm = crypto_alloc_kpp(kpp_name, 0, 0);
251 		if (IS_ERR(chap->dh_tfm)) {
252 			int ret = PTR_ERR(chap->dh_tfm);
253 
254 			dev_warn(ctrl->device,
255 				 "qid %d: error %d initializing DH group %s\n",
256 				 chap->qid, ret, gid_name);
257 			chap->status = NVME_AUTH_DHCHAP_FAILURE_DHGROUP_UNUSABLE;
258 			chap->dh_tfm = NULL;
259 			return ret;
260 		}
261 		dev_dbg(ctrl->device, "qid %d: selected DH group %s\n",
262 			chap->qid, gid_name);
263 	} else if (dhvlen != 0) {
264 		dev_warn(ctrl->device,
265 			 "qid %d: invalid DH value for NULL DH\n",
266 			 chap->qid);
267 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
268 		return -EPROTO;
269 	}
270 	chap->dhgroup_id = data->dhgid;
271 
272 skip_kpp:
273 	chap->s1 = le32_to_cpu(data->seqnum);
274 	memcpy(chap->c1, data->cval, chap->hash_len);
275 	if (dhvlen) {
276 		chap->ctrl_key = kmalloc(dhvlen, GFP_KERNEL);
277 		if (!chap->ctrl_key) {
278 			chap->status = NVME_AUTH_DHCHAP_FAILURE_FAILED;
279 			return -ENOMEM;
280 		}
281 		chap->ctrl_key_len = dhvlen;
282 		memcpy(chap->ctrl_key, data->cval + chap->hash_len,
283 		       dhvlen);
284 		dev_dbg(ctrl->device, "ctrl public key %*ph\n",
285 			 (int)chap->ctrl_key_len, chap->ctrl_key);
286 	}
287 
288 	return 0;
289 }
290 
291 static int nvme_auth_set_dhchap_reply_data(struct nvme_ctrl *ctrl,
292 		struct nvme_dhchap_queue_context *chap)
293 {
294 	struct nvmf_auth_dhchap_reply_data *data = chap->buf;
295 	size_t size = sizeof(*data);
296 
297 	size += 2 * chap->hash_len;
298 
299 	if (chap->host_key_len)
300 		size += chap->host_key_len;
301 
302 	if (size > CHAP_BUF_SIZE) {
303 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
304 		return -EINVAL;
305 	}
306 
307 	memset(chap->buf, 0, size);
308 	data->auth_type = NVME_AUTH_DHCHAP_MESSAGES;
309 	data->auth_id = NVME_AUTH_DHCHAP_MESSAGE_REPLY;
310 	data->t_id = cpu_to_le16(chap->transaction);
311 	data->hl = chap->hash_len;
312 	data->dhvlen = cpu_to_le16(chap->host_key_len);
313 	memcpy(data->rval, chap->response, chap->hash_len);
314 	if (ctrl->ctrl_key) {
315 		get_random_bytes(chap->c2, chap->hash_len);
316 		data->cvalid = 1;
317 		chap->s2 = nvme_auth_get_seqnum();
318 		memcpy(data->rval + chap->hash_len, chap->c2,
319 		       chap->hash_len);
320 		dev_dbg(ctrl->device, "%s: qid %d ctrl challenge %*ph\n",
321 			__func__, chap->qid, (int)chap->hash_len, chap->c2);
322 	} else {
323 		memset(chap->c2, 0, chap->hash_len);
324 		chap->s2 = 0;
325 	}
326 	data->seqnum = cpu_to_le32(chap->s2);
327 	if (chap->host_key_len) {
328 		dev_dbg(ctrl->device, "%s: qid %d host public key %*ph\n",
329 			__func__, chap->qid,
330 			chap->host_key_len, chap->host_key);
331 		memcpy(data->rval + 2 * chap->hash_len, chap->host_key,
332 		       chap->host_key_len);
333 	}
334 
335 	return size;
336 }
337 
338 static int nvme_auth_process_dhchap_success1(struct nvme_ctrl *ctrl,
339 		struct nvme_dhchap_queue_context *chap)
340 {
341 	struct nvmf_auth_dhchap_success1_data *data = chap->buf;
342 	size_t size = sizeof(*data);
343 
344 	if (chap->ctrl_key)
345 		size += chap->hash_len;
346 
347 	if (size > CHAP_BUF_SIZE) {
348 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
349 		return -EINVAL;
350 	}
351 
352 	if (data->hl != chap->hash_len) {
353 		dev_warn(ctrl->device,
354 			 "qid %d: invalid hash length %u\n",
355 			 chap->qid, data->hl);
356 		chap->status = NVME_AUTH_DHCHAP_FAILURE_HASH_UNUSABLE;
357 		return -EPROTO;
358 	}
359 
360 	/* Just print out information for the admin queue */
361 	if (chap->qid == 0)
362 		dev_info(ctrl->device,
363 			 "qid 0: authenticated with hash %s dhgroup %s\n",
364 			 nvme_auth_hmac_name(chap->hash_id),
365 			 nvme_auth_dhgroup_name(chap->dhgroup_id));
366 
367 	if (!data->rvalid)
368 		return 0;
369 
370 	/* Validate controller response */
371 	if (memcmp(chap->response, data->rval, data->hl)) {
372 		dev_dbg(ctrl->device, "%s: qid %d ctrl response %*ph\n",
373 			__func__, chap->qid, (int)chap->hash_len, data->rval);
374 		dev_dbg(ctrl->device, "%s: qid %d host response %*ph\n",
375 			__func__, chap->qid, (int)chap->hash_len,
376 			chap->response);
377 		dev_warn(ctrl->device,
378 			 "qid %d: controller authentication failed\n",
379 			 chap->qid);
380 		chap->status = NVME_AUTH_DHCHAP_FAILURE_FAILED;
381 		return -ECONNREFUSED;
382 	}
383 
384 	/* Just print out information for the admin queue */
385 	if (chap->qid == 0)
386 		dev_info(ctrl->device,
387 			 "qid 0: controller authenticated\n");
388 	return 0;
389 }
390 
391 static int nvme_auth_set_dhchap_success2_data(struct nvme_ctrl *ctrl,
392 		struct nvme_dhchap_queue_context *chap)
393 {
394 	struct nvmf_auth_dhchap_success2_data *data = chap->buf;
395 	size_t size = sizeof(*data);
396 
397 	memset(chap->buf, 0, size);
398 	data->auth_type = NVME_AUTH_DHCHAP_MESSAGES;
399 	data->auth_id = NVME_AUTH_DHCHAP_MESSAGE_SUCCESS2;
400 	data->t_id = cpu_to_le16(chap->transaction);
401 
402 	return size;
403 }
404 
405 static int nvme_auth_set_dhchap_failure2_data(struct nvme_ctrl *ctrl,
406 		struct nvme_dhchap_queue_context *chap)
407 {
408 	struct nvmf_auth_dhchap_failure_data *data = chap->buf;
409 	size_t size = sizeof(*data);
410 
411 	memset(chap->buf, 0, size);
412 	data->auth_type = NVME_AUTH_COMMON_MESSAGES;
413 	data->auth_id = NVME_AUTH_DHCHAP_MESSAGE_FAILURE2;
414 	data->t_id = cpu_to_le16(chap->transaction);
415 	data->rescode = NVME_AUTH_DHCHAP_FAILURE_REASON_FAILED;
416 	data->rescode_exp = chap->status;
417 
418 	return size;
419 }
420 
421 static int nvme_auth_dhchap_setup_host_response(struct nvme_ctrl *ctrl,
422 		struct nvme_dhchap_queue_context *chap)
423 {
424 	SHASH_DESC_ON_STACK(shash, chap->shash_tfm);
425 	u8 buf[4], *challenge = chap->c1;
426 	int ret;
427 
428 	dev_dbg(ctrl->device, "%s: qid %d host response seq %u transaction %d\n",
429 		__func__, chap->qid, chap->s1, chap->transaction);
430 
431 	if (!chap->host_response) {
432 		chap->host_response = nvme_auth_transform_key(ctrl->host_key,
433 						ctrl->opts->host->nqn);
434 		if (IS_ERR(chap->host_response)) {
435 			ret = PTR_ERR(chap->host_response);
436 			chap->host_response = NULL;
437 			return ret;
438 		}
439 	} else {
440 		dev_dbg(ctrl->device, "%s: qid %d re-using host response\n",
441 			__func__, chap->qid);
442 	}
443 
444 	ret = crypto_shash_setkey(chap->shash_tfm,
445 			chap->host_response, ctrl->host_key->len);
446 	if (ret) {
447 		dev_warn(ctrl->device, "qid %d: failed to set key, error %d\n",
448 			 chap->qid, ret);
449 		goto out;
450 	}
451 
452 	if (chap->dh_tfm) {
453 		challenge = kmalloc(chap->hash_len, GFP_KERNEL);
454 		if (!challenge) {
455 			ret = -ENOMEM;
456 			goto out;
457 		}
458 		ret = nvme_auth_augmented_challenge(chap->hash_id,
459 						    chap->sess_key,
460 						    chap->sess_key_len,
461 						    chap->c1, challenge,
462 						    chap->hash_len);
463 		if (ret)
464 			goto out;
465 	}
466 
467 	shash->tfm = chap->shash_tfm;
468 	ret = crypto_shash_init(shash);
469 	if (ret)
470 		goto out;
471 	ret = crypto_shash_update(shash, challenge, chap->hash_len);
472 	if (ret)
473 		goto out;
474 	put_unaligned_le32(chap->s1, buf);
475 	ret = crypto_shash_update(shash, buf, 4);
476 	if (ret)
477 		goto out;
478 	put_unaligned_le16(chap->transaction, buf);
479 	ret = crypto_shash_update(shash, buf, 2);
480 	if (ret)
481 		goto out;
482 	memset(buf, 0, sizeof(buf));
483 	ret = crypto_shash_update(shash, buf, 1);
484 	if (ret)
485 		goto out;
486 	ret = crypto_shash_update(shash, "HostHost", 8);
487 	if (ret)
488 		goto out;
489 	ret = crypto_shash_update(shash, ctrl->opts->host->nqn,
490 				  strlen(ctrl->opts->host->nqn));
491 	if (ret)
492 		goto out;
493 	ret = crypto_shash_update(shash, buf, 1);
494 	if (ret)
495 		goto out;
496 	ret = crypto_shash_update(shash, ctrl->opts->subsysnqn,
497 			    strlen(ctrl->opts->subsysnqn));
498 	if (ret)
499 		goto out;
500 	ret = crypto_shash_final(shash, chap->response);
501 out:
502 	if (challenge != chap->c1)
503 		kfree(challenge);
504 	return ret;
505 }
506 
507 static int nvme_auth_dhchap_setup_ctrl_response(struct nvme_ctrl *ctrl,
508 		struct nvme_dhchap_queue_context *chap)
509 {
510 	SHASH_DESC_ON_STACK(shash, chap->shash_tfm);
511 	u8 *ctrl_response;
512 	u8 buf[4], *challenge = chap->c2;
513 	int ret;
514 
515 	ctrl_response = nvme_auth_transform_key(ctrl->ctrl_key,
516 				ctrl->opts->subsysnqn);
517 	if (IS_ERR(ctrl_response)) {
518 		ret = PTR_ERR(ctrl_response);
519 		return ret;
520 	}
521 
522 	ret = crypto_shash_setkey(chap->shash_tfm,
523 			ctrl_response, ctrl->ctrl_key->len);
524 	if (ret) {
525 		dev_warn(ctrl->device, "qid %d: failed to set key, error %d\n",
526 			 chap->qid, ret);
527 		goto out;
528 	}
529 
530 	if (chap->dh_tfm) {
531 		challenge = kmalloc(chap->hash_len, GFP_KERNEL);
532 		if (!challenge) {
533 			ret = -ENOMEM;
534 			goto out;
535 		}
536 		ret = nvme_auth_augmented_challenge(chap->hash_id,
537 						    chap->sess_key,
538 						    chap->sess_key_len,
539 						    chap->c2, challenge,
540 						    chap->hash_len);
541 		if (ret)
542 			goto out;
543 	}
544 	dev_dbg(ctrl->device, "%s: qid %d ctrl response seq %u transaction %d\n",
545 		__func__, chap->qid, chap->s2, chap->transaction);
546 	dev_dbg(ctrl->device, "%s: qid %d challenge %*ph\n",
547 		__func__, chap->qid, (int)chap->hash_len, challenge);
548 	dev_dbg(ctrl->device, "%s: qid %d subsysnqn %s\n",
549 		__func__, chap->qid, ctrl->opts->subsysnqn);
550 	dev_dbg(ctrl->device, "%s: qid %d hostnqn %s\n",
551 		__func__, chap->qid, ctrl->opts->host->nqn);
552 	shash->tfm = chap->shash_tfm;
553 	ret = crypto_shash_init(shash);
554 	if (ret)
555 		goto out;
556 	ret = crypto_shash_update(shash, challenge, chap->hash_len);
557 	if (ret)
558 		goto out;
559 	put_unaligned_le32(chap->s2, buf);
560 	ret = crypto_shash_update(shash, buf, 4);
561 	if (ret)
562 		goto out;
563 	put_unaligned_le16(chap->transaction, buf);
564 	ret = crypto_shash_update(shash, buf, 2);
565 	if (ret)
566 		goto out;
567 	memset(buf, 0, 4);
568 	ret = crypto_shash_update(shash, buf, 1);
569 	if (ret)
570 		goto out;
571 	ret = crypto_shash_update(shash, "Controller", 10);
572 	if (ret)
573 		goto out;
574 	ret = crypto_shash_update(shash, ctrl->opts->subsysnqn,
575 				  strlen(ctrl->opts->subsysnqn));
576 	if (ret)
577 		goto out;
578 	ret = crypto_shash_update(shash, buf, 1);
579 	if (ret)
580 		goto out;
581 	ret = crypto_shash_update(shash, ctrl->opts->host->nqn,
582 				  strlen(ctrl->opts->host->nqn));
583 	if (ret)
584 		goto out;
585 	ret = crypto_shash_final(shash, chap->response);
586 out:
587 	if (challenge != chap->c2)
588 		kfree(challenge);
589 	kfree(ctrl_response);
590 	return ret;
591 }
592 
593 static int nvme_auth_dhchap_exponential(struct nvme_ctrl *ctrl,
594 		struct nvme_dhchap_queue_context *chap)
595 {
596 	int ret;
597 
598 	if (chap->host_key && chap->host_key_len) {
599 		dev_dbg(ctrl->device,
600 			"qid %d: reusing host key\n", chap->qid);
601 		goto gen_sesskey;
602 	}
603 	ret = nvme_auth_gen_privkey(chap->dh_tfm, chap->dhgroup_id);
604 	if (ret < 0) {
605 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
606 		return ret;
607 	}
608 
609 	chap->host_key_len = crypto_kpp_maxsize(chap->dh_tfm);
610 
611 	chap->host_key = kzalloc(chap->host_key_len, GFP_KERNEL);
612 	if (!chap->host_key) {
613 		chap->host_key_len = 0;
614 		chap->status = NVME_AUTH_DHCHAP_FAILURE_FAILED;
615 		return -ENOMEM;
616 	}
617 	ret = nvme_auth_gen_pubkey(chap->dh_tfm,
618 				   chap->host_key, chap->host_key_len);
619 	if (ret) {
620 		dev_dbg(ctrl->device,
621 			"failed to generate public key, error %d\n", ret);
622 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
623 		return ret;
624 	}
625 
626 gen_sesskey:
627 	chap->sess_key_len = chap->host_key_len;
628 	chap->sess_key = kmalloc(chap->sess_key_len, GFP_KERNEL);
629 	if (!chap->sess_key) {
630 		chap->sess_key_len = 0;
631 		chap->status = NVME_AUTH_DHCHAP_FAILURE_FAILED;
632 		return -ENOMEM;
633 	}
634 
635 	ret = nvme_auth_gen_shared_secret(chap->dh_tfm,
636 					  chap->ctrl_key, chap->ctrl_key_len,
637 					  chap->sess_key, chap->sess_key_len);
638 	if (ret) {
639 		dev_dbg(ctrl->device,
640 			"failed to generate shared secret, error %d\n", ret);
641 		chap->status = NVME_AUTH_DHCHAP_FAILURE_INCORRECT_PAYLOAD;
642 		return ret;
643 	}
644 	dev_dbg(ctrl->device, "shared secret %*ph\n",
645 		(int)chap->sess_key_len, chap->sess_key);
646 	return 0;
647 }
648 
649 static void nvme_auth_reset_dhchap(struct nvme_dhchap_queue_context *chap)
650 {
651 	kfree_sensitive(chap->host_response);
652 	chap->host_response = NULL;
653 	kfree_sensitive(chap->host_key);
654 	chap->host_key = NULL;
655 	chap->host_key_len = 0;
656 	kfree_sensitive(chap->ctrl_key);
657 	chap->ctrl_key = NULL;
658 	chap->ctrl_key_len = 0;
659 	kfree_sensitive(chap->sess_key);
660 	chap->sess_key = NULL;
661 	chap->sess_key_len = 0;
662 	chap->status = 0;
663 	chap->error = 0;
664 	chap->s1 = 0;
665 	chap->s2 = 0;
666 	chap->transaction = 0;
667 	memset(chap->c1, 0, sizeof(chap->c1));
668 	memset(chap->c2, 0, sizeof(chap->c2));
669 	mempool_free(chap->buf, nvme_chap_buf_pool);
670 	chap->buf = NULL;
671 }
672 
673 static void nvme_auth_free_dhchap(struct nvme_dhchap_queue_context *chap)
674 {
675 	nvme_auth_reset_dhchap(chap);
676 	if (chap->shash_tfm)
677 		crypto_free_shash(chap->shash_tfm);
678 	if (chap->dh_tfm)
679 		crypto_free_kpp(chap->dh_tfm);
680 }
681 
682 static void nvme_queue_auth_work(struct work_struct *work)
683 {
684 	struct nvme_dhchap_queue_context *chap =
685 		container_of(work, struct nvme_dhchap_queue_context, auth_work);
686 	struct nvme_ctrl *ctrl = chap->ctrl;
687 	size_t tl;
688 	int ret = 0;
689 
690 	/*
691 	 * Allocate a large enough buffer for the entire negotiation:
692 	 * 4k is enough to ffdhe8192.
693 	 */
694 	chap->buf = mempool_alloc(nvme_chap_buf_pool, GFP_KERNEL);
695 	if (!chap->buf) {
696 		chap->error = -ENOMEM;
697 		return;
698 	}
699 
700 	chap->transaction = ctrl->transaction++;
701 
702 	/* DH-HMAC-CHAP Step 1: send negotiate */
703 	dev_dbg(ctrl->device, "%s: qid %d send negotiate\n",
704 		__func__, chap->qid);
705 	ret = nvme_auth_set_dhchap_negotiate_data(ctrl, chap);
706 	if (ret < 0) {
707 		chap->error = ret;
708 		return;
709 	}
710 	tl = ret;
711 	ret = nvme_auth_submit(ctrl, chap->qid, chap->buf, tl, true);
712 	if (ret) {
713 		chap->error = ret;
714 		return;
715 	}
716 
717 	/* DH-HMAC-CHAP Step 2: receive challenge */
718 	dev_dbg(ctrl->device, "%s: qid %d receive challenge\n",
719 		__func__, chap->qid);
720 
721 	memset(chap->buf, 0, CHAP_BUF_SIZE);
722 	ret = nvme_auth_submit(ctrl, chap->qid, chap->buf, CHAP_BUF_SIZE,
723 			       false);
724 	if (ret) {
725 		dev_warn(ctrl->device,
726 			 "qid %d failed to receive challenge, %s %d\n",
727 			 chap->qid, ret < 0 ? "error" : "nvme status", ret);
728 		chap->error = ret;
729 		return;
730 	}
731 	ret = nvme_auth_receive_validate(ctrl, chap->qid, chap->buf, chap->transaction,
732 					 NVME_AUTH_DHCHAP_MESSAGE_CHALLENGE);
733 	if (ret) {
734 		chap->status = ret;
735 		chap->error = -ECONNREFUSED;
736 		return;
737 	}
738 
739 	ret = nvme_auth_process_dhchap_challenge(ctrl, chap);
740 	if (ret) {
741 		/* Invalid challenge parameters */
742 		chap->error = ret;
743 		goto fail2;
744 	}
745 
746 	if (chap->ctrl_key_len) {
747 		dev_dbg(ctrl->device,
748 			"%s: qid %d DH exponential\n",
749 			__func__, chap->qid);
750 		ret = nvme_auth_dhchap_exponential(ctrl, chap);
751 		if (ret) {
752 			chap->error = ret;
753 			goto fail2;
754 		}
755 	}
756 
757 	dev_dbg(ctrl->device, "%s: qid %d host response\n",
758 		__func__, chap->qid);
759 	mutex_lock(&ctrl->dhchap_auth_mutex);
760 	ret = nvme_auth_dhchap_setup_host_response(ctrl, chap);
761 	if (ret) {
762 		mutex_unlock(&ctrl->dhchap_auth_mutex);
763 		chap->error = ret;
764 		goto fail2;
765 	}
766 	mutex_unlock(&ctrl->dhchap_auth_mutex);
767 
768 	/* DH-HMAC-CHAP Step 3: send reply */
769 	dev_dbg(ctrl->device, "%s: qid %d send reply\n",
770 		__func__, chap->qid);
771 	ret = nvme_auth_set_dhchap_reply_data(ctrl, chap);
772 	if (ret < 0) {
773 		chap->error = ret;
774 		goto fail2;
775 	}
776 
777 	tl = ret;
778 	ret = nvme_auth_submit(ctrl, chap->qid, chap->buf, tl, true);
779 	if (ret) {
780 		chap->error = ret;
781 		goto fail2;
782 	}
783 
784 	/* DH-HMAC-CHAP Step 4: receive success1 */
785 	dev_dbg(ctrl->device, "%s: qid %d receive success1\n",
786 		__func__, chap->qid);
787 
788 	memset(chap->buf, 0, CHAP_BUF_SIZE);
789 	ret = nvme_auth_submit(ctrl, chap->qid, chap->buf, CHAP_BUF_SIZE,
790 			       false);
791 	if (ret) {
792 		dev_warn(ctrl->device,
793 			 "qid %d failed to receive success1, %s %d\n",
794 			 chap->qid, ret < 0 ? "error" : "nvme status", ret);
795 		chap->error = ret;
796 		return;
797 	}
798 	ret = nvme_auth_receive_validate(ctrl, chap->qid,
799 					 chap->buf, chap->transaction,
800 					 NVME_AUTH_DHCHAP_MESSAGE_SUCCESS1);
801 	if (ret) {
802 		chap->status = ret;
803 		chap->error = -ECONNREFUSED;
804 		return;
805 	}
806 
807 	mutex_lock(&ctrl->dhchap_auth_mutex);
808 	if (ctrl->ctrl_key) {
809 		dev_dbg(ctrl->device,
810 			"%s: qid %d controller response\n",
811 			__func__, chap->qid);
812 		ret = nvme_auth_dhchap_setup_ctrl_response(ctrl, chap);
813 		if (ret) {
814 			mutex_unlock(&ctrl->dhchap_auth_mutex);
815 			chap->error = ret;
816 			goto fail2;
817 		}
818 	}
819 	mutex_unlock(&ctrl->dhchap_auth_mutex);
820 
821 	ret = nvme_auth_process_dhchap_success1(ctrl, chap);
822 	if (ret) {
823 		/* Controller authentication failed */
824 		chap->error = -ECONNREFUSED;
825 		goto fail2;
826 	}
827 
828 	if (chap->ctrl_key) {
829 		/* DH-HMAC-CHAP Step 5: send success2 */
830 		dev_dbg(ctrl->device, "%s: qid %d send success2\n",
831 			__func__, chap->qid);
832 		tl = nvme_auth_set_dhchap_success2_data(ctrl, chap);
833 		ret = nvme_auth_submit(ctrl, chap->qid, chap->buf, tl, true);
834 		if (ret)
835 			chap->error = ret;
836 	}
837 	if (!ret) {
838 		chap->error = 0;
839 		return;
840 	}
841 
842 fail2:
843 	dev_dbg(ctrl->device, "%s: qid %d send failure2, status %x\n",
844 		__func__, chap->qid, chap->status);
845 	tl = nvme_auth_set_dhchap_failure2_data(ctrl, chap);
846 	ret = nvme_auth_submit(ctrl, chap->qid, chap->buf, tl, true);
847 	/*
848 	 * only update error if send failure2 failed and no other
849 	 * error had been set during authentication.
850 	 */
851 	if (ret && !chap->error)
852 		chap->error = ret;
853 }
854 
855 int nvme_auth_negotiate(struct nvme_ctrl *ctrl, int qid)
856 {
857 	struct nvme_dhchap_queue_context *chap;
858 
859 	if (!ctrl->host_key) {
860 		dev_warn(ctrl->device, "qid %d: no key\n", qid);
861 		return -ENOKEY;
862 	}
863 
864 	if (ctrl->opts->dhchap_ctrl_secret && !ctrl->ctrl_key) {
865 		dev_warn(ctrl->device, "qid %d: invalid ctrl key\n", qid);
866 		return -ENOKEY;
867 	}
868 
869 	chap = &ctrl->dhchap_ctxs[qid];
870 	cancel_work_sync(&chap->auth_work);
871 	queue_work(nvme_auth_wq, &chap->auth_work);
872 	return 0;
873 }
874 EXPORT_SYMBOL_GPL(nvme_auth_negotiate);
875 
876 int nvme_auth_wait(struct nvme_ctrl *ctrl, int qid)
877 {
878 	struct nvme_dhchap_queue_context *chap;
879 	int ret;
880 
881 	chap = &ctrl->dhchap_ctxs[qid];
882 	flush_work(&chap->auth_work);
883 	ret = chap->error;
884 	/* clear sensitive info */
885 	nvme_auth_reset_dhchap(chap);
886 	return ret;
887 }
888 EXPORT_SYMBOL_GPL(nvme_auth_wait);
889 
890 static void nvme_ctrl_auth_work(struct work_struct *work)
891 {
892 	struct nvme_ctrl *ctrl =
893 		container_of(work, struct nvme_ctrl, dhchap_auth_work);
894 	int ret, q;
895 
896 	/*
897 	 * If the ctrl is no connected, bail as reconnect will handle
898 	 * authentication.
899 	 */
900 	if (ctrl->state != NVME_CTRL_LIVE)
901 		return;
902 
903 	/* Authenticate admin queue first */
904 	ret = nvme_auth_negotiate(ctrl, 0);
905 	if (ret) {
906 		dev_warn(ctrl->device,
907 			 "qid 0: error %d setting up authentication\n", ret);
908 		return;
909 	}
910 	ret = nvme_auth_wait(ctrl, 0);
911 	if (ret) {
912 		dev_warn(ctrl->device,
913 			 "qid 0: authentication failed\n");
914 		return;
915 	}
916 
917 	for (q = 1; q < ctrl->queue_count; q++) {
918 		ret = nvme_auth_negotiate(ctrl, q);
919 		if (ret) {
920 			dev_warn(ctrl->device,
921 				 "qid %d: error %d setting up authentication\n",
922 				 q, ret);
923 			break;
924 		}
925 	}
926 
927 	/*
928 	 * Failure is a soft-state; credentials remain valid until
929 	 * the controller terminates the connection.
930 	 */
931 	for (q = 1; q < ctrl->queue_count; q++) {
932 		ret = nvme_auth_wait(ctrl, q);
933 		if (ret)
934 			dev_warn(ctrl->device,
935 				 "qid %d: authentication failed\n", q);
936 	}
937 }
938 
939 int nvme_auth_init_ctrl(struct nvme_ctrl *ctrl)
940 {
941 	struct nvme_dhchap_queue_context *chap;
942 	int i, ret;
943 
944 	mutex_init(&ctrl->dhchap_auth_mutex);
945 	INIT_WORK(&ctrl->dhchap_auth_work, nvme_ctrl_auth_work);
946 	if (!ctrl->opts)
947 		return 0;
948 	ret = nvme_auth_generate_key(ctrl->opts->dhchap_secret,
949 			&ctrl->host_key);
950 	if (ret)
951 		return ret;
952 	ret = nvme_auth_generate_key(ctrl->opts->dhchap_ctrl_secret,
953 			&ctrl->ctrl_key);
954 	if (ret)
955 		goto err_free_dhchap_secret;
956 
957 	if (!ctrl->opts->dhchap_secret && !ctrl->opts->dhchap_ctrl_secret)
958 		return 0;
959 
960 	ctrl->dhchap_ctxs = kvcalloc(ctrl_max_dhchaps(ctrl),
961 				sizeof(*chap), GFP_KERNEL);
962 	if (!ctrl->dhchap_ctxs) {
963 		ret = -ENOMEM;
964 		goto err_free_dhchap_ctrl_secret;
965 	}
966 
967 	for (i = 0; i < ctrl_max_dhchaps(ctrl); i++) {
968 		chap = &ctrl->dhchap_ctxs[i];
969 		chap->qid = i;
970 		chap->ctrl = ctrl;
971 		INIT_WORK(&chap->auth_work, nvme_queue_auth_work);
972 	}
973 
974 	return 0;
975 err_free_dhchap_ctrl_secret:
976 	nvme_auth_free_key(ctrl->ctrl_key);
977 	ctrl->ctrl_key = NULL;
978 err_free_dhchap_secret:
979 	nvme_auth_free_key(ctrl->host_key);
980 	ctrl->host_key = NULL;
981 	return ret;
982 }
983 EXPORT_SYMBOL_GPL(nvme_auth_init_ctrl);
984 
985 void nvme_auth_stop(struct nvme_ctrl *ctrl)
986 {
987 	cancel_work_sync(&ctrl->dhchap_auth_work);
988 }
989 EXPORT_SYMBOL_GPL(nvme_auth_stop);
990 
991 void nvme_auth_free(struct nvme_ctrl *ctrl)
992 {
993 	int i;
994 
995 	if (ctrl->dhchap_ctxs) {
996 		for (i = 0; i < ctrl_max_dhchaps(ctrl); i++)
997 			nvme_auth_free_dhchap(&ctrl->dhchap_ctxs[i]);
998 		kfree(ctrl->dhchap_ctxs);
999 	}
1000 	if (ctrl->host_key) {
1001 		nvme_auth_free_key(ctrl->host_key);
1002 		ctrl->host_key = NULL;
1003 	}
1004 	if (ctrl->ctrl_key) {
1005 		nvme_auth_free_key(ctrl->ctrl_key);
1006 		ctrl->ctrl_key = NULL;
1007 	}
1008 }
1009 EXPORT_SYMBOL_GPL(nvme_auth_free);
1010 
1011 int __init nvme_init_auth(void)
1012 {
1013 	nvme_auth_wq = alloc_workqueue("nvme-auth-wq",
1014 			       WQ_UNBOUND | WQ_MEM_RECLAIM | WQ_SYSFS, 0);
1015 	if (!nvme_auth_wq)
1016 		return -ENOMEM;
1017 
1018 	nvme_chap_buf_cache = kmem_cache_create("nvme-chap-buf-cache",
1019 				CHAP_BUF_SIZE, 0, SLAB_HWCACHE_ALIGN, NULL);
1020 	if (!nvme_chap_buf_cache)
1021 		goto err_destroy_workqueue;
1022 
1023 	nvme_chap_buf_pool = mempool_create(16, mempool_alloc_slab,
1024 			mempool_free_slab, nvme_chap_buf_cache);
1025 	if (!nvme_chap_buf_pool)
1026 		goto err_destroy_chap_buf_cache;
1027 
1028 	return 0;
1029 err_destroy_chap_buf_cache:
1030 	kmem_cache_destroy(nvme_chap_buf_cache);
1031 err_destroy_workqueue:
1032 	destroy_workqueue(nvme_auth_wq);
1033 	return -ENOMEM;
1034 }
1035 
1036 void __exit nvme_exit_auth(void)
1037 {
1038 	mempool_destroy(nvme_chap_buf_pool);
1039 	kmem_cache_destroy(nvme_chap_buf_cache);
1040 	destroy_workqueue(nvme_auth_wq);
1041 }
1042