1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * sun4i-ss-cipher.c - hardware cryptographic accelerator for Allwinner A20 SoC 4 * 5 * Copyright (C) 2013-2015 Corentin LABBE <clabbe.montjoie@gmail.com> 6 * 7 * This file add support for AES cipher with 128,192,256 bits 8 * keysize in CBC and ECB mode. 9 * Add support also for DES and 3DES in CBC and ECB mode. 10 * 11 * You could find the datasheet in Documentation/arch/arm/sunxi.rst 12 */ 13 #include "sun4i-ss.h" 14 15 static int noinline_for_stack sun4i_ss_opti_poll(struct skcipher_request *areq) 16 { 17 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 18 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 19 struct sun4i_ss_ctx *ss = op->ss; 20 unsigned int ivsize = crypto_skcipher_ivsize(tfm); 21 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq); 22 u32 mode = ctx->mode; 23 /* when activating SS, the default FIFO space is SS_RX_DEFAULT(32) */ 24 u32 rx_cnt = SS_RX_DEFAULT; 25 u32 tx_cnt = 0; 26 u32 spaces; 27 u32 v; 28 int err = 0; 29 unsigned int i; 30 unsigned int ileft = areq->cryptlen; 31 unsigned int oleft = areq->cryptlen; 32 unsigned int todo; 33 unsigned long pi = 0, po = 0; /* progress for in and out */ 34 bool miter_err; 35 struct sg_mapping_iter mi, mo; 36 unsigned int oi, oo; /* offset for in and out */ 37 unsigned long flags; 38 struct skcipher_alg *alg = crypto_skcipher_alg(tfm); 39 struct sun4i_ss_alg_template *algt; 40 41 if (!areq->cryptlen) 42 return 0; 43 44 if (!areq->src || !areq->dst) { 45 dev_err_ratelimited(ss->dev, "ERROR: Some SGs are NULL\n"); 46 return -EINVAL; 47 } 48 49 if (areq->iv && ivsize > 0 && mode & SS_DECRYPTION) { 50 scatterwalk_map_and_copy(ctx->backup_iv, areq->src, 51 areq->cryptlen - ivsize, ivsize, 0); 52 } 53 54 if (IS_ENABLED(CONFIG_CRYPTO_DEV_SUN4I_SS_DEBUG)) { 55 algt = container_of(alg, struct sun4i_ss_alg_template, alg.crypto); 56 algt->stat_opti++; 57 algt->stat_bytes += areq->cryptlen; 58 } 59 60 spin_lock_irqsave(&ss->slock, flags); 61 62 for (i = 0; i < op->keylen / 4; i++) 63 writesl(ss->base + SS_KEY0 + i * 4, &op->key[i], 1); 64 65 if (areq->iv) { 66 for (i = 0; i < 4 && i < ivsize / 4; i++) { 67 v = *(u32 *)(areq->iv + i * 4); 68 writesl(ss->base + SS_IV0 + i * 4, &v, 1); 69 } 70 } 71 writel(mode, ss->base + SS_CTL); 72 73 74 ileft = areq->cryptlen / 4; 75 oleft = areq->cryptlen / 4; 76 oi = 0; 77 oo = 0; 78 do { 79 if (ileft) { 80 sg_miter_start(&mi, areq->src, sg_nents(areq->src), 81 SG_MITER_FROM_SG | SG_MITER_ATOMIC); 82 if (pi) 83 sg_miter_skip(&mi, pi); 84 miter_err = sg_miter_next(&mi); 85 if (!miter_err || !mi.addr) { 86 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n"); 87 err = -EINVAL; 88 goto release_ss; 89 } 90 todo = min(rx_cnt, ileft); 91 todo = min_t(size_t, todo, (mi.length - oi) / 4); 92 if (todo) { 93 ileft -= todo; 94 writesl(ss->base + SS_RXFIFO, mi.addr + oi, todo); 95 oi += todo * 4; 96 } 97 if (oi == mi.length) { 98 pi += mi.length; 99 oi = 0; 100 } 101 sg_miter_stop(&mi); 102 } 103 104 spaces = readl(ss->base + SS_FCSR); 105 rx_cnt = SS_RXFIFO_SPACES(spaces); 106 tx_cnt = SS_TXFIFO_SPACES(spaces); 107 108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), 109 SG_MITER_TO_SG | SG_MITER_ATOMIC); 110 if (po) 111 sg_miter_skip(&mo, po); 112 miter_err = sg_miter_next(&mo); 113 if (!miter_err || !mo.addr) { 114 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n"); 115 err = -EINVAL; 116 goto release_ss; 117 } 118 todo = min(tx_cnt, oleft); 119 todo = min_t(size_t, todo, (mo.length - oo) / 4); 120 if (todo) { 121 oleft -= todo; 122 readsl(ss->base + SS_TXFIFO, mo.addr + oo, todo); 123 oo += todo * 4; 124 } 125 if (oo == mo.length) { 126 oo = 0; 127 po += mo.length; 128 } 129 sg_miter_stop(&mo); 130 } while (oleft); 131 132 if (areq->iv) { 133 if (mode & SS_DECRYPTION) { 134 memcpy(areq->iv, ctx->backup_iv, ivsize); 135 memzero_explicit(ctx->backup_iv, ivsize); 136 } else { 137 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize, 138 ivsize, 0); 139 } 140 } 141 142 release_ss: 143 writel(0, ss->base + SS_CTL); 144 spin_unlock_irqrestore(&ss->slock, flags); 145 return err; 146 } 147 148 static int noinline_for_stack sun4i_ss_cipher_poll_fallback(struct skcipher_request *areq) 149 { 150 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 151 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 152 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq); 153 int err; 154 struct skcipher_alg *alg = crypto_skcipher_alg(tfm); 155 struct sun4i_ss_alg_template *algt; 156 157 if (IS_ENABLED(CONFIG_CRYPTO_DEV_SUN4I_SS_DEBUG)) { 158 algt = container_of(alg, struct sun4i_ss_alg_template, alg.crypto); 159 algt->stat_fb++; 160 } 161 162 skcipher_request_set_tfm(&ctx->fallback_req, op->fallback_tfm); 163 skcipher_request_set_callback(&ctx->fallback_req, areq->base.flags, 164 areq->base.complete, areq->base.data); 165 skcipher_request_set_crypt(&ctx->fallback_req, areq->src, areq->dst, 166 areq->cryptlen, areq->iv); 167 if (ctx->mode & SS_DECRYPTION) 168 err = crypto_skcipher_decrypt(&ctx->fallback_req); 169 else 170 err = crypto_skcipher_encrypt(&ctx->fallback_req); 171 172 return err; 173 } 174 175 /* Generic function that support SG with size not multiple of 4 */ 176 static int sun4i_ss_cipher_poll(struct skcipher_request *areq) 177 { 178 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 179 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 180 struct sun4i_ss_ctx *ss = op->ss; 181 int no_chunk = 1; 182 struct scatterlist *in_sg = areq->src; 183 struct scatterlist *out_sg = areq->dst; 184 unsigned int ivsize = crypto_skcipher_ivsize(tfm); 185 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq); 186 struct skcipher_alg *alg = crypto_skcipher_alg(tfm); 187 struct sun4i_ss_alg_template *algt; 188 u32 mode = ctx->mode; 189 /* when activating SS, the default FIFO space is SS_RX_DEFAULT(32) */ 190 u32 rx_cnt = SS_RX_DEFAULT; 191 u32 tx_cnt = 0; 192 u32 v; 193 u32 spaces; 194 int err = 0; 195 unsigned int i; 196 unsigned int ileft = areq->cryptlen; 197 unsigned int oleft = areq->cryptlen; 198 unsigned int todo; 199 struct sg_mapping_iter mi, mo; 200 unsigned long pi = 0, po = 0; /* progress for in and out */ 201 bool miter_err; 202 unsigned int oi, oo; /* offset for in and out */ 203 unsigned int ob = 0; /* offset in buf */ 204 unsigned int obo = 0; /* offset in bufo*/ 205 unsigned int obl = 0; /* length of data in bufo */ 206 unsigned long flags; 207 bool need_fallback = false; 208 209 if (!areq->cryptlen) 210 return 0; 211 212 if (!areq->src || !areq->dst) { 213 dev_err_ratelimited(ss->dev, "ERROR: Some SGs are NULL\n"); 214 return -EINVAL; 215 } 216 217 algt = container_of(alg, struct sun4i_ss_alg_template, alg.crypto); 218 if (areq->cryptlen % algt->alg.crypto.base.cra_blocksize) 219 need_fallback = true; 220 221 /* 222 * if we have only SGs with size multiple of 4, 223 * we can use the SS optimized function 224 */ 225 while (in_sg && no_chunk == 1) { 226 if ((in_sg->length | in_sg->offset) & 3u) 227 no_chunk = 0; 228 in_sg = sg_next(in_sg); 229 } 230 while (out_sg && no_chunk == 1) { 231 if ((out_sg->length | out_sg->offset) & 3u) 232 no_chunk = 0; 233 out_sg = sg_next(out_sg); 234 } 235 236 if (no_chunk == 1 && !need_fallback) 237 return sun4i_ss_opti_poll(areq); 238 239 if (need_fallback) 240 return sun4i_ss_cipher_poll_fallback(areq); 241 242 if (areq->iv && ivsize > 0 && mode & SS_DECRYPTION) { 243 scatterwalk_map_and_copy(ctx->backup_iv, areq->src, 244 areq->cryptlen - ivsize, ivsize, 0); 245 } 246 247 if (IS_ENABLED(CONFIG_CRYPTO_DEV_SUN4I_SS_DEBUG)) { 248 algt->stat_req++; 249 algt->stat_bytes += areq->cryptlen; 250 } 251 252 spin_lock_irqsave(&ss->slock, flags); 253 254 for (i = 0; i < op->keylen / 4; i++) 255 writesl(ss->base + SS_KEY0 + i * 4, &op->key[i], 1); 256 257 if (areq->iv) { 258 for (i = 0; i < 4 && i < ivsize / 4; i++) { 259 v = *(u32 *)(areq->iv + i * 4); 260 writesl(ss->base + SS_IV0 + i * 4, &v, 1); 261 } 262 } 263 writel(mode, ss->base + SS_CTL); 264 265 ileft = areq->cryptlen; 266 oleft = areq->cryptlen; 267 oi = 0; 268 oo = 0; 269 270 while (oleft) { 271 if (ileft) { 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), 273 SG_MITER_FROM_SG | SG_MITER_ATOMIC); 274 if (pi) 275 sg_miter_skip(&mi, pi); 276 miter_err = sg_miter_next(&mi); 277 if (!miter_err || !mi.addr) { 278 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n"); 279 err = -EINVAL; 280 goto release_ss; 281 } 282 /* 283 * todo is the number of consecutive 4byte word that we 284 * can read from current SG 285 */ 286 todo = min(rx_cnt, ileft / 4); 287 todo = min_t(size_t, todo, (mi.length - oi) / 4); 288 if (todo && !ob) { 289 writesl(ss->base + SS_RXFIFO, mi.addr + oi, 290 todo); 291 ileft -= todo * 4; 292 oi += todo * 4; 293 } else { 294 /* 295 * not enough consecutive bytes, so we need to 296 * linearize in buf. todo is in bytes 297 * After that copy, if we have a multiple of 4 298 * we need to be able to write all buf in one 299 * pass, so it is why we min() with rx_cnt 300 */ 301 todo = min(rx_cnt * 4 - ob, ileft); 302 todo = min_t(size_t, todo, mi.length - oi); 303 memcpy(ss->buf + ob, mi.addr + oi, todo); 304 ileft -= todo; 305 oi += todo; 306 ob += todo; 307 if (!(ob % 4)) { 308 writesl(ss->base + SS_RXFIFO, ss->buf, 309 ob / 4); 310 ob = 0; 311 } 312 } 313 if (oi == mi.length) { 314 pi += mi.length; 315 oi = 0; 316 } 317 sg_miter_stop(&mi); 318 } 319 320 spaces = readl(ss->base + SS_FCSR); 321 rx_cnt = SS_RXFIFO_SPACES(spaces); 322 tx_cnt = SS_TXFIFO_SPACES(spaces); 323 324 if (!tx_cnt) 325 continue; 326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), 327 SG_MITER_TO_SG | SG_MITER_ATOMIC); 328 if (po) 329 sg_miter_skip(&mo, po); 330 miter_err = sg_miter_next(&mo); 331 if (!miter_err || !mo.addr) { 332 dev_err_ratelimited(ss->dev, "ERROR: sg_miter return null\n"); 333 err = -EINVAL; 334 goto release_ss; 335 } 336 /* todo in 4bytes word */ 337 todo = min(tx_cnt, oleft / 4); 338 todo = min_t(size_t, todo, (mo.length - oo) / 4); 339 340 if (todo) { 341 readsl(ss->base + SS_TXFIFO, mo.addr + oo, todo); 342 oleft -= todo * 4; 343 oo += todo * 4; 344 if (oo == mo.length) { 345 po += mo.length; 346 oo = 0; 347 } 348 } else { 349 /* 350 * read obl bytes in bufo, we read at maximum for 351 * emptying the device 352 */ 353 readsl(ss->base + SS_TXFIFO, ss->bufo, tx_cnt); 354 obl = tx_cnt * 4; 355 obo = 0; 356 do { 357 /* 358 * how many bytes we can copy ? 359 * no more than remaining SG size 360 * no more than remaining buffer 361 * no need to test against oleft 362 */ 363 todo = min_t(size_t, 364 mo.length - oo, obl - obo); 365 memcpy(mo.addr + oo, ss->bufo + obo, todo); 366 oleft -= todo; 367 obo += todo; 368 oo += todo; 369 if (oo == mo.length) { 370 po += mo.length; 371 sg_miter_next(&mo); 372 oo = 0; 373 } 374 } while (obo < obl); 375 /* bufo must be fully used here */ 376 } 377 sg_miter_stop(&mo); 378 } 379 if (areq->iv) { 380 if (mode & SS_DECRYPTION) { 381 memcpy(areq->iv, ctx->backup_iv, ivsize); 382 memzero_explicit(ctx->backup_iv, ivsize); 383 } else { 384 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize, 385 ivsize, 0); 386 } 387 } 388 389 release_ss: 390 writel(0, ss->base + SS_CTL); 391 spin_unlock_irqrestore(&ss->slock, flags); 392 393 return err; 394 } 395 396 /* CBC AES */ 397 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq) 398 { 399 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 400 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 401 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 402 403 rctx->mode = SS_OP_AES | SS_CBC | SS_ENABLED | SS_ENCRYPTION | 404 op->keymode; 405 return sun4i_ss_cipher_poll(areq); 406 } 407 408 int sun4i_ss_cbc_aes_decrypt(struct skcipher_request *areq) 409 { 410 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 411 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 412 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 413 414 rctx->mode = SS_OP_AES | SS_CBC | SS_ENABLED | SS_DECRYPTION | 415 op->keymode; 416 return sun4i_ss_cipher_poll(areq); 417 } 418 419 /* ECB AES */ 420 int sun4i_ss_ecb_aes_encrypt(struct skcipher_request *areq) 421 { 422 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 423 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 424 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 425 426 rctx->mode = SS_OP_AES | SS_ECB | SS_ENABLED | SS_ENCRYPTION | 427 op->keymode; 428 return sun4i_ss_cipher_poll(areq); 429 } 430 431 int sun4i_ss_ecb_aes_decrypt(struct skcipher_request *areq) 432 { 433 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 434 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 435 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 436 437 rctx->mode = SS_OP_AES | SS_ECB | SS_ENABLED | SS_DECRYPTION | 438 op->keymode; 439 return sun4i_ss_cipher_poll(areq); 440 } 441 442 /* CBC DES */ 443 int sun4i_ss_cbc_des_encrypt(struct skcipher_request *areq) 444 { 445 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 446 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 447 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 448 449 rctx->mode = SS_OP_DES | SS_CBC | SS_ENABLED | SS_ENCRYPTION | 450 op->keymode; 451 return sun4i_ss_cipher_poll(areq); 452 } 453 454 int sun4i_ss_cbc_des_decrypt(struct skcipher_request *areq) 455 { 456 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 457 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 458 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 459 460 rctx->mode = SS_OP_DES | SS_CBC | SS_ENABLED | SS_DECRYPTION | 461 op->keymode; 462 return sun4i_ss_cipher_poll(areq); 463 } 464 465 /* ECB DES */ 466 int sun4i_ss_ecb_des_encrypt(struct skcipher_request *areq) 467 { 468 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 469 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 470 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 471 472 rctx->mode = SS_OP_DES | SS_ECB | SS_ENABLED | SS_ENCRYPTION | 473 op->keymode; 474 return sun4i_ss_cipher_poll(areq); 475 } 476 477 int sun4i_ss_ecb_des_decrypt(struct skcipher_request *areq) 478 { 479 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 480 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 481 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 482 483 rctx->mode = SS_OP_DES | SS_ECB | SS_ENABLED | SS_DECRYPTION | 484 op->keymode; 485 return sun4i_ss_cipher_poll(areq); 486 } 487 488 /* CBC 3DES */ 489 int sun4i_ss_cbc_des3_encrypt(struct skcipher_request *areq) 490 { 491 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 492 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 493 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 494 495 rctx->mode = SS_OP_3DES | SS_CBC | SS_ENABLED | SS_ENCRYPTION | 496 op->keymode; 497 return sun4i_ss_cipher_poll(areq); 498 } 499 500 int sun4i_ss_cbc_des3_decrypt(struct skcipher_request *areq) 501 { 502 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 503 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 504 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 505 506 rctx->mode = SS_OP_3DES | SS_CBC | SS_ENABLED | SS_DECRYPTION | 507 op->keymode; 508 return sun4i_ss_cipher_poll(areq); 509 } 510 511 /* ECB 3DES */ 512 int sun4i_ss_ecb_des3_encrypt(struct skcipher_request *areq) 513 { 514 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 515 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 516 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 517 518 rctx->mode = SS_OP_3DES | SS_ECB | SS_ENABLED | SS_ENCRYPTION | 519 op->keymode; 520 return sun4i_ss_cipher_poll(areq); 521 } 522 523 int sun4i_ss_ecb_des3_decrypt(struct skcipher_request *areq) 524 { 525 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); 526 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 527 struct sun4i_cipher_req_ctx *rctx = skcipher_request_ctx(areq); 528 529 rctx->mode = SS_OP_3DES | SS_ECB | SS_ENABLED | SS_DECRYPTION | 530 op->keymode; 531 return sun4i_ss_cipher_poll(areq); 532 } 533 534 int sun4i_ss_cipher_init(struct crypto_tfm *tfm) 535 { 536 struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm); 537 struct sun4i_ss_alg_template *algt; 538 const char *name = crypto_tfm_alg_name(tfm); 539 int err; 540 541 memset(op, 0, sizeof(struct sun4i_tfm_ctx)); 542 543 algt = container_of(tfm->__crt_alg, struct sun4i_ss_alg_template, 544 alg.crypto.base); 545 op->ss = algt->ss; 546 547 op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK); 548 if (IS_ERR(op->fallback_tfm)) { 549 dev_err(op->ss->dev, "ERROR: Cannot allocate fallback for %s %ld\n", 550 name, PTR_ERR(op->fallback_tfm)); 551 return PTR_ERR(op->fallback_tfm); 552 } 553 554 crypto_skcipher_set_reqsize(__crypto_skcipher_cast(tfm), 555 sizeof(struct sun4i_cipher_req_ctx) + 556 crypto_skcipher_reqsize(op->fallback_tfm)); 557 558 err = pm_runtime_resume_and_get(op->ss->dev); 559 if (err < 0) 560 goto error_pm; 561 562 return 0; 563 error_pm: 564 crypto_free_skcipher(op->fallback_tfm); 565 return err; 566 } 567 568 void sun4i_ss_cipher_exit(struct crypto_tfm *tfm) 569 { 570 struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm); 571 572 crypto_free_skcipher(op->fallback_tfm); 573 pm_runtime_put(op->ss->dev); 574 } 575 576 /* check and set the AES key, prepare the mode to be used */ 577 int sun4i_ss_aes_setkey(struct crypto_skcipher *tfm, const u8 *key, 578 unsigned int keylen) 579 { 580 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 581 struct sun4i_ss_ctx *ss = op->ss; 582 583 switch (keylen) { 584 case 128 / 8: 585 op->keymode = SS_AES_128BITS; 586 break; 587 case 192 / 8: 588 op->keymode = SS_AES_192BITS; 589 break; 590 case 256 / 8: 591 op->keymode = SS_AES_256BITS; 592 break; 593 default: 594 dev_dbg(ss->dev, "ERROR: Invalid keylen %u\n", keylen); 595 return -EINVAL; 596 } 597 op->keylen = keylen; 598 memcpy(op->key, key, keylen); 599 600 crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK); 601 crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK); 602 603 return crypto_skcipher_setkey(op->fallback_tfm, key, keylen); 604 } 605 606 /* check and set the DES key, prepare the mode to be used */ 607 int sun4i_ss_des_setkey(struct crypto_skcipher *tfm, const u8 *key, 608 unsigned int keylen) 609 { 610 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 611 int err; 612 613 err = verify_skcipher_des_key(tfm, key); 614 if (err) 615 return err; 616 617 op->keylen = keylen; 618 memcpy(op->key, key, keylen); 619 620 crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK); 621 crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK); 622 623 return crypto_skcipher_setkey(op->fallback_tfm, key, keylen); 624 } 625 626 /* check and set the 3DES key, prepare the mode to be used */ 627 int sun4i_ss_des3_setkey(struct crypto_skcipher *tfm, const u8 *key, 628 unsigned int keylen) 629 { 630 struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm); 631 int err; 632 633 err = verify_skcipher_des3_key(tfm, key); 634 if (err) 635 return err; 636 637 op->keylen = keylen; 638 memcpy(op->key, key, keylen); 639 640 crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK); 641 crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK); 642 643 return crypto_skcipher_setkey(op->fallback_tfm, key, keylen); 644 } 645