1 /* 2 * Cryptographic API for algorithms (i.e., low-level API). 3 * 4 * Copyright (c) 2006 Herbert Xu <herbert@gondor.apana.org.au> 5 * 6 * This program is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License as published by the Free 8 * Software Foundation; either version 2 of the License, or (at your option) 9 * any later version. 10 * 11 */ 12 13 #include <linux/err.h> 14 #include <linux/errno.h> 15 #include <linux/fips.h> 16 #include <linux/init.h> 17 #include <linux/kernel.h> 18 #include <linux/list.h> 19 #include <linux/module.h> 20 #include <linux/rtnetlink.h> 21 #include <linux/slab.h> 22 #include <linux/string.h> 23 24 #include "internal.h" 25 26 static LIST_HEAD(crypto_template_list); 27 28 static inline int crypto_set_driver_name(struct crypto_alg *alg) 29 { 30 static const char suffix[] = "-generic"; 31 char *driver_name = alg->cra_driver_name; 32 int len; 33 34 if (*driver_name) 35 return 0; 36 37 len = strlcpy(driver_name, alg->cra_name, CRYPTO_MAX_ALG_NAME); 38 if (len + sizeof(suffix) > CRYPTO_MAX_ALG_NAME) 39 return -ENAMETOOLONG; 40 41 memcpy(driver_name + len, suffix, sizeof(suffix)); 42 return 0; 43 } 44 45 static inline void crypto_check_module_sig(struct module *mod) 46 { 47 if (fips_enabled && mod && !module_sig_ok(mod)) 48 panic("Module %s signature verification failed in FIPS mode\n", 49 module_name(mod)); 50 } 51 52 static int crypto_check_alg(struct crypto_alg *alg) 53 { 54 crypto_check_module_sig(alg->cra_module); 55 56 if (alg->cra_alignmask & (alg->cra_alignmask + 1)) 57 return -EINVAL; 58 59 if (alg->cra_blocksize > PAGE_SIZE / 8) 60 return -EINVAL; 61 62 if (alg->cra_priority < 0) 63 return -EINVAL; 64 65 refcount_set(&alg->cra_refcnt, 1); 66 67 return crypto_set_driver_name(alg); 68 } 69 70 static void crypto_free_instance(struct crypto_instance *inst) 71 { 72 if (!inst->alg.cra_type->free) { 73 inst->tmpl->free(inst); 74 return; 75 } 76 77 inst->alg.cra_type->free(inst); 78 } 79 80 static void crypto_destroy_instance(struct crypto_alg *alg) 81 { 82 struct crypto_instance *inst = (void *)alg; 83 struct crypto_template *tmpl = inst->tmpl; 84 85 crypto_free_instance(inst); 86 crypto_tmpl_put(tmpl); 87 } 88 89 static struct list_head *crypto_more_spawns(struct crypto_alg *alg, 90 struct list_head *stack, 91 struct list_head *top, 92 struct list_head *secondary_spawns) 93 { 94 struct crypto_spawn *spawn, *n; 95 96 spawn = list_first_entry_or_null(stack, struct crypto_spawn, list); 97 if (!spawn) 98 return NULL; 99 100 n = list_next_entry(spawn, list); 101 102 if (spawn->alg && &n->list != stack && !n->alg) 103 n->alg = (n->list.next == stack) ? alg : 104 &list_next_entry(n, list)->inst->alg; 105 106 list_move(&spawn->list, secondary_spawns); 107 108 return &n->list == stack ? top : &n->inst->alg.cra_users; 109 } 110 111 static void crypto_remove_instance(struct crypto_instance *inst, 112 struct list_head *list) 113 { 114 struct crypto_template *tmpl = inst->tmpl; 115 116 if (crypto_is_dead(&inst->alg)) 117 return; 118 119 inst->alg.cra_flags |= CRYPTO_ALG_DEAD; 120 if (hlist_unhashed(&inst->list)) 121 return; 122 123 if (!tmpl || !crypto_tmpl_get(tmpl)) 124 return; 125 126 list_move(&inst->alg.cra_list, list); 127 hlist_del(&inst->list); 128 inst->alg.cra_destroy = crypto_destroy_instance; 129 130 BUG_ON(!list_empty(&inst->alg.cra_users)); 131 } 132 133 void crypto_remove_spawns(struct crypto_alg *alg, struct list_head *list, 134 struct crypto_alg *nalg) 135 { 136 u32 new_type = (nalg ?: alg)->cra_flags; 137 struct crypto_spawn *spawn, *n; 138 LIST_HEAD(secondary_spawns); 139 struct list_head *spawns; 140 LIST_HEAD(stack); 141 LIST_HEAD(top); 142 143 spawns = &alg->cra_users; 144 list_for_each_entry_safe(spawn, n, spawns, list) { 145 if ((spawn->alg->cra_flags ^ new_type) & spawn->mask) 146 continue; 147 148 list_move(&spawn->list, &top); 149 } 150 151 spawns = ⊤ 152 do { 153 while (!list_empty(spawns)) { 154 struct crypto_instance *inst; 155 156 spawn = list_first_entry(spawns, struct crypto_spawn, 157 list); 158 inst = spawn->inst; 159 160 BUG_ON(&inst->alg == alg); 161 162 list_move(&spawn->list, &stack); 163 164 if (&inst->alg == nalg) 165 break; 166 167 spawn->alg = NULL; 168 spawns = &inst->alg.cra_users; 169 170 /* 171 * We may encounter an unregistered instance here, since 172 * an instance's spawns are set up prior to the instance 173 * being registered. An unregistered instance will have 174 * NULL ->cra_users.next, since ->cra_users isn't 175 * properly initialized until registration. But an 176 * unregistered instance cannot have any users, so treat 177 * it the same as ->cra_users being empty. 178 */ 179 if (spawns->next == NULL) 180 break; 181 } 182 } while ((spawns = crypto_more_spawns(alg, &stack, &top, 183 &secondary_spawns))); 184 185 list_for_each_entry_safe(spawn, n, &secondary_spawns, list) { 186 if (spawn->alg) 187 list_move(&spawn->list, &spawn->alg->cra_users); 188 else 189 crypto_remove_instance(spawn->inst, list); 190 } 191 } 192 EXPORT_SYMBOL_GPL(crypto_remove_spawns); 193 194 static struct crypto_larval *__crypto_register_alg(struct crypto_alg *alg) 195 { 196 struct crypto_alg *q; 197 struct crypto_larval *larval; 198 int ret = -EAGAIN; 199 200 if (crypto_is_dead(alg)) 201 goto err; 202 203 INIT_LIST_HEAD(&alg->cra_users); 204 205 /* No cheating! */ 206 alg->cra_flags &= ~CRYPTO_ALG_TESTED; 207 208 ret = -EEXIST; 209 210 list_for_each_entry(q, &crypto_alg_list, cra_list) { 211 if (q == alg) 212 goto err; 213 214 if (crypto_is_moribund(q)) 215 continue; 216 217 if (crypto_is_larval(q)) { 218 if (!strcmp(alg->cra_driver_name, q->cra_driver_name)) 219 goto err; 220 continue; 221 } 222 223 if (!strcmp(q->cra_driver_name, alg->cra_name) || 224 !strcmp(q->cra_name, alg->cra_driver_name)) 225 goto err; 226 } 227 228 larval = crypto_larval_alloc(alg->cra_name, 229 alg->cra_flags | CRYPTO_ALG_TESTED, 0); 230 if (IS_ERR(larval)) 231 goto out; 232 233 ret = -ENOENT; 234 larval->adult = crypto_mod_get(alg); 235 if (!larval->adult) 236 goto free_larval; 237 238 refcount_set(&larval->alg.cra_refcnt, 1); 239 memcpy(larval->alg.cra_driver_name, alg->cra_driver_name, 240 CRYPTO_MAX_ALG_NAME); 241 larval->alg.cra_priority = alg->cra_priority; 242 243 list_add(&alg->cra_list, &crypto_alg_list); 244 list_add(&larval->alg.cra_list, &crypto_alg_list); 245 246 out: 247 return larval; 248 249 free_larval: 250 kfree(larval); 251 err: 252 larval = ERR_PTR(ret); 253 goto out; 254 } 255 256 void crypto_alg_tested(const char *name, int err) 257 { 258 struct crypto_larval *test; 259 struct crypto_alg *alg; 260 struct crypto_alg *q; 261 LIST_HEAD(list); 262 263 down_write(&crypto_alg_sem); 264 list_for_each_entry(q, &crypto_alg_list, cra_list) { 265 if (crypto_is_moribund(q) || !crypto_is_larval(q)) 266 continue; 267 268 test = (struct crypto_larval *)q; 269 270 if (!strcmp(q->cra_driver_name, name)) 271 goto found; 272 } 273 274 pr_err("alg: Unexpected test result for %s: %d\n", name, err); 275 goto unlock; 276 277 found: 278 q->cra_flags |= CRYPTO_ALG_DEAD; 279 alg = test->adult; 280 if (err || list_empty(&alg->cra_list)) 281 goto complete; 282 283 alg->cra_flags |= CRYPTO_ALG_TESTED; 284 285 list_for_each_entry(q, &crypto_alg_list, cra_list) { 286 if (q == alg) 287 continue; 288 289 if (crypto_is_moribund(q)) 290 continue; 291 292 if (crypto_is_larval(q)) { 293 struct crypto_larval *larval = (void *)q; 294 295 /* 296 * Check to see if either our generic name or 297 * specific name can satisfy the name requested 298 * by the larval entry q. 299 */ 300 if (strcmp(alg->cra_name, q->cra_name) && 301 strcmp(alg->cra_driver_name, q->cra_name)) 302 continue; 303 304 if (larval->adult) 305 continue; 306 if ((q->cra_flags ^ alg->cra_flags) & larval->mask) 307 continue; 308 if (!crypto_mod_get(alg)) 309 continue; 310 311 larval->adult = alg; 312 continue; 313 } 314 315 if (strcmp(alg->cra_name, q->cra_name)) 316 continue; 317 318 if (strcmp(alg->cra_driver_name, q->cra_driver_name) && 319 q->cra_priority > alg->cra_priority) 320 continue; 321 322 crypto_remove_spawns(q, &list, alg); 323 } 324 325 complete: 326 complete_all(&test->completion); 327 328 unlock: 329 up_write(&crypto_alg_sem); 330 331 crypto_remove_final(&list); 332 } 333 EXPORT_SYMBOL_GPL(crypto_alg_tested); 334 335 void crypto_remove_final(struct list_head *list) 336 { 337 struct crypto_alg *alg; 338 struct crypto_alg *n; 339 340 list_for_each_entry_safe(alg, n, list, cra_list) { 341 list_del_init(&alg->cra_list); 342 crypto_alg_put(alg); 343 } 344 } 345 EXPORT_SYMBOL_GPL(crypto_remove_final); 346 347 static void crypto_wait_for_test(struct crypto_larval *larval) 348 { 349 int err; 350 351 err = crypto_probing_notify(CRYPTO_MSG_ALG_REGISTER, larval->adult); 352 if (err != NOTIFY_STOP) { 353 if (WARN_ON(err != NOTIFY_DONE)) 354 goto out; 355 crypto_alg_tested(larval->alg.cra_driver_name, 0); 356 } 357 358 err = wait_for_completion_killable(&larval->completion); 359 WARN_ON(err); 360 361 out: 362 crypto_larval_kill(&larval->alg); 363 } 364 365 int crypto_register_alg(struct crypto_alg *alg) 366 { 367 struct crypto_larval *larval; 368 int err; 369 370 alg->cra_flags &= ~CRYPTO_ALG_DEAD; 371 err = crypto_check_alg(alg); 372 if (err) 373 return err; 374 375 down_write(&crypto_alg_sem); 376 larval = __crypto_register_alg(alg); 377 up_write(&crypto_alg_sem); 378 379 if (IS_ERR(larval)) 380 return PTR_ERR(larval); 381 382 crypto_wait_for_test(larval); 383 return 0; 384 } 385 EXPORT_SYMBOL_GPL(crypto_register_alg); 386 387 static int crypto_remove_alg(struct crypto_alg *alg, struct list_head *list) 388 { 389 if (unlikely(list_empty(&alg->cra_list))) 390 return -ENOENT; 391 392 alg->cra_flags |= CRYPTO_ALG_DEAD; 393 394 list_del_init(&alg->cra_list); 395 crypto_remove_spawns(alg, list, NULL); 396 397 return 0; 398 } 399 400 int crypto_unregister_alg(struct crypto_alg *alg) 401 { 402 int ret; 403 LIST_HEAD(list); 404 405 down_write(&crypto_alg_sem); 406 ret = crypto_remove_alg(alg, &list); 407 up_write(&crypto_alg_sem); 408 409 if (ret) 410 return ret; 411 412 BUG_ON(refcount_read(&alg->cra_refcnt) != 1); 413 if (alg->cra_destroy) 414 alg->cra_destroy(alg); 415 416 crypto_remove_final(&list); 417 return 0; 418 } 419 EXPORT_SYMBOL_GPL(crypto_unregister_alg); 420 421 int crypto_register_algs(struct crypto_alg *algs, int count) 422 { 423 int i, ret; 424 425 for (i = 0; i < count; i++) { 426 ret = crypto_register_alg(&algs[i]); 427 if (ret) 428 goto err; 429 } 430 431 return 0; 432 433 err: 434 for (--i; i >= 0; --i) 435 crypto_unregister_alg(&algs[i]); 436 437 return ret; 438 } 439 EXPORT_SYMBOL_GPL(crypto_register_algs); 440 441 int crypto_unregister_algs(struct crypto_alg *algs, int count) 442 { 443 int i, ret; 444 445 for (i = 0; i < count; i++) { 446 ret = crypto_unregister_alg(&algs[i]); 447 if (ret) 448 pr_err("Failed to unregister %s %s: %d\n", 449 algs[i].cra_driver_name, algs[i].cra_name, ret); 450 } 451 452 return 0; 453 } 454 EXPORT_SYMBOL_GPL(crypto_unregister_algs); 455 456 int crypto_register_template(struct crypto_template *tmpl) 457 { 458 struct crypto_template *q; 459 int err = -EEXIST; 460 461 down_write(&crypto_alg_sem); 462 463 crypto_check_module_sig(tmpl->module); 464 465 list_for_each_entry(q, &crypto_template_list, list) { 466 if (q == tmpl) 467 goto out; 468 } 469 470 list_add(&tmpl->list, &crypto_template_list); 471 err = 0; 472 out: 473 up_write(&crypto_alg_sem); 474 return err; 475 } 476 EXPORT_SYMBOL_GPL(crypto_register_template); 477 478 void crypto_unregister_template(struct crypto_template *tmpl) 479 { 480 struct crypto_instance *inst; 481 struct hlist_node *n; 482 struct hlist_head *list; 483 LIST_HEAD(users); 484 485 down_write(&crypto_alg_sem); 486 487 BUG_ON(list_empty(&tmpl->list)); 488 list_del_init(&tmpl->list); 489 490 list = &tmpl->instances; 491 hlist_for_each_entry(inst, list, list) { 492 int err = crypto_remove_alg(&inst->alg, &users); 493 494 BUG_ON(err); 495 } 496 497 up_write(&crypto_alg_sem); 498 499 hlist_for_each_entry_safe(inst, n, list, list) { 500 BUG_ON(refcount_read(&inst->alg.cra_refcnt) != 1); 501 crypto_free_instance(inst); 502 } 503 crypto_remove_final(&users); 504 } 505 EXPORT_SYMBOL_GPL(crypto_unregister_template); 506 507 static struct crypto_template *__crypto_lookup_template(const char *name) 508 { 509 struct crypto_template *q, *tmpl = NULL; 510 511 down_read(&crypto_alg_sem); 512 list_for_each_entry(q, &crypto_template_list, list) { 513 if (strcmp(q->name, name)) 514 continue; 515 if (unlikely(!crypto_tmpl_get(q))) 516 continue; 517 518 tmpl = q; 519 break; 520 } 521 up_read(&crypto_alg_sem); 522 523 return tmpl; 524 } 525 526 struct crypto_template *crypto_lookup_template(const char *name) 527 { 528 return try_then_request_module(__crypto_lookup_template(name), 529 "crypto-%s", name); 530 } 531 EXPORT_SYMBOL_GPL(crypto_lookup_template); 532 533 int crypto_register_instance(struct crypto_template *tmpl, 534 struct crypto_instance *inst) 535 { 536 struct crypto_larval *larval; 537 int err; 538 539 err = crypto_check_alg(&inst->alg); 540 if (err) 541 return err; 542 543 inst->alg.cra_module = tmpl->module; 544 inst->alg.cra_flags |= CRYPTO_ALG_INSTANCE; 545 546 if (unlikely(!crypto_mod_get(&inst->alg))) 547 return -EAGAIN; 548 549 down_write(&crypto_alg_sem); 550 551 larval = __crypto_register_alg(&inst->alg); 552 if (IS_ERR(larval)) 553 goto unlock; 554 555 hlist_add_head(&inst->list, &tmpl->instances); 556 inst->tmpl = tmpl; 557 558 unlock: 559 up_write(&crypto_alg_sem); 560 561 err = PTR_ERR(larval); 562 if (IS_ERR(larval)) 563 goto err; 564 565 crypto_wait_for_test(larval); 566 567 /* Remove instance if test failed */ 568 if (!(inst->alg.cra_flags & CRYPTO_ALG_TESTED)) 569 crypto_unregister_instance(inst); 570 err = 0; 571 572 err: 573 crypto_mod_put(&inst->alg); 574 return err; 575 } 576 EXPORT_SYMBOL_GPL(crypto_register_instance); 577 578 int crypto_unregister_instance(struct crypto_instance *inst) 579 { 580 LIST_HEAD(list); 581 582 down_write(&crypto_alg_sem); 583 584 crypto_remove_spawns(&inst->alg, &list, NULL); 585 crypto_remove_instance(inst, &list); 586 587 up_write(&crypto_alg_sem); 588 589 crypto_remove_final(&list); 590 591 return 0; 592 } 593 EXPORT_SYMBOL_GPL(crypto_unregister_instance); 594 595 int crypto_init_spawn(struct crypto_spawn *spawn, struct crypto_alg *alg, 596 struct crypto_instance *inst, u32 mask) 597 { 598 int err = -EAGAIN; 599 600 spawn->inst = inst; 601 spawn->mask = mask; 602 603 down_write(&crypto_alg_sem); 604 if (!crypto_is_moribund(alg)) { 605 list_add(&spawn->list, &alg->cra_users); 606 spawn->alg = alg; 607 err = 0; 608 } 609 up_write(&crypto_alg_sem); 610 611 return err; 612 } 613 EXPORT_SYMBOL_GPL(crypto_init_spawn); 614 615 int crypto_init_spawn2(struct crypto_spawn *spawn, struct crypto_alg *alg, 616 struct crypto_instance *inst, 617 const struct crypto_type *frontend) 618 { 619 int err = -EINVAL; 620 621 if ((alg->cra_flags ^ frontend->type) & frontend->maskset) 622 goto out; 623 624 spawn->frontend = frontend; 625 err = crypto_init_spawn(spawn, alg, inst, frontend->maskset); 626 627 out: 628 return err; 629 } 630 EXPORT_SYMBOL_GPL(crypto_init_spawn2); 631 632 int crypto_grab_spawn(struct crypto_spawn *spawn, const char *name, 633 u32 type, u32 mask) 634 { 635 struct crypto_alg *alg; 636 int err; 637 638 alg = crypto_find_alg(name, spawn->frontend, type, mask); 639 if (IS_ERR(alg)) 640 return PTR_ERR(alg); 641 642 err = crypto_init_spawn(spawn, alg, spawn->inst, mask); 643 crypto_mod_put(alg); 644 return err; 645 } 646 EXPORT_SYMBOL_GPL(crypto_grab_spawn); 647 648 void crypto_drop_spawn(struct crypto_spawn *spawn) 649 { 650 if (!spawn->alg) 651 return; 652 653 down_write(&crypto_alg_sem); 654 list_del(&spawn->list); 655 up_write(&crypto_alg_sem); 656 } 657 EXPORT_SYMBOL_GPL(crypto_drop_spawn); 658 659 static struct crypto_alg *crypto_spawn_alg(struct crypto_spawn *spawn) 660 { 661 struct crypto_alg *alg; 662 struct crypto_alg *alg2; 663 664 down_read(&crypto_alg_sem); 665 alg = spawn->alg; 666 alg2 = alg; 667 if (alg2) 668 alg2 = crypto_mod_get(alg2); 669 up_read(&crypto_alg_sem); 670 671 if (!alg2) { 672 if (alg) 673 crypto_shoot_alg(alg); 674 return ERR_PTR(-EAGAIN); 675 } 676 677 return alg; 678 } 679 680 struct crypto_tfm *crypto_spawn_tfm(struct crypto_spawn *spawn, u32 type, 681 u32 mask) 682 { 683 struct crypto_alg *alg; 684 struct crypto_tfm *tfm; 685 686 alg = crypto_spawn_alg(spawn); 687 if (IS_ERR(alg)) 688 return ERR_CAST(alg); 689 690 tfm = ERR_PTR(-EINVAL); 691 if (unlikely((alg->cra_flags ^ type) & mask)) 692 goto out_put_alg; 693 694 tfm = __crypto_alloc_tfm(alg, type, mask); 695 if (IS_ERR(tfm)) 696 goto out_put_alg; 697 698 return tfm; 699 700 out_put_alg: 701 crypto_mod_put(alg); 702 return tfm; 703 } 704 EXPORT_SYMBOL_GPL(crypto_spawn_tfm); 705 706 void *crypto_spawn_tfm2(struct crypto_spawn *spawn) 707 { 708 struct crypto_alg *alg; 709 struct crypto_tfm *tfm; 710 711 alg = crypto_spawn_alg(spawn); 712 if (IS_ERR(alg)) 713 return ERR_CAST(alg); 714 715 tfm = crypto_create_tfm(alg, spawn->frontend); 716 if (IS_ERR(tfm)) 717 goto out_put_alg; 718 719 return tfm; 720 721 out_put_alg: 722 crypto_mod_put(alg); 723 return tfm; 724 } 725 EXPORT_SYMBOL_GPL(crypto_spawn_tfm2); 726 727 int crypto_register_notifier(struct notifier_block *nb) 728 { 729 return blocking_notifier_chain_register(&crypto_chain, nb); 730 } 731 EXPORT_SYMBOL_GPL(crypto_register_notifier); 732 733 int crypto_unregister_notifier(struct notifier_block *nb) 734 { 735 return blocking_notifier_chain_unregister(&crypto_chain, nb); 736 } 737 EXPORT_SYMBOL_GPL(crypto_unregister_notifier); 738 739 struct crypto_attr_type *crypto_get_attr_type(struct rtattr **tb) 740 { 741 struct rtattr *rta = tb[0]; 742 struct crypto_attr_type *algt; 743 744 if (!rta) 745 return ERR_PTR(-ENOENT); 746 if (RTA_PAYLOAD(rta) < sizeof(*algt)) 747 return ERR_PTR(-EINVAL); 748 if (rta->rta_type != CRYPTOA_TYPE) 749 return ERR_PTR(-EINVAL); 750 751 algt = RTA_DATA(rta); 752 753 return algt; 754 } 755 EXPORT_SYMBOL_GPL(crypto_get_attr_type); 756 757 int crypto_check_attr_type(struct rtattr **tb, u32 type) 758 { 759 struct crypto_attr_type *algt; 760 761 algt = crypto_get_attr_type(tb); 762 if (IS_ERR(algt)) 763 return PTR_ERR(algt); 764 765 if ((algt->type ^ type) & algt->mask) 766 return -EINVAL; 767 768 return 0; 769 } 770 EXPORT_SYMBOL_GPL(crypto_check_attr_type); 771 772 const char *crypto_attr_alg_name(struct rtattr *rta) 773 { 774 struct crypto_attr_alg *alga; 775 776 if (!rta) 777 return ERR_PTR(-ENOENT); 778 if (RTA_PAYLOAD(rta) < sizeof(*alga)) 779 return ERR_PTR(-EINVAL); 780 if (rta->rta_type != CRYPTOA_ALG) 781 return ERR_PTR(-EINVAL); 782 783 alga = RTA_DATA(rta); 784 alga->name[CRYPTO_MAX_ALG_NAME - 1] = 0; 785 786 return alga->name; 787 } 788 EXPORT_SYMBOL_GPL(crypto_attr_alg_name); 789 790 struct crypto_alg *crypto_attr_alg2(struct rtattr *rta, 791 const struct crypto_type *frontend, 792 u32 type, u32 mask) 793 { 794 const char *name; 795 796 name = crypto_attr_alg_name(rta); 797 if (IS_ERR(name)) 798 return ERR_CAST(name); 799 800 return crypto_find_alg(name, frontend, type, mask); 801 } 802 EXPORT_SYMBOL_GPL(crypto_attr_alg2); 803 804 int crypto_attr_u32(struct rtattr *rta, u32 *num) 805 { 806 struct crypto_attr_u32 *nu32; 807 808 if (!rta) 809 return -ENOENT; 810 if (RTA_PAYLOAD(rta) < sizeof(*nu32)) 811 return -EINVAL; 812 if (rta->rta_type != CRYPTOA_U32) 813 return -EINVAL; 814 815 nu32 = RTA_DATA(rta); 816 *num = nu32->num; 817 818 return 0; 819 } 820 EXPORT_SYMBOL_GPL(crypto_attr_u32); 821 822 int crypto_inst_setname(struct crypto_instance *inst, const char *name, 823 struct crypto_alg *alg) 824 { 825 if (snprintf(inst->alg.cra_name, CRYPTO_MAX_ALG_NAME, "%s(%s)", name, 826 alg->cra_name) >= CRYPTO_MAX_ALG_NAME) 827 return -ENAMETOOLONG; 828 829 if (snprintf(inst->alg.cra_driver_name, CRYPTO_MAX_ALG_NAME, "%s(%s)", 830 name, alg->cra_driver_name) >= CRYPTO_MAX_ALG_NAME) 831 return -ENAMETOOLONG; 832 833 return 0; 834 } 835 EXPORT_SYMBOL_GPL(crypto_inst_setname); 836 837 void *crypto_alloc_instance2(const char *name, struct crypto_alg *alg, 838 unsigned int head) 839 { 840 struct crypto_instance *inst; 841 char *p; 842 int err; 843 844 p = kzalloc(head + sizeof(*inst) + sizeof(struct crypto_spawn), 845 GFP_KERNEL); 846 if (!p) 847 return ERR_PTR(-ENOMEM); 848 849 inst = (void *)(p + head); 850 851 err = crypto_inst_setname(inst, name, alg); 852 if (err) 853 goto err_free_inst; 854 855 return p; 856 857 err_free_inst: 858 kfree(p); 859 return ERR_PTR(err); 860 } 861 EXPORT_SYMBOL_GPL(crypto_alloc_instance2); 862 863 struct crypto_instance *crypto_alloc_instance(const char *name, 864 struct crypto_alg *alg) 865 { 866 struct crypto_instance *inst; 867 struct crypto_spawn *spawn; 868 int err; 869 870 inst = crypto_alloc_instance2(name, alg, 0); 871 if (IS_ERR(inst)) 872 goto out; 873 874 spawn = crypto_instance_ctx(inst); 875 err = crypto_init_spawn(spawn, alg, inst, 876 CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC); 877 878 if (err) 879 goto err_free_inst; 880 881 return inst; 882 883 err_free_inst: 884 kfree(inst); 885 inst = ERR_PTR(err); 886 887 out: 888 return inst; 889 } 890 EXPORT_SYMBOL_GPL(crypto_alloc_instance); 891 892 void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen) 893 { 894 INIT_LIST_HEAD(&queue->list); 895 queue->backlog = &queue->list; 896 queue->qlen = 0; 897 queue->max_qlen = max_qlen; 898 } 899 EXPORT_SYMBOL_GPL(crypto_init_queue); 900 901 int crypto_enqueue_request(struct crypto_queue *queue, 902 struct crypto_async_request *request) 903 { 904 int err = -EINPROGRESS; 905 906 if (unlikely(queue->qlen >= queue->max_qlen)) { 907 if (!(request->flags & CRYPTO_TFM_REQ_MAY_BACKLOG)) { 908 err = -ENOSPC; 909 goto out; 910 } 911 err = -EBUSY; 912 if (queue->backlog == &queue->list) 913 queue->backlog = &request->list; 914 } 915 916 queue->qlen++; 917 list_add_tail(&request->list, &queue->list); 918 919 out: 920 return err; 921 } 922 EXPORT_SYMBOL_GPL(crypto_enqueue_request); 923 924 struct crypto_async_request *crypto_dequeue_request(struct crypto_queue *queue) 925 { 926 struct list_head *request; 927 928 if (unlikely(!queue->qlen)) 929 return NULL; 930 931 queue->qlen--; 932 933 if (queue->backlog != &queue->list) 934 queue->backlog = queue->backlog->next; 935 936 request = queue->list.next; 937 list_del(request); 938 939 return list_entry(request, struct crypto_async_request, list); 940 } 941 EXPORT_SYMBOL_GPL(crypto_dequeue_request); 942 943 int crypto_tfm_in_queue(struct crypto_queue *queue, struct crypto_tfm *tfm) 944 { 945 struct crypto_async_request *req; 946 947 list_for_each_entry(req, &queue->list, list) { 948 if (req->tfm == tfm) 949 return 1; 950 } 951 952 return 0; 953 } 954 EXPORT_SYMBOL_GPL(crypto_tfm_in_queue); 955 956 static inline void crypto_inc_byte(u8 *a, unsigned int size) 957 { 958 u8 *b = (a + size); 959 u8 c; 960 961 for (; size; size--) { 962 c = *--b + 1; 963 *b = c; 964 if (c) 965 break; 966 } 967 } 968 969 void crypto_inc(u8 *a, unsigned int size) 970 { 971 __be32 *b = (__be32 *)(a + size); 972 u32 c; 973 974 if (IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) || 975 IS_ALIGNED((unsigned long)b, __alignof__(*b))) 976 for (; size >= 4; size -= 4) { 977 c = be32_to_cpu(*--b) + 1; 978 *b = cpu_to_be32(c); 979 if (likely(c)) 980 return; 981 } 982 983 crypto_inc_byte(a, size); 984 } 985 EXPORT_SYMBOL_GPL(crypto_inc); 986 987 void __crypto_xor(u8 *dst, const u8 *src1, const u8 *src2, unsigned int len) 988 { 989 int relalign = 0; 990 991 if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) { 992 int size = sizeof(unsigned long); 993 int d = (((unsigned long)dst ^ (unsigned long)src1) | 994 ((unsigned long)dst ^ (unsigned long)src2)) & 995 (size - 1); 996 997 relalign = d ? 1 << __ffs(d) : size; 998 999 /* 1000 * If we care about alignment, process as many bytes as 1001 * needed to advance dst and src to values whose alignments 1002 * equal their relative alignment. This will allow us to 1003 * process the remainder of the input using optimal strides. 1004 */ 1005 while (((unsigned long)dst & (relalign - 1)) && len > 0) { 1006 *dst++ = *src1++ ^ *src2++; 1007 len--; 1008 } 1009 } 1010 1011 while (IS_ENABLED(CONFIG_64BIT) && len >= 8 && !(relalign & 7)) { 1012 *(u64 *)dst = *(u64 *)src1 ^ *(u64 *)src2; 1013 dst += 8; 1014 src1 += 8; 1015 src2 += 8; 1016 len -= 8; 1017 } 1018 1019 while (len >= 4 && !(relalign & 3)) { 1020 *(u32 *)dst = *(u32 *)src1 ^ *(u32 *)src2; 1021 dst += 4; 1022 src1 += 4; 1023 src2 += 4; 1024 len -= 4; 1025 } 1026 1027 while (len >= 2 && !(relalign & 1)) { 1028 *(u16 *)dst = *(u16 *)src1 ^ *(u16 *)src2; 1029 dst += 2; 1030 src1 += 2; 1031 src2 += 2; 1032 len -= 2; 1033 } 1034 1035 while (len--) 1036 *dst++ = *src1++ ^ *src2++; 1037 } 1038 EXPORT_SYMBOL_GPL(__crypto_xor); 1039 1040 unsigned int crypto_alg_extsize(struct crypto_alg *alg) 1041 { 1042 return alg->cra_ctxsize + 1043 (alg->cra_alignmask & ~(crypto_tfm_ctx_alignment() - 1)); 1044 } 1045 EXPORT_SYMBOL_GPL(crypto_alg_extsize); 1046 1047 int crypto_type_has_alg(const char *name, const struct crypto_type *frontend, 1048 u32 type, u32 mask) 1049 { 1050 int ret = 0; 1051 struct crypto_alg *alg = crypto_find_alg(name, frontend, type, mask); 1052 1053 if (!IS_ERR(alg)) { 1054 crypto_mod_put(alg); 1055 ret = 1; 1056 } 1057 1058 return ret; 1059 } 1060 EXPORT_SYMBOL_GPL(crypto_type_has_alg); 1061 1062 static int __init crypto_algapi_init(void) 1063 { 1064 crypto_init_proc(); 1065 return 0; 1066 } 1067 1068 static void __exit crypto_algapi_exit(void) 1069 { 1070 crypto_exit_proc(); 1071 } 1072 1073 module_init(crypto_algapi_init); 1074 module_exit(crypto_algapi_exit); 1075 1076 MODULE_LICENSE("GPL"); 1077 MODULE_DESCRIPTION("Cryptographic algorithms API"); 1078