1 /** 2 * AES routines supporting VMX instructions on the Power 8 3 * 4 * Copyright (C) 2015 International Business Machines Inc. 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License as published by 8 * the Free Software Foundation; version 2 only. 9 * 10 * This program is distributed in the hope that it will be useful, 11 * but WITHOUT ANY WARRANTY; without even the implied warranty of 12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 13 * GNU General Public License for more details. 14 * 15 * You should have received a copy of the GNU General Public License 16 * along with this program; if not, write to the Free Software 17 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 18 * 19 * Author: Marcelo Henrique Cerri <mhcerri@br.ibm.com> 20 */ 21 22 #include <linux/types.h> 23 #include <linux/err.h> 24 #include <linux/crypto.h> 25 #include <linux/delay.h> 26 #include <asm/simd.h> 27 #include <asm/switch_to.h> 28 #include <crypto/aes.h> 29 #include <crypto/internal/simd.h> 30 31 #include "aesp8-ppc.h" 32 33 struct p8_aes_ctx { 34 struct crypto_cipher *fallback; 35 struct aes_key enc_key; 36 struct aes_key dec_key; 37 }; 38 39 static int p8_aes_init(struct crypto_tfm *tfm) 40 { 41 const char *alg = crypto_tfm_alg_name(tfm); 42 struct crypto_cipher *fallback; 43 struct p8_aes_ctx *ctx = crypto_tfm_ctx(tfm); 44 45 fallback = crypto_alloc_cipher(alg, 0, CRYPTO_ALG_NEED_FALLBACK); 46 if (IS_ERR(fallback)) { 47 printk(KERN_ERR 48 "Failed to allocate transformation for '%s': %ld\n", 49 alg, PTR_ERR(fallback)); 50 return PTR_ERR(fallback); 51 } 52 53 crypto_cipher_set_flags(fallback, 54 crypto_cipher_get_flags((struct 55 crypto_cipher *) 56 tfm)); 57 ctx->fallback = fallback; 58 59 return 0; 60 } 61 62 static void p8_aes_exit(struct crypto_tfm *tfm) 63 { 64 struct p8_aes_ctx *ctx = crypto_tfm_ctx(tfm); 65 66 if (ctx->fallback) { 67 crypto_free_cipher(ctx->fallback); 68 ctx->fallback = NULL; 69 } 70 } 71 72 static int p8_aes_setkey(struct crypto_tfm *tfm, const u8 *key, 73 unsigned int keylen) 74 { 75 int ret; 76 struct p8_aes_ctx *ctx = crypto_tfm_ctx(tfm); 77 78 preempt_disable(); 79 pagefault_disable(); 80 enable_kernel_vsx(); 81 ret = aes_p8_set_encrypt_key(key, keylen * 8, &ctx->enc_key); 82 ret |= aes_p8_set_decrypt_key(key, keylen * 8, &ctx->dec_key); 83 disable_kernel_vsx(); 84 pagefault_enable(); 85 preempt_enable(); 86 87 ret |= crypto_cipher_setkey(ctx->fallback, key, keylen); 88 89 return ret ? -EINVAL : 0; 90 } 91 92 static void p8_aes_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) 93 { 94 struct p8_aes_ctx *ctx = crypto_tfm_ctx(tfm); 95 96 if (!crypto_simd_usable()) { 97 crypto_cipher_encrypt_one(ctx->fallback, dst, src); 98 } else { 99 preempt_disable(); 100 pagefault_disable(); 101 enable_kernel_vsx(); 102 aes_p8_encrypt(src, dst, &ctx->enc_key); 103 disable_kernel_vsx(); 104 pagefault_enable(); 105 preempt_enable(); 106 } 107 } 108 109 static void p8_aes_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) 110 { 111 struct p8_aes_ctx *ctx = crypto_tfm_ctx(tfm); 112 113 if (!crypto_simd_usable()) { 114 crypto_cipher_decrypt_one(ctx->fallback, dst, src); 115 } else { 116 preempt_disable(); 117 pagefault_disable(); 118 enable_kernel_vsx(); 119 aes_p8_decrypt(src, dst, &ctx->dec_key); 120 disable_kernel_vsx(); 121 pagefault_enable(); 122 preempt_enable(); 123 } 124 } 125 126 struct crypto_alg p8_aes_alg = { 127 .cra_name = "aes", 128 .cra_driver_name = "p8_aes", 129 .cra_module = THIS_MODULE, 130 .cra_priority = 1000, 131 .cra_type = NULL, 132 .cra_flags = CRYPTO_ALG_TYPE_CIPHER | CRYPTO_ALG_NEED_FALLBACK, 133 .cra_alignmask = 0, 134 .cra_blocksize = AES_BLOCK_SIZE, 135 .cra_ctxsize = sizeof(struct p8_aes_ctx), 136 .cra_init = p8_aes_init, 137 .cra_exit = p8_aes_exit, 138 .cra_cipher = { 139 .cia_min_keysize = AES_MIN_KEY_SIZE, 140 .cia_max_keysize = AES_MAX_KEY_SIZE, 141 .cia_setkey = p8_aes_setkey, 142 .cia_encrypt = p8_aes_encrypt, 143 .cia_decrypt = p8_aes_decrypt, 144 }, 145 }; 146