1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Cryptographic API. 4 * 5 * SHA-256, as specified in 6 * http://csrc.nist.gov/groups/STM/cavp/documents/shs/sha256-384-512.pdf 7 * 8 * SHA-256 code by Jean-Luc Cooke <jlcooke@certainkey.com>. 9 * 10 * Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com> 11 * Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk> 12 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au> 13 * SHA224 Support Copyright 2007 Intel Corporation <jonathan.lynch@intel.com> 14 */ 15 #include <crypto/internal/hash.h> 16 #include <linux/init.h> 17 #include <linux/module.h> 18 #include <linux/mm.h> 19 #include <linux/types.h> 20 #include <crypto/sha.h> 21 #include <crypto/sha256_base.h> 22 #include <asm/byteorder.h> 23 #include <asm/unaligned.h> 24 25 const u8 sha224_zero_message_hash[SHA224_DIGEST_SIZE] = { 26 0xd1, 0x4a, 0x02, 0x8c, 0x2a, 0x3a, 0x2b, 0xc9, 0x47, 27 0x61, 0x02, 0xbb, 0x28, 0x82, 0x34, 0xc4, 0x15, 0xa2, 28 0xb0, 0x1f, 0x82, 0x8e, 0xa6, 0x2a, 0xc5, 0xb3, 0xe4, 29 0x2f 30 }; 31 EXPORT_SYMBOL_GPL(sha224_zero_message_hash); 32 33 const u8 sha256_zero_message_hash[SHA256_DIGEST_SIZE] = { 34 0xe3, 0xb0, 0xc4, 0x42, 0x98, 0xfc, 0x1c, 0x14, 35 0x9a, 0xfb, 0xf4, 0xc8, 0x99, 0x6f, 0xb9, 0x24, 36 0x27, 0xae, 0x41, 0xe4, 0x64, 0x9b, 0x93, 0x4c, 37 0xa4, 0x95, 0x99, 0x1b, 0x78, 0x52, 0xb8, 0x55 38 }; 39 EXPORT_SYMBOL_GPL(sha256_zero_message_hash); 40 41 static inline u32 Ch(u32 x, u32 y, u32 z) 42 { 43 return z ^ (x & (y ^ z)); 44 } 45 46 static inline u32 Maj(u32 x, u32 y, u32 z) 47 { 48 return (x & y) | (z & (x | y)); 49 } 50 51 #define e0(x) (ror32(x, 2) ^ ror32(x,13) ^ ror32(x,22)) 52 #define e1(x) (ror32(x, 6) ^ ror32(x,11) ^ ror32(x,25)) 53 #define s0(x) (ror32(x, 7) ^ ror32(x,18) ^ (x >> 3)) 54 #define s1(x) (ror32(x,17) ^ ror32(x,19) ^ (x >> 10)) 55 56 static inline void LOAD_OP(int I, u32 *W, const u8 *input) 57 { 58 W[I] = get_unaligned_be32((__u32 *)input + I); 59 } 60 61 static inline void BLEND_OP(int I, u32 *W) 62 { 63 W[I] = s1(W[I-2]) + W[I-7] + s0(W[I-15]) + W[I-16]; 64 } 65 66 static void sha256_transform(u32 *state, const u8 *input) 67 { 68 u32 a, b, c, d, e, f, g, h, t1, t2; 69 u32 W[64]; 70 int i; 71 72 /* load the input */ 73 for (i = 0; i < 16; i++) 74 LOAD_OP(i, W, input); 75 76 /* now blend */ 77 for (i = 16; i < 64; i++) 78 BLEND_OP(i, W); 79 80 /* load the state into our registers */ 81 a=state[0]; b=state[1]; c=state[2]; d=state[3]; 82 e=state[4]; f=state[5]; g=state[6]; h=state[7]; 83 84 /* now iterate */ 85 t1 = h + e1(e) + Ch(e,f,g) + 0x428a2f98 + W[ 0]; 86 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 87 t1 = g + e1(d) + Ch(d,e,f) + 0x71374491 + W[ 1]; 88 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 89 t1 = f + e1(c) + Ch(c,d,e) + 0xb5c0fbcf + W[ 2]; 90 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 91 t1 = e + e1(b) + Ch(b,c,d) + 0xe9b5dba5 + W[ 3]; 92 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 93 t1 = d + e1(a) + Ch(a,b,c) + 0x3956c25b + W[ 4]; 94 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 95 t1 = c + e1(h) + Ch(h,a,b) + 0x59f111f1 + W[ 5]; 96 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 97 t1 = b + e1(g) + Ch(g,h,a) + 0x923f82a4 + W[ 6]; 98 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 99 t1 = a + e1(f) + Ch(f,g,h) + 0xab1c5ed5 + W[ 7]; 100 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 101 102 t1 = h + e1(e) + Ch(e,f,g) + 0xd807aa98 + W[ 8]; 103 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 104 t1 = g + e1(d) + Ch(d,e,f) + 0x12835b01 + W[ 9]; 105 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 106 t1 = f + e1(c) + Ch(c,d,e) + 0x243185be + W[10]; 107 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 108 t1 = e + e1(b) + Ch(b,c,d) + 0x550c7dc3 + W[11]; 109 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 110 t1 = d + e1(a) + Ch(a,b,c) + 0x72be5d74 + W[12]; 111 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 112 t1 = c + e1(h) + Ch(h,a,b) + 0x80deb1fe + W[13]; 113 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 114 t1 = b + e1(g) + Ch(g,h,a) + 0x9bdc06a7 + W[14]; 115 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 116 t1 = a + e1(f) + Ch(f,g,h) + 0xc19bf174 + W[15]; 117 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 118 119 t1 = h + e1(e) + Ch(e,f,g) + 0xe49b69c1 + W[16]; 120 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 121 t1 = g + e1(d) + Ch(d,e,f) + 0xefbe4786 + W[17]; 122 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 123 t1 = f + e1(c) + Ch(c,d,e) + 0x0fc19dc6 + W[18]; 124 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 125 t1 = e + e1(b) + Ch(b,c,d) + 0x240ca1cc + W[19]; 126 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 127 t1 = d + e1(a) + Ch(a,b,c) + 0x2de92c6f + W[20]; 128 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 129 t1 = c + e1(h) + Ch(h,a,b) + 0x4a7484aa + W[21]; 130 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 131 t1 = b + e1(g) + Ch(g,h,a) + 0x5cb0a9dc + W[22]; 132 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 133 t1 = a + e1(f) + Ch(f,g,h) + 0x76f988da + W[23]; 134 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 135 136 t1 = h + e1(e) + Ch(e,f,g) + 0x983e5152 + W[24]; 137 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 138 t1 = g + e1(d) + Ch(d,e,f) + 0xa831c66d + W[25]; 139 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 140 t1 = f + e1(c) + Ch(c,d,e) + 0xb00327c8 + W[26]; 141 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 142 t1 = e + e1(b) + Ch(b,c,d) + 0xbf597fc7 + W[27]; 143 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 144 t1 = d + e1(a) + Ch(a,b,c) + 0xc6e00bf3 + W[28]; 145 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 146 t1 = c + e1(h) + Ch(h,a,b) + 0xd5a79147 + W[29]; 147 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 148 t1 = b + e1(g) + Ch(g,h,a) + 0x06ca6351 + W[30]; 149 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 150 t1 = a + e1(f) + Ch(f,g,h) + 0x14292967 + W[31]; 151 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 152 153 t1 = h + e1(e) + Ch(e,f,g) + 0x27b70a85 + W[32]; 154 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 155 t1 = g + e1(d) + Ch(d,e,f) + 0x2e1b2138 + W[33]; 156 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 157 t1 = f + e1(c) + Ch(c,d,e) + 0x4d2c6dfc + W[34]; 158 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 159 t1 = e + e1(b) + Ch(b,c,d) + 0x53380d13 + W[35]; 160 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 161 t1 = d + e1(a) + Ch(a,b,c) + 0x650a7354 + W[36]; 162 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 163 t1 = c + e1(h) + Ch(h,a,b) + 0x766a0abb + W[37]; 164 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 165 t1 = b + e1(g) + Ch(g,h,a) + 0x81c2c92e + W[38]; 166 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 167 t1 = a + e1(f) + Ch(f,g,h) + 0x92722c85 + W[39]; 168 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 169 170 t1 = h + e1(e) + Ch(e,f,g) + 0xa2bfe8a1 + W[40]; 171 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 172 t1 = g + e1(d) + Ch(d,e,f) + 0xa81a664b + W[41]; 173 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 174 t1 = f + e1(c) + Ch(c,d,e) + 0xc24b8b70 + W[42]; 175 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 176 t1 = e + e1(b) + Ch(b,c,d) + 0xc76c51a3 + W[43]; 177 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 178 t1 = d + e1(a) + Ch(a,b,c) + 0xd192e819 + W[44]; 179 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 180 t1 = c + e1(h) + Ch(h,a,b) + 0xd6990624 + W[45]; 181 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 182 t1 = b + e1(g) + Ch(g,h,a) + 0xf40e3585 + W[46]; 183 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 184 t1 = a + e1(f) + Ch(f,g,h) + 0x106aa070 + W[47]; 185 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 186 187 t1 = h + e1(e) + Ch(e,f,g) + 0x19a4c116 + W[48]; 188 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 189 t1 = g + e1(d) + Ch(d,e,f) + 0x1e376c08 + W[49]; 190 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 191 t1 = f + e1(c) + Ch(c,d,e) + 0x2748774c + W[50]; 192 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 193 t1 = e + e1(b) + Ch(b,c,d) + 0x34b0bcb5 + W[51]; 194 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 195 t1 = d + e1(a) + Ch(a,b,c) + 0x391c0cb3 + W[52]; 196 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 197 t1 = c + e1(h) + Ch(h,a,b) + 0x4ed8aa4a + W[53]; 198 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 199 t1 = b + e1(g) + Ch(g,h,a) + 0x5b9cca4f + W[54]; 200 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 201 t1 = a + e1(f) + Ch(f,g,h) + 0x682e6ff3 + W[55]; 202 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 203 204 t1 = h + e1(e) + Ch(e,f,g) + 0x748f82ee + W[56]; 205 t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; 206 t1 = g + e1(d) + Ch(d,e,f) + 0x78a5636f + W[57]; 207 t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; 208 t1 = f + e1(c) + Ch(c,d,e) + 0x84c87814 + W[58]; 209 t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; 210 t1 = e + e1(b) + Ch(b,c,d) + 0x8cc70208 + W[59]; 211 t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; 212 t1 = d + e1(a) + Ch(a,b,c) + 0x90befffa + W[60]; 213 t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; 214 t1 = c + e1(h) + Ch(h,a,b) + 0xa4506ceb + W[61]; 215 t2 = e0(d) + Maj(d,e,f); g+=t1; c=t1+t2; 216 t1 = b + e1(g) + Ch(g,h,a) + 0xbef9a3f7 + W[62]; 217 t2 = e0(c) + Maj(c,d,e); f+=t1; b=t1+t2; 218 t1 = a + e1(f) + Ch(f,g,h) + 0xc67178f2 + W[63]; 219 t2 = e0(b) + Maj(b,c,d); e+=t1; a=t1+t2; 220 221 state[0] += a; state[1] += b; state[2] += c; state[3] += d; 222 state[4] += e; state[5] += f; state[6] += g; state[7] += h; 223 224 /* clear any sensitive info... */ 225 a = b = c = d = e = f = g = h = t1 = t2 = 0; 226 memzero_explicit(W, 64 * sizeof(u32)); 227 } 228 229 static void sha256_generic_block_fn(struct sha256_state *sst, u8 const *src, 230 int blocks) 231 { 232 while (blocks--) { 233 sha256_transform(sst->state, src); 234 src += SHA256_BLOCK_SIZE; 235 } 236 } 237 238 int crypto_sha256_update(struct shash_desc *desc, const u8 *data, 239 unsigned int len) 240 { 241 return sha256_base_do_update(desc, data, len, sha256_generic_block_fn); 242 } 243 EXPORT_SYMBOL(crypto_sha256_update); 244 245 static int sha256_final(struct shash_desc *desc, u8 *out) 246 { 247 sha256_base_do_finalize(desc, sha256_generic_block_fn); 248 return sha256_base_finish(desc, out); 249 } 250 251 int crypto_sha256_finup(struct shash_desc *desc, const u8 *data, 252 unsigned int len, u8 *hash) 253 { 254 sha256_base_do_update(desc, data, len, sha256_generic_block_fn); 255 return sha256_final(desc, hash); 256 } 257 EXPORT_SYMBOL(crypto_sha256_finup); 258 259 static struct shash_alg sha256_algs[2] = { { 260 .digestsize = SHA256_DIGEST_SIZE, 261 .init = sha256_base_init, 262 .update = crypto_sha256_update, 263 .final = sha256_final, 264 .finup = crypto_sha256_finup, 265 .descsize = sizeof(struct sha256_state), 266 .base = { 267 .cra_name = "sha256", 268 .cra_driver_name= "sha256-generic", 269 .cra_priority = 100, 270 .cra_blocksize = SHA256_BLOCK_SIZE, 271 .cra_module = THIS_MODULE, 272 } 273 }, { 274 .digestsize = SHA224_DIGEST_SIZE, 275 .init = sha224_base_init, 276 .update = crypto_sha256_update, 277 .final = sha256_final, 278 .finup = crypto_sha256_finup, 279 .descsize = sizeof(struct sha256_state), 280 .base = { 281 .cra_name = "sha224", 282 .cra_driver_name= "sha224-generic", 283 .cra_priority = 100, 284 .cra_blocksize = SHA224_BLOCK_SIZE, 285 .cra_module = THIS_MODULE, 286 } 287 } }; 288 289 static int __init sha256_generic_mod_init(void) 290 { 291 return crypto_register_shashes(sha256_algs, ARRAY_SIZE(sha256_algs)); 292 } 293 294 static void __exit sha256_generic_mod_fini(void) 295 { 296 crypto_unregister_shashes(sha256_algs, ARRAY_SIZE(sha256_algs)); 297 } 298 299 subsys_initcall(sha256_generic_mod_init); 300 module_exit(sha256_generic_mod_fini); 301 302 MODULE_LICENSE("GPL"); 303 MODULE_DESCRIPTION("SHA-224 and SHA-256 Secure Hash Algorithm"); 304 305 MODULE_ALIAS_CRYPTO("sha224"); 306 MODULE_ALIAS_CRYPTO("sha224-generic"); 307 MODULE_ALIAS_CRYPTO("sha256"); 308 MODULE_ALIAS_CRYPTO("sha256-generic"); 309