xref: /openbmc/linux/arch/s390/crypto/crc32-vx.c (revision bcb63314)
1 /*
2  * Crypto-API module for CRC-32 algorithms implemented with the
3  * z/Architecture Vector Extension Facility.
4  *
5  * Copyright IBM Corp. 2015
6  * Author(s): Hendrik Brueckner <brueckner@linux.vnet.ibm.com>
7  */
8 #define KMSG_COMPONENT	"crc32-vx"
9 #define pr_fmt(fmt)	KMSG_COMPONENT ": " fmt
10 
11 #include <linux/module.h>
12 #include <linux/cpufeature.h>
13 #include <linux/crc32.h>
14 #include <crypto/internal/hash.h>
15 #include <asm/fpu/api.h>
16 
17 
18 #define CRC32_BLOCK_SIZE	1
19 #define CRC32_DIGEST_SIZE	4
20 
21 #define VX_MIN_LEN		64
22 #define VX_ALIGNMENT		16L
23 #define VX_ALIGN_MASK		(VX_ALIGNMENT - 1)
24 
25 struct crc_ctx {
26 	u32 key;
27 };
28 
29 struct crc_desc_ctx {
30 	u32 crc;
31 };
32 
33 /* Prototypes for functions in assembly files */
34 u32 crc32_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
35 u32 crc32_be_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
36 u32 crc32c_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
37 
38 /*
39  * DEFINE_CRC32_VX() - Define a CRC-32 function using the vector extension
40  *
41  * Creates a function to perform a particular CRC-32 computation. Depending
42  * on the message buffer, the hardware-accelerated or software implementation
43  * is used.   Note that the message buffer is aligned to improve fetch
44  * operations of VECTOR LOAD MULTIPLE instructions.
45  *
46  */
47 #define DEFINE_CRC32_VX(___fname, ___crc32_vx, ___crc32_sw)		    \
48 	static u32 __pure ___fname(u32 crc,				    \
49 				unsigned char const *data, size_t datalen)  \
50 	{								    \
51 		struct kernel_fpu vxstate;				    \
52 		unsigned long prealign, aligned, remaining;		    \
53 									    \
54 		if (datalen < VX_MIN_LEN + VX_ALIGN_MASK)		    \
55 			return ___crc32_sw(crc, data, datalen);		    \
56 									    \
57 		if ((unsigned long)data & VX_ALIGN_MASK) {		    \
58 			prealign = VX_ALIGNMENT -			    \
59 				  ((unsigned long)data & VX_ALIGN_MASK);    \
60 			datalen -= prealign;				    \
61 			crc = ___crc32_sw(crc, data, prealign);		    \
62 			data = (void *)((unsigned long)data + prealign);    \
63 		}							    \
64 									    \
65 		aligned = datalen & ~VX_ALIGN_MASK;			    \
66 		remaining = datalen & VX_ALIGN_MASK;			    \
67 									    \
68 		kernel_fpu_begin(&vxstate, KERNEL_VXR_LOW);		    \
69 		crc = ___crc32_vx(crc, data, aligned);			    \
70 		kernel_fpu_end(&vxstate, KERNEL_VXR_LOW);		    \
71 									    \
72 		if (remaining)						    \
73 			crc = ___crc32_sw(crc, data + aligned, remaining);  \
74 									    \
75 		return crc;						    \
76 	}
77 
78 DEFINE_CRC32_VX(crc32_le_vx, crc32_le_vgfm_16, crc32_le)
79 DEFINE_CRC32_VX(crc32_be_vx, crc32_be_vgfm_16, crc32_be)
80 DEFINE_CRC32_VX(crc32c_le_vx, crc32c_le_vgfm_16, __crc32c_le)
81 
82 
83 static int crc32_vx_cra_init_zero(struct crypto_tfm *tfm)
84 {
85 	struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
86 
87 	mctx->key = 0;
88 	return 0;
89 }
90 
91 static int crc32_vx_cra_init_invert(struct crypto_tfm *tfm)
92 {
93 	struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
94 
95 	mctx->key = ~0;
96 	return 0;
97 }
98 
99 static int crc32_vx_init(struct shash_desc *desc)
100 {
101 	struct crc_ctx *mctx = crypto_shash_ctx(desc->tfm);
102 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
103 
104 	ctx->crc = mctx->key;
105 	return 0;
106 }
107 
108 static int crc32_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
109 			   unsigned int newkeylen)
110 {
111 	struct crc_ctx *mctx = crypto_shash_ctx(tfm);
112 
113 	if (newkeylen != sizeof(mctx->key)) {
114 		crypto_shash_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
115 		return -EINVAL;
116 	}
117 	mctx->key = le32_to_cpu(*(__le32 *)newkey);
118 	return 0;
119 }
120 
121 static int crc32be_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
122 			     unsigned int newkeylen)
123 {
124 	struct crc_ctx *mctx = crypto_shash_ctx(tfm);
125 
126 	if (newkeylen != sizeof(mctx->key)) {
127 		crypto_shash_set_flags(tfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
128 		return -EINVAL;
129 	}
130 	mctx->key = be32_to_cpu(*(__be32 *)newkey);
131 	return 0;
132 }
133 
134 static int crc32le_vx_final(struct shash_desc *desc, u8 *out)
135 {
136 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
137 
138 	*(__le32 *)out = cpu_to_le32p(&ctx->crc);
139 	return 0;
140 }
141 
142 static int crc32be_vx_final(struct shash_desc *desc, u8 *out)
143 {
144 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
145 
146 	*(__be32 *)out = cpu_to_be32p(&ctx->crc);
147 	return 0;
148 }
149 
150 static int crc32c_vx_final(struct shash_desc *desc, u8 *out)
151 {
152 	struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
153 
154 	/*
155 	 * Perform a final XOR with 0xFFFFFFFF to be in sync
156 	 * with the generic crc32c shash implementation.
157 	 */
158 	*(__le32 *)out = ~cpu_to_le32p(&ctx->crc);
159 	return 0;
160 }
161 
162 static int __crc32le_vx_finup(u32 *crc, const u8 *data, unsigned int len,
163 			      u8 *out)
164 {
165 	*(__le32 *)out = cpu_to_le32(crc32_le_vx(*crc, data, len));
166 	return 0;
167 }
168 
169 static int __crc32be_vx_finup(u32 *crc, const u8 *data, unsigned int len,
170 			      u8 *out)
171 {
172 	*(__be32 *)out = cpu_to_be32(crc32_be_vx(*crc, data, len));
173 	return 0;
174 }
175 
176 static int __crc32c_vx_finup(u32 *crc, const u8 *data, unsigned int len,
177 			     u8 *out)
178 {
179 	/*
180 	 * Perform a final XOR with 0xFFFFFFFF to be in sync
181 	 * with the generic crc32c shash implementation.
182 	 */
183 	*(__le32 *)out = ~cpu_to_le32(crc32c_le_vx(*crc, data, len));
184 	return 0;
185 }
186 
187 
188 #define CRC32_VX_FINUP(alg, func)					      \
189 	static int alg ## _vx_finup(struct shash_desc *desc, const u8 *data,  \
190 				   unsigned int datalen, u8 *out)	      \
191 	{								      \
192 		return __ ## alg ## _vx_finup(shash_desc_ctx(desc),	      \
193 					      data, datalen, out);	      \
194 	}
195 
196 CRC32_VX_FINUP(crc32le, crc32_le_vx)
197 CRC32_VX_FINUP(crc32be, crc32_be_vx)
198 CRC32_VX_FINUP(crc32c, crc32c_le_vx)
199 
200 #define CRC32_VX_DIGEST(alg, func)					      \
201 	static int alg ## _vx_digest(struct shash_desc *desc, const u8 *data, \
202 				     unsigned int len, u8 *out)		      \
203 	{								      \
204 		return __ ## alg ## _vx_finup(crypto_shash_ctx(desc->tfm),    \
205 					      data, len, out);		      \
206 	}
207 
208 CRC32_VX_DIGEST(crc32le, crc32_le_vx)
209 CRC32_VX_DIGEST(crc32be, crc32_be_vx)
210 CRC32_VX_DIGEST(crc32c, crc32c_le_vx)
211 
212 #define CRC32_VX_UPDATE(alg, func)					      \
213 	static int alg ## _vx_update(struct shash_desc *desc, const u8 *data, \
214 				     unsigned int datalen)		      \
215 	{								      \
216 		struct crc_desc_ctx *ctx = shash_desc_ctx(desc);	      \
217 		ctx->crc = func(ctx->crc, data, datalen);		      \
218 		return 0;						      \
219 	}
220 
221 CRC32_VX_UPDATE(crc32le, crc32_le_vx)
222 CRC32_VX_UPDATE(crc32be, crc32_be_vx)
223 CRC32_VX_UPDATE(crc32c, crc32c_le_vx)
224 
225 
226 static struct shash_alg crc32_vx_algs[] = {
227 	/* CRC-32 LE */
228 	{
229 		.init		=	crc32_vx_init,
230 		.setkey		=	crc32_vx_setkey,
231 		.update		=	crc32le_vx_update,
232 		.final		=	crc32le_vx_final,
233 		.finup		=	crc32le_vx_finup,
234 		.digest		=	crc32le_vx_digest,
235 		.descsize	=	sizeof(struct crc_desc_ctx),
236 		.digestsize	=	CRC32_DIGEST_SIZE,
237 		.base		=	{
238 			.cra_name	 = "crc32",
239 			.cra_driver_name = "crc32-vx",
240 			.cra_priority	 = 200,
241 			.cra_blocksize	 = CRC32_BLOCK_SIZE,
242 			.cra_ctxsize	 = sizeof(struct crc_ctx),
243 			.cra_module	 = THIS_MODULE,
244 			.cra_init	 = crc32_vx_cra_init_zero,
245 		},
246 	},
247 	/* CRC-32 BE */
248 	{
249 		.init		=	crc32_vx_init,
250 		.setkey		=	crc32be_vx_setkey,
251 		.update		=	crc32be_vx_update,
252 		.final		=	crc32be_vx_final,
253 		.finup		=	crc32be_vx_finup,
254 		.digest		=	crc32be_vx_digest,
255 		.descsize	=	sizeof(struct crc_desc_ctx),
256 		.digestsize	=	CRC32_DIGEST_SIZE,
257 		.base		=	{
258 			.cra_name	 = "crc32be",
259 			.cra_driver_name = "crc32be-vx",
260 			.cra_priority	 = 200,
261 			.cra_blocksize	 = CRC32_BLOCK_SIZE,
262 			.cra_ctxsize	 = sizeof(struct crc_ctx),
263 			.cra_module	 = THIS_MODULE,
264 			.cra_init	 = crc32_vx_cra_init_zero,
265 		},
266 	},
267 	/* CRC-32C LE */
268 	{
269 		.init		=	crc32_vx_init,
270 		.setkey		=	crc32_vx_setkey,
271 		.update		=	crc32c_vx_update,
272 		.final		=	crc32c_vx_final,
273 		.finup		=	crc32c_vx_finup,
274 		.digest		=	crc32c_vx_digest,
275 		.descsize	=	sizeof(struct crc_desc_ctx),
276 		.digestsize	=	CRC32_DIGEST_SIZE,
277 		.base		=	{
278 			.cra_name	 = "crc32c",
279 			.cra_driver_name = "crc32c-vx",
280 			.cra_priority	 = 200,
281 			.cra_blocksize	 = CRC32_BLOCK_SIZE,
282 			.cra_ctxsize	 = sizeof(struct crc_ctx),
283 			.cra_module	 = THIS_MODULE,
284 			.cra_init	 = crc32_vx_cra_init_invert,
285 		},
286 	},
287 };
288 
289 
290 static int __init crc_vx_mod_init(void)
291 {
292 	return crypto_register_shashes(crc32_vx_algs,
293 				       ARRAY_SIZE(crc32_vx_algs));
294 }
295 
296 static void __exit crc_vx_mod_exit(void)
297 {
298 	crypto_unregister_shashes(crc32_vx_algs, ARRAY_SIZE(crc32_vx_algs));
299 }
300 
301 module_cpu_feature_match(VXRS, crc_vx_mod_init);
302 module_exit(crc_vx_mod_exit);
303 
304 MODULE_AUTHOR("Hendrik Brueckner <brueckner@linux.vnet.ibm.com>");
305 MODULE_LICENSE("GPL");
306 
307 MODULE_ALIAS_CRYPTO("crc32");
308 MODULE_ALIAS_CRYPTO("crc32-vx");
309 MODULE_ALIAS_CRYPTO("crc32c");
310 MODULE_ALIAS_CRYPTO("crc32c-vx");
311