xref: /openbmc/linux/crypto/acompress.c (revision 0c6dfa75)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * Asynchronous Compression operations
4  *
5  * Copyright (c) 2016, Intel Corporation
6  * Authors: Weigang Li <weigang.li@intel.com>
7  *          Giovanni Cabiddu <giovanni.cabiddu@intel.com>
8  */
9 
10 #include <crypto/internal/acompress.h>
11 #include <linux/cryptouser.h>
12 #include <linux/errno.h>
13 #include <linux/kernel.h>
14 #include <linux/module.h>
15 #include <linux/seq_file.h>
16 #include <linux/slab.h>
17 #include <linux/string.h>
18 #include <net/netlink.h>
19 
20 #include "compress.h"
21 
22 struct crypto_scomp;
23 
24 static const struct crypto_type crypto_acomp_type;
25 
26 static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg)
27 {
28 	return container_of(alg, struct acomp_alg, calg.base);
29 }
30 
31 static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm)
32 {
33 	return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg);
34 }
35 
36 static int __maybe_unused crypto_acomp_report(
37 	struct sk_buff *skb, struct crypto_alg *alg)
38 {
39 	struct crypto_report_acomp racomp;
40 
41 	memset(&racomp, 0, sizeof(racomp));
42 
43 	strscpy(racomp.type, "acomp", sizeof(racomp.type));
44 
45 	return nla_put(skb, CRYPTOCFGA_REPORT_ACOMP, sizeof(racomp), &racomp);
46 }
47 
48 static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
49 	__maybe_unused;
50 
51 static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
52 {
53 	seq_puts(m, "type         : acomp\n");
54 }
55 
56 static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm)
57 {
58 	struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
59 	struct acomp_alg *alg = crypto_acomp_alg(acomp);
60 
61 	alg->exit(acomp);
62 }
63 
64 static int crypto_acomp_init_tfm(struct crypto_tfm *tfm)
65 {
66 	struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
67 	struct acomp_alg *alg = crypto_acomp_alg(acomp);
68 
69 	if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
70 		return crypto_init_scomp_ops_async(tfm);
71 
72 	acomp->compress = alg->compress;
73 	acomp->decompress = alg->decompress;
74 	acomp->dst_free = alg->dst_free;
75 	acomp->reqsize = alg->reqsize;
76 
77 	if (alg->exit)
78 		acomp->base.exit = crypto_acomp_exit_tfm;
79 
80 	if (alg->init)
81 		return alg->init(acomp);
82 
83 	return 0;
84 }
85 
86 static unsigned int crypto_acomp_extsize(struct crypto_alg *alg)
87 {
88 	int extsize = crypto_alg_extsize(alg);
89 
90 	if (alg->cra_type != &crypto_acomp_type)
91 		extsize += sizeof(struct crypto_scomp *);
92 
93 	return extsize;
94 }
95 
96 static inline int __crypto_acomp_report_stat(struct sk_buff *skb,
97 					     struct crypto_alg *alg)
98 {
99 	struct comp_alg_common *calg = __crypto_comp_alg_common(alg);
100 	struct crypto_istat_compress *istat = comp_get_stat(calg);
101 	struct crypto_stat_compress racomp;
102 
103 	memset(&racomp, 0, sizeof(racomp));
104 
105 	strscpy(racomp.type, "acomp", sizeof(racomp.type));
106 	racomp.stat_compress_cnt = atomic64_read(&istat->compress_cnt);
107 	racomp.stat_compress_tlen = atomic64_read(&istat->compress_tlen);
108 	racomp.stat_decompress_cnt =  atomic64_read(&istat->decompress_cnt);
109 	racomp.stat_decompress_tlen = atomic64_read(&istat->decompress_tlen);
110 	racomp.stat_err_cnt = atomic64_read(&istat->err_cnt);
111 
112 	return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp);
113 }
114 
115 #ifdef CONFIG_CRYPTO_STATS
116 int crypto_acomp_report_stat(struct sk_buff *skb, struct crypto_alg *alg)
117 {
118 	return __crypto_acomp_report_stat(skb, alg);
119 }
120 #endif
121 
122 static const struct crypto_type crypto_acomp_type = {
123 	.extsize = crypto_acomp_extsize,
124 	.init_tfm = crypto_acomp_init_tfm,
125 #ifdef CONFIG_PROC_FS
126 	.show = crypto_acomp_show,
127 #endif
128 #if IS_ENABLED(CONFIG_CRYPTO_USER)
129 	.report = crypto_acomp_report,
130 #endif
131 #ifdef CONFIG_CRYPTO_STATS
132 	.report_stat = crypto_acomp_report_stat,
133 #endif
134 	.maskclear = ~CRYPTO_ALG_TYPE_MASK,
135 	.maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK,
136 	.type = CRYPTO_ALG_TYPE_ACOMPRESS,
137 	.tfmsize = offsetof(struct crypto_acomp, base),
138 };
139 
140 struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
141 					u32 mask)
142 {
143 	return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask);
144 }
145 EXPORT_SYMBOL_GPL(crypto_alloc_acomp);
146 
147 struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
148 					u32 mask, int node)
149 {
150 	return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask,
151 				node);
152 }
153 EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node);
154 
155 struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp)
156 {
157 	struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
158 	struct acomp_req *req;
159 
160 	req = __acomp_request_alloc(acomp);
161 	if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type))
162 		return crypto_acomp_scomp_alloc_ctx(req);
163 
164 	return req;
165 }
166 EXPORT_SYMBOL_GPL(acomp_request_alloc);
167 
168 void acomp_request_free(struct acomp_req *req)
169 {
170 	struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
171 	struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
172 
173 	if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
174 		crypto_acomp_scomp_free_ctx(req);
175 
176 	if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) {
177 		acomp->dst_free(req->dst);
178 		req->dst = NULL;
179 	}
180 
181 	__acomp_request_free(req);
182 }
183 EXPORT_SYMBOL_GPL(acomp_request_free);
184 
185 void comp_prepare_alg(struct comp_alg_common *alg)
186 {
187 	struct crypto_istat_compress *istat = comp_get_stat(alg);
188 	struct crypto_alg *base = &alg->base;
189 
190 	base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK;
191 
192 	if (IS_ENABLED(CONFIG_CRYPTO_STATS))
193 		memset(istat, 0, sizeof(*istat));
194 }
195 
196 int crypto_register_acomp(struct acomp_alg *alg)
197 {
198 	struct crypto_alg *base = &alg->calg.base;
199 
200 	comp_prepare_alg(&alg->calg);
201 
202 	base->cra_type = &crypto_acomp_type;
203 	base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS;
204 
205 	return crypto_register_alg(base);
206 }
207 EXPORT_SYMBOL_GPL(crypto_register_acomp);
208 
209 void crypto_unregister_acomp(struct acomp_alg *alg)
210 {
211 	crypto_unregister_alg(&alg->base);
212 }
213 EXPORT_SYMBOL_GPL(crypto_unregister_acomp);
214 
215 int crypto_register_acomps(struct acomp_alg *algs, int count)
216 {
217 	int i, ret;
218 
219 	for (i = 0; i < count; i++) {
220 		ret = crypto_register_acomp(&algs[i]);
221 		if (ret)
222 			goto err;
223 	}
224 
225 	return 0;
226 
227 err:
228 	for (--i; i >= 0; --i)
229 		crypto_unregister_acomp(&algs[i]);
230 
231 	return ret;
232 }
233 EXPORT_SYMBOL_GPL(crypto_register_acomps);
234 
235 void crypto_unregister_acomps(struct acomp_alg *algs, int count)
236 {
237 	int i;
238 
239 	for (i = count - 1; i >= 0; --i)
240 		crypto_unregister_acomp(&algs[i]);
241 }
242 EXPORT_SYMBOL_GPL(crypto_unregister_acomps);
243 
244 MODULE_LICENSE("GPL");
245 MODULE_DESCRIPTION("Asynchronous compression type");
246