xref: /openbmc/linux/drivers/crypto/omap-crypto.c (revision 07d9a767)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * OMAP Crypto driver common support routines.
4  *
5  * Copyright (c) 2017 Texas Instruments Incorporated
6  *   Tero Kristo <t-kristo@ti.com>
7  */
8 
9 #include <linux/module.h>
10 #include <linux/kernel.h>
11 #include <linux/scatterlist.h>
12 #include <crypto/scatterwalk.h>
13 
14 #include "omap-crypto.h"
15 
16 static int omap_crypto_copy_sg_lists(int total, int bs,
17 				     struct scatterlist **sg,
18 				     struct scatterlist *new_sg, u16 flags)
19 {
20 	int n = sg_nents(*sg);
21 	struct scatterlist *tmp;
22 
23 	if (!(flags & OMAP_CRYPTO_FORCE_SINGLE_ENTRY)) {
24 		new_sg = kmalloc_array(n, sizeof(*sg), GFP_KERNEL);
25 		if (!new_sg)
26 			return -ENOMEM;
27 
28 		sg_init_table(new_sg, n);
29 	}
30 
31 	tmp = new_sg;
32 
33 	while (*sg && total) {
34 		int len = (*sg)->length;
35 
36 		if (total < len)
37 			len = total;
38 
39 		if (len > 0) {
40 			total -= len;
41 			sg_set_page(tmp, sg_page(*sg), len, (*sg)->offset);
42 			if (total <= 0)
43 				sg_mark_end(tmp);
44 			tmp = sg_next(tmp);
45 		}
46 
47 		*sg = sg_next(*sg);
48 	}
49 
50 	*sg = new_sg;
51 
52 	return 0;
53 }
54 
55 static int omap_crypto_copy_sgs(int total, int bs, struct scatterlist **sg,
56 				struct scatterlist *new_sg, u16 flags)
57 {
58 	void *buf;
59 	int pages;
60 	int new_len;
61 
62 	new_len = ALIGN(total, bs);
63 	pages = get_order(new_len);
64 
65 	buf = (void *)__get_free_pages(GFP_ATOMIC, pages);
66 	if (!buf) {
67 		pr_err("%s: Couldn't allocate pages for unaligned cases.\n",
68 		       __func__);
69 		return -ENOMEM;
70 	}
71 
72 	if (flags & OMAP_CRYPTO_COPY_DATA) {
73 		scatterwalk_map_and_copy(buf, *sg, 0, total, 0);
74 		if (flags & OMAP_CRYPTO_ZERO_BUF)
75 			memset(buf + total, 0, new_len - total);
76 	}
77 
78 	if (!(flags & OMAP_CRYPTO_FORCE_SINGLE_ENTRY))
79 		sg_init_table(new_sg, 1);
80 
81 	sg_set_buf(new_sg, buf, new_len);
82 
83 	*sg = new_sg;
84 
85 	return 0;
86 }
87 
88 static int omap_crypto_check_sg(struct scatterlist *sg, int total, int bs,
89 				u16 flags)
90 {
91 	int len = 0;
92 	int num_sg = 0;
93 
94 	if (!IS_ALIGNED(total, bs))
95 		return OMAP_CRYPTO_NOT_ALIGNED;
96 
97 	while (sg) {
98 		num_sg++;
99 
100 		if (!IS_ALIGNED(sg->offset, 4))
101 			return OMAP_CRYPTO_NOT_ALIGNED;
102 		if (!IS_ALIGNED(sg->length, bs))
103 			return OMAP_CRYPTO_NOT_ALIGNED;
104 #ifdef CONFIG_ZONE_DMA
105 		if (page_zonenum(sg_page(sg)) != ZONE_DMA)
106 			return OMAP_CRYPTO_NOT_ALIGNED;
107 #endif
108 
109 		len += sg->length;
110 		sg = sg_next(sg);
111 
112 		if (len >= total)
113 			break;
114 	}
115 
116 	if ((flags & OMAP_CRYPTO_FORCE_SINGLE_ENTRY) && num_sg > 1)
117 		return OMAP_CRYPTO_NOT_ALIGNED;
118 
119 	if (len != total)
120 		return OMAP_CRYPTO_BAD_DATA_LENGTH;
121 
122 	return 0;
123 }
124 
125 int omap_crypto_align_sg(struct scatterlist **sg, int total, int bs,
126 			 struct scatterlist *new_sg, u16 flags,
127 			 u8 flags_shift, unsigned long *dd_flags)
128 {
129 	int ret;
130 
131 	*dd_flags &= ~(OMAP_CRYPTO_COPY_MASK << flags_shift);
132 
133 	if (flags & OMAP_CRYPTO_FORCE_COPY)
134 		ret = OMAP_CRYPTO_NOT_ALIGNED;
135 	else
136 		ret = omap_crypto_check_sg(*sg, total, bs, flags);
137 
138 	if (ret == OMAP_CRYPTO_NOT_ALIGNED) {
139 		ret = omap_crypto_copy_sgs(total, bs, sg, new_sg, flags);
140 		if (ret)
141 			return ret;
142 		*dd_flags |= OMAP_CRYPTO_DATA_COPIED << flags_shift;
143 	} else if (ret == OMAP_CRYPTO_BAD_DATA_LENGTH) {
144 		ret = omap_crypto_copy_sg_lists(total, bs, sg, new_sg, flags);
145 		if (ret)
146 			return ret;
147 		if (!(flags & OMAP_CRYPTO_FORCE_SINGLE_ENTRY))
148 			*dd_flags |= OMAP_CRYPTO_SG_COPIED << flags_shift;
149 	} else if (flags & OMAP_CRYPTO_FORCE_SINGLE_ENTRY) {
150 		sg_set_buf(new_sg, sg_virt(*sg), (*sg)->length);
151 	}
152 
153 	return 0;
154 }
155 EXPORT_SYMBOL_GPL(omap_crypto_align_sg);
156 
157 static void omap_crypto_copy_data(struct scatterlist *src,
158 				  struct scatterlist *dst,
159 				  int offset, int len)
160 {
161 	int amt;
162 	void *srcb, *dstb;
163 	int srco = 0, dsto = offset;
164 
165 	while (src && dst && len) {
166 		if (srco >= src->length) {
167 			srco -= src->length;
168 			src = sg_next(src);
169 			continue;
170 		}
171 
172 		if (dsto >= dst->length) {
173 			dsto -= dst->length;
174 			dst = sg_next(dst);
175 			continue;
176 		}
177 
178 		amt = min(src->length - srco, dst->length - dsto);
179 		amt = min(len, amt);
180 
181 		srcb = kmap_atomic(sg_page(src)) + srco + src->offset;
182 		dstb = kmap_atomic(sg_page(dst)) + dsto + dst->offset;
183 
184 		memcpy(dstb, srcb, amt);
185 
186 		if (!PageSlab(sg_page(dst)))
187 			flush_kernel_dcache_page(sg_page(dst));
188 
189 		kunmap_atomic(srcb);
190 		kunmap_atomic(dstb);
191 
192 		srco += amt;
193 		dsto += amt;
194 		len -= amt;
195 	}
196 }
197 
198 void omap_crypto_cleanup(struct scatterlist *sg, struct scatterlist *orig,
199 			 int offset, int len, u8 flags_shift,
200 			 unsigned long flags)
201 {
202 	void *buf;
203 	int pages;
204 
205 	flags >>= flags_shift;
206 	flags &= OMAP_CRYPTO_COPY_MASK;
207 
208 	if (!flags)
209 		return;
210 
211 	buf = sg_virt(sg);
212 	pages = get_order(len);
213 
214 	if (orig && (flags & OMAP_CRYPTO_COPY_MASK))
215 		omap_crypto_copy_data(sg, orig, offset, len);
216 
217 	if (flags & OMAP_CRYPTO_DATA_COPIED)
218 		free_pages((unsigned long)buf, pages);
219 	else if (flags & OMAP_CRYPTO_SG_COPIED)
220 		kfree(sg);
221 }
222 EXPORT_SYMBOL_GPL(omap_crypto_cleanup);
223 
224 MODULE_DESCRIPTION("OMAP crypto support library.");
225 MODULE_LICENSE("GPL v2");
226 MODULE_AUTHOR("Tero Kristo <t-kristo@ti.com>");
227