xref: /openbmc/linux/drivers/crypto/atmel-tdes.c (revision 7c0afcad)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Cryptographic API.
4  *
5  * Support for ATMEL DES/TDES HW acceleration.
6  *
7  * Copyright (c) 2012 Eukréa Electromatique - ATMEL
8  * Author: Nicolas Royer <nicolas@eukrea.com>
9  *
10  * Some ideas are from omap-aes.c drivers.
11  */
12 
13 
14 #include <linux/kernel.h>
15 #include <linux/module.h>
16 #include <linux/slab.h>
17 #include <linux/err.h>
18 #include <linux/clk.h>
19 #include <linux/io.h>
20 #include <linux/hw_random.h>
21 #include <linux/platform_device.h>
22 
23 #include <linux/device.h>
24 #include <linux/dmaengine.h>
25 #include <linux/init.h>
26 #include <linux/errno.h>
27 #include <linux/interrupt.h>
28 #include <linux/irq.h>
29 #include <linux/scatterlist.h>
30 #include <linux/dma-mapping.h>
31 #include <linux/of_device.h>
32 #include <linux/delay.h>
33 #include <linux/crypto.h>
34 #include <crypto/scatterwalk.h>
35 #include <crypto/algapi.h>
36 #include <crypto/internal/des.h>
37 #include <crypto/internal/skcipher.h>
38 #include "atmel-tdes-regs.h"
39 
40 #define ATMEL_TDES_PRIORITY	300
41 
42 /* TDES flags  */
43 /* Reserve bits [17:16], [13:12], [2:0] for AES Mode Register */
44 #define TDES_FLAGS_ENCRYPT	TDES_MR_CYPHER_ENC
45 #define TDES_FLAGS_OPMODE_MASK	(TDES_MR_OPMOD_MASK | TDES_MR_CFBS_MASK)
46 #define TDES_FLAGS_ECB		TDES_MR_OPMOD_ECB
47 #define TDES_FLAGS_CBC		TDES_MR_OPMOD_CBC
48 #define TDES_FLAGS_OFB		TDES_MR_OPMOD_OFB
49 #define TDES_FLAGS_CFB64	(TDES_MR_OPMOD_CFB | TDES_MR_CFBS_64b)
50 #define TDES_FLAGS_CFB32	(TDES_MR_OPMOD_CFB | TDES_MR_CFBS_32b)
51 #define TDES_FLAGS_CFB16	(TDES_MR_OPMOD_CFB | TDES_MR_CFBS_16b)
52 #define TDES_FLAGS_CFB8		(TDES_MR_OPMOD_CFB | TDES_MR_CFBS_8b)
53 
54 #define TDES_FLAGS_MODE_MASK	(TDES_FLAGS_OPMODE_MASK | TDES_FLAGS_ENCRYPT)
55 
56 #define TDES_FLAGS_INIT		BIT(3)
57 #define TDES_FLAGS_FAST		BIT(4)
58 #define TDES_FLAGS_BUSY		BIT(5)
59 #define TDES_FLAGS_DMA		BIT(6)
60 
61 #define ATMEL_TDES_QUEUE_LENGTH	50
62 
63 #define CFB8_BLOCK_SIZE		1
64 #define CFB16_BLOCK_SIZE	2
65 #define CFB32_BLOCK_SIZE	4
66 
67 struct atmel_tdes_caps {
68 	bool	has_dma;
69 	u32		has_cfb_3keys;
70 };
71 
72 struct atmel_tdes_dev;
73 
74 struct atmel_tdes_ctx {
75 	struct atmel_tdes_dev *dd;
76 
77 	int		keylen;
78 	u32		key[DES3_EDE_KEY_SIZE / sizeof(u32)];
79 	unsigned long	flags;
80 
81 	u16		block_size;
82 };
83 
84 struct atmel_tdes_reqctx {
85 	unsigned long mode;
86 	u8 lastc[DES_BLOCK_SIZE];
87 };
88 
89 struct atmel_tdes_dma {
90 	struct dma_chan			*chan;
91 	struct dma_slave_config dma_conf;
92 };
93 
94 struct atmel_tdes_dev {
95 	struct list_head	list;
96 	unsigned long		phys_base;
97 	void __iomem		*io_base;
98 
99 	struct atmel_tdes_ctx	*ctx;
100 	struct device		*dev;
101 	struct clk			*iclk;
102 	int					irq;
103 
104 	unsigned long		flags;
105 
106 	spinlock_t		lock;
107 	struct crypto_queue	queue;
108 
109 	struct tasklet_struct	done_task;
110 	struct tasklet_struct	queue_task;
111 
112 	struct skcipher_request	*req;
113 	size_t				total;
114 
115 	struct scatterlist	*in_sg;
116 	unsigned int		nb_in_sg;
117 	size_t				in_offset;
118 	struct scatterlist	*out_sg;
119 	unsigned int		nb_out_sg;
120 	size_t				out_offset;
121 
122 	size_t	buflen;
123 	size_t	dma_size;
124 
125 	void	*buf_in;
126 	int		dma_in;
127 	dma_addr_t	dma_addr_in;
128 	struct atmel_tdes_dma	dma_lch_in;
129 
130 	void	*buf_out;
131 	int		dma_out;
132 	dma_addr_t	dma_addr_out;
133 	struct atmel_tdes_dma	dma_lch_out;
134 
135 	struct atmel_tdes_caps	caps;
136 
137 	u32	hw_version;
138 };
139 
140 struct atmel_tdes_drv {
141 	struct list_head	dev_list;
142 	spinlock_t		lock;
143 };
144 
145 static struct atmel_tdes_drv atmel_tdes = {
146 	.dev_list = LIST_HEAD_INIT(atmel_tdes.dev_list),
147 	.lock = __SPIN_LOCK_UNLOCKED(atmel_tdes.lock),
148 };
149 
150 static int atmel_tdes_sg_copy(struct scatterlist **sg, size_t *offset,
151 			void *buf, size_t buflen, size_t total, int out)
152 {
153 	size_t count, off = 0;
154 
155 	while (buflen && total) {
156 		count = min((*sg)->length - *offset, total);
157 		count = min(count, buflen);
158 
159 		if (!count)
160 			return off;
161 
162 		scatterwalk_map_and_copy(buf + off, *sg, *offset, count, out);
163 
164 		off += count;
165 		buflen -= count;
166 		*offset += count;
167 		total -= count;
168 
169 		if (*offset == (*sg)->length) {
170 			*sg = sg_next(*sg);
171 			if (*sg)
172 				*offset = 0;
173 			else
174 				total = 0;
175 		}
176 	}
177 
178 	return off;
179 }
180 
181 static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset)
182 {
183 	return readl_relaxed(dd->io_base + offset);
184 }
185 
186 static inline void atmel_tdes_write(struct atmel_tdes_dev *dd,
187 					u32 offset, u32 value)
188 {
189 	writel_relaxed(value, dd->io_base + offset);
190 }
191 
192 static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset,
193 			       const u32 *value, int count)
194 {
195 	for (; count--; value++, offset += 4)
196 		atmel_tdes_write(dd, offset, *value);
197 }
198 
199 static struct atmel_tdes_dev *atmel_tdes_find_dev(struct atmel_tdes_ctx *ctx)
200 {
201 	struct atmel_tdes_dev *tdes_dd = NULL;
202 	struct atmel_tdes_dev *tmp;
203 
204 	spin_lock_bh(&atmel_tdes.lock);
205 	if (!ctx->dd) {
206 		list_for_each_entry(tmp, &atmel_tdes.dev_list, list) {
207 			tdes_dd = tmp;
208 			break;
209 		}
210 		ctx->dd = tdes_dd;
211 	} else {
212 		tdes_dd = ctx->dd;
213 	}
214 	spin_unlock_bh(&atmel_tdes.lock);
215 
216 	return tdes_dd;
217 }
218 
219 static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd)
220 {
221 	int err;
222 
223 	err = clk_prepare_enable(dd->iclk);
224 	if (err)
225 		return err;
226 
227 	if (!(dd->flags & TDES_FLAGS_INIT)) {
228 		atmel_tdes_write(dd, TDES_CR, TDES_CR_SWRST);
229 		dd->flags |= TDES_FLAGS_INIT;
230 	}
231 
232 	return 0;
233 }
234 
235 static inline unsigned int atmel_tdes_get_version(struct atmel_tdes_dev *dd)
236 {
237 	return atmel_tdes_read(dd, TDES_HW_VERSION) & 0x00000fff;
238 }
239 
240 static int atmel_tdes_hw_version_init(struct atmel_tdes_dev *dd)
241 {
242 	int err;
243 
244 	err = atmel_tdes_hw_init(dd);
245 	if (err)
246 		return err;
247 
248 	dd->hw_version = atmel_tdes_get_version(dd);
249 
250 	dev_info(dd->dev,
251 			"version: 0x%x\n", dd->hw_version);
252 
253 	clk_disable_unprepare(dd->iclk);
254 
255 	return 0;
256 }
257 
258 static void atmel_tdes_dma_callback(void *data)
259 {
260 	struct atmel_tdes_dev *dd = data;
261 
262 	/* dma_lch_out - completed */
263 	tasklet_schedule(&dd->done_task);
264 }
265 
266 static int atmel_tdes_write_ctrl(struct atmel_tdes_dev *dd)
267 {
268 	int err;
269 	u32 valmr = TDES_MR_SMOD_PDC;
270 
271 	err = atmel_tdes_hw_init(dd);
272 
273 	if (err)
274 		return err;
275 
276 	if (!dd->caps.has_dma)
277 		atmel_tdes_write(dd, TDES_PTCR,
278 			TDES_PTCR_TXTDIS | TDES_PTCR_RXTDIS);
279 
280 	/* MR register must be set before IV registers */
281 	if (dd->ctx->keylen > (DES_KEY_SIZE << 1)) {
282 		valmr |= TDES_MR_KEYMOD_3KEY;
283 		valmr |= TDES_MR_TDESMOD_TDES;
284 	} else if (dd->ctx->keylen > DES_KEY_SIZE) {
285 		valmr |= TDES_MR_KEYMOD_2KEY;
286 		valmr |= TDES_MR_TDESMOD_TDES;
287 	} else {
288 		valmr |= TDES_MR_TDESMOD_DES;
289 	}
290 
291 	valmr |= dd->flags & TDES_FLAGS_MODE_MASK;
292 
293 	atmel_tdes_write(dd, TDES_MR, valmr);
294 
295 	atmel_tdes_write_n(dd, TDES_KEY1W1R, dd->ctx->key,
296 						dd->ctx->keylen >> 2);
297 
298 	if (dd->req->iv && (valmr & TDES_MR_OPMOD_MASK) != TDES_MR_OPMOD_ECB)
299 		atmel_tdes_write_n(dd, TDES_IV1R, (void *)dd->req->iv, 2);
300 
301 	return 0;
302 }
303 
304 static int atmel_tdes_crypt_pdc_stop(struct atmel_tdes_dev *dd)
305 {
306 	int err = 0;
307 	size_t count;
308 
309 	atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
310 
311 	if (dd->flags & TDES_FLAGS_FAST) {
312 		dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
313 		dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
314 	} else {
315 		dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
316 					   dd->dma_size, DMA_FROM_DEVICE);
317 
318 		/* copy data */
319 		count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
320 				dd->buf_out, dd->buflen, dd->dma_size, 1);
321 		if (count != dd->dma_size) {
322 			err = -EINVAL;
323 			pr_err("not all data converted: %zu\n", count);
324 		}
325 	}
326 
327 	return err;
328 }
329 
330 static int atmel_tdes_buff_init(struct atmel_tdes_dev *dd)
331 {
332 	int err = -ENOMEM;
333 
334 	dd->buf_in = (void *)__get_free_pages(GFP_KERNEL, 0);
335 	dd->buf_out = (void *)__get_free_pages(GFP_KERNEL, 0);
336 	dd->buflen = PAGE_SIZE;
337 	dd->buflen &= ~(DES_BLOCK_SIZE - 1);
338 
339 	if (!dd->buf_in || !dd->buf_out) {
340 		dev_err(dd->dev, "unable to alloc pages.\n");
341 		goto err_alloc;
342 	}
343 
344 	/* MAP here */
345 	dd->dma_addr_in = dma_map_single(dd->dev, dd->buf_in,
346 					dd->buflen, DMA_TO_DEVICE);
347 	if (dma_mapping_error(dd->dev, dd->dma_addr_in)) {
348 		dev_err(dd->dev, "dma %zd bytes error\n", dd->buflen);
349 		err = -EINVAL;
350 		goto err_map_in;
351 	}
352 
353 	dd->dma_addr_out = dma_map_single(dd->dev, dd->buf_out,
354 					dd->buflen, DMA_FROM_DEVICE);
355 	if (dma_mapping_error(dd->dev, dd->dma_addr_out)) {
356 		dev_err(dd->dev, "dma %zd bytes error\n", dd->buflen);
357 		err = -EINVAL;
358 		goto err_map_out;
359 	}
360 
361 	return 0;
362 
363 err_map_out:
364 	dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
365 		DMA_TO_DEVICE);
366 err_map_in:
367 err_alloc:
368 	free_page((unsigned long)dd->buf_out);
369 	free_page((unsigned long)dd->buf_in);
370 	if (err)
371 		pr_err("error: %d\n", err);
372 	return err;
373 }
374 
375 static void atmel_tdes_buff_cleanup(struct atmel_tdes_dev *dd)
376 {
377 	dma_unmap_single(dd->dev, dd->dma_addr_out, dd->buflen,
378 			 DMA_FROM_DEVICE);
379 	dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen,
380 		DMA_TO_DEVICE);
381 	free_page((unsigned long)dd->buf_out);
382 	free_page((unsigned long)dd->buf_in);
383 }
384 
385 static int atmel_tdes_crypt_pdc(struct atmel_tdes_dev *dd,
386 				dma_addr_t dma_addr_in,
387 				dma_addr_t dma_addr_out, int length)
388 {
389 	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(dd->req);
390 	int len32;
391 
392 	dd->dma_size = length;
393 
394 	if (!(dd->flags & TDES_FLAGS_FAST)) {
395 		dma_sync_single_for_device(dd->dev, dma_addr_in, length,
396 					   DMA_TO_DEVICE);
397 	}
398 
399 	switch (rctx->mode & TDES_FLAGS_OPMODE_MASK) {
400 	case TDES_FLAGS_CFB8:
401 		len32 = DIV_ROUND_UP(length, sizeof(u8));
402 		break;
403 
404 	case TDES_FLAGS_CFB16:
405 		len32 = DIV_ROUND_UP(length, sizeof(u16));
406 		break;
407 
408 	default:
409 		len32 = DIV_ROUND_UP(length, sizeof(u32));
410 		break;
411 	}
412 
413 	atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS);
414 	atmel_tdes_write(dd, TDES_TPR, dma_addr_in);
415 	atmel_tdes_write(dd, TDES_TCR, len32);
416 	atmel_tdes_write(dd, TDES_RPR, dma_addr_out);
417 	atmel_tdes_write(dd, TDES_RCR, len32);
418 
419 	/* Enable Interrupt */
420 	atmel_tdes_write(dd, TDES_IER, TDES_INT_ENDRX);
421 
422 	/* Start DMA transfer */
423 	atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTEN | TDES_PTCR_RXTEN);
424 
425 	return 0;
426 }
427 
428 static int atmel_tdes_crypt_dma(struct atmel_tdes_dev *dd,
429 				dma_addr_t dma_addr_in,
430 				dma_addr_t dma_addr_out, int length)
431 {
432 	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(dd->req);
433 	struct scatterlist sg[2];
434 	struct dma_async_tx_descriptor	*in_desc, *out_desc;
435 	enum dma_slave_buswidth addr_width;
436 
437 	dd->dma_size = length;
438 
439 	if (!(dd->flags & TDES_FLAGS_FAST)) {
440 		dma_sync_single_for_device(dd->dev, dma_addr_in, length,
441 					   DMA_TO_DEVICE);
442 	}
443 
444 	switch (rctx->mode & TDES_FLAGS_OPMODE_MASK) {
445 	case TDES_FLAGS_CFB8:
446 		addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
447 		break;
448 
449 	case TDES_FLAGS_CFB16:
450 		addr_width = DMA_SLAVE_BUSWIDTH_2_BYTES;
451 		break;
452 
453 	default:
454 		addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
455 		break;
456 	}
457 
458 	dd->dma_lch_in.dma_conf.dst_addr_width = addr_width;
459 	dd->dma_lch_out.dma_conf.src_addr_width = addr_width;
460 
461 	dmaengine_slave_config(dd->dma_lch_in.chan, &dd->dma_lch_in.dma_conf);
462 	dmaengine_slave_config(dd->dma_lch_out.chan, &dd->dma_lch_out.dma_conf);
463 
464 	dd->flags |= TDES_FLAGS_DMA;
465 
466 	sg_init_table(&sg[0], 1);
467 	sg_dma_address(&sg[0]) = dma_addr_in;
468 	sg_dma_len(&sg[0]) = length;
469 
470 	sg_init_table(&sg[1], 1);
471 	sg_dma_address(&sg[1]) = dma_addr_out;
472 	sg_dma_len(&sg[1]) = length;
473 
474 	in_desc = dmaengine_prep_slave_sg(dd->dma_lch_in.chan, &sg[0],
475 				1, DMA_MEM_TO_DEV,
476 				DMA_PREP_INTERRUPT  |  DMA_CTRL_ACK);
477 	if (!in_desc)
478 		return -EINVAL;
479 
480 	out_desc = dmaengine_prep_slave_sg(dd->dma_lch_out.chan, &sg[1],
481 				1, DMA_DEV_TO_MEM,
482 				DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
483 	if (!out_desc)
484 		return -EINVAL;
485 
486 	out_desc->callback = atmel_tdes_dma_callback;
487 	out_desc->callback_param = dd;
488 
489 	dmaengine_submit(out_desc);
490 	dma_async_issue_pending(dd->dma_lch_out.chan);
491 
492 	dmaengine_submit(in_desc);
493 	dma_async_issue_pending(dd->dma_lch_in.chan);
494 
495 	return 0;
496 }
497 
498 static int atmel_tdes_crypt_start(struct atmel_tdes_dev *dd)
499 {
500 	int err, fast = 0, in, out;
501 	size_t count;
502 	dma_addr_t addr_in, addr_out;
503 
504 	if ((!dd->in_offset) && (!dd->out_offset)) {
505 		/* check for alignment */
506 		in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) &&
507 			IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size);
508 		out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) &&
509 			IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size);
510 		fast = in && out;
511 
512 		if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg))
513 			fast = 0;
514 	}
515 
516 
517 	if (fast)  {
518 		count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg));
519 		count = min_t(size_t, count, sg_dma_len(dd->out_sg));
520 
521 		err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
522 		if (!err) {
523 			dev_err(dd->dev, "dma_map_sg() error\n");
524 			return -EINVAL;
525 		}
526 
527 		err = dma_map_sg(dd->dev, dd->out_sg, 1,
528 				DMA_FROM_DEVICE);
529 		if (!err) {
530 			dev_err(dd->dev, "dma_map_sg() error\n");
531 			dma_unmap_sg(dd->dev, dd->in_sg, 1,
532 				DMA_TO_DEVICE);
533 			return -EINVAL;
534 		}
535 
536 		addr_in = sg_dma_address(dd->in_sg);
537 		addr_out = sg_dma_address(dd->out_sg);
538 
539 		dd->flags |= TDES_FLAGS_FAST;
540 
541 	} else {
542 		/* use cache buffers */
543 		count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset,
544 				dd->buf_in, dd->buflen, dd->total, 0);
545 
546 		addr_in = dd->dma_addr_in;
547 		addr_out = dd->dma_addr_out;
548 
549 		dd->flags &= ~TDES_FLAGS_FAST;
550 	}
551 
552 	dd->total -= count;
553 
554 	if (dd->caps.has_dma)
555 		err = atmel_tdes_crypt_dma(dd, addr_in, addr_out, count);
556 	else
557 		err = atmel_tdes_crypt_pdc(dd, addr_in, addr_out, count);
558 
559 	if (err && (dd->flags & TDES_FLAGS_FAST)) {
560 		dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
561 		dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE);
562 	}
563 
564 	return err;
565 }
566 
567 static void
568 atmel_tdes_set_iv_as_last_ciphertext_block(struct atmel_tdes_dev *dd)
569 {
570 	struct skcipher_request *req = dd->req;
571 	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
572 	struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
573 	unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
574 
575 	if (req->cryptlen < ivsize)
576 		return;
577 
578 	if (rctx->mode & TDES_FLAGS_ENCRYPT) {
579 		scatterwalk_map_and_copy(req->iv, req->dst,
580 					 req->cryptlen - ivsize, ivsize, 0);
581 	} else {
582 		if (req->src == req->dst)
583 			memcpy(req->iv, rctx->lastc, ivsize);
584 		else
585 			scatterwalk_map_and_copy(req->iv, req->src,
586 						 req->cryptlen - ivsize,
587 						 ivsize, 0);
588 	}
589 }
590 
591 static void atmel_tdes_finish_req(struct atmel_tdes_dev *dd, int err)
592 {
593 	struct skcipher_request *req = dd->req;
594 	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
595 
596 	clk_disable_unprepare(dd->iclk);
597 
598 	dd->flags &= ~TDES_FLAGS_BUSY;
599 
600 	if (!err && (rctx->mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB)
601 		atmel_tdes_set_iv_as_last_ciphertext_block(dd);
602 
603 	req->base.complete(&req->base, err);
604 }
605 
606 static int atmel_tdes_handle_queue(struct atmel_tdes_dev *dd,
607 			       struct skcipher_request *req)
608 {
609 	struct crypto_async_request *async_req, *backlog;
610 	struct atmel_tdes_ctx *ctx;
611 	struct atmel_tdes_reqctx *rctx;
612 	unsigned long flags;
613 	int err, ret = 0;
614 
615 	spin_lock_irqsave(&dd->lock, flags);
616 	if (req)
617 		ret = crypto_enqueue_request(&dd->queue, &req->base);
618 	if (dd->flags & TDES_FLAGS_BUSY) {
619 		spin_unlock_irqrestore(&dd->lock, flags);
620 		return ret;
621 	}
622 	backlog = crypto_get_backlog(&dd->queue);
623 	async_req = crypto_dequeue_request(&dd->queue);
624 	if (async_req)
625 		dd->flags |= TDES_FLAGS_BUSY;
626 	spin_unlock_irqrestore(&dd->lock, flags);
627 
628 	if (!async_req)
629 		return ret;
630 
631 	if (backlog)
632 		backlog->complete(backlog, -EINPROGRESS);
633 
634 	req = skcipher_request_cast(async_req);
635 
636 	/* assign new request to device */
637 	dd->req = req;
638 	dd->total = req->cryptlen;
639 	dd->in_offset = 0;
640 	dd->in_sg = req->src;
641 	dd->out_offset = 0;
642 	dd->out_sg = req->dst;
643 
644 	rctx = skcipher_request_ctx(req);
645 	ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req));
646 	rctx->mode &= TDES_FLAGS_MODE_MASK;
647 	dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode;
648 	dd->ctx = ctx;
649 	ctx->dd = dd;
650 
651 	err = atmel_tdes_write_ctrl(dd);
652 	if (!err)
653 		err = atmel_tdes_crypt_start(dd);
654 	if (err) {
655 		/* des_task will not finish it, so do it here */
656 		atmel_tdes_finish_req(dd, err);
657 		tasklet_schedule(&dd->queue_task);
658 	}
659 
660 	return ret;
661 }
662 
663 static int atmel_tdes_crypt_dma_stop(struct atmel_tdes_dev *dd)
664 {
665 	int err = -EINVAL;
666 	size_t count;
667 
668 	if (dd->flags & TDES_FLAGS_DMA) {
669 		err = 0;
670 		if  (dd->flags & TDES_FLAGS_FAST) {
671 			dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE);
672 			dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE);
673 		} else {
674 			dma_sync_single_for_device(dd->dev, dd->dma_addr_out,
675 				dd->dma_size, DMA_FROM_DEVICE);
676 
677 			/* copy data */
678 			count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset,
679 				dd->buf_out, dd->buflen, dd->dma_size, 1);
680 			if (count != dd->dma_size) {
681 				err = -EINVAL;
682 				pr_err("not all data converted: %zu\n", count);
683 			}
684 		}
685 	}
686 	return err;
687 }
688 
689 static int atmel_tdes_crypt(struct skcipher_request *req, unsigned long mode)
690 {
691 	struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req);
692 	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(skcipher);
693 	struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req);
694 
695 	switch (mode & TDES_FLAGS_OPMODE_MASK) {
696 	case TDES_FLAGS_CFB8:
697 		if (!IS_ALIGNED(req->cryptlen, CFB8_BLOCK_SIZE)) {
698 			pr_err("request size is not exact amount of CFB8 blocks\n");
699 			return -EINVAL;
700 		}
701 		ctx->block_size = CFB8_BLOCK_SIZE;
702 		break;
703 
704 	case TDES_FLAGS_CFB16:
705 		if (!IS_ALIGNED(req->cryptlen, CFB16_BLOCK_SIZE)) {
706 			pr_err("request size is not exact amount of CFB16 blocks\n");
707 			return -EINVAL;
708 		}
709 		ctx->block_size = CFB16_BLOCK_SIZE;
710 		break;
711 
712 	case TDES_FLAGS_CFB32:
713 		if (!IS_ALIGNED(req->cryptlen, CFB32_BLOCK_SIZE)) {
714 			pr_err("request size is not exact amount of CFB32 blocks\n");
715 			return -EINVAL;
716 		}
717 		ctx->block_size = CFB32_BLOCK_SIZE;
718 		break;
719 
720 	default:
721 		if (!IS_ALIGNED(req->cryptlen, DES_BLOCK_SIZE)) {
722 			pr_err("request size is not exact amount of DES blocks\n");
723 			return -EINVAL;
724 		}
725 		ctx->block_size = DES_BLOCK_SIZE;
726 		break;
727 	}
728 
729 	rctx->mode = mode;
730 
731 	if ((mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB &&
732 	    !(mode & TDES_FLAGS_ENCRYPT) && req->src == req->dst) {
733 		unsigned int ivsize = crypto_skcipher_ivsize(skcipher);
734 
735 		if (req->cryptlen >= ivsize)
736 			scatterwalk_map_and_copy(rctx->lastc, req->src,
737 						 req->cryptlen - ivsize,
738 						 ivsize, 0);
739 	}
740 
741 	return atmel_tdes_handle_queue(ctx->dd, req);
742 }
743 
744 static int atmel_tdes_dma_init(struct atmel_tdes_dev *dd)
745 {
746 	int ret;
747 
748 	/* Try to grab 2 DMA channels */
749 	dd->dma_lch_in.chan = dma_request_chan(dd->dev, "tx");
750 	if (IS_ERR(dd->dma_lch_in.chan)) {
751 		ret = PTR_ERR(dd->dma_lch_in.chan);
752 		goto err_dma_in;
753 	}
754 
755 	dd->dma_lch_in.dma_conf.dst_addr = dd->phys_base +
756 		TDES_IDATA1R;
757 	dd->dma_lch_in.dma_conf.src_maxburst = 1;
758 	dd->dma_lch_in.dma_conf.src_addr_width =
759 		DMA_SLAVE_BUSWIDTH_4_BYTES;
760 	dd->dma_lch_in.dma_conf.dst_maxburst = 1;
761 	dd->dma_lch_in.dma_conf.dst_addr_width =
762 		DMA_SLAVE_BUSWIDTH_4_BYTES;
763 	dd->dma_lch_in.dma_conf.device_fc = false;
764 
765 	dd->dma_lch_out.chan = dma_request_chan(dd->dev, "rx");
766 	if (IS_ERR(dd->dma_lch_out.chan)) {
767 		ret = PTR_ERR(dd->dma_lch_out.chan);
768 		goto err_dma_out;
769 	}
770 
771 	dd->dma_lch_out.dma_conf.src_addr = dd->phys_base +
772 		TDES_ODATA1R;
773 	dd->dma_lch_out.dma_conf.src_maxburst = 1;
774 	dd->dma_lch_out.dma_conf.src_addr_width =
775 		DMA_SLAVE_BUSWIDTH_4_BYTES;
776 	dd->dma_lch_out.dma_conf.dst_maxburst = 1;
777 	dd->dma_lch_out.dma_conf.dst_addr_width =
778 		DMA_SLAVE_BUSWIDTH_4_BYTES;
779 	dd->dma_lch_out.dma_conf.device_fc = false;
780 
781 	return 0;
782 
783 err_dma_out:
784 	dma_release_channel(dd->dma_lch_in.chan);
785 err_dma_in:
786 	dev_err(dd->dev, "no DMA channel available\n");
787 	return ret;
788 }
789 
790 static void atmel_tdes_dma_cleanup(struct atmel_tdes_dev *dd)
791 {
792 	dma_release_channel(dd->dma_lch_in.chan);
793 	dma_release_channel(dd->dma_lch_out.chan);
794 }
795 
796 static int atmel_des_setkey(struct crypto_skcipher *tfm, const u8 *key,
797 			   unsigned int keylen)
798 {
799 	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
800 	int err;
801 
802 	err = verify_skcipher_des_key(tfm, key);
803 	if (err)
804 		return err;
805 
806 	memcpy(ctx->key, key, keylen);
807 	ctx->keylen = keylen;
808 
809 	return 0;
810 }
811 
812 static int atmel_tdes_setkey(struct crypto_skcipher *tfm, const u8 *key,
813 			   unsigned int keylen)
814 {
815 	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
816 	int err;
817 
818 	err = verify_skcipher_des3_key(tfm, key);
819 	if (err)
820 		return err;
821 
822 	memcpy(ctx->key, key, keylen);
823 	ctx->keylen = keylen;
824 
825 	return 0;
826 }
827 
828 static int atmel_tdes_ecb_encrypt(struct skcipher_request *req)
829 {
830 	return atmel_tdes_crypt(req, TDES_FLAGS_ECB | TDES_FLAGS_ENCRYPT);
831 }
832 
833 static int atmel_tdes_ecb_decrypt(struct skcipher_request *req)
834 {
835 	return atmel_tdes_crypt(req, TDES_FLAGS_ECB);
836 }
837 
838 static int atmel_tdes_cbc_encrypt(struct skcipher_request *req)
839 {
840 	return atmel_tdes_crypt(req, TDES_FLAGS_CBC | TDES_FLAGS_ENCRYPT);
841 }
842 
843 static int atmel_tdes_cbc_decrypt(struct skcipher_request *req)
844 {
845 	return atmel_tdes_crypt(req, TDES_FLAGS_CBC);
846 }
847 static int atmel_tdes_cfb_encrypt(struct skcipher_request *req)
848 {
849 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB64 | TDES_FLAGS_ENCRYPT);
850 }
851 
852 static int atmel_tdes_cfb_decrypt(struct skcipher_request *req)
853 {
854 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB64);
855 }
856 
857 static int atmel_tdes_cfb8_encrypt(struct skcipher_request *req)
858 {
859 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB8 | TDES_FLAGS_ENCRYPT);
860 }
861 
862 static int atmel_tdes_cfb8_decrypt(struct skcipher_request *req)
863 {
864 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB8);
865 }
866 
867 static int atmel_tdes_cfb16_encrypt(struct skcipher_request *req)
868 {
869 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB16 | TDES_FLAGS_ENCRYPT);
870 }
871 
872 static int atmel_tdes_cfb16_decrypt(struct skcipher_request *req)
873 {
874 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB16);
875 }
876 
877 static int atmel_tdes_cfb32_encrypt(struct skcipher_request *req)
878 {
879 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB32 | TDES_FLAGS_ENCRYPT);
880 }
881 
882 static int atmel_tdes_cfb32_decrypt(struct skcipher_request *req)
883 {
884 	return atmel_tdes_crypt(req, TDES_FLAGS_CFB32);
885 }
886 
887 static int atmel_tdes_ofb_encrypt(struct skcipher_request *req)
888 {
889 	return atmel_tdes_crypt(req, TDES_FLAGS_OFB | TDES_FLAGS_ENCRYPT);
890 }
891 
892 static int atmel_tdes_ofb_decrypt(struct skcipher_request *req)
893 {
894 	return atmel_tdes_crypt(req, TDES_FLAGS_OFB);
895 }
896 
897 static int atmel_tdes_init_tfm(struct crypto_skcipher *tfm)
898 {
899 	struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm);
900 	struct atmel_tdes_dev *dd;
901 
902 	crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_tdes_reqctx));
903 
904 	dd = atmel_tdes_find_dev(ctx);
905 	if (!dd)
906 		return -ENODEV;
907 
908 	return 0;
909 }
910 
911 static void atmel_tdes_skcipher_alg_init(struct skcipher_alg *alg)
912 {
913 	alg->base.cra_priority = ATMEL_TDES_PRIORITY;
914 	alg->base.cra_flags = CRYPTO_ALG_ASYNC;
915 	alg->base.cra_ctxsize = sizeof(struct atmel_tdes_ctx);
916 	alg->base.cra_module = THIS_MODULE;
917 
918 	alg->init = atmel_tdes_init_tfm;
919 }
920 
921 static struct skcipher_alg tdes_algs[] = {
922 {
923 	.base.cra_name		= "ecb(des)",
924 	.base.cra_driver_name	= "atmel-ecb-des",
925 	.base.cra_blocksize	= DES_BLOCK_SIZE,
926 	.base.cra_alignmask	= 0x7,
927 
928 	.min_keysize		= DES_KEY_SIZE,
929 	.max_keysize		= DES_KEY_SIZE,
930 	.setkey			= atmel_des_setkey,
931 	.encrypt		= atmel_tdes_ecb_encrypt,
932 	.decrypt		= atmel_tdes_ecb_decrypt,
933 },
934 {
935 	.base.cra_name		= "cbc(des)",
936 	.base.cra_driver_name	= "atmel-cbc-des",
937 	.base.cra_blocksize	= DES_BLOCK_SIZE,
938 	.base.cra_alignmask	= 0x7,
939 
940 	.min_keysize		= DES_KEY_SIZE,
941 	.max_keysize		= DES_KEY_SIZE,
942 	.ivsize			= DES_BLOCK_SIZE,
943 	.setkey			= atmel_des_setkey,
944 	.encrypt		= atmel_tdes_cbc_encrypt,
945 	.decrypt		= atmel_tdes_cbc_decrypt,
946 },
947 {
948 	.base.cra_name		= "cfb(des)",
949 	.base.cra_driver_name	= "atmel-cfb-des",
950 	.base.cra_blocksize	= DES_BLOCK_SIZE,
951 	.base.cra_alignmask	= 0x7,
952 
953 	.min_keysize		= DES_KEY_SIZE,
954 	.max_keysize		= DES_KEY_SIZE,
955 	.ivsize			= DES_BLOCK_SIZE,
956 	.setkey			= atmel_des_setkey,
957 	.encrypt		= atmel_tdes_cfb_encrypt,
958 	.decrypt		= atmel_tdes_cfb_decrypt,
959 },
960 {
961 	.base.cra_name		= "cfb8(des)",
962 	.base.cra_driver_name	= "atmel-cfb8-des",
963 	.base.cra_blocksize	= CFB8_BLOCK_SIZE,
964 	.base.cra_alignmask	= 0,
965 
966 	.min_keysize		= DES_KEY_SIZE,
967 	.max_keysize		= DES_KEY_SIZE,
968 	.ivsize			= DES_BLOCK_SIZE,
969 	.setkey			= atmel_des_setkey,
970 	.encrypt		= atmel_tdes_cfb8_encrypt,
971 	.decrypt		= atmel_tdes_cfb8_decrypt,
972 },
973 {
974 	.base.cra_name		= "cfb16(des)",
975 	.base.cra_driver_name	= "atmel-cfb16-des",
976 	.base.cra_blocksize	= CFB16_BLOCK_SIZE,
977 	.base.cra_alignmask	= 0x1,
978 
979 	.min_keysize		= DES_KEY_SIZE,
980 	.max_keysize		= DES_KEY_SIZE,
981 	.ivsize			= DES_BLOCK_SIZE,
982 	.setkey			= atmel_des_setkey,
983 	.encrypt		= atmel_tdes_cfb16_encrypt,
984 	.decrypt		= atmel_tdes_cfb16_decrypt,
985 },
986 {
987 	.base.cra_name		= "cfb32(des)",
988 	.base.cra_driver_name	= "atmel-cfb32-des",
989 	.base.cra_blocksize	= CFB32_BLOCK_SIZE,
990 	.base.cra_alignmask	= 0x3,
991 
992 	.min_keysize		= DES_KEY_SIZE,
993 	.max_keysize		= DES_KEY_SIZE,
994 	.ivsize			= DES_BLOCK_SIZE,
995 	.setkey			= atmel_des_setkey,
996 	.encrypt		= atmel_tdes_cfb32_encrypt,
997 	.decrypt		= atmel_tdes_cfb32_decrypt,
998 },
999 {
1000 	.base.cra_name		= "ofb(des)",
1001 	.base.cra_driver_name	= "atmel-ofb-des",
1002 	.base.cra_blocksize	= DES_BLOCK_SIZE,
1003 	.base.cra_alignmask	= 0x7,
1004 
1005 	.min_keysize		= DES_KEY_SIZE,
1006 	.max_keysize		= DES_KEY_SIZE,
1007 	.ivsize			= DES_BLOCK_SIZE,
1008 	.setkey			= atmel_des_setkey,
1009 	.encrypt		= atmel_tdes_ofb_encrypt,
1010 	.decrypt		= atmel_tdes_ofb_decrypt,
1011 },
1012 {
1013 	.base.cra_name		= "ecb(des3_ede)",
1014 	.base.cra_driver_name	= "atmel-ecb-tdes",
1015 	.base.cra_blocksize	= DES_BLOCK_SIZE,
1016 	.base.cra_alignmask	= 0x7,
1017 
1018 	.min_keysize		= DES3_EDE_KEY_SIZE,
1019 	.max_keysize		= DES3_EDE_KEY_SIZE,
1020 	.setkey			= atmel_tdes_setkey,
1021 	.encrypt		= atmel_tdes_ecb_encrypt,
1022 	.decrypt		= atmel_tdes_ecb_decrypt,
1023 },
1024 {
1025 	.base.cra_name		= "cbc(des3_ede)",
1026 	.base.cra_driver_name	= "atmel-cbc-tdes",
1027 	.base.cra_blocksize	= DES_BLOCK_SIZE,
1028 	.base.cra_alignmask	= 0x7,
1029 
1030 	.min_keysize		= DES3_EDE_KEY_SIZE,
1031 	.max_keysize		= DES3_EDE_KEY_SIZE,
1032 	.setkey			= atmel_tdes_setkey,
1033 	.encrypt		= atmel_tdes_cbc_encrypt,
1034 	.decrypt		= atmel_tdes_cbc_decrypt,
1035 	.ivsize			= DES_BLOCK_SIZE,
1036 },
1037 {
1038 	.base.cra_name		= "ofb(des3_ede)",
1039 	.base.cra_driver_name	= "atmel-ofb-tdes",
1040 	.base.cra_blocksize	= DES_BLOCK_SIZE,
1041 	.base.cra_alignmask	= 0x7,
1042 
1043 	.min_keysize		= DES3_EDE_KEY_SIZE,
1044 	.max_keysize		= DES3_EDE_KEY_SIZE,
1045 	.setkey			= atmel_tdes_setkey,
1046 	.encrypt		= atmel_tdes_ofb_encrypt,
1047 	.decrypt		= atmel_tdes_ofb_decrypt,
1048 	.ivsize			= DES_BLOCK_SIZE,
1049 },
1050 };
1051 
1052 static void atmel_tdes_queue_task(unsigned long data)
1053 {
1054 	struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *)data;
1055 
1056 	atmel_tdes_handle_queue(dd, NULL);
1057 }
1058 
1059 static void atmel_tdes_done_task(unsigned long data)
1060 {
1061 	struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *) data;
1062 	int err;
1063 
1064 	if (!(dd->flags & TDES_FLAGS_DMA))
1065 		err = atmel_tdes_crypt_pdc_stop(dd);
1066 	else
1067 		err = atmel_tdes_crypt_dma_stop(dd);
1068 
1069 	if (dd->total && !err) {
1070 		if (dd->flags & TDES_FLAGS_FAST) {
1071 			dd->in_sg = sg_next(dd->in_sg);
1072 			dd->out_sg = sg_next(dd->out_sg);
1073 			if (!dd->in_sg || !dd->out_sg)
1074 				err = -EINVAL;
1075 		}
1076 		if (!err)
1077 			err = atmel_tdes_crypt_start(dd);
1078 		if (!err)
1079 			return; /* DMA started. Not fininishing. */
1080 	}
1081 
1082 	atmel_tdes_finish_req(dd, err);
1083 	atmel_tdes_handle_queue(dd, NULL);
1084 }
1085 
1086 static irqreturn_t atmel_tdes_irq(int irq, void *dev_id)
1087 {
1088 	struct atmel_tdes_dev *tdes_dd = dev_id;
1089 	u32 reg;
1090 
1091 	reg = atmel_tdes_read(tdes_dd, TDES_ISR);
1092 	if (reg & atmel_tdes_read(tdes_dd, TDES_IMR)) {
1093 		atmel_tdes_write(tdes_dd, TDES_IDR, reg);
1094 		if (TDES_FLAGS_BUSY & tdes_dd->flags)
1095 			tasklet_schedule(&tdes_dd->done_task);
1096 		else
1097 			dev_warn(tdes_dd->dev, "TDES interrupt when no active requests.\n");
1098 		return IRQ_HANDLED;
1099 	}
1100 
1101 	return IRQ_NONE;
1102 }
1103 
1104 static void atmel_tdes_unregister_algs(struct atmel_tdes_dev *dd)
1105 {
1106 	int i;
1107 
1108 	for (i = 0; i < ARRAY_SIZE(tdes_algs); i++)
1109 		crypto_unregister_skcipher(&tdes_algs[i]);
1110 }
1111 
1112 static int atmel_tdes_register_algs(struct atmel_tdes_dev *dd)
1113 {
1114 	int err, i, j;
1115 
1116 	for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) {
1117 		atmel_tdes_skcipher_alg_init(&tdes_algs[i]);
1118 
1119 		err = crypto_register_skcipher(&tdes_algs[i]);
1120 		if (err)
1121 			goto err_tdes_algs;
1122 	}
1123 
1124 	return 0;
1125 
1126 err_tdes_algs:
1127 	for (j = 0; j < i; j++)
1128 		crypto_unregister_skcipher(&tdes_algs[j]);
1129 
1130 	return err;
1131 }
1132 
1133 static void atmel_tdes_get_cap(struct atmel_tdes_dev *dd)
1134 {
1135 
1136 	dd->caps.has_dma = 0;
1137 	dd->caps.has_cfb_3keys = 0;
1138 
1139 	/* keep only major version number */
1140 	switch (dd->hw_version & 0xf00) {
1141 	case 0x700:
1142 		dd->caps.has_dma = 1;
1143 		dd->caps.has_cfb_3keys = 1;
1144 		break;
1145 	case 0x600:
1146 		break;
1147 	default:
1148 		dev_warn(dd->dev,
1149 				"Unmanaged tdes version, set minimum capabilities\n");
1150 		break;
1151 	}
1152 }
1153 
1154 #if defined(CONFIG_OF)
1155 static const struct of_device_id atmel_tdes_dt_ids[] = {
1156 	{ .compatible = "atmel,at91sam9g46-tdes" },
1157 	{ /* sentinel */ }
1158 };
1159 MODULE_DEVICE_TABLE(of, atmel_tdes_dt_ids);
1160 #endif
1161 
1162 static int atmel_tdes_probe(struct platform_device *pdev)
1163 {
1164 	struct atmel_tdes_dev *tdes_dd;
1165 	struct device *dev = &pdev->dev;
1166 	struct resource *tdes_res;
1167 	int err;
1168 
1169 	tdes_dd = devm_kmalloc(&pdev->dev, sizeof(*tdes_dd), GFP_KERNEL);
1170 	if (!tdes_dd)
1171 		return -ENOMEM;
1172 
1173 	tdes_dd->dev = dev;
1174 
1175 	platform_set_drvdata(pdev, tdes_dd);
1176 
1177 	INIT_LIST_HEAD(&tdes_dd->list);
1178 	spin_lock_init(&tdes_dd->lock);
1179 
1180 	tasklet_init(&tdes_dd->done_task, atmel_tdes_done_task,
1181 					(unsigned long)tdes_dd);
1182 	tasklet_init(&tdes_dd->queue_task, atmel_tdes_queue_task,
1183 					(unsigned long)tdes_dd);
1184 
1185 	crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH);
1186 
1187 	/* Get the base address */
1188 	tdes_res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1189 	if (!tdes_res) {
1190 		dev_err(dev, "no MEM resource info\n");
1191 		err = -ENODEV;
1192 		goto err_tasklet_kill;
1193 	}
1194 	tdes_dd->phys_base = tdes_res->start;
1195 
1196 	/* Get the IRQ */
1197 	tdes_dd->irq = platform_get_irq(pdev,  0);
1198 	if (tdes_dd->irq < 0) {
1199 		err = tdes_dd->irq;
1200 		goto err_tasklet_kill;
1201 	}
1202 
1203 	err = devm_request_irq(&pdev->dev, tdes_dd->irq, atmel_tdes_irq,
1204 			       IRQF_SHARED, "atmel-tdes", tdes_dd);
1205 	if (err) {
1206 		dev_err(dev, "unable to request tdes irq.\n");
1207 		goto err_tasklet_kill;
1208 	}
1209 
1210 	/* Initializing the clock */
1211 	tdes_dd->iclk = devm_clk_get(&pdev->dev, "tdes_clk");
1212 	if (IS_ERR(tdes_dd->iclk)) {
1213 		dev_err(dev, "clock initialization failed.\n");
1214 		err = PTR_ERR(tdes_dd->iclk);
1215 		goto err_tasklet_kill;
1216 	}
1217 
1218 	tdes_dd->io_base = devm_ioremap_resource(&pdev->dev, tdes_res);
1219 	if (IS_ERR(tdes_dd->io_base)) {
1220 		dev_err(dev, "can't ioremap\n");
1221 		err = PTR_ERR(tdes_dd->io_base);
1222 		goto err_tasklet_kill;
1223 	}
1224 
1225 	err = atmel_tdes_hw_version_init(tdes_dd);
1226 	if (err)
1227 		goto err_tasklet_kill;
1228 
1229 	atmel_tdes_get_cap(tdes_dd);
1230 
1231 	err = atmel_tdes_buff_init(tdes_dd);
1232 	if (err)
1233 		goto err_tasklet_kill;
1234 
1235 	if (tdes_dd->caps.has_dma) {
1236 		err = atmel_tdes_dma_init(tdes_dd);
1237 		if (err)
1238 			goto err_buff_cleanup;
1239 
1240 		dev_info(dev, "using %s, %s for DMA transfers\n",
1241 				dma_chan_name(tdes_dd->dma_lch_in.chan),
1242 				dma_chan_name(tdes_dd->dma_lch_out.chan));
1243 	}
1244 
1245 	spin_lock(&atmel_tdes.lock);
1246 	list_add_tail(&tdes_dd->list, &atmel_tdes.dev_list);
1247 	spin_unlock(&atmel_tdes.lock);
1248 
1249 	err = atmel_tdes_register_algs(tdes_dd);
1250 	if (err)
1251 		goto err_algs;
1252 
1253 	dev_info(dev, "Atmel DES/TDES\n");
1254 
1255 	return 0;
1256 
1257 err_algs:
1258 	spin_lock(&atmel_tdes.lock);
1259 	list_del(&tdes_dd->list);
1260 	spin_unlock(&atmel_tdes.lock);
1261 	if (tdes_dd->caps.has_dma)
1262 		atmel_tdes_dma_cleanup(tdes_dd);
1263 err_buff_cleanup:
1264 	atmel_tdes_buff_cleanup(tdes_dd);
1265 err_tasklet_kill:
1266 	tasklet_kill(&tdes_dd->done_task);
1267 	tasklet_kill(&tdes_dd->queue_task);
1268 
1269 	return err;
1270 }
1271 
1272 static int atmel_tdes_remove(struct platform_device *pdev)
1273 {
1274 	struct atmel_tdes_dev *tdes_dd;
1275 
1276 	tdes_dd = platform_get_drvdata(pdev);
1277 	if (!tdes_dd)
1278 		return -ENODEV;
1279 	spin_lock(&atmel_tdes.lock);
1280 	list_del(&tdes_dd->list);
1281 	spin_unlock(&atmel_tdes.lock);
1282 
1283 	atmel_tdes_unregister_algs(tdes_dd);
1284 
1285 	tasklet_kill(&tdes_dd->done_task);
1286 	tasklet_kill(&tdes_dd->queue_task);
1287 
1288 	if (tdes_dd->caps.has_dma)
1289 		atmel_tdes_dma_cleanup(tdes_dd);
1290 
1291 	atmel_tdes_buff_cleanup(tdes_dd);
1292 
1293 	return 0;
1294 }
1295 
1296 static struct platform_driver atmel_tdes_driver = {
1297 	.probe		= atmel_tdes_probe,
1298 	.remove		= atmel_tdes_remove,
1299 	.driver		= {
1300 		.name	= "atmel_tdes",
1301 		.of_match_table = of_match_ptr(atmel_tdes_dt_ids),
1302 	},
1303 };
1304 
1305 module_platform_driver(atmel_tdes_driver);
1306 
1307 MODULE_DESCRIPTION("Atmel DES/TDES hw acceleration support.");
1308 MODULE_LICENSE("GPL v2");
1309 MODULE_AUTHOR("Nicolas Royer - Eukréa Electromatique");
1310