1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Cryptographic API. 4 * 5 * Support for ATMEL DES/TDES HW acceleration. 6 * 7 * Copyright (c) 2012 Eukréa Electromatique - ATMEL 8 * Author: Nicolas Royer <nicolas@eukrea.com> 9 * 10 * Some ideas are from omap-aes.c drivers. 11 */ 12 13 14 #include <linux/kernel.h> 15 #include <linux/module.h> 16 #include <linux/slab.h> 17 #include <linux/err.h> 18 #include <linux/clk.h> 19 #include <linux/io.h> 20 #include <linux/hw_random.h> 21 #include <linux/platform_device.h> 22 23 #include <linux/device.h> 24 #include <linux/dmaengine.h> 25 #include <linux/init.h> 26 #include <linux/errno.h> 27 #include <linux/interrupt.h> 28 #include <linux/irq.h> 29 #include <linux/scatterlist.h> 30 #include <linux/dma-mapping.h> 31 #include <linux/mod_devicetable.h> 32 #include <linux/delay.h> 33 #include <linux/crypto.h> 34 #include <crypto/scatterwalk.h> 35 #include <crypto/algapi.h> 36 #include <crypto/internal/des.h> 37 #include <crypto/internal/skcipher.h> 38 #include "atmel-tdes-regs.h" 39 40 #define ATMEL_TDES_PRIORITY 300 41 42 /* TDES flags */ 43 /* Reserve bits [17:16], [13:12], [2:0] for AES Mode Register */ 44 #define TDES_FLAGS_ENCRYPT TDES_MR_CYPHER_ENC 45 #define TDES_FLAGS_OPMODE_MASK (TDES_MR_OPMOD_MASK | TDES_MR_CFBS_MASK) 46 #define TDES_FLAGS_ECB TDES_MR_OPMOD_ECB 47 #define TDES_FLAGS_CBC TDES_MR_OPMOD_CBC 48 #define TDES_FLAGS_OFB TDES_MR_OPMOD_OFB 49 #define TDES_FLAGS_CFB64 (TDES_MR_OPMOD_CFB | TDES_MR_CFBS_64b) 50 #define TDES_FLAGS_CFB32 (TDES_MR_OPMOD_CFB | TDES_MR_CFBS_32b) 51 #define TDES_FLAGS_CFB16 (TDES_MR_OPMOD_CFB | TDES_MR_CFBS_16b) 52 #define TDES_FLAGS_CFB8 (TDES_MR_OPMOD_CFB | TDES_MR_CFBS_8b) 53 54 #define TDES_FLAGS_MODE_MASK (TDES_FLAGS_OPMODE_MASK | TDES_FLAGS_ENCRYPT) 55 56 #define TDES_FLAGS_INIT BIT(3) 57 #define TDES_FLAGS_FAST BIT(4) 58 #define TDES_FLAGS_BUSY BIT(5) 59 #define TDES_FLAGS_DMA BIT(6) 60 61 #define ATMEL_TDES_QUEUE_LENGTH 50 62 63 #define CFB8_BLOCK_SIZE 1 64 #define CFB16_BLOCK_SIZE 2 65 #define CFB32_BLOCK_SIZE 4 66 67 struct atmel_tdes_caps { 68 bool has_dma; 69 u32 has_cfb_3keys; 70 }; 71 72 struct atmel_tdes_dev; 73 74 struct atmel_tdes_ctx { 75 struct atmel_tdes_dev *dd; 76 77 int keylen; 78 u32 key[DES3_EDE_KEY_SIZE / sizeof(u32)]; 79 unsigned long flags; 80 81 u16 block_size; 82 }; 83 84 struct atmel_tdes_reqctx { 85 unsigned long mode; 86 u8 lastc[DES_BLOCK_SIZE]; 87 }; 88 89 struct atmel_tdes_dma { 90 struct dma_chan *chan; 91 struct dma_slave_config dma_conf; 92 }; 93 94 struct atmel_tdes_dev { 95 struct list_head list; 96 unsigned long phys_base; 97 void __iomem *io_base; 98 99 struct atmel_tdes_ctx *ctx; 100 struct device *dev; 101 struct clk *iclk; 102 int irq; 103 104 unsigned long flags; 105 106 spinlock_t lock; 107 struct crypto_queue queue; 108 109 struct tasklet_struct done_task; 110 struct tasklet_struct queue_task; 111 112 struct skcipher_request *req; 113 size_t total; 114 115 struct scatterlist *in_sg; 116 unsigned int nb_in_sg; 117 size_t in_offset; 118 struct scatterlist *out_sg; 119 unsigned int nb_out_sg; 120 size_t out_offset; 121 122 size_t buflen; 123 size_t dma_size; 124 125 void *buf_in; 126 int dma_in; 127 dma_addr_t dma_addr_in; 128 struct atmel_tdes_dma dma_lch_in; 129 130 void *buf_out; 131 int dma_out; 132 dma_addr_t dma_addr_out; 133 struct atmel_tdes_dma dma_lch_out; 134 135 struct atmel_tdes_caps caps; 136 137 u32 hw_version; 138 }; 139 140 struct atmel_tdes_drv { 141 struct list_head dev_list; 142 spinlock_t lock; 143 }; 144 145 static struct atmel_tdes_drv atmel_tdes = { 146 .dev_list = LIST_HEAD_INIT(atmel_tdes.dev_list), 147 .lock = __SPIN_LOCK_UNLOCKED(atmel_tdes.lock), 148 }; 149 150 static int atmel_tdes_sg_copy(struct scatterlist **sg, size_t *offset, 151 void *buf, size_t buflen, size_t total, int out) 152 { 153 size_t count, off = 0; 154 155 while (buflen && total) { 156 count = min((*sg)->length - *offset, total); 157 count = min(count, buflen); 158 159 if (!count) 160 return off; 161 162 scatterwalk_map_and_copy(buf + off, *sg, *offset, count, out); 163 164 off += count; 165 buflen -= count; 166 *offset += count; 167 total -= count; 168 169 if (*offset == (*sg)->length) { 170 *sg = sg_next(*sg); 171 if (*sg) 172 *offset = 0; 173 else 174 total = 0; 175 } 176 } 177 178 return off; 179 } 180 181 static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset) 182 { 183 return readl_relaxed(dd->io_base + offset); 184 } 185 186 static inline void atmel_tdes_write(struct atmel_tdes_dev *dd, 187 u32 offset, u32 value) 188 { 189 writel_relaxed(value, dd->io_base + offset); 190 } 191 192 static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset, 193 const u32 *value, int count) 194 { 195 for (; count--; value++, offset += 4) 196 atmel_tdes_write(dd, offset, *value); 197 } 198 199 static struct atmel_tdes_dev *atmel_tdes_dev_alloc(void) 200 { 201 struct atmel_tdes_dev *tdes_dd; 202 203 spin_lock_bh(&atmel_tdes.lock); 204 /* One TDES IP per SoC. */ 205 tdes_dd = list_first_entry_or_null(&atmel_tdes.dev_list, 206 struct atmel_tdes_dev, list); 207 spin_unlock_bh(&atmel_tdes.lock); 208 return tdes_dd; 209 } 210 211 static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd) 212 { 213 int err; 214 215 err = clk_prepare_enable(dd->iclk); 216 if (err) 217 return err; 218 219 if (!(dd->flags & TDES_FLAGS_INIT)) { 220 atmel_tdes_write(dd, TDES_CR, TDES_CR_SWRST); 221 dd->flags |= TDES_FLAGS_INIT; 222 } 223 224 return 0; 225 } 226 227 static inline unsigned int atmel_tdes_get_version(struct atmel_tdes_dev *dd) 228 { 229 return atmel_tdes_read(dd, TDES_HW_VERSION) & 0x00000fff; 230 } 231 232 static int atmel_tdes_hw_version_init(struct atmel_tdes_dev *dd) 233 { 234 int err; 235 236 err = atmel_tdes_hw_init(dd); 237 if (err) 238 return err; 239 240 dd->hw_version = atmel_tdes_get_version(dd); 241 242 dev_info(dd->dev, 243 "version: 0x%x\n", dd->hw_version); 244 245 clk_disable_unprepare(dd->iclk); 246 247 return 0; 248 } 249 250 static void atmel_tdes_dma_callback(void *data) 251 { 252 struct atmel_tdes_dev *dd = data; 253 254 /* dma_lch_out - completed */ 255 tasklet_schedule(&dd->done_task); 256 } 257 258 static int atmel_tdes_write_ctrl(struct atmel_tdes_dev *dd) 259 { 260 int err; 261 u32 valmr = TDES_MR_SMOD_PDC; 262 263 err = atmel_tdes_hw_init(dd); 264 265 if (err) 266 return err; 267 268 if (!dd->caps.has_dma) 269 atmel_tdes_write(dd, TDES_PTCR, 270 TDES_PTCR_TXTDIS | TDES_PTCR_RXTDIS); 271 272 /* MR register must be set before IV registers */ 273 if (dd->ctx->keylen > (DES_KEY_SIZE << 1)) { 274 valmr |= TDES_MR_KEYMOD_3KEY; 275 valmr |= TDES_MR_TDESMOD_TDES; 276 } else if (dd->ctx->keylen > DES_KEY_SIZE) { 277 valmr |= TDES_MR_KEYMOD_2KEY; 278 valmr |= TDES_MR_TDESMOD_TDES; 279 } else { 280 valmr |= TDES_MR_TDESMOD_DES; 281 } 282 283 valmr |= dd->flags & TDES_FLAGS_MODE_MASK; 284 285 atmel_tdes_write(dd, TDES_MR, valmr); 286 287 atmel_tdes_write_n(dd, TDES_KEY1W1R, dd->ctx->key, 288 dd->ctx->keylen >> 2); 289 290 if (dd->req->iv && (valmr & TDES_MR_OPMOD_MASK) != TDES_MR_OPMOD_ECB) 291 atmel_tdes_write_n(dd, TDES_IV1R, (void *)dd->req->iv, 2); 292 293 return 0; 294 } 295 296 static int atmel_tdes_crypt_pdc_stop(struct atmel_tdes_dev *dd) 297 { 298 int err = 0; 299 size_t count; 300 301 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS); 302 303 if (dd->flags & TDES_FLAGS_FAST) { 304 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); 305 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 306 } else { 307 dma_sync_single_for_device(dd->dev, dd->dma_addr_out, 308 dd->dma_size, DMA_FROM_DEVICE); 309 310 /* copy data */ 311 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, 312 dd->buf_out, dd->buflen, dd->dma_size, 1); 313 if (count != dd->dma_size) { 314 err = -EINVAL; 315 dev_dbg(dd->dev, "not all data converted: %zu\n", count); 316 } 317 } 318 319 return err; 320 } 321 322 static int atmel_tdes_buff_init(struct atmel_tdes_dev *dd) 323 { 324 int err = -ENOMEM; 325 326 dd->buf_in = (void *)__get_free_pages(GFP_KERNEL, 0); 327 dd->buf_out = (void *)__get_free_pages(GFP_KERNEL, 0); 328 dd->buflen = PAGE_SIZE; 329 dd->buflen &= ~(DES_BLOCK_SIZE - 1); 330 331 if (!dd->buf_in || !dd->buf_out) { 332 dev_dbg(dd->dev, "unable to alloc pages.\n"); 333 goto err_alloc; 334 } 335 336 /* MAP here */ 337 dd->dma_addr_in = dma_map_single(dd->dev, dd->buf_in, 338 dd->buflen, DMA_TO_DEVICE); 339 err = dma_mapping_error(dd->dev, dd->dma_addr_in); 340 if (err) { 341 dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen); 342 goto err_map_in; 343 } 344 345 dd->dma_addr_out = dma_map_single(dd->dev, dd->buf_out, 346 dd->buflen, DMA_FROM_DEVICE); 347 err = dma_mapping_error(dd->dev, dd->dma_addr_out); 348 if (err) { 349 dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen); 350 goto err_map_out; 351 } 352 353 return 0; 354 355 err_map_out: 356 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen, 357 DMA_TO_DEVICE); 358 err_map_in: 359 err_alloc: 360 free_page((unsigned long)dd->buf_out); 361 free_page((unsigned long)dd->buf_in); 362 return err; 363 } 364 365 static void atmel_tdes_buff_cleanup(struct atmel_tdes_dev *dd) 366 { 367 dma_unmap_single(dd->dev, dd->dma_addr_out, dd->buflen, 368 DMA_FROM_DEVICE); 369 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen, 370 DMA_TO_DEVICE); 371 free_page((unsigned long)dd->buf_out); 372 free_page((unsigned long)dd->buf_in); 373 } 374 375 static int atmel_tdes_crypt_pdc(struct atmel_tdes_dev *dd, 376 dma_addr_t dma_addr_in, 377 dma_addr_t dma_addr_out, int length) 378 { 379 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(dd->req); 380 int len32; 381 382 dd->dma_size = length; 383 384 if (!(dd->flags & TDES_FLAGS_FAST)) { 385 dma_sync_single_for_device(dd->dev, dma_addr_in, length, 386 DMA_TO_DEVICE); 387 } 388 389 switch (rctx->mode & TDES_FLAGS_OPMODE_MASK) { 390 case TDES_FLAGS_CFB8: 391 len32 = DIV_ROUND_UP(length, sizeof(u8)); 392 break; 393 394 case TDES_FLAGS_CFB16: 395 len32 = DIV_ROUND_UP(length, sizeof(u16)); 396 break; 397 398 default: 399 len32 = DIV_ROUND_UP(length, sizeof(u32)); 400 break; 401 } 402 403 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS); 404 atmel_tdes_write(dd, TDES_TPR, dma_addr_in); 405 atmel_tdes_write(dd, TDES_TCR, len32); 406 atmel_tdes_write(dd, TDES_RPR, dma_addr_out); 407 atmel_tdes_write(dd, TDES_RCR, len32); 408 409 /* Enable Interrupt */ 410 atmel_tdes_write(dd, TDES_IER, TDES_INT_ENDRX); 411 412 /* Start DMA transfer */ 413 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTEN | TDES_PTCR_RXTEN); 414 415 return 0; 416 } 417 418 static int atmel_tdes_crypt_dma(struct atmel_tdes_dev *dd, 419 dma_addr_t dma_addr_in, 420 dma_addr_t dma_addr_out, int length) 421 { 422 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(dd->req); 423 struct scatterlist sg[2]; 424 struct dma_async_tx_descriptor *in_desc, *out_desc; 425 enum dma_slave_buswidth addr_width; 426 427 dd->dma_size = length; 428 429 if (!(dd->flags & TDES_FLAGS_FAST)) { 430 dma_sync_single_for_device(dd->dev, dma_addr_in, length, 431 DMA_TO_DEVICE); 432 } 433 434 switch (rctx->mode & TDES_FLAGS_OPMODE_MASK) { 435 case TDES_FLAGS_CFB8: 436 addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; 437 break; 438 439 case TDES_FLAGS_CFB16: 440 addr_width = DMA_SLAVE_BUSWIDTH_2_BYTES; 441 break; 442 443 default: 444 addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; 445 break; 446 } 447 448 dd->dma_lch_in.dma_conf.dst_addr_width = addr_width; 449 dd->dma_lch_out.dma_conf.src_addr_width = addr_width; 450 451 dmaengine_slave_config(dd->dma_lch_in.chan, &dd->dma_lch_in.dma_conf); 452 dmaengine_slave_config(dd->dma_lch_out.chan, &dd->dma_lch_out.dma_conf); 453 454 dd->flags |= TDES_FLAGS_DMA; 455 456 sg_init_table(&sg[0], 1); 457 sg_dma_address(&sg[0]) = dma_addr_in; 458 sg_dma_len(&sg[0]) = length; 459 460 sg_init_table(&sg[1], 1); 461 sg_dma_address(&sg[1]) = dma_addr_out; 462 sg_dma_len(&sg[1]) = length; 463 464 in_desc = dmaengine_prep_slave_sg(dd->dma_lch_in.chan, &sg[0], 465 1, DMA_MEM_TO_DEV, 466 DMA_PREP_INTERRUPT | DMA_CTRL_ACK); 467 if (!in_desc) 468 return -EINVAL; 469 470 out_desc = dmaengine_prep_slave_sg(dd->dma_lch_out.chan, &sg[1], 471 1, DMA_DEV_TO_MEM, 472 DMA_PREP_INTERRUPT | DMA_CTRL_ACK); 473 if (!out_desc) 474 return -EINVAL; 475 476 out_desc->callback = atmel_tdes_dma_callback; 477 out_desc->callback_param = dd; 478 479 dmaengine_submit(out_desc); 480 dma_async_issue_pending(dd->dma_lch_out.chan); 481 482 dmaengine_submit(in_desc); 483 dma_async_issue_pending(dd->dma_lch_in.chan); 484 485 return 0; 486 } 487 488 static int atmel_tdes_crypt_start(struct atmel_tdes_dev *dd) 489 { 490 int err, fast = 0, in, out; 491 size_t count; 492 dma_addr_t addr_in, addr_out; 493 494 if ((!dd->in_offset) && (!dd->out_offset)) { 495 /* check for alignment */ 496 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) && 497 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size); 498 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && 499 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); 500 fast = in && out; 501 502 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) 503 fast = 0; 504 } 505 506 507 if (fast) { 508 count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg)); 509 count = min_t(size_t, count, sg_dma_len(dd->out_sg)); 510 511 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 512 if (!err) { 513 dev_dbg(dd->dev, "dma_map_sg() error\n"); 514 return -EINVAL; 515 } 516 517 err = dma_map_sg(dd->dev, dd->out_sg, 1, 518 DMA_FROM_DEVICE); 519 if (!err) { 520 dev_dbg(dd->dev, "dma_map_sg() error\n"); 521 dma_unmap_sg(dd->dev, dd->in_sg, 1, 522 DMA_TO_DEVICE); 523 return -EINVAL; 524 } 525 526 addr_in = sg_dma_address(dd->in_sg); 527 addr_out = sg_dma_address(dd->out_sg); 528 529 dd->flags |= TDES_FLAGS_FAST; 530 531 } else { 532 /* use cache buffers */ 533 count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset, 534 dd->buf_in, dd->buflen, dd->total, 0); 535 536 addr_in = dd->dma_addr_in; 537 addr_out = dd->dma_addr_out; 538 539 dd->flags &= ~TDES_FLAGS_FAST; 540 } 541 542 dd->total -= count; 543 544 if (dd->caps.has_dma) 545 err = atmel_tdes_crypt_dma(dd, addr_in, addr_out, count); 546 else 547 err = atmel_tdes_crypt_pdc(dd, addr_in, addr_out, count); 548 549 if (err && (dd->flags & TDES_FLAGS_FAST)) { 550 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 551 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); 552 } 553 554 return err; 555 } 556 557 static void 558 atmel_tdes_set_iv_as_last_ciphertext_block(struct atmel_tdes_dev *dd) 559 { 560 struct skcipher_request *req = dd->req; 561 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req); 562 struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); 563 unsigned int ivsize = crypto_skcipher_ivsize(skcipher); 564 565 if (req->cryptlen < ivsize) 566 return; 567 568 if (rctx->mode & TDES_FLAGS_ENCRYPT) 569 scatterwalk_map_and_copy(req->iv, req->dst, 570 req->cryptlen - ivsize, ivsize, 0); 571 else 572 memcpy(req->iv, rctx->lastc, ivsize); 573 574 } 575 576 static void atmel_tdes_finish_req(struct atmel_tdes_dev *dd, int err) 577 { 578 struct skcipher_request *req = dd->req; 579 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req); 580 581 clk_disable_unprepare(dd->iclk); 582 583 dd->flags &= ~TDES_FLAGS_BUSY; 584 585 if (!err && (rctx->mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB) 586 atmel_tdes_set_iv_as_last_ciphertext_block(dd); 587 588 skcipher_request_complete(req, err); 589 } 590 591 static int atmel_tdes_handle_queue(struct atmel_tdes_dev *dd, 592 struct skcipher_request *req) 593 { 594 struct crypto_async_request *async_req, *backlog; 595 struct atmel_tdes_ctx *ctx; 596 struct atmel_tdes_reqctx *rctx; 597 unsigned long flags; 598 int err, ret = 0; 599 600 spin_lock_irqsave(&dd->lock, flags); 601 if (req) 602 ret = crypto_enqueue_request(&dd->queue, &req->base); 603 if (dd->flags & TDES_FLAGS_BUSY) { 604 spin_unlock_irqrestore(&dd->lock, flags); 605 return ret; 606 } 607 backlog = crypto_get_backlog(&dd->queue); 608 async_req = crypto_dequeue_request(&dd->queue); 609 if (async_req) 610 dd->flags |= TDES_FLAGS_BUSY; 611 spin_unlock_irqrestore(&dd->lock, flags); 612 613 if (!async_req) 614 return ret; 615 616 if (backlog) 617 crypto_request_complete(backlog, -EINPROGRESS); 618 619 req = skcipher_request_cast(async_req); 620 621 /* assign new request to device */ 622 dd->req = req; 623 dd->total = req->cryptlen; 624 dd->in_offset = 0; 625 dd->in_sg = req->src; 626 dd->out_offset = 0; 627 dd->out_sg = req->dst; 628 629 rctx = skcipher_request_ctx(req); 630 ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req)); 631 rctx->mode &= TDES_FLAGS_MODE_MASK; 632 dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode; 633 dd->ctx = ctx; 634 635 err = atmel_tdes_write_ctrl(dd); 636 if (!err) 637 err = atmel_tdes_crypt_start(dd); 638 if (err) { 639 /* des_task will not finish it, so do it here */ 640 atmel_tdes_finish_req(dd, err); 641 tasklet_schedule(&dd->queue_task); 642 } 643 644 return ret; 645 } 646 647 static int atmel_tdes_crypt_dma_stop(struct atmel_tdes_dev *dd) 648 { 649 int err = -EINVAL; 650 size_t count; 651 652 if (dd->flags & TDES_FLAGS_DMA) { 653 err = 0; 654 if (dd->flags & TDES_FLAGS_FAST) { 655 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); 656 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 657 } else { 658 dma_sync_single_for_device(dd->dev, dd->dma_addr_out, 659 dd->dma_size, DMA_FROM_DEVICE); 660 661 /* copy data */ 662 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, 663 dd->buf_out, dd->buflen, dd->dma_size, 1); 664 if (count != dd->dma_size) { 665 err = -EINVAL; 666 dev_dbg(dd->dev, "not all data converted: %zu\n", count); 667 } 668 } 669 } 670 return err; 671 } 672 673 static int atmel_tdes_crypt(struct skcipher_request *req, unsigned long mode) 674 { 675 struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); 676 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(skcipher); 677 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req); 678 struct device *dev = ctx->dd->dev; 679 680 if (!req->cryptlen) 681 return 0; 682 683 switch (mode & TDES_FLAGS_OPMODE_MASK) { 684 case TDES_FLAGS_CFB8: 685 if (!IS_ALIGNED(req->cryptlen, CFB8_BLOCK_SIZE)) { 686 dev_dbg(dev, "request size is not exact amount of CFB8 blocks\n"); 687 return -EINVAL; 688 } 689 ctx->block_size = CFB8_BLOCK_SIZE; 690 break; 691 692 case TDES_FLAGS_CFB16: 693 if (!IS_ALIGNED(req->cryptlen, CFB16_BLOCK_SIZE)) { 694 dev_dbg(dev, "request size is not exact amount of CFB16 blocks\n"); 695 return -EINVAL; 696 } 697 ctx->block_size = CFB16_BLOCK_SIZE; 698 break; 699 700 case TDES_FLAGS_CFB32: 701 if (!IS_ALIGNED(req->cryptlen, CFB32_BLOCK_SIZE)) { 702 dev_dbg(dev, "request size is not exact amount of CFB32 blocks\n"); 703 return -EINVAL; 704 } 705 ctx->block_size = CFB32_BLOCK_SIZE; 706 break; 707 708 default: 709 if (!IS_ALIGNED(req->cryptlen, DES_BLOCK_SIZE)) { 710 dev_dbg(dev, "request size is not exact amount of DES blocks\n"); 711 return -EINVAL; 712 } 713 ctx->block_size = DES_BLOCK_SIZE; 714 break; 715 } 716 717 rctx->mode = mode; 718 719 if ((mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB && 720 !(mode & TDES_FLAGS_ENCRYPT)) { 721 unsigned int ivsize = crypto_skcipher_ivsize(skcipher); 722 723 if (req->cryptlen >= ivsize) 724 scatterwalk_map_and_copy(rctx->lastc, req->src, 725 req->cryptlen - ivsize, 726 ivsize, 0); 727 } 728 729 return atmel_tdes_handle_queue(ctx->dd, req); 730 } 731 732 static int atmel_tdes_dma_init(struct atmel_tdes_dev *dd) 733 { 734 int ret; 735 736 /* Try to grab 2 DMA channels */ 737 dd->dma_lch_in.chan = dma_request_chan(dd->dev, "tx"); 738 if (IS_ERR(dd->dma_lch_in.chan)) { 739 ret = PTR_ERR(dd->dma_lch_in.chan); 740 goto err_dma_in; 741 } 742 743 dd->dma_lch_in.dma_conf.dst_addr = dd->phys_base + 744 TDES_IDATA1R; 745 dd->dma_lch_in.dma_conf.src_maxburst = 1; 746 dd->dma_lch_in.dma_conf.src_addr_width = 747 DMA_SLAVE_BUSWIDTH_4_BYTES; 748 dd->dma_lch_in.dma_conf.dst_maxburst = 1; 749 dd->dma_lch_in.dma_conf.dst_addr_width = 750 DMA_SLAVE_BUSWIDTH_4_BYTES; 751 dd->dma_lch_in.dma_conf.device_fc = false; 752 753 dd->dma_lch_out.chan = dma_request_chan(dd->dev, "rx"); 754 if (IS_ERR(dd->dma_lch_out.chan)) { 755 ret = PTR_ERR(dd->dma_lch_out.chan); 756 goto err_dma_out; 757 } 758 759 dd->dma_lch_out.dma_conf.src_addr = dd->phys_base + 760 TDES_ODATA1R; 761 dd->dma_lch_out.dma_conf.src_maxburst = 1; 762 dd->dma_lch_out.dma_conf.src_addr_width = 763 DMA_SLAVE_BUSWIDTH_4_BYTES; 764 dd->dma_lch_out.dma_conf.dst_maxburst = 1; 765 dd->dma_lch_out.dma_conf.dst_addr_width = 766 DMA_SLAVE_BUSWIDTH_4_BYTES; 767 dd->dma_lch_out.dma_conf.device_fc = false; 768 769 return 0; 770 771 err_dma_out: 772 dma_release_channel(dd->dma_lch_in.chan); 773 err_dma_in: 774 dev_err(dd->dev, "no DMA channel available\n"); 775 return ret; 776 } 777 778 static void atmel_tdes_dma_cleanup(struct atmel_tdes_dev *dd) 779 { 780 dma_release_channel(dd->dma_lch_in.chan); 781 dma_release_channel(dd->dma_lch_out.chan); 782 } 783 784 static int atmel_des_setkey(struct crypto_skcipher *tfm, const u8 *key, 785 unsigned int keylen) 786 { 787 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm); 788 int err; 789 790 err = verify_skcipher_des_key(tfm, key); 791 if (err) 792 return err; 793 794 memcpy(ctx->key, key, keylen); 795 ctx->keylen = keylen; 796 797 return 0; 798 } 799 800 static int atmel_tdes_setkey(struct crypto_skcipher *tfm, const u8 *key, 801 unsigned int keylen) 802 { 803 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm); 804 int err; 805 806 err = verify_skcipher_des3_key(tfm, key); 807 if (err) 808 return err; 809 810 memcpy(ctx->key, key, keylen); 811 ctx->keylen = keylen; 812 813 return 0; 814 } 815 816 static int atmel_tdes_ecb_encrypt(struct skcipher_request *req) 817 { 818 return atmel_tdes_crypt(req, TDES_FLAGS_ECB | TDES_FLAGS_ENCRYPT); 819 } 820 821 static int atmel_tdes_ecb_decrypt(struct skcipher_request *req) 822 { 823 return atmel_tdes_crypt(req, TDES_FLAGS_ECB); 824 } 825 826 static int atmel_tdes_cbc_encrypt(struct skcipher_request *req) 827 { 828 return atmel_tdes_crypt(req, TDES_FLAGS_CBC | TDES_FLAGS_ENCRYPT); 829 } 830 831 static int atmel_tdes_cbc_decrypt(struct skcipher_request *req) 832 { 833 return atmel_tdes_crypt(req, TDES_FLAGS_CBC); 834 } 835 static int atmel_tdes_cfb_encrypt(struct skcipher_request *req) 836 { 837 return atmel_tdes_crypt(req, TDES_FLAGS_CFB64 | TDES_FLAGS_ENCRYPT); 838 } 839 840 static int atmel_tdes_cfb_decrypt(struct skcipher_request *req) 841 { 842 return atmel_tdes_crypt(req, TDES_FLAGS_CFB64); 843 } 844 845 static int atmel_tdes_cfb8_encrypt(struct skcipher_request *req) 846 { 847 return atmel_tdes_crypt(req, TDES_FLAGS_CFB8 | TDES_FLAGS_ENCRYPT); 848 } 849 850 static int atmel_tdes_cfb8_decrypt(struct skcipher_request *req) 851 { 852 return atmel_tdes_crypt(req, TDES_FLAGS_CFB8); 853 } 854 855 static int atmel_tdes_cfb16_encrypt(struct skcipher_request *req) 856 { 857 return atmel_tdes_crypt(req, TDES_FLAGS_CFB16 | TDES_FLAGS_ENCRYPT); 858 } 859 860 static int atmel_tdes_cfb16_decrypt(struct skcipher_request *req) 861 { 862 return atmel_tdes_crypt(req, TDES_FLAGS_CFB16); 863 } 864 865 static int atmel_tdes_cfb32_encrypt(struct skcipher_request *req) 866 { 867 return atmel_tdes_crypt(req, TDES_FLAGS_CFB32 | TDES_FLAGS_ENCRYPT); 868 } 869 870 static int atmel_tdes_cfb32_decrypt(struct skcipher_request *req) 871 { 872 return atmel_tdes_crypt(req, TDES_FLAGS_CFB32); 873 } 874 875 static int atmel_tdes_ofb_encrypt(struct skcipher_request *req) 876 { 877 return atmel_tdes_crypt(req, TDES_FLAGS_OFB | TDES_FLAGS_ENCRYPT); 878 } 879 880 static int atmel_tdes_ofb_decrypt(struct skcipher_request *req) 881 { 882 return atmel_tdes_crypt(req, TDES_FLAGS_OFB); 883 } 884 885 static int atmel_tdes_init_tfm(struct crypto_skcipher *tfm) 886 { 887 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm); 888 889 ctx->dd = atmel_tdes_dev_alloc(); 890 if (!ctx->dd) 891 return -ENODEV; 892 893 crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_tdes_reqctx)); 894 895 return 0; 896 } 897 898 static void atmel_tdes_skcipher_alg_init(struct skcipher_alg *alg) 899 { 900 alg->base.cra_priority = ATMEL_TDES_PRIORITY; 901 alg->base.cra_flags = CRYPTO_ALG_ASYNC; 902 alg->base.cra_ctxsize = sizeof(struct atmel_tdes_ctx); 903 alg->base.cra_module = THIS_MODULE; 904 905 alg->init = atmel_tdes_init_tfm; 906 } 907 908 static struct skcipher_alg tdes_algs[] = { 909 { 910 .base.cra_name = "ecb(des)", 911 .base.cra_driver_name = "atmel-ecb-des", 912 .base.cra_blocksize = DES_BLOCK_SIZE, 913 .base.cra_alignmask = 0x7, 914 915 .min_keysize = DES_KEY_SIZE, 916 .max_keysize = DES_KEY_SIZE, 917 .setkey = atmel_des_setkey, 918 .encrypt = atmel_tdes_ecb_encrypt, 919 .decrypt = atmel_tdes_ecb_decrypt, 920 }, 921 { 922 .base.cra_name = "cbc(des)", 923 .base.cra_driver_name = "atmel-cbc-des", 924 .base.cra_blocksize = DES_BLOCK_SIZE, 925 .base.cra_alignmask = 0x7, 926 927 .min_keysize = DES_KEY_SIZE, 928 .max_keysize = DES_KEY_SIZE, 929 .ivsize = DES_BLOCK_SIZE, 930 .setkey = atmel_des_setkey, 931 .encrypt = atmel_tdes_cbc_encrypt, 932 .decrypt = atmel_tdes_cbc_decrypt, 933 }, 934 { 935 .base.cra_name = "cfb(des)", 936 .base.cra_driver_name = "atmel-cfb-des", 937 .base.cra_blocksize = DES_BLOCK_SIZE, 938 .base.cra_alignmask = 0x7, 939 940 .min_keysize = DES_KEY_SIZE, 941 .max_keysize = DES_KEY_SIZE, 942 .ivsize = DES_BLOCK_SIZE, 943 .setkey = atmel_des_setkey, 944 .encrypt = atmel_tdes_cfb_encrypt, 945 .decrypt = atmel_tdes_cfb_decrypt, 946 }, 947 { 948 .base.cra_name = "cfb8(des)", 949 .base.cra_driver_name = "atmel-cfb8-des", 950 .base.cra_blocksize = CFB8_BLOCK_SIZE, 951 .base.cra_alignmask = 0, 952 953 .min_keysize = DES_KEY_SIZE, 954 .max_keysize = DES_KEY_SIZE, 955 .ivsize = DES_BLOCK_SIZE, 956 .setkey = atmel_des_setkey, 957 .encrypt = atmel_tdes_cfb8_encrypt, 958 .decrypt = atmel_tdes_cfb8_decrypt, 959 }, 960 { 961 .base.cra_name = "cfb16(des)", 962 .base.cra_driver_name = "atmel-cfb16-des", 963 .base.cra_blocksize = CFB16_BLOCK_SIZE, 964 .base.cra_alignmask = 0x1, 965 966 .min_keysize = DES_KEY_SIZE, 967 .max_keysize = DES_KEY_SIZE, 968 .ivsize = DES_BLOCK_SIZE, 969 .setkey = atmel_des_setkey, 970 .encrypt = atmel_tdes_cfb16_encrypt, 971 .decrypt = atmel_tdes_cfb16_decrypt, 972 }, 973 { 974 .base.cra_name = "cfb32(des)", 975 .base.cra_driver_name = "atmel-cfb32-des", 976 .base.cra_blocksize = CFB32_BLOCK_SIZE, 977 .base.cra_alignmask = 0x3, 978 979 .min_keysize = DES_KEY_SIZE, 980 .max_keysize = DES_KEY_SIZE, 981 .ivsize = DES_BLOCK_SIZE, 982 .setkey = atmel_des_setkey, 983 .encrypt = atmel_tdes_cfb32_encrypt, 984 .decrypt = atmel_tdes_cfb32_decrypt, 985 }, 986 { 987 .base.cra_name = "ofb(des)", 988 .base.cra_driver_name = "atmel-ofb-des", 989 .base.cra_blocksize = 1, 990 .base.cra_alignmask = 0x7, 991 992 .min_keysize = DES_KEY_SIZE, 993 .max_keysize = DES_KEY_SIZE, 994 .ivsize = DES_BLOCK_SIZE, 995 .setkey = atmel_des_setkey, 996 .encrypt = atmel_tdes_ofb_encrypt, 997 .decrypt = atmel_tdes_ofb_decrypt, 998 }, 999 { 1000 .base.cra_name = "ecb(des3_ede)", 1001 .base.cra_driver_name = "atmel-ecb-tdes", 1002 .base.cra_blocksize = DES_BLOCK_SIZE, 1003 .base.cra_alignmask = 0x7, 1004 1005 .min_keysize = DES3_EDE_KEY_SIZE, 1006 .max_keysize = DES3_EDE_KEY_SIZE, 1007 .setkey = atmel_tdes_setkey, 1008 .encrypt = atmel_tdes_ecb_encrypt, 1009 .decrypt = atmel_tdes_ecb_decrypt, 1010 }, 1011 { 1012 .base.cra_name = "cbc(des3_ede)", 1013 .base.cra_driver_name = "atmel-cbc-tdes", 1014 .base.cra_blocksize = DES_BLOCK_SIZE, 1015 .base.cra_alignmask = 0x7, 1016 1017 .min_keysize = DES3_EDE_KEY_SIZE, 1018 .max_keysize = DES3_EDE_KEY_SIZE, 1019 .setkey = atmel_tdes_setkey, 1020 .encrypt = atmel_tdes_cbc_encrypt, 1021 .decrypt = atmel_tdes_cbc_decrypt, 1022 .ivsize = DES_BLOCK_SIZE, 1023 }, 1024 { 1025 .base.cra_name = "ofb(des3_ede)", 1026 .base.cra_driver_name = "atmel-ofb-tdes", 1027 .base.cra_blocksize = DES_BLOCK_SIZE, 1028 .base.cra_alignmask = 0x7, 1029 1030 .min_keysize = DES3_EDE_KEY_SIZE, 1031 .max_keysize = DES3_EDE_KEY_SIZE, 1032 .setkey = atmel_tdes_setkey, 1033 .encrypt = atmel_tdes_ofb_encrypt, 1034 .decrypt = atmel_tdes_ofb_decrypt, 1035 .ivsize = DES_BLOCK_SIZE, 1036 }, 1037 }; 1038 1039 static void atmel_tdes_queue_task(unsigned long data) 1040 { 1041 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *)data; 1042 1043 atmel_tdes_handle_queue(dd, NULL); 1044 } 1045 1046 static void atmel_tdes_done_task(unsigned long data) 1047 { 1048 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *) data; 1049 int err; 1050 1051 if (!(dd->flags & TDES_FLAGS_DMA)) 1052 err = atmel_tdes_crypt_pdc_stop(dd); 1053 else 1054 err = atmel_tdes_crypt_dma_stop(dd); 1055 1056 if (dd->total && !err) { 1057 if (dd->flags & TDES_FLAGS_FAST) { 1058 dd->in_sg = sg_next(dd->in_sg); 1059 dd->out_sg = sg_next(dd->out_sg); 1060 if (!dd->in_sg || !dd->out_sg) 1061 err = -EINVAL; 1062 } 1063 if (!err) 1064 err = atmel_tdes_crypt_start(dd); 1065 if (!err) 1066 return; /* DMA started. Not fininishing. */ 1067 } 1068 1069 atmel_tdes_finish_req(dd, err); 1070 atmel_tdes_handle_queue(dd, NULL); 1071 } 1072 1073 static irqreturn_t atmel_tdes_irq(int irq, void *dev_id) 1074 { 1075 struct atmel_tdes_dev *tdes_dd = dev_id; 1076 u32 reg; 1077 1078 reg = atmel_tdes_read(tdes_dd, TDES_ISR); 1079 if (reg & atmel_tdes_read(tdes_dd, TDES_IMR)) { 1080 atmel_tdes_write(tdes_dd, TDES_IDR, reg); 1081 if (TDES_FLAGS_BUSY & tdes_dd->flags) 1082 tasklet_schedule(&tdes_dd->done_task); 1083 else 1084 dev_warn(tdes_dd->dev, "TDES interrupt when no active requests.\n"); 1085 return IRQ_HANDLED; 1086 } 1087 1088 return IRQ_NONE; 1089 } 1090 1091 static void atmel_tdes_unregister_algs(struct atmel_tdes_dev *dd) 1092 { 1093 int i; 1094 1095 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) 1096 crypto_unregister_skcipher(&tdes_algs[i]); 1097 } 1098 1099 static int atmel_tdes_register_algs(struct atmel_tdes_dev *dd) 1100 { 1101 int err, i, j; 1102 1103 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) { 1104 atmel_tdes_skcipher_alg_init(&tdes_algs[i]); 1105 1106 err = crypto_register_skcipher(&tdes_algs[i]); 1107 if (err) 1108 goto err_tdes_algs; 1109 } 1110 1111 return 0; 1112 1113 err_tdes_algs: 1114 for (j = 0; j < i; j++) 1115 crypto_unregister_skcipher(&tdes_algs[j]); 1116 1117 return err; 1118 } 1119 1120 static void atmel_tdes_get_cap(struct atmel_tdes_dev *dd) 1121 { 1122 1123 dd->caps.has_dma = 0; 1124 dd->caps.has_cfb_3keys = 0; 1125 1126 /* keep only major version number */ 1127 switch (dd->hw_version & 0xf00) { 1128 case 0x800: 1129 case 0x700: 1130 dd->caps.has_dma = 1; 1131 dd->caps.has_cfb_3keys = 1; 1132 break; 1133 case 0x600: 1134 break; 1135 default: 1136 dev_warn(dd->dev, 1137 "Unmanaged tdes version, set minimum capabilities\n"); 1138 break; 1139 } 1140 } 1141 1142 static const struct of_device_id atmel_tdes_dt_ids[] = { 1143 { .compatible = "atmel,at91sam9g46-tdes" }, 1144 { /* sentinel */ } 1145 }; 1146 MODULE_DEVICE_TABLE(of, atmel_tdes_dt_ids); 1147 1148 static int atmel_tdes_probe(struct platform_device *pdev) 1149 { 1150 struct atmel_tdes_dev *tdes_dd; 1151 struct device *dev = &pdev->dev; 1152 struct resource *tdes_res; 1153 int err; 1154 1155 tdes_dd = devm_kmalloc(&pdev->dev, sizeof(*tdes_dd), GFP_KERNEL); 1156 if (!tdes_dd) 1157 return -ENOMEM; 1158 1159 tdes_dd->dev = dev; 1160 1161 platform_set_drvdata(pdev, tdes_dd); 1162 1163 INIT_LIST_HEAD(&tdes_dd->list); 1164 spin_lock_init(&tdes_dd->lock); 1165 1166 tasklet_init(&tdes_dd->done_task, atmel_tdes_done_task, 1167 (unsigned long)tdes_dd); 1168 tasklet_init(&tdes_dd->queue_task, atmel_tdes_queue_task, 1169 (unsigned long)tdes_dd); 1170 1171 crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH); 1172 1173 tdes_dd->io_base = devm_platform_get_and_ioremap_resource(pdev, 0, &tdes_res); 1174 if (IS_ERR(tdes_dd->io_base)) { 1175 err = PTR_ERR(tdes_dd->io_base); 1176 goto err_tasklet_kill; 1177 } 1178 tdes_dd->phys_base = tdes_res->start; 1179 1180 /* Get the IRQ */ 1181 tdes_dd->irq = platform_get_irq(pdev, 0); 1182 if (tdes_dd->irq < 0) { 1183 err = tdes_dd->irq; 1184 goto err_tasklet_kill; 1185 } 1186 1187 err = devm_request_irq(&pdev->dev, tdes_dd->irq, atmel_tdes_irq, 1188 IRQF_SHARED, "atmel-tdes", tdes_dd); 1189 if (err) { 1190 dev_err(dev, "unable to request tdes irq.\n"); 1191 goto err_tasklet_kill; 1192 } 1193 1194 /* Initializing the clock */ 1195 tdes_dd->iclk = devm_clk_get(&pdev->dev, "tdes_clk"); 1196 if (IS_ERR(tdes_dd->iclk)) { 1197 dev_err(dev, "clock initialization failed.\n"); 1198 err = PTR_ERR(tdes_dd->iclk); 1199 goto err_tasklet_kill; 1200 } 1201 1202 err = atmel_tdes_hw_version_init(tdes_dd); 1203 if (err) 1204 goto err_tasklet_kill; 1205 1206 atmel_tdes_get_cap(tdes_dd); 1207 1208 err = atmel_tdes_buff_init(tdes_dd); 1209 if (err) 1210 goto err_tasklet_kill; 1211 1212 if (tdes_dd->caps.has_dma) { 1213 err = atmel_tdes_dma_init(tdes_dd); 1214 if (err) 1215 goto err_buff_cleanup; 1216 1217 dev_info(dev, "using %s, %s for DMA transfers\n", 1218 dma_chan_name(tdes_dd->dma_lch_in.chan), 1219 dma_chan_name(tdes_dd->dma_lch_out.chan)); 1220 } 1221 1222 spin_lock(&atmel_tdes.lock); 1223 list_add_tail(&tdes_dd->list, &atmel_tdes.dev_list); 1224 spin_unlock(&atmel_tdes.lock); 1225 1226 err = atmel_tdes_register_algs(tdes_dd); 1227 if (err) 1228 goto err_algs; 1229 1230 dev_info(dev, "Atmel DES/TDES\n"); 1231 1232 return 0; 1233 1234 err_algs: 1235 spin_lock(&atmel_tdes.lock); 1236 list_del(&tdes_dd->list); 1237 spin_unlock(&atmel_tdes.lock); 1238 if (tdes_dd->caps.has_dma) 1239 atmel_tdes_dma_cleanup(tdes_dd); 1240 err_buff_cleanup: 1241 atmel_tdes_buff_cleanup(tdes_dd); 1242 err_tasklet_kill: 1243 tasklet_kill(&tdes_dd->done_task); 1244 tasklet_kill(&tdes_dd->queue_task); 1245 1246 return err; 1247 } 1248 1249 static int atmel_tdes_remove(struct platform_device *pdev) 1250 { 1251 struct atmel_tdes_dev *tdes_dd = platform_get_drvdata(pdev); 1252 1253 spin_lock(&atmel_tdes.lock); 1254 list_del(&tdes_dd->list); 1255 spin_unlock(&atmel_tdes.lock); 1256 1257 atmel_tdes_unregister_algs(tdes_dd); 1258 1259 tasklet_kill(&tdes_dd->done_task); 1260 tasklet_kill(&tdes_dd->queue_task); 1261 1262 if (tdes_dd->caps.has_dma) 1263 atmel_tdes_dma_cleanup(tdes_dd); 1264 1265 atmel_tdes_buff_cleanup(tdes_dd); 1266 1267 return 0; 1268 } 1269 1270 static struct platform_driver atmel_tdes_driver = { 1271 .probe = atmel_tdes_probe, 1272 .remove = atmel_tdes_remove, 1273 .driver = { 1274 .name = "atmel_tdes", 1275 .of_match_table = atmel_tdes_dt_ids, 1276 }, 1277 }; 1278 1279 module_platform_driver(atmel_tdes_driver); 1280 1281 MODULE_DESCRIPTION("Atmel DES/TDES hw acceleration support."); 1282 MODULE_LICENSE("GPL v2"); 1283 MODULE_AUTHOR("Nicolas Royer - Eukréa Electromatique"); 1284