Lines Matching refs:AES_BLOCK_SIZE

233 		n = nbytes & ~(AES_BLOCK_SIZE - 1);  in ecb_paes_crypt()
263 .base.cra_blocksize = AES_BLOCK_SIZE,
335 u8 iv[AES_BLOCK_SIZE]; in cbc_paes_crypt()
343 memcpy(param.iv, walk.iv, AES_BLOCK_SIZE); in cbc_paes_crypt()
350 n = nbytes & ~(AES_BLOCK_SIZE - 1); in cbc_paes_crypt()
354 memcpy(walk.iv, param.iv, AES_BLOCK_SIZE); in cbc_paes_crypt()
382 .base.cra_blocksize = AES_BLOCK_SIZE,
390 .ivsize = AES_BLOCK_SIZE,
530 n = nbytes & ~(AES_BLOCK_SIZE - 1); in xts_paes_crypt()
562 .base.cra_blocksize = AES_BLOCK_SIZE,
570 .ivsize = AES_BLOCK_SIZE,
633 memcpy(ctrptr, iv, AES_BLOCK_SIZE); in __ctrblk_init()
634 n = (nbytes > PAGE_SIZE) ? PAGE_SIZE : nbytes & ~(AES_BLOCK_SIZE - 1); in __ctrblk_init()
635 for (i = (n / AES_BLOCK_SIZE) - 1; i > 0; i--) { in __ctrblk_init()
636 memcpy(ctrptr + AES_BLOCK_SIZE, ctrptr, AES_BLOCK_SIZE); in __ctrblk_init()
637 crypto_inc(ctrptr + AES_BLOCK_SIZE, AES_BLOCK_SIZE); in __ctrblk_init()
638 ctrptr += AES_BLOCK_SIZE; in __ctrblk_init()
647 u8 buf[AES_BLOCK_SIZE], *ctrptr; in ctr_paes_crypt()
665 while ((nbytes = walk.nbytes) >= AES_BLOCK_SIZE) { in ctr_paes_crypt()
666 n = AES_BLOCK_SIZE; in ctr_paes_crypt()
667 if (nbytes >= 2*AES_BLOCK_SIZE && locked) in ctr_paes_crypt()
669 ctrptr = (n > AES_BLOCK_SIZE) ? ctrblk : walk.iv; in ctr_paes_crypt()
674 memcpy(walk.iv, ctrptr + k - AES_BLOCK_SIZE, in ctr_paes_crypt()
675 AES_BLOCK_SIZE); in ctr_paes_crypt()
676 crypto_inc(walk.iv, AES_BLOCK_SIZE); in ctr_paes_crypt()
696 memset(buf, 0, AES_BLOCK_SIZE); in ctr_paes_crypt()
700 buf, AES_BLOCK_SIZE, in ctr_paes_crypt()
701 walk.iv) == AES_BLOCK_SIZE) in ctr_paes_crypt()
710 crypto_inc(walk.iv, AES_BLOCK_SIZE); in ctr_paes_crypt()
729 .ivsize = AES_BLOCK_SIZE,
733 .chunksize = AES_BLOCK_SIZE,