1004a403cSLoc Ho /* 2004a403cSLoc Ho * Asynchronous Cryptographic Hash operations. 3004a403cSLoc Ho * 4004a403cSLoc Ho * This is the asynchronous version of hash.c with notification of 5004a403cSLoc Ho * completion via a callback. 6004a403cSLoc Ho * 7004a403cSLoc Ho * Copyright (c) 2008 Loc Ho <lho@amcc.com> 8004a403cSLoc Ho * 9004a403cSLoc Ho * This program is free software; you can redistribute it and/or modify it 10004a403cSLoc Ho * under the terms of the GNU General Public License as published by the Free 11004a403cSLoc Ho * Software Foundation; either version 2 of the License, or (at your option) 12004a403cSLoc Ho * any later version. 13004a403cSLoc Ho * 14004a403cSLoc Ho */ 15004a403cSLoc Ho 1620036252SHerbert Xu #include <crypto/internal/hash.h> 1720036252SHerbert Xu #include <crypto/scatterwalk.h> 1875ecb231SHerbert Xu #include <linux/bug.h> 19004a403cSLoc Ho #include <linux/err.h> 20004a403cSLoc Ho #include <linux/kernel.h> 21004a403cSLoc Ho #include <linux/module.h> 22004a403cSLoc Ho #include <linux/sched.h> 23004a403cSLoc Ho #include <linux/slab.h> 24004a403cSLoc Ho #include <linux/seq_file.h> 256238cbaeSSteffen Klassert #include <linux/cryptouser.h> 266238cbaeSSteffen Klassert #include <net/netlink.h> 27004a403cSLoc Ho 28004a403cSLoc Ho #include "internal.h" 29004a403cSLoc Ho 3066f6ce5eSHerbert Xu struct ahash_request_priv { 3166f6ce5eSHerbert Xu crypto_completion_t complete; 3266f6ce5eSHerbert Xu void *data; 3366f6ce5eSHerbert Xu u8 *result; 3466f6ce5eSHerbert Xu void *ubuf[] CRYPTO_MINALIGN_ATTR; 3566f6ce5eSHerbert Xu }; 3666f6ce5eSHerbert Xu 3788056ec3SHerbert Xu static inline struct ahash_alg *crypto_ahash_alg(struct crypto_ahash *hash) 3888056ec3SHerbert Xu { 3988056ec3SHerbert Xu return container_of(crypto_hash_alg_common(hash), struct ahash_alg, 4088056ec3SHerbert Xu halg); 4188056ec3SHerbert Xu } 4288056ec3SHerbert Xu 4320036252SHerbert Xu static int hash_walk_next(struct crypto_hash_walk *walk) 4420036252SHerbert Xu { 4520036252SHerbert Xu unsigned int alignmask = walk->alignmask; 4620036252SHerbert Xu unsigned int offset = walk->offset; 4720036252SHerbert Xu unsigned int nbytes = min(walk->entrylen, 4820036252SHerbert Xu ((unsigned int)(PAGE_SIZE)) - offset); 4920036252SHerbert Xu 5075ecb231SHerbert Xu if (walk->flags & CRYPTO_ALG_ASYNC) 5175ecb231SHerbert Xu walk->data = kmap(walk->pg); 5275ecb231SHerbert Xu else 53f0dfc0b0SCong Wang walk->data = kmap_atomic(walk->pg); 5420036252SHerbert Xu walk->data += offset; 5520036252SHerbert Xu 5623a75eeeSSzilveszter Ördög if (offset & alignmask) { 5723a75eeeSSzilveszter Ördög unsigned int unaligned = alignmask + 1 - (offset & alignmask); 5823a75eeeSSzilveszter Ördög if (nbytes > unaligned) 5923a75eeeSSzilveszter Ördög nbytes = unaligned; 6023a75eeeSSzilveszter Ördög } 6120036252SHerbert Xu 6220036252SHerbert Xu walk->entrylen -= nbytes; 6320036252SHerbert Xu return nbytes; 6420036252SHerbert Xu } 6520036252SHerbert Xu 6620036252SHerbert Xu static int hash_walk_new_entry(struct crypto_hash_walk *walk) 6720036252SHerbert Xu { 6820036252SHerbert Xu struct scatterlist *sg; 6920036252SHerbert Xu 7020036252SHerbert Xu sg = walk->sg; 7120036252SHerbert Xu walk->pg = sg_page(sg); 7220036252SHerbert Xu walk->offset = sg->offset; 7320036252SHerbert Xu walk->entrylen = sg->length; 7420036252SHerbert Xu 7520036252SHerbert Xu if (walk->entrylen > walk->total) 7620036252SHerbert Xu walk->entrylen = walk->total; 7720036252SHerbert Xu walk->total -= walk->entrylen; 7820036252SHerbert Xu 7920036252SHerbert Xu return hash_walk_next(walk); 8020036252SHerbert Xu } 8120036252SHerbert Xu 8220036252SHerbert Xu int crypto_hash_walk_done(struct crypto_hash_walk *walk, int err) 8320036252SHerbert Xu { 8420036252SHerbert Xu unsigned int alignmask = walk->alignmask; 8520036252SHerbert Xu unsigned int nbytes = walk->entrylen; 8620036252SHerbert Xu 8720036252SHerbert Xu walk->data -= walk->offset; 8820036252SHerbert Xu 8920036252SHerbert Xu if (nbytes && walk->offset & alignmask && !err) { 9020036252SHerbert Xu walk->offset = ALIGN(walk->offset, alignmask + 1); 9120036252SHerbert Xu walk->data += walk->offset; 9220036252SHerbert Xu 9320036252SHerbert Xu nbytes = min(nbytes, 9420036252SHerbert Xu ((unsigned int)(PAGE_SIZE)) - walk->offset); 9520036252SHerbert Xu walk->entrylen -= nbytes; 9620036252SHerbert Xu 9720036252SHerbert Xu return nbytes; 9820036252SHerbert Xu } 9920036252SHerbert Xu 10075ecb231SHerbert Xu if (walk->flags & CRYPTO_ALG_ASYNC) 10175ecb231SHerbert Xu kunmap(walk->pg); 10275ecb231SHerbert Xu else { 103f0dfc0b0SCong Wang kunmap_atomic(walk->data); 10475ecb231SHerbert Xu /* 10575ecb231SHerbert Xu * The may sleep test only makes sense for sync users. 10675ecb231SHerbert Xu * Async users don't need to sleep here anyway. 10775ecb231SHerbert Xu */ 10820036252SHerbert Xu crypto_yield(walk->flags); 10975ecb231SHerbert Xu } 11020036252SHerbert Xu 11120036252SHerbert Xu if (err) 11220036252SHerbert Xu return err; 11320036252SHerbert Xu 114d315a0e0SHerbert Xu if (nbytes) { 11520036252SHerbert Xu walk->offset = 0; 116d315a0e0SHerbert Xu walk->pg++; 11720036252SHerbert Xu return hash_walk_next(walk); 118d315a0e0SHerbert Xu } 11920036252SHerbert Xu 12020036252SHerbert Xu if (!walk->total) 12120036252SHerbert Xu return 0; 12220036252SHerbert Xu 12320036252SHerbert Xu walk->sg = scatterwalk_sg_next(walk->sg); 12420036252SHerbert Xu 12520036252SHerbert Xu return hash_walk_new_entry(walk); 12620036252SHerbert Xu } 12720036252SHerbert Xu EXPORT_SYMBOL_GPL(crypto_hash_walk_done); 12820036252SHerbert Xu 12920036252SHerbert Xu int crypto_hash_walk_first(struct ahash_request *req, 13020036252SHerbert Xu struct crypto_hash_walk *walk) 13120036252SHerbert Xu { 13220036252SHerbert Xu walk->total = req->nbytes; 13320036252SHerbert Xu 134*6d9529c5STim Chen if (!walk->total) { 135*6d9529c5STim Chen walk->entrylen = 0; 13620036252SHerbert Xu return 0; 137*6d9529c5STim Chen } 13820036252SHerbert Xu 13920036252SHerbert Xu walk->alignmask = crypto_ahash_alignmask(crypto_ahash_reqtfm(req)); 14020036252SHerbert Xu walk->sg = req->src; 14175ecb231SHerbert Xu walk->flags = req->base.flags & CRYPTO_TFM_REQ_MASK; 14220036252SHerbert Xu 14320036252SHerbert Xu return hash_walk_new_entry(walk); 14420036252SHerbert Xu } 14520036252SHerbert Xu EXPORT_SYMBOL_GPL(crypto_hash_walk_first); 14620036252SHerbert Xu 14775ecb231SHerbert Xu int crypto_ahash_walk_first(struct ahash_request *req, 14875ecb231SHerbert Xu struct crypto_hash_walk *walk) 14975ecb231SHerbert Xu { 15075ecb231SHerbert Xu walk->total = req->nbytes; 15175ecb231SHerbert Xu 152*6d9529c5STim Chen if (!walk->total) { 153*6d9529c5STim Chen walk->entrylen = 0; 15475ecb231SHerbert Xu return 0; 155*6d9529c5STim Chen } 15675ecb231SHerbert Xu 15775ecb231SHerbert Xu walk->alignmask = crypto_ahash_alignmask(crypto_ahash_reqtfm(req)); 15875ecb231SHerbert Xu walk->sg = req->src; 15975ecb231SHerbert Xu walk->flags = req->base.flags & CRYPTO_TFM_REQ_MASK; 16075ecb231SHerbert Xu walk->flags |= CRYPTO_ALG_ASYNC; 16175ecb231SHerbert Xu 16275ecb231SHerbert Xu BUILD_BUG_ON(CRYPTO_TFM_REQ_MASK & CRYPTO_ALG_ASYNC); 16375ecb231SHerbert Xu 16475ecb231SHerbert Xu return hash_walk_new_entry(walk); 16575ecb231SHerbert Xu } 16675ecb231SHerbert Xu EXPORT_SYMBOL_GPL(crypto_ahash_walk_first); 16775ecb231SHerbert Xu 1685f7082edSHerbert Xu int crypto_hash_walk_first_compat(struct hash_desc *hdesc, 1695f7082edSHerbert Xu struct crypto_hash_walk *walk, 1705f7082edSHerbert Xu struct scatterlist *sg, unsigned int len) 1715f7082edSHerbert Xu { 1725f7082edSHerbert Xu walk->total = len; 1735f7082edSHerbert Xu 174*6d9529c5STim Chen if (!walk->total) { 175*6d9529c5STim Chen walk->entrylen = 0; 1765f7082edSHerbert Xu return 0; 177*6d9529c5STim Chen } 1785f7082edSHerbert Xu 1795f7082edSHerbert Xu walk->alignmask = crypto_hash_alignmask(hdesc->tfm); 1805f7082edSHerbert Xu walk->sg = sg; 18175ecb231SHerbert Xu walk->flags = hdesc->flags & CRYPTO_TFM_REQ_MASK; 1825f7082edSHerbert Xu 1835f7082edSHerbert Xu return hash_walk_new_entry(walk); 1845f7082edSHerbert Xu } 1855f7082edSHerbert Xu 186004a403cSLoc Ho static int ahash_setkey_unaligned(struct crypto_ahash *tfm, const u8 *key, 187004a403cSLoc Ho unsigned int keylen) 188004a403cSLoc Ho { 189004a403cSLoc Ho unsigned long alignmask = crypto_ahash_alignmask(tfm); 190004a403cSLoc Ho int ret; 191004a403cSLoc Ho u8 *buffer, *alignbuffer; 192004a403cSLoc Ho unsigned long absize; 193004a403cSLoc Ho 194004a403cSLoc Ho absize = keylen + alignmask; 195093900c2SHerbert Xu buffer = kmalloc(absize, GFP_KERNEL); 196004a403cSLoc Ho if (!buffer) 197004a403cSLoc Ho return -ENOMEM; 198004a403cSLoc Ho 199004a403cSLoc Ho alignbuffer = (u8 *)ALIGN((unsigned long)buffer, alignmask + 1); 200004a403cSLoc Ho memcpy(alignbuffer, key, keylen); 201a70c5225SHerbert Xu ret = tfm->setkey(tfm, alignbuffer, keylen); 2028c32c516SHerbert Xu kzfree(buffer); 203004a403cSLoc Ho return ret; 204004a403cSLoc Ho } 205004a403cSLoc Ho 20666f6ce5eSHerbert Xu int crypto_ahash_setkey(struct crypto_ahash *tfm, const u8 *key, 207004a403cSLoc Ho unsigned int keylen) 208004a403cSLoc Ho { 209004a403cSLoc Ho unsigned long alignmask = crypto_ahash_alignmask(tfm); 210004a403cSLoc Ho 211004a403cSLoc Ho if ((unsigned long)key & alignmask) 212004a403cSLoc Ho return ahash_setkey_unaligned(tfm, key, keylen); 213004a403cSLoc Ho 214a70c5225SHerbert Xu return tfm->setkey(tfm, key, keylen); 215004a403cSLoc Ho } 21666f6ce5eSHerbert Xu EXPORT_SYMBOL_GPL(crypto_ahash_setkey); 217004a403cSLoc Ho 2183751f402SHerbert Xu static int ahash_nosetkey(struct crypto_ahash *tfm, const u8 *key, 2193751f402SHerbert Xu unsigned int keylen) 2203751f402SHerbert Xu { 2213751f402SHerbert Xu return -ENOSYS; 2223751f402SHerbert Xu } 2233751f402SHerbert Xu 22466f6ce5eSHerbert Xu static inline unsigned int ahash_align_buffer_size(unsigned len, 22566f6ce5eSHerbert Xu unsigned long mask) 22666f6ce5eSHerbert Xu { 22766f6ce5eSHerbert Xu return len + (mask & ~(crypto_tfm_ctx_alignment() - 1)); 22866f6ce5eSHerbert Xu } 22966f6ce5eSHerbert Xu 2301ffc9fbdSMarek Vasut static int ahash_save_req(struct ahash_request *req, crypto_completion_t cplt) 23166f6ce5eSHerbert Xu { 23266f6ce5eSHerbert Xu struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); 23366f6ce5eSHerbert Xu unsigned long alignmask = crypto_ahash_alignmask(tfm); 23466f6ce5eSHerbert Xu unsigned int ds = crypto_ahash_digestsize(tfm); 23566f6ce5eSHerbert Xu struct ahash_request_priv *priv; 23666f6ce5eSHerbert Xu 23766f6ce5eSHerbert Xu priv = kmalloc(sizeof(*priv) + ahash_align_buffer_size(ds, alignmask), 23866f6ce5eSHerbert Xu (req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? 2395befbd5aSSteffen Klassert GFP_KERNEL : GFP_ATOMIC); 24066f6ce5eSHerbert Xu if (!priv) 24166f6ce5eSHerbert Xu return -ENOMEM; 24266f6ce5eSHerbert Xu 243ab6bf4e5SMarek Vasut /* 244ab6bf4e5SMarek Vasut * WARNING: Voodoo programming below! 245ab6bf4e5SMarek Vasut * 246ab6bf4e5SMarek Vasut * The code below is obscure and hard to understand, thus explanation 247ab6bf4e5SMarek Vasut * is necessary. See include/crypto/hash.h and include/linux/crypto.h 248ab6bf4e5SMarek Vasut * to understand the layout of structures used here! 249ab6bf4e5SMarek Vasut * 250ab6bf4e5SMarek Vasut * The code here will replace portions of the ORIGINAL request with 251ab6bf4e5SMarek Vasut * pointers to new code and buffers so the hashing operation can store 252ab6bf4e5SMarek Vasut * the result in aligned buffer. We will call the modified request 253ab6bf4e5SMarek Vasut * an ADJUSTED request. 254ab6bf4e5SMarek Vasut * 255ab6bf4e5SMarek Vasut * The newly mangled request will look as such: 256ab6bf4e5SMarek Vasut * 257ab6bf4e5SMarek Vasut * req { 258ab6bf4e5SMarek Vasut * .result = ADJUSTED[new aligned buffer] 259ab6bf4e5SMarek Vasut * .base.complete = ADJUSTED[pointer to completion function] 260ab6bf4e5SMarek Vasut * .base.data = ADJUSTED[*req (pointer to self)] 261ab6bf4e5SMarek Vasut * .priv = ADJUSTED[new priv] { 262ab6bf4e5SMarek Vasut * .result = ORIGINAL(result) 263ab6bf4e5SMarek Vasut * .complete = ORIGINAL(base.complete) 264ab6bf4e5SMarek Vasut * .data = ORIGINAL(base.data) 265ab6bf4e5SMarek Vasut * } 266ab6bf4e5SMarek Vasut */ 267ab6bf4e5SMarek Vasut 26866f6ce5eSHerbert Xu priv->result = req->result; 26966f6ce5eSHerbert Xu priv->complete = req->base.complete; 27066f6ce5eSHerbert Xu priv->data = req->base.data; 271ab6bf4e5SMarek Vasut /* 272ab6bf4e5SMarek Vasut * WARNING: We do not backup req->priv here! The req->priv 273ab6bf4e5SMarek Vasut * is for internal use of the Crypto API and the 274ab6bf4e5SMarek Vasut * user must _NOT_ _EVER_ depend on it's content! 275ab6bf4e5SMarek Vasut */ 27666f6ce5eSHerbert Xu 27766f6ce5eSHerbert Xu req->result = PTR_ALIGN((u8 *)priv->ubuf, alignmask + 1); 2781ffc9fbdSMarek Vasut req->base.complete = cplt; 27966f6ce5eSHerbert Xu req->base.data = req; 28066f6ce5eSHerbert Xu req->priv = priv; 28166f6ce5eSHerbert Xu 2821ffc9fbdSMarek Vasut return 0; 2831ffc9fbdSMarek Vasut } 2841ffc9fbdSMarek Vasut 2851ffc9fbdSMarek Vasut static void ahash_restore_req(struct ahash_request *req) 2861ffc9fbdSMarek Vasut { 2871ffc9fbdSMarek Vasut struct ahash_request_priv *priv = req->priv; 2881ffc9fbdSMarek Vasut 2891ffc9fbdSMarek Vasut /* Restore the original crypto request. */ 2901ffc9fbdSMarek Vasut req->result = priv->result; 2911ffc9fbdSMarek Vasut req->base.complete = priv->complete; 2921ffc9fbdSMarek Vasut req->base.data = priv->data; 2931ffc9fbdSMarek Vasut req->priv = NULL; 2941ffc9fbdSMarek Vasut 2951ffc9fbdSMarek Vasut /* Free the req->priv.priv from the ADJUSTED request. */ 2961ffc9fbdSMarek Vasut kzfree(priv); 2971ffc9fbdSMarek Vasut } 2981ffc9fbdSMarek Vasut 2991ffc9fbdSMarek Vasut static void ahash_op_unaligned_finish(struct ahash_request *req, int err) 3001ffc9fbdSMarek Vasut { 3011ffc9fbdSMarek Vasut struct ahash_request_priv *priv = req->priv; 3021ffc9fbdSMarek Vasut 3031ffc9fbdSMarek Vasut if (err == -EINPROGRESS) 3041ffc9fbdSMarek Vasut return; 3051ffc9fbdSMarek Vasut 3061ffc9fbdSMarek Vasut if (!err) 3071ffc9fbdSMarek Vasut memcpy(priv->result, req->result, 3081ffc9fbdSMarek Vasut crypto_ahash_digestsize(crypto_ahash_reqtfm(req))); 3091ffc9fbdSMarek Vasut 3101ffc9fbdSMarek Vasut ahash_restore_req(req); 3111ffc9fbdSMarek Vasut } 3121ffc9fbdSMarek Vasut 3131ffc9fbdSMarek Vasut static void ahash_op_unaligned_done(struct crypto_async_request *req, int err) 3141ffc9fbdSMarek Vasut { 3151ffc9fbdSMarek Vasut struct ahash_request *areq = req->data; 3161ffc9fbdSMarek Vasut 3171ffc9fbdSMarek Vasut /* 3181ffc9fbdSMarek Vasut * Restore the original request, see ahash_op_unaligned() for what 3191ffc9fbdSMarek Vasut * goes where. 3201ffc9fbdSMarek Vasut * 3211ffc9fbdSMarek Vasut * The "struct ahash_request *req" here is in fact the "req.base" 3221ffc9fbdSMarek Vasut * from the ADJUSTED request from ahash_op_unaligned(), thus as it 3231ffc9fbdSMarek Vasut * is a pointer to self, it is also the ADJUSTED "req" . 3241ffc9fbdSMarek Vasut */ 3251ffc9fbdSMarek Vasut 3261ffc9fbdSMarek Vasut /* First copy req->result into req->priv.result */ 3271ffc9fbdSMarek Vasut ahash_op_unaligned_finish(areq, err); 3281ffc9fbdSMarek Vasut 3291ffc9fbdSMarek Vasut /* Complete the ORIGINAL request. */ 3301ffc9fbdSMarek Vasut areq->base.complete(&areq->base, err); 3311ffc9fbdSMarek Vasut } 3321ffc9fbdSMarek Vasut 3331ffc9fbdSMarek Vasut static int ahash_op_unaligned(struct ahash_request *req, 3341ffc9fbdSMarek Vasut int (*op)(struct ahash_request *)) 3351ffc9fbdSMarek Vasut { 3361ffc9fbdSMarek Vasut int err; 3371ffc9fbdSMarek Vasut 3381ffc9fbdSMarek Vasut err = ahash_save_req(req, ahash_op_unaligned_done); 3391ffc9fbdSMarek Vasut if (err) 3401ffc9fbdSMarek Vasut return err; 3411ffc9fbdSMarek Vasut 34266f6ce5eSHerbert Xu err = op(req); 34366f6ce5eSHerbert Xu ahash_op_unaligned_finish(req, err); 34466f6ce5eSHerbert Xu 34566f6ce5eSHerbert Xu return err; 34666f6ce5eSHerbert Xu } 34766f6ce5eSHerbert Xu 34866f6ce5eSHerbert Xu static int crypto_ahash_op(struct ahash_request *req, 34966f6ce5eSHerbert Xu int (*op)(struct ahash_request *)) 35066f6ce5eSHerbert Xu { 35166f6ce5eSHerbert Xu struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); 35266f6ce5eSHerbert Xu unsigned long alignmask = crypto_ahash_alignmask(tfm); 35366f6ce5eSHerbert Xu 35466f6ce5eSHerbert Xu if ((unsigned long)req->result & alignmask) 35566f6ce5eSHerbert Xu return ahash_op_unaligned(req, op); 35666f6ce5eSHerbert Xu 35766f6ce5eSHerbert Xu return op(req); 35866f6ce5eSHerbert Xu } 35966f6ce5eSHerbert Xu 36066f6ce5eSHerbert Xu int crypto_ahash_final(struct ahash_request *req) 36166f6ce5eSHerbert Xu { 36266f6ce5eSHerbert Xu return crypto_ahash_op(req, crypto_ahash_reqtfm(req)->final); 36366f6ce5eSHerbert Xu } 36466f6ce5eSHerbert Xu EXPORT_SYMBOL_GPL(crypto_ahash_final); 36566f6ce5eSHerbert Xu 36666f6ce5eSHerbert Xu int crypto_ahash_finup(struct ahash_request *req) 36766f6ce5eSHerbert Xu { 36866f6ce5eSHerbert Xu return crypto_ahash_op(req, crypto_ahash_reqtfm(req)->finup); 36966f6ce5eSHerbert Xu } 37066f6ce5eSHerbert Xu EXPORT_SYMBOL_GPL(crypto_ahash_finup); 37166f6ce5eSHerbert Xu 37266f6ce5eSHerbert Xu int crypto_ahash_digest(struct ahash_request *req) 37366f6ce5eSHerbert Xu { 37466f6ce5eSHerbert Xu return crypto_ahash_op(req, crypto_ahash_reqtfm(req)->digest); 37566f6ce5eSHerbert Xu } 37666f6ce5eSHerbert Xu EXPORT_SYMBOL_GPL(crypto_ahash_digest); 37766f6ce5eSHerbert Xu 37866f6ce5eSHerbert Xu static void ahash_def_finup_finish2(struct ahash_request *req, int err) 37966f6ce5eSHerbert Xu { 38066f6ce5eSHerbert Xu struct ahash_request_priv *priv = req->priv; 38166f6ce5eSHerbert Xu 38266f6ce5eSHerbert Xu if (err == -EINPROGRESS) 38366f6ce5eSHerbert Xu return; 38466f6ce5eSHerbert Xu 38566f6ce5eSHerbert Xu if (!err) 38666f6ce5eSHerbert Xu memcpy(priv->result, req->result, 38766f6ce5eSHerbert Xu crypto_ahash_digestsize(crypto_ahash_reqtfm(req))); 38866f6ce5eSHerbert Xu 389d4a7a0fbSMarek Vasut ahash_restore_req(req); 39066f6ce5eSHerbert Xu } 39166f6ce5eSHerbert Xu 39266f6ce5eSHerbert Xu static void ahash_def_finup_done2(struct crypto_async_request *req, int err) 39366f6ce5eSHerbert Xu { 39466f6ce5eSHerbert Xu struct ahash_request *areq = req->data; 39566f6ce5eSHerbert Xu 39666f6ce5eSHerbert Xu ahash_def_finup_finish2(areq, err); 39766f6ce5eSHerbert Xu 398d4a7a0fbSMarek Vasut areq->base.complete(&areq->base, err); 39966f6ce5eSHerbert Xu } 40066f6ce5eSHerbert Xu 40166f6ce5eSHerbert Xu static int ahash_def_finup_finish1(struct ahash_request *req, int err) 40266f6ce5eSHerbert Xu { 40366f6ce5eSHerbert Xu if (err) 40466f6ce5eSHerbert Xu goto out; 40566f6ce5eSHerbert Xu 40666f6ce5eSHerbert Xu req->base.complete = ahash_def_finup_done2; 40766f6ce5eSHerbert Xu req->base.flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; 40866f6ce5eSHerbert Xu err = crypto_ahash_reqtfm(req)->final(req); 40966f6ce5eSHerbert Xu 41066f6ce5eSHerbert Xu out: 41166f6ce5eSHerbert Xu ahash_def_finup_finish2(req, err); 41266f6ce5eSHerbert Xu return err; 41366f6ce5eSHerbert Xu } 41466f6ce5eSHerbert Xu 41566f6ce5eSHerbert Xu static void ahash_def_finup_done1(struct crypto_async_request *req, int err) 41666f6ce5eSHerbert Xu { 41766f6ce5eSHerbert Xu struct ahash_request *areq = req->data; 41866f6ce5eSHerbert Xu 41966f6ce5eSHerbert Xu err = ahash_def_finup_finish1(areq, err); 42066f6ce5eSHerbert Xu 421d4a7a0fbSMarek Vasut areq->base.complete(&areq->base, err); 42266f6ce5eSHerbert Xu } 42366f6ce5eSHerbert Xu 42466f6ce5eSHerbert Xu static int ahash_def_finup(struct ahash_request *req) 42566f6ce5eSHerbert Xu { 42666f6ce5eSHerbert Xu struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); 427d4a7a0fbSMarek Vasut int err; 42866f6ce5eSHerbert Xu 429d4a7a0fbSMarek Vasut err = ahash_save_req(req, ahash_def_finup_done1); 430d4a7a0fbSMarek Vasut if (err) 431d4a7a0fbSMarek Vasut return err; 43266f6ce5eSHerbert Xu 433d4a7a0fbSMarek Vasut err = tfm->update(req); 434d4a7a0fbSMarek Vasut return ahash_def_finup_finish1(req, err); 43566f6ce5eSHerbert Xu } 43666f6ce5eSHerbert Xu 43766f6ce5eSHerbert Xu static int ahash_no_export(struct ahash_request *req, void *out) 43866f6ce5eSHerbert Xu { 43966f6ce5eSHerbert Xu return -ENOSYS; 44066f6ce5eSHerbert Xu } 44166f6ce5eSHerbert Xu 44266f6ce5eSHerbert Xu static int ahash_no_import(struct ahash_request *req, const void *in) 44366f6ce5eSHerbert Xu { 44466f6ce5eSHerbert Xu return -ENOSYS; 44566f6ce5eSHerbert Xu } 44666f6ce5eSHerbert Xu 44788056ec3SHerbert Xu static int crypto_ahash_init_tfm(struct crypto_tfm *tfm) 44888056ec3SHerbert Xu { 44988056ec3SHerbert Xu struct crypto_ahash *hash = __crypto_ahash_cast(tfm); 45088056ec3SHerbert Xu struct ahash_alg *alg = crypto_ahash_alg(hash); 45188056ec3SHerbert Xu 45266f6ce5eSHerbert Xu hash->setkey = ahash_nosetkey; 45366f6ce5eSHerbert Xu hash->export = ahash_no_export; 45466f6ce5eSHerbert Xu hash->import = ahash_no_import; 45566f6ce5eSHerbert Xu 45688056ec3SHerbert Xu if (tfm->__crt_alg->cra_type != &crypto_ahash_type) 45788056ec3SHerbert Xu return crypto_init_shash_ops_async(tfm); 45888056ec3SHerbert Xu 45988056ec3SHerbert Xu hash->init = alg->init; 46088056ec3SHerbert Xu hash->update = alg->update; 46188056ec3SHerbert Xu hash->final = alg->final; 46266f6ce5eSHerbert Xu hash->finup = alg->finup ?: ahash_def_finup; 46388056ec3SHerbert Xu hash->digest = alg->digest; 46466f6ce5eSHerbert Xu 46566f6ce5eSHerbert Xu if (alg->setkey) 46666f6ce5eSHerbert Xu hash->setkey = alg->setkey; 46766f6ce5eSHerbert Xu if (alg->export) 46866f6ce5eSHerbert Xu hash->export = alg->export; 46966f6ce5eSHerbert Xu if (alg->import) 47066f6ce5eSHerbert Xu hash->import = alg->import; 47188056ec3SHerbert Xu 47288056ec3SHerbert Xu return 0; 47388056ec3SHerbert Xu } 47488056ec3SHerbert Xu 47588056ec3SHerbert Xu static unsigned int crypto_ahash_extsize(struct crypto_alg *alg) 47688056ec3SHerbert Xu { 47788056ec3SHerbert Xu if (alg->cra_type == &crypto_ahash_type) 47888056ec3SHerbert Xu return alg->cra_ctxsize; 47988056ec3SHerbert Xu 48088056ec3SHerbert Xu return sizeof(struct crypto_shash *); 48188056ec3SHerbert Xu } 48288056ec3SHerbert Xu 4833acc8473SHerbert Xu #ifdef CONFIG_NET 4846238cbaeSSteffen Klassert static int crypto_ahash_report(struct sk_buff *skb, struct crypto_alg *alg) 4856238cbaeSSteffen Klassert { 4866238cbaeSSteffen Klassert struct crypto_report_hash rhash; 4876238cbaeSSteffen Klassert 4889a5467bfSMathias Krause strncpy(rhash.type, "ahash", sizeof(rhash.type)); 4896238cbaeSSteffen Klassert 4906238cbaeSSteffen Klassert rhash.blocksize = alg->cra_blocksize; 4916238cbaeSSteffen Klassert rhash.digestsize = __crypto_hash_alg_common(alg)->digestsize; 4926238cbaeSSteffen Klassert 4936662df33SDavid S. Miller if (nla_put(skb, CRYPTOCFGA_REPORT_HASH, 4946662df33SDavid S. Miller sizeof(struct crypto_report_hash), &rhash)) 4956662df33SDavid S. Miller goto nla_put_failure; 4966238cbaeSSteffen Klassert return 0; 4976238cbaeSSteffen Klassert 4986238cbaeSSteffen Klassert nla_put_failure: 4996238cbaeSSteffen Klassert return -EMSGSIZE; 5006238cbaeSSteffen Klassert } 5013acc8473SHerbert Xu #else 5023acc8473SHerbert Xu static int crypto_ahash_report(struct sk_buff *skb, struct crypto_alg *alg) 5033acc8473SHerbert Xu { 5043acc8473SHerbert Xu return -ENOSYS; 5053acc8473SHerbert Xu } 5063acc8473SHerbert Xu #endif 5076238cbaeSSteffen Klassert 508004a403cSLoc Ho static void crypto_ahash_show(struct seq_file *m, struct crypto_alg *alg) 509004a403cSLoc Ho __attribute__ ((unused)); 510004a403cSLoc Ho static void crypto_ahash_show(struct seq_file *m, struct crypto_alg *alg) 511004a403cSLoc Ho { 512004a403cSLoc Ho seq_printf(m, "type : ahash\n"); 513004a403cSLoc Ho seq_printf(m, "async : %s\n", alg->cra_flags & CRYPTO_ALG_ASYNC ? 514004a403cSLoc Ho "yes" : "no"); 515004a403cSLoc Ho seq_printf(m, "blocksize : %u\n", alg->cra_blocksize); 51688056ec3SHerbert Xu seq_printf(m, "digestsize : %u\n", 51788056ec3SHerbert Xu __crypto_hash_alg_common(alg)->digestsize); 518004a403cSLoc Ho } 519004a403cSLoc Ho 520004a403cSLoc Ho const struct crypto_type crypto_ahash_type = { 52188056ec3SHerbert Xu .extsize = crypto_ahash_extsize, 52288056ec3SHerbert Xu .init_tfm = crypto_ahash_init_tfm, 523004a403cSLoc Ho #ifdef CONFIG_PROC_FS 524004a403cSLoc Ho .show = crypto_ahash_show, 525004a403cSLoc Ho #endif 5266238cbaeSSteffen Klassert .report = crypto_ahash_report, 52788056ec3SHerbert Xu .maskclear = ~CRYPTO_ALG_TYPE_MASK, 52888056ec3SHerbert Xu .maskset = CRYPTO_ALG_TYPE_AHASH_MASK, 52988056ec3SHerbert Xu .type = CRYPTO_ALG_TYPE_AHASH, 53088056ec3SHerbert Xu .tfmsize = offsetof(struct crypto_ahash, base), 531004a403cSLoc Ho }; 532004a403cSLoc Ho EXPORT_SYMBOL_GPL(crypto_ahash_type); 533004a403cSLoc Ho 53488056ec3SHerbert Xu struct crypto_ahash *crypto_alloc_ahash(const char *alg_name, u32 type, 53588056ec3SHerbert Xu u32 mask) 53688056ec3SHerbert Xu { 53788056ec3SHerbert Xu return crypto_alloc_tfm(alg_name, &crypto_ahash_type, type, mask); 53888056ec3SHerbert Xu } 53988056ec3SHerbert Xu EXPORT_SYMBOL_GPL(crypto_alloc_ahash); 54088056ec3SHerbert Xu 54101c2deceSHerbert Xu static int ahash_prepare_alg(struct ahash_alg *alg) 54201c2deceSHerbert Xu { 54301c2deceSHerbert Xu struct crypto_alg *base = &alg->halg.base; 54401c2deceSHerbert Xu 54501c2deceSHerbert Xu if (alg->halg.digestsize > PAGE_SIZE / 8 || 54601c2deceSHerbert Xu alg->halg.statesize > PAGE_SIZE / 8) 54701c2deceSHerbert Xu return -EINVAL; 54801c2deceSHerbert Xu 54901c2deceSHerbert Xu base->cra_type = &crypto_ahash_type; 55001c2deceSHerbert Xu base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; 55101c2deceSHerbert Xu base->cra_flags |= CRYPTO_ALG_TYPE_AHASH; 55201c2deceSHerbert Xu 55301c2deceSHerbert Xu return 0; 55401c2deceSHerbert Xu } 55501c2deceSHerbert Xu 55601c2deceSHerbert Xu int crypto_register_ahash(struct ahash_alg *alg) 55701c2deceSHerbert Xu { 55801c2deceSHerbert Xu struct crypto_alg *base = &alg->halg.base; 55901c2deceSHerbert Xu int err; 56001c2deceSHerbert Xu 56101c2deceSHerbert Xu err = ahash_prepare_alg(alg); 56201c2deceSHerbert Xu if (err) 56301c2deceSHerbert Xu return err; 56401c2deceSHerbert Xu 56501c2deceSHerbert Xu return crypto_register_alg(base); 56601c2deceSHerbert Xu } 56701c2deceSHerbert Xu EXPORT_SYMBOL_GPL(crypto_register_ahash); 56801c2deceSHerbert Xu 56901c2deceSHerbert Xu int crypto_unregister_ahash(struct ahash_alg *alg) 57001c2deceSHerbert Xu { 57101c2deceSHerbert Xu return crypto_unregister_alg(&alg->halg.base); 57201c2deceSHerbert Xu } 57301c2deceSHerbert Xu EXPORT_SYMBOL_GPL(crypto_unregister_ahash); 57401c2deceSHerbert Xu 57501c2deceSHerbert Xu int ahash_register_instance(struct crypto_template *tmpl, 57601c2deceSHerbert Xu struct ahash_instance *inst) 57701c2deceSHerbert Xu { 57801c2deceSHerbert Xu int err; 57901c2deceSHerbert Xu 58001c2deceSHerbert Xu err = ahash_prepare_alg(&inst->alg); 58101c2deceSHerbert Xu if (err) 58201c2deceSHerbert Xu return err; 58301c2deceSHerbert Xu 58401c2deceSHerbert Xu return crypto_register_instance(tmpl, ahash_crypto_instance(inst)); 58501c2deceSHerbert Xu } 58601c2deceSHerbert Xu EXPORT_SYMBOL_GPL(ahash_register_instance); 58701c2deceSHerbert Xu 58801c2deceSHerbert Xu void ahash_free_instance(struct crypto_instance *inst) 58901c2deceSHerbert Xu { 59001c2deceSHerbert Xu crypto_drop_spawn(crypto_instance_ctx(inst)); 59101c2deceSHerbert Xu kfree(ahash_instance(inst)); 59201c2deceSHerbert Xu } 59301c2deceSHerbert Xu EXPORT_SYMBOL_GPL(ahash_free_instance); 59401c2deceSHerbert Xu 59501c2deceSHerbert Xu int crypto_init_ahash_spawn(struct crypto_ahash_spawn *spawn, 59601c2deceSHerbert Xu struct hash_alg_common *alg, 59701c2deceSHerbert Xu struct crypto_instance *inst) 59801c2deceSHerbert Xu { 59901c2deceSHerbert Xu return crypto_init_spawn2(&spawn->base, &alg->base, inst, 60001c2deceSHerbert Xu &crypto_ahash_type); 60101c2deceSHerbert Xu } 60201c2deceSHerbert Xu EXPORT_SYMBOL_GPL(crypto_init_ahash_spawn); 60301c2deceSHerbert Xu 60401c2deceSHerbert Xu struct hash_alg_common *ahash_attr_alg(struct rtattr *rta, u32 type, u32 mask) 60501c2deceSHerbert Xu { 60601c2deceSHerbert Xu struct crypto_alg *alg; 60701c2deceSHerbert Xu 60801c2deceSHerbert Xu alg = crypto_attr_alg2(rta, &crypto_ahash_type, type, mask); 60901c2deceSHerbert Xu return IS_ERR(alg) ? ERR_CAST(alg) : __crypto_hash_alg_common(alg); 61001c2deceSHerbert Xu } 61101c2deceSHerbert Xu EXPORT_SYMBOL_GPL(ahash_attr_alg); 61201c2deceSHerbert Xu 613004a403cSLoc Ho MODULE_LICENSE("GPL"); 614004a403cSLoc Ho MODULE_DESCRIPTION("Asynchronous cryptographic hash type"); 615