#include <crypto/aes.h>
 #include <crypto/des.h>
+#include <crypto/internal/hash.h>
 #include <crypto/internal/skcipher.h>
 #include <crypto/scatterwalk.h>
+#include <crypto/sha.h>
 
 #include "sa2ul.h"
 
 
 /**
  * struct sa_cmdl_cfg - Command label configuration descriptor
+ * @aalg: authentication algorithm ID
  * @enc_eng_id: Encryption Engine ID supported by the SA hardware
+ * @auth_eng_id: Authentication Engine ID
  * @iv_size: Initialization Vector size
+ * @akey: Authentication key
+ * @akey_len: Authentication key length
  */
 struct sa_cmdl_cfg {
+       int aalg;
        u8 enc_eng_id;
+       u8 auth_eng_id;
        u8 iv_size;
+       const u8 *akey;
+       u16 akey_len;
 };
 
 /**
  * struct algo_data - Crypto algorithm specific data
  * @enc_eng: Encryption engine info structure
+ * @auth_eng: Authentication engine info structure
+ * @auth_ctrl: Authentication control word
+ * @hash_size: Size of digest
  * @iv_idx: iv index in psdata
  * @iv_out_size: iv out size
  * @ealg_id: Encryption Algorithm ID
+ * @aalg_id: Authentication algorithm ID
  * @mci_enc: Mode Control Instruction for Encryption algorithm
  * @mci_dec: Mode Control Instruction for Decryption
  * @inv_key: Whether the encryption algorithm demands key inversion
  */
 struct algo_data {
        struct sa_eng_info enc_eng;
+       struct sa_eng_info auth_eng;
+       u8 auth_ctrl;
+       u8 hash_size;
        u8 iv_idx;
        u8 iv_out_size;
        u8 ealg_id;
+       u8 aalg_id;
        u8 *mci_enc;
        u8 *mci_dec;
        bool inv_key;
        u32 type;               /* CRYPTO_ALG_TYPE from <linux/crypto.h> */
        union {
                struct skcipher_alg skcipher;
+               struct ahash_alg ahash;
        } alg;
        bool registered;
 };
  * @enc_offset: offset of cipher data
  * @enc_size: data to be passed to cipher engine
  * @enc_iv: cipher IV
+ * @auth_offset: offset of the authentication data
+ * @auth_size: size of the authentication data
+ * @auth_iv: authentication IV
  * @type: algorithm type for the request
  * @cmdl: command label pointer
  * @base: pointer to the base request
        u8 enc_offset;
        u16 enc_size;
        u8 *enc_iv;
+       u8 auth_offset;
+       u16 auth_size;
+       u8 *auth_iv;
        u32 type;
        u32 *cmdl;
        struct crypto_async_request *base;
        return 0;
 }
 
+/* Set Security context for the authentication engine */
+static void sa_set_sc_auth(struct algo_data *ad, const u8 *key, u16 key_sz,
+                          u8 *sc_buf)
+{
+       /* Set Authentication mode selector to hash processing */
+       sc_buf[0] = SA_HASH_PROCESSING;
+       /* Auth SW ctrl word: bit[6]=1 (upload computed hash to TLR section) */
+       sc_buf[1] = SA_UPLOAD_HASH_TO_TLR;
+       sc_buf[1] |= ad->auth_ctrl;
+
+       /* basic hash */
+       sc_buf[1] |= SA_BASIC_HASH;
+}
+
 static inline void sa_copy_iv(__be32 *out, const u8 *iv, bool size16)
 {
        int j;
 static int sa_format_cmdl_gen(struct sa_cmdl_cfg *cfg, u8 *cmdl,
                              struct sa_cmdl_upd_info *upd_info)
 {
-       u8 enc_offset = 0, total = 0;
+       u8 enc_offset = 0, auth_offset = 0, total = 0;
        u8 enc_next_eng = SA_ENG_ID_OUTPORT2;
+       u8 auth_next_eng = SA_ENG_ID_OUTPORT2;
        u32 *word_ptr = (u32 *)cmdl;
        int i;
 
        /* Iniialize the command update structure */
        memzero_explicit(upd_info, sizeof(*upd_info));
 
-       if (cfg->enc_eng_id != SA_ENG_ID_NONE)
+       if (cfg->enc_eng_id)
+               total = SA_CMDL_HEADER_SIZE_BYTES;
+
+       if (cfg->auth_eng_id)
                total = SA_CMDL_HEADER_SIZE_BYTES;
 
        if (cfg->iv_size)
 
        enc_next_eng = SA_ENG_ID_OUTPORT2;
 
-       if (cfg->enc_eng_id != SA_ENG_ID_NONE) {
+       if (cfg->enc_eng_id) {
                upd_info->flags |= SA_CMDL_UPD_ENC;
                upd_info->enc_size.index = enc_offset >> 2;
                upd_info->enc_offset.index = upd_info->enc_size.index + 1;
                }
        }
 
+       if (cfg->auth_eng_id) {
+               upd_info->flags |= SA_CMDL_UPD_AUTH;
+               upd_info->auth_size.index = auth_offset >> 2;
+               upd_info->auth_offset.index = upd_info->auth_size.index + 1;
+               cmdl[auth_offset + SA_CMDL_OFFSET_NESC] = auth_next_eng;
+               cmdl[auth_offset + SA_CMDL_OFFSET_LABEL_LEN] =
+                       SA_CMDL_HEADER_SIZE_BYTES;
+               total += SA_CMDL_HEADER_SIZE_BYTES;
+       }
+
        total = roundup(total, 8);
 
        for (i = 0; i < total / 4; i++)
                        }
                }
        }
+
+       if (likely(upd_info->flags & SA_CMDL_UPD_AUTH)) {
+               cmdl[upd_info->auth_size.index] &= ~SA_CMDL_PAYLOAD_LENGTH_MASK;
+               cmdl[upd_info->auth_size.index] |= req->auth_size;
+               cmdl[upd_info->auth_offset.index] &=
+                       ~SA_CMDL_SOP_BYPASS_LEN_MASK;
+               cmdl[upd_info->auth_offset.index] |=
+                       ((u32)req->auth_offset <<
+                        __ffs(SA_CMDL_SOP_BYPASS_LEN_MASK));
+               if (upd_info->flags & SA_CMDL_UPD_AUTH_IV) {
+                       sa_copy_iv((void *)&cmdl[upd_info->auth_iv.index],
+                                  req->auth_iv,
+                                  (upd_info->auth_iv.size > 8));
+               }
+               if (upd_info->flags & SA_CMDL_UPD_AUX_KEY) {
+                       int offset = (req->auth_size & 0xF) ? 4 : 0;
+
+                       memcpy(&cmdl[upd_info->aux_key_info.index],
+                              &upd_info->aux_key[offset], 16);
+               }
+       }
 }
 
 /* Format SWINFO words to be sent to SA */
 
 static
 int sa_init_sc(struct sa_ctx_info *ctx, const u8 *enc_key,
-              u16 enc_key_sz, struct algo_data *ad, u8 enc, u32 *swinfo)
+              u16 enc_key_sz, const u8 *auth_key, u16 auth_key_sz,
+              struct algo_data *ad, u8 enc, u32 *swinfo)
 {
        int enc_sc_offset = 0;
+       int auth_sc_offset = 0;
        u8 *sc_buf = ctx->sc;
        u16 sc_id = ctx->sc_id;
        u8 first_engine;
 
        memzero_explicit(sc_buf, SA_CTX_MAX_SZ);
 
-       enc_sc_offset = SA_CTX_PHP_PE_CTX_SZ;
+       if (ad->enc_eng.eng_id) {
+               enc_sc_offset = SA_CTX_PHP_PE_CTX_SZ;
+               first_engine = ad->enc_eng.eng_id;
+               sc_buf[1] = SA_SCCTL_FE_ENC;
+               ad->hash_size = ad->iv_out_size;
+       } else {
+               enc_sc_offset = SA_CTX_PHP_PE_CTX_SZ;
+               auth_sc_offset = enc_sc_offset + ad->enc_eng.sc_size;
+               first_engine = ad->auth_eng.eng_id;
+               sc_buf[1] = SA_SCCTL_FE_AUTH_ENC;
+               if (!ad->hash_size)
+                       return -EINVAL;
+               ad->hash_size = roundup(ad->hash_size, 8);
+       }
 
        /* SCCTL Owner info: 0=host, 1=CP_ACE */
        sc_buf[SA_CTX_SCCTL_OWNER_OFFSET] = 0;
-       /* SCCTL F/E control */
-       sc_buf[1] = SA_SCCTL_FE_ENC;
        memcpy(&sc_buf[2], &sc_id, 2);
        sc_buf[4] = 0x0;
        sc_buf[5] = PRIV_ID;
                        return -EINVAL;
        }
 
+       /* Prepare context for authentication engine */
+       if (ad->auth_eng.sc_size)
+               sa_set_sc_auth(ad, auth_key, auth_key_sz,
+                              &sc_buf[auth_sc_offset]);
+
        /* Set the ownership of context to CP_ACE */
        sc_buf[SA_CTX_SCCTL_OWNER_OFFSET] = 0x80;
 
        /* swizzle the security context */
        sa_swiz_128(sc_buf, SA_CTX_MAX_SZ);
-       /* Setup SWINFO */
-       first_engine = ad->enc_eng.eng_id;
 
        sa_set_swinfo(first_engine, ctx->sc_id, ctx->sc_phys, 1, 0,
-                     SA_SW_INFO_FLAG_EVICT, ad->iv_out_size, swinfo);
+                     SA_SW_INFO_FLAG_EVICT, ad->hash_size, swinfo);
 
        sa_dump_sc(sc_buf, ctx->sc_phys);
 
                return ret;
 
        /* Setup Encryption Security Context & Command label template */
-       if (sa_init_sc(&ctx->enc, key, keylen, ad, 1, &ctx->enc.epib[1]))
+       if (sa_init_sc(&ctx->enc, key, keylen, NULL, 0, ad, 1,
+                      &ctx->enc.epib[1]))
                goto badkey;
 
        cmdl_len = sa_format_cmdl_gen(&cfg,
        ctx->enc.cmdl_size = cmdl_len;
 
        /* Setup Decryption Security Context & Command label template */
-       if (sa_init_sc(&ctx->dec, key, keylen, ad, 0, &ctx->dec.epib[1]))
+       if (sa_init_sc(&ctx->dec, key, keylen, NULL, 0, ad, 0,
+                      &ctx->dec.epib[1]))
                goto badkey;
 
        cfg.enc_eng_id = ad->enc_eng.eng_id;
        return sa_cipher_run(req, req->iv, 0);
 }
 
+static void sa_sha_dma_in_callback(void *data)
+{
+       struct sa_rx_data *rxd = (struct sa_rx_data *)data;
+       struct ahash_request *req;
+       struct crypto_ahash *tfm;
+       unsigned int authsize;
+       int i, sg_nents;
+       size_t ml, pl;
+       u32 *result;
+       __be32 *mdptr;
+
+       req = container_of(rxd->req, struct ahash_request, base);
+       tfm = crypto_ahash_reqtfm(req);
+       authsize = crypto_ahash_digestsize(tfm);
+
+       mdptr = (__be32 *)dmaengine_desc_get_metadata_ptr(rxd->tx_in, &pl, &ml);
+       result = (u32 *)req->result;
+
+       for (i = 0; i < (authsize / 4); i++)
+               result[i] = be32_to_cpu(mdptr[i + 4]);
+
+       sg_nents = sg_nents_for_len(req->src, req->nbytes);
+       dma_unmap_sg(rxd->ddev, req->src, sg_nents, DMA_FROM_DEVICE);
+
+       kfree(rxd->split_src_sg);
+
+       kfree(rxd);
+
+       ahash_request_complete(req, 0);
+}
+
+static int zero_message_process(struct ahash_request *req)
+{
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       int sa_digest_size = crypto_ahash_digestsize(tfm);
+
+       switch (sa_digest_size) {
+       case SHA1_DIGEST_SIZE:
+               memcpy(req->result, sha1_zero_message_hash, sa_digest_size);
+               break;
+       case SHA256_DIGEST_SIZE:
+               memcpy(req->result, sha256_zero_message_hash, sa_digest_size);
+               break;
+       case SHA512_DIGEST_SIZE:
+               memcpy(req->result, sha512_zero_message_hash, sa_digest_size);
+               break;
+       default:
+               return -EINVAL;
+       }
+
+       return 0;
+}
+
+static int sa_sha_run(struct ahash_request *req)
+{
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(crypto_ahash_reqtfm(req));
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct sa_req sa_req = { 0 };
+       size_t auth_len;
+
+       auth_len = req->nbytes;
+
+       if (!auth_len)
+               return zero_message_process(req);
+
+       if (auth_len > SA_MAX_DATA_SZ ||
+           (auth_len >= SA_UNSAFE_DATA_SZ_MIN &&
+            auth_len <= SA_UNSAFE_DATA_SZ_MAX)) {
+               struct ahash_request *subreq = &rctx->fallback_req;
+               int ret = 0;
+
+               ahash_request_set_tfm(subreq, ctx->fallback.ahash);
+               subreq->base.flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP;
+
+               crypto_ahash_init(subreq);
+
+               subreq->nbytes = auth_len;
+               subreq->src = req->src;
+               subreq->result = req->result;
+
+               ret |= crypto_ahash_update(subreq);
+
+               subreq->nbytes = 0;
+
+               ret |= crypto_ahash_final(subreq);
+
+               return ret;
+       }
+
+       sa_req.size = auth_len;
+       sa_req.auth_size = auth_len;
+       sa_req.src = req->src;
+       sa_req.dst = req->src;
+       sa_req.enc = true;
+       sa_req.type = CRYPTO_ALG_TYPE_AHASH;
+       sa_req.callback = sa_sha_dma_in_callback;
+       sa_req.mdata_size = 28;
+       sa_req.ctx = ctx;
+       sa_req.base = &req->base;
+
+       return sa_run(&sa_req);
+}
+
+static int sa_sha_setup(struct sa_tfm_ctx *ctx, struct  algo_data *ad)
+{
+       int bs = crypto_shash_blocksize(ctx->shash);
+       int cmdl_len;
+       struct sa_cmdl_cfg cfg;
+
+       ad->enc_eng.sc_size = SA_CTX_ENC_TYPE1_SZ;
+       ad->auth_eng.eng_id = SA_ENG_ID_AM1;
+       ad->auth_eng.sc_size = SA_CTX_AUTH_TYPE2_SZ;
+
+       memset(ctx->authkey, 0, bs);
+       memset(&cfg, 0, sizeof(cfg));
+       cfg.aalg = ad->aalg_id;
+       cfg.enc_eng_id = ad->enc_eng.eng_id;
+       cfg.auth_eng_id = ad->auth_eng.eng_id;
+       cfg.iv_size = 0;
+       cfg.akey = NULL;
+       cfg.akey_len = 0;
+
+       /* Setup Encryption Security Context & Command label template */
+       if (sa_init_sc(&ctx->enc, NULL, 0, NULL, 0, ad, 0,
+                      &ctx->enc.epib[1]))
+               goto badkey;
+
+       cmdl_len = sa_format_cmdl_gen(&cfg,
+                                     (u8 *)ctx->enc.cmdl,
+                                     &ctx->enc.cmdl_upd_info);
+       if (cmdl_len <= 0 || (cmdl_len > SA_MAX_CMDL_WORDS * sizeof(u32)))
+               goto badkey;
+
+       ctx->enc.cmdl_size = cmdl_len;
+
+       return 0;
+
+badkey:
+       dev_err(sa_k3_dev, "%s: badkey\n", __func__);
+       return -EINVAL;
+}
+
+static int sa_sha_cra_init_alg(struct crypto_tfm *tfm, const char *alg_base)
+{
+       struct sa_tfm_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct sa_crypto_data *data = dev_get_drvdata(sa_k3_dev);
+       int ret;
+
+       memset(ctx, 0, sizeof(*ctx));
+       ctx->dev_data = data;
+       ret = sa_init_ctx_info(&ctx->enc, data);
+       if (ret)
+               return ret;
+
+       if (alg_base) {
+               ctx->shash = crypto_alloc_shash(alg_base, 0,
+                                               CRYPTO_ALG_NEED_FALLBACK);
+               if (IS_ERR(ctx->shash)) {
+                       dev_err(sa_k3_dev, "base driver %s couldn't be loaded\n",
+                               alg_base);
+                       return PTR_ERR(ctx->shash);
+               }
+               /* for fallback */
+               ctx->fallback.ahash =
+                       crypto_alloc_ahash(alg_base, 0,
+                                          CRYPTO_ALG_NEED_FALLBACK);
+               if (IS_ERR(ctx->fallback.ahash)) {
+                       dev_err(ctx->dev_data->dev,
+                               "Could not load fallback driver\n");
+                       return PTR_ERR(ctx->fallback.ahash);
+               }
+       }
+
+       dev_dbg(sa_k3_dev, "%s(0x%p) sc-ids(0x%x(0x%pad), 0x%x(0x%pad))\n",
+               __func__, tfm, ctx->enc.sc_id, &ctx->enc.sc_phys,
+               ctx->dec.sc_id, &ctx->dec.sc_phys);
+
+       crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
+                                sizeof(struct sa_sha_req_ctx) +
+                                crypto_ahash_reqsize(ctx->fallback.ahash));
+
+       return 0;
+}
+
+static int sa_sha_digest(struct ahash_request *req)
+{
+       return sa_sha_run(req);
+}
+
+static int sa_sha_init(struct ahash_request *req)
+{
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(tfm);
+
+       dev_dbg(sa_k3_dev, "init: digest size: %d, rctx=%llx\n",
+               crypto_ahash_digestsize(tfm), (u64)rctx);
+
+       ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback.ahash);
+       rctx->fallback_req.base.flags =
+               req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP;
+
+       return crypto_ahash_init(&rctx->fallback_req);
+}
+
+static int sa_sha_update(struct ahash_request *req)
+{
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(tfm);
+
+       ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback.ahash);
+       rctx->fallback_req.base.flags =
+               req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP;
+       rctx->fallback_req.nbytes = req->nbytes;
+       rctx->fallback_req.src = req->src;
+
+       return crypto_ahash_update(&rctx->fallback_req);
+}
+
+static int sa_sha_final(struct ahash_request *req)
+{
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(tfm);
+
+       ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback.ahash);
+       rctx->fallback_req.base.flags =
+               req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP;
+       rctx->fallback_req.result = req->result;
+
+       return crypto_ahash_final(&rctx->fallback_req);
+}
+
+static int sa_sha_finup(struct ahash_request *req)
+{
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(tfm);
+
+       ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback.ahash);
+       rctx->fallback_req.base.flags =
+               req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP;
+
+       rctx->fallback_req.nbytes = req->nbytes;
+       rctx->fallback_req.src = req->src;
+       rctx->fallback_req.result = req->result;
+
+       return crypto_ahash_finup(&rctx->fallback_req);
+}
+
+static int sa_sha_import(struct ahash_request *req, const void *in)
+{
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(tfm);
+
+       ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback.ahash);
+       rctx->fallback_req.base.flags = req->base.flags &
+               CRYPTO_TFM_REQ_MAY_SLEEP;
+
+       return crypto_ahash_import(&rctx->fallback_req, in);
+}
+
+static int sa_sha_export(struct ahash_request *req, void *out)
+{
+       struct sa_sha_req_ctx *rctx = ahash_request_ctx(req);
+       struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
+       struct sa_tfm_ctx *ctx = crypto_ahash_ctx(tfm);
+       struct ahash_request *subreq = &rctx->fallback_req;
+
+       ahash_request_set_tfm(subreq, ctx->fallback.ahash);
+       subreq->base.flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP;
+
+       return crypto_ahash_export(subreq, out);
+}
+
+static int sa_sha1_cra_init(struct crypto_tfm *tfm)
+{
+       struct algo_data ad = { 0 };
+       struct sa_tfm_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       sa_sha_cra_init_alg(tfm, "sha1");
+
+       ad.aalg_id = SA_AALG_ID_SHA1;
+       ad.hash_size = SHA1_DIGEST_SIZE;
+       ad.auth_ctrl = SA_AUTH_SW_CTRL_SHA1;
+
+       sa_sha_setup(ctx, &ad);
+
+       return 0;
+}
+
+static int sa_sha256_cra_init(struct crypto_tfm *tfm)
+{
+       struct algo_data ad = { 0 };
+       struct sa_tfm_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       sa_sha_cra_init_alg(tfm, "sha256");
+
+       ad.aalg_id = SA_AALG_ID_SHA2_256;
+       ad.hash_size = SHA256_DIGEST_SIZE;
+       ad.auth_ctrl = SA_AUTH_SW_CTRL_SHA256;
+
+       sa_sha_setup(ctx, &ad);
+
+       return 0;
+}
+
+static int sa_sha512_cra_init(struct crypto_tfm *tfm)
+{
+       struct algo_data ad = { 0 };
+       struct sa_tfm_ctx *ctx = crypto_tfm_ctx(tfm);
+
+       sa_sha_cra_init_alg(tfm, "sha512");
+
+       ad.aalg_id = SA_AALG_ID_SHA2_512;
+       ad.hash_size = SHA512_DIGEST_SIZE;
+       ad.auth_ctrl = SA_AUTH_SW_CTRL_SHA512;
+
+       sa_sha_setup(ctx, &ad);
+
+       return 0;
+}
+
+static void sa_sha_cra_exit(struct crypto_tfm *tfm)
+{
+       struct sa_tfm_ctx *ctx = crypto_tfm_ctx(tfm);
+       struct sa_crypto_data *data = dev_get_drvdata(sa_k3_dev);
+
+       dev_dbg(sa_k3_dev, "%s(0x%p) sc-ids(0x%x(0x%pad), 0x%x(0x%pad))\n",
+               __func__, tfm, ctx->enc.sc_id, &ctx->enc.sc_phys,
+               ctx->dec.sc_id, &ctx->dec.sc_phys);
+
+       if (crypto_tfm_alg_type(tfm) == CRYPTO_ALG_TYPE_AHASH)
+               sa_free_ctx_info(&ctx->enc, data);
+
+       crypto_free_shash(ctx->shash);
+       crypto_free_ahash(ctx->fallback.ahash);
+}
+
 static struct sa_alg_tmpl sa_algs[] = {
        {
                .type = CRYPTO_ALG_TYPE_SKCIPHER,
                        .decrypt                = sa_decrypt,
                }
        },
+       {
+               .type = CRYPTO_ALG_TYPE_AHASH,
+               .alg.ahash = {
+                       .halg.base = {
+                               .cra_name       = "sha1",
+                               .cra_driver_name        = "sha1-sa2ul",
+                               .cra_priority   = 400,
+                               .cra_flags      = CRYPTO_ALG_TYPE_AHASH |
+                                                 CRYPTO_ALG_ASYNC |
+                                                 CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                                 CRYPTO_ALG_NEED_FALLBACK,
+                               .cra_blocksize  = SHA1_BLOCK_SIZE,
+                               .cra_ctxsize    = sizeof(struct sa_tfm_ctx),
+                               .cra_module     = THIS_MODULE,
+                               .cra_init       = sa_sha1_cra_init,
+                               .cra_exit       = sa_sha_cra_exit,
+                       },
+                       .halg.digestsize        = SHA1_DIGEST_SIZE,
+                       .halg.statesize         = sizeof(struct sa_sha_req_ctx) +
+                                                 sizeof(struct sha1_state),
+                       .init                   = sa_sha_init,
+                       .update                 = sa_sha_update,
+                       .final                  = sa_sha_final,
+                       .finup                  = sa_sha_finup,
+                       .digest                 = sa_sha_digest,
+                       .export                 = sa_sha_export,
+                       .import                 = sa_sha_import,
+               },
+       },
+       {
+               .type = CRYPTO_ALG_TYPE_AHASH,
+               .alg.ahash = {
+                       .halg.base = {
+                               .cra_name       = "sha256",
+                               .cra_driver_name        = "sha256-sa2ul",
+                               .cra_priority   = 400,
+                               .cra_flags      = CRYPTO_ALG_TYPE_AHASH |
+                                                 CRYPTO_ALG_ASYNC |
+                                                 CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                                 CRYPTO_ALG_NEED_FALLBACK,
+                               .cra_blocksize  = SHA256_BLOCK_SIZE,
+                               .cra_ctxsize    = sizeof(struct sa_tfm_ctx),
+                               .cra_module     = THIS_MODULE,
+                               .cra_init       = sa_sha256_cra_init,
+                               .cra_exit       = sa_sha_cra_exit,
+                       },
+                       .halg.digestsize        = SHA256_DIGEST_SIZE,
+                       .halg.statesize         = sizeof(struct sa_sha_req_ctx) +
+                                                 sizeof(struct sha256_state),
+                       .init                   = sa_sha_init,
+                       .update                 = sa_sha_update,
+                       .final                  = sa_sha_final,
+                       .finup                  = sa_sha_finup,
+                       .digest                 = sa_sha_digest,
+                       .export                 = sa_sha_export,
+                       .import                 = sa_sha_import,
+               },
+       },
+       {
+               .type = CRYPTO_ALG_TYPE_AHASH,
+               .alg.ahash = {
+                       .halg.base = {
+                               .cra_name       = "sha512",
+                               .cra_driver_name        = "sha512-sa2ul",
+                               .cra_priority   = 400,
+                               .cra_flags      = CRYPTO_ALG_TYPE_AHASH |
+                                                 CRYPTO_ALG_ASYNC |
+                                                 CRYPTO_ALG_KERN_DRIVER_ONLY |
+                                                 CRYPTO_ALG_NEED_FALLBACK,
+                               .cra_blocksize  = SHA512_BLOCK_SIZE,
+                               .cra_ctxsize    = sizeof(struct sa_tfm_ctx),
+                               .cra_module     = THIS_MODULE,
+                               .cra_init       = sa_sha512_cra_init,
+                               .cra_exit       = sa_sha_cra_exit,
+                       },
+                       .halg.digestsize        = SHA512_DIGEST_SIZE,
+                       .halg.statesize         = sizeof(struct sa_sha_req_ctx) +
+                                                 sizeof(struct sha512_state),
+                       .init                   = sa_sha_init,
+                       .update                 = sa_sha_update,
+                       .final                  = sa_sha_final,
+                       .finup                  = sa_sha_finup,
+                       .digest                 = sa_sha_digest,
+                       .export                 = sa_sha_export,
+                       .import                 = sa_sha_import,
+               },
+       },
 };
 
 /* Register the algorithms in crypto framework */
                if (type == CRYPTO_ALG_TYPE_SKCIPHER) {
                        alg_name = sa_algs[i].alg.skcipher.base.cra_name;
                        err = crypto_register_skcipher(&sa_algs[i].alg.skcipher);
+               } else if (type == CRYPTO_ALG_TYPE_AHASH) {
+                       alg_name = sa_algs[i].alg.ahash.halg.base.cra_name;
+                       err = crypto_register_ahash(&sa_algs[i].alg.ahash);
                } else {
                        dev_err(dev,
                                "un-supported crypto algorithm (%d)",
                        continue;
                if (type == CRYPTO_ALG_TYPE_SKCIPHER)
                        crypto_unregister_skcipher(&sa_algs[i].alg.skcipher);
+               else if (type == CRYPTO_ALG_TYPE_AHASH)
+                       crypto_unregister_ahash(&sa_algs[i].alg.ahash);
 
                sa_algs[i].registered = false;
        }