};
 
 struct iv_essiv_private {
-       struct crypto_ahash *hash_tfm;
+       struct crypto_shash *hash_tfm;
        u8 *salt;
 };
 
 static int crypt_iv_essiv_init(struct crypt_config *cc)
 {
        struct iv_essiv_private *essiv = &cc->iv_gen_private.essiv;
-       AHASH_REQUEST_ON_STACK(req, essiv->hash_tfm);
-       struct scatterlist sg;
+       SHASH_DESC_ON_STACK(desc, essiv->hash_tfm);
        struct crypto_cipher *essiv_tfm;
        int err;
 
-       sg_init_one(&sg, cc->key, cc->key_size);
-       ahash_request_set_tfm(req, essiv->hash_tfm);
-       ahash_request_set_callback(req, CRYPTO_TFM_REQ_MAY_SLEEP, NULL, NULL);
-       ahash_request_set_crypt(req, &sg, essiv->salt, cc->key_size);
+       desc->tfm = essiv->hash_tfm;
+       desc->flags = CRYPTO_TFM_REQ_MAY_SLEEP;
 
-       err = crypto_ahash_digest(req);
-       ahash_request_zero(req);
+       err = crypto_shash_digest(desc, cc->key, cc->key_size, essiv->salt);
+       shash_desc_zero(desc);
        if (err)
                return err;
 
        essiv_tfm = cc->iv_private;
 
        err = crypto_cipher_setkey(essiv_tfm, essiv->salt,
-                           crypto_ahash_digestsize(essiv->hash_tfm));
+                           crypto_shash_digestsize(essiv->hash_tfm));
        if (err)
                return err;
 
 static int crypt_iv_essiv_wipe(struct crypt_config *cc)
 {
        struct iv_essiv_private *essiv = &cc->iv_gen_private.essiv;
-       unsigned salt_size = crypto_ahash_digestsize(essiv->hash_tfm);
+       unsigned salt_size = crypto_shash_digestsize(essiv->hash_tfm);
        struct crypto_cipher *essiv_tfm;
        int r, err = 0;
 
        struct crypto_cipher *essiv_tfm;
        struct iv_essiv_private *essiv = &cc->iv_gen_private.essiv;
 
-       crypto_free_ahash(essiv->hash_tfm);
+       crypto_free_shash(essiv->hash_tfm);
        essiv->hash_tfm = NULL;
 
        kzfree(essiv->salt);
                              const char *opts)
 {
        struct crypto_cipher *essiv_tfm = NULL;
-       struct crypto_ahash *hash_tfm = NULL;
+       struct crypto_shash *hash_tfm = NULL;
        u8 *salt = NULL;
        int err;
 
        }
 
        /* Allocate hash algorithm */
-       hash_tfm = crypto_alloc_ahash(opts, 0, CRYPTO_ALG_ASYNC);
+       hash_tfm = crypto_alloc_shash(opts, 0, 0);
        if (IS_ERR(hash_tfm)) {
                ti->error = "Error initializing ESSIV hash";
                err = PTR_ERR(hash_tfm);
                goto bad;
        }
 
-       salt = kzalloc(crypto_ahash_digestsize(hash_tfm), GFP_KERNEL);
+       salt = kzalloc(crypto_shash_digestsize(hash_tfm), GFP_KERNEL);
        if (!salt) {
                ti->error = "Error kmallocing salt storage in ESSIV";
                err = -ENOMEM;
        cc->iv_gen_private.essiv.hash_tfm = hash_tfm;
 
        essiv_tfm = alloc_essiv_cipher(cc, ti, salt,
-                                      crypto_ahash_digestsize(hash_tfm));
+                                      crypto_shash_digestsize(hash_tfm));
        if (IS_ERR(essiv_tfm)) {
                crypt_iv_essiv_dtr(cc);
                return PTR_ERR(essiv_tfm);
 
 bad:
        if (hash_tfm && !IS_ERR(hash_tfm))
-               crypto_free_ahash(hash_tfm);
+               crypto_free_shash(hash_tfm);
        kfree(salt);
        return err;
 }