* Encryption (AES128)
  */
 static int mxs_dcp_run_aes(struct dcp_async_ctx *actx,
-                          struct ablkcipher_request *req, int init)
+                          struct skcipher_request *req, int init)
 {
        struct dcp *sdcp = global_sdcp;
        struct dcp_dma_desc *desc = &sdcp->coh->desc[actx->chan];
-       struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req);
+       struct dcp_aes_req_ctx *rctx = skcipher_request_ctx(req);
        int ret;
 
        dma_addr_t key_phys = dma_map_single(sdcp->dev, sdcp->coh->aes_key,
 {
        struct dcp *sdcp = global_sdcp;
 
-       struct ablkcipher_request *req = ablkcipher_request_cast(arq);
+       struct skcipher_request *req = skcipher_request_cast(arq);
        struct dcp_async_ctx *actx = crypto_tfm_ctx(arq->tfm);
-       struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req);
+       struct dcp_aes_req_ctx *rctx = skcipher_request_ctx(req);
 
        struct scatterlist *dst = req->dst;
        struct scatterlist *src = req->src;
 
        if (!rctx->ecb) {
                /* Copy the CBC IV just past the key. */
-               memcpy(key + AES_KEYSIZE_128, req->info, AES_KEYSIZE_128);
+               memcpy(key + AES_KEYSIZE_128, req->iv, AES_KEYSIZE_128);
                /* CBC needs the INIT set. */
                init = 1;
        } else {
                src_buf = sg_virt(src);
                len = sg_dma_len(src);
                tlen += len;
-               limit_hit = tlen > req->nbytes;
+               limit_hit = tlen > req->cryptlen;
 
                if (limit_hit)
-                       len = req->nbytes - (tlen - len);
+                       len = req->cryptlen - (tlen - len);
 
                do {
                        if (actx->fill + len > out_off)
        /* Copy the IV for CBC for chaining */
        if (!rctx->ecb) {
                if (rctx->enc)
-                       memcpy(req->info, out_buf+(last_out_len-AES_BLOCK_SIZE),
+                       memcpy(req->iv, out_buf+(last_out_len-AES_BLOCK_SIZE),
                                AES_BLOCK_SIZE);
                else
-                       memcpy(req->info, in_buf+(last_out_len-AES_BLOCK_SIZE),
+                       memcpy(req->iv, in_buf+(last_out_len-AES_BLOCK_SIZE),
                                AES_BLOCK_SIZE);
        }
 
        return 0;
 }
 
-static int mxs_dcp_block_fallback(struct ablkcipher_request *req, int enc)
+static int mxs_dcp_block_fallback(struct skcipher_request *req, int enc)
 {
-       struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
-       struct dcp_async_ctx *ctx = crypto_ablkcipher_ctx(tfm);
+       struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
+       struct dcp_async_ctx *ctx = crypto_skcipher_ctx(tfm);
        SYNC_SKCIPHER_REQUEST_ON_STACK(subreq, ctx->fallback);
        int ret;
 
        skcipher_request_set_sync_tfm(subreq, ctx->fallback);
        skcipher_request_set_callback(subreq, req->base.flags, NULL, NULL);
        skcipher_request_set_crypt(subreq, req->src, req->dst,
-                                  req->nbytes, req->info);
+                                  req->cryptlen, req->iv);
 
        if (enc)
                ret = crypto_skcipher_encrypt(subreq);
        return ret;
 }
 
-static int mxs_dcp_aes_enqueue(struct ablkcipher_request *req, int enc, int ecb)
+static int mxs_dcp_aes_enqueue(struct skcipher_request *req, int enc, int ecb)
 {
        struct dcp *sdcp = global_sdcp;
        struct crypto_async_request *arq = &req->base;
        struct dcp_async_ctx *actx = crypto_tfm_ctx(arq->tfm);
-       struct dcp_aes_req_ctx *rctx = ablkcipher_request_ctx(req);
+       struct dcp_aes_req_ctx *rctx = skcipher_request_ctx(req);
        int ret;
 
        if (unlikely(actx->key_len != AES_KEYSIZE_128))
        return ret;
 }
 
-static int mxs_dcp_aes_ecb_decrypt(struct ablkcipher_request *req)
+static int mxs_dcp_aes_ecb_decrypt(struct skcipher_request *req)
 {
        return mxs_dcp_aes_enqueue(req, 0, 1);
 }
 
-static int mxs_dcp_aes_ecb_encrypt(struct ablkcipher_request *req)
+static int mxs_dcp_aes_ecb_encrypt(struct skcipher_request *req)
 {
        return mxs_dcp_aes_enqueue(req, 1, 1);
 }
 
-static int mxs_dcp_aes_cbc_decrypt(struct ablkcipher_request *req)
+static int mxs_dcp_aes_cbc_decrypt(struct skcipher_request *req)
 {
        return mxs_dcp_aes_enqueue(req, 0, 0);
 }
 
-static int mxs_dcp_aes_cbc_encrypt(struct ablkcipher_request *req)
+static int mxs_dcp_aes_cbc_encrypt(struct skcipher_request *req)
 {
        return mxs_dcp_aes_enqueue(req, 1, 0);
 }
 
-static int mxs_dcp_aes_setkey(struct crypto_ablkcipher *tfm, const u8 *key,
+static int mxs_dcp_aes_setkey(struct crypto_skcipher *tfm, const u8 *key,
                              unsigned int len)
 {
-       struct dcp_async_ctx *actx = crypto_ablkcipher_ctx(tfm);
+       struct dcp_async_ctx *actx = crypto_skcipher_ctx(tfm);
        unsigned int ret;
 
        /*
        return ret;
 }
 
-static int mxs_dcp_aes_fallback_init(struct crypto_tfm *tfm)
+static int mxs_dcp_aes_fallback_init_tfm(struct crypto_skcipher *tfm)
 {
-       const char *name = crypto_tfm_alg_name(tfm);
-       struct dcp_async_ctx *actx = crypto_tfm_ctx(tfm);
+       const char *name = crypto_tfm_alg_name(crypto_skcipher_tfm(tfm));
+       struct dcp_async_ctx *actx = crypto_skcipher_ctx(tfm);
        struct crypto_sync_skcipher *blk;
 
        blk = crypto_alloc_sync_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
                return PTR_ERR(blk);
 
        actx->fallback = blk;
-       tfm->crt_ablkcipher.reqsize = sizeof(struct dcp_aes_req_ctx);
+       crypto_skcipher_set_reqsize(tfm, sizeof(struct dcp_aes_req_ctx));
        return 0;
 }
 
-static void mxs_dcp_aes_fallback_exit(struct crypto_tfm *tfm)
+static void mxs_dcp_aes_fallback_exit_tfm(struct crypto_skcipher *tfm)
 {
-       struct dcp_async_ctx *actx = crypto_tfm_ctx(tfm);
+       struct dcp_async_ctx *actx = crypto_skcipher_ctx(tfm);
 
        crypto_free_sync_skcipher(actx->fallback);
 }
 }
 
 /* AES 128 ECB and AES 128 CBC */
-static struct crypto_alg dcp_aes_algs[] = {
+static struct skcipher_alg dcp_aes_algs[] = {
        {
-               .cra_name               = "ecb(aes)",
-               .cra_driver_name        = "ecb-aes-dcp",
-               .cra_priority           = 400,
-               .cra_alignmask          = 15,
-               .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
-                                         CRYPTO_ALG_ASYNC |
+               .base.cra_name          = "ecb(aes)",
+               .base.cra_driver_name   = "ecb-aes-dcp",
+               .base.cra_priority      = 400,
+               .base.cra_alignmask     = 15,
+               .base.cra_flags         = CRYPTO_ALG_ASYNC |
                                          CRYPTO_ALG_NEED_FALLBACK,
-               .cra_init               = mxs_dcp_aes_fallback_init,
-               .cra_exit               = mxs_dcp_aes_fallback_exit,
-               .cra_blocksize          = AES_BLOCK_SIZE,
-               .cra_ctxsize            = sizeof(struct dcp_async_ctx),
-               .cra_type               = &crypto_ablkcipher_type,
-               .cra_module             = THIS_MODULE,
-               .cra_u  = {
-                       .ablkcipher = {
-                               .min_keysize    = AES_MIN_KEY_SIZE,
-                               .max_keysize    = AES_MAX_KEY_SIZE,
-                               .setkey         = mxs_dcp_aes_setkey,
-                               .encrypt        = mxs_dcp_aes_ecb_encrypt,
-                               .decrypt        = mxs_dcp_aes_ecb_decrypt
-                       },
-               },
+               .base.cra_blocksize     = AES_BLOCK_SIZE,
+               .base.cra_ctxsize       = sizeof(struct dcp_async_ctx),
+               .base.cra_module        = THIS_MODULE,
+
+               .min_keysize            = AES_MIN_KEY_SIZE,
+               .max_keysize            = AES_MAX_KEY_SIZE,
+               .setkey                 = mxs_dcp_aes_setkey,
+               .encrypt                = mxs_dcp_aes_ecb_encrypt,
+               .decrypt                = mxs_dcp_aes_ecb_decrypt,
+               .init                   = mxs_dcp_aes_fallback_init_tfm,
+               .exit                   = mxs_dcp_aes_fallback_exit_tfm,
        }, {
-               .cra_name               = "cbc(aes)",
-               .cra_driver_name        = "cbc-aes-dcp",
-               .cra_priority           = 400,
-               .cra_alignmask          = 15,
-               .cra_flags              = CRYPTO_ALG_TYPE_ABLKCIPHER |
-                                         CRYPTO_ALG_ASYNC |
+               .base.cra_name          = "cbc(aes)",
+               .base.cra_driver_name   = "cbc-aes-dcp",
+               .base.cra_priority      = 400,
+               .base.cra_alignmask     = 15,
+               .base.cra_flags         = CRYPTO_ALG_ASYNC |
                                          CRYPTO_ALG_NEED_FALLBACK,
-               .cra_init               = mxs_dcp_aes_fallback_init,
-               .cra_exit               = mxs_dcp_aes_fallback_exit,
-               .cra_blocksize          = AES_BLOCK_SIZE,
-               .cra_ctxsize            = sizeof(struct dcp_async_ctx),
-               .cra_type               = &crypto_ablkcipher_type,
-               .cra_module             = THIS_MODULE,
-               .cra_u = {
-                       .ablkcipher = {
-                               .min_keysize    = AES_MIN_KEY_SIZE,
-                               .max_keysize    = AES_MAX_KEY_SIZE,
-                               .setkey         = mxs_dcp_aes_setkey,
-                               .encrypt        = mxs_dcp_aes_cbc_encrypt,
-                               .decrypt        = mxs_dcp_aes_cbc_decrypt,
-                               .ivsize         = AES_BLOCK_SIZE,
-                       },
-               },
+               .base.cra_blocksize     = AES_BLOCK_SIZE,
+               .base.cra_ctxsize       = sizeof(struct dcp_async_ctx),
+               .base.cra_module        = THIS_MODULE,
+
+               .min_keysize            = AES_MIN_KEY_SIZE,
+               .max_keysize            = AES_MAX_KEY_SIZE,
+               .setkey                 = mxs_dcp_aes_setkey,
+               .encrypt                = mxs_dcp_aes_cbc_encrypt,
+               .decrypt                = mxs_dcp_aes_cbc_decrypt,
+               .ivsize                 = AES_BLOCK_SIZE,
+               .init                   = mxs_dcp_aes_fallback_init_tfm,
+               .exit                   = mxs_dcp_aes_fallback_exit_tfm,
        },
 };
 
        sdcp->caps = readl(sdcp->base + MXS_DCP_CAPABILITY1);
 
        if (sdcp->caps & MXS_DCP_CAPABILITY1_AES128) {
-               ret = crypto_register_algs(dcp_aes_algs,
-                                          ARRAY_SIZE(dcp_aes_algs));
+               ret = crypto_register_skciphers(dcp_aes_algs,
+                                               ARRAY_SIZE(dcp_aes_algs));
                if (ret) {
                        /* Failed to register algorithm. */
                        dev_err(dev, "Failed to register AES crypto!\n");
 
 err_unregister_aes:
        if (sdcp->caps & MXS_DCP_CAPABILITY1_AES128)
-               crypto_unregister_algs(dcp_aes_algs, ARRAY_SIZE(dcp_aes_algs));
+               crypto_unregister_skciphers(dcp_aes_algs, ARRAY_SIZE(dcp_aes_algs));
 
 err_destroy_aes_thread:
        kthread_stop(sdcp->thread[DCP_CHAN_CRYPTO]);
                crypto_unregister_ahash(&dcp_sha1_alg);
 
        if (sdcp->caps & MXS_DCP_CAPABILITY1_AES128)
-               crypto_unregister_algs(dcp_aes_algs, ARRAY_SIZE(dcp_aes_algs));
+               crypto_unregister_skciphers(dcp_aes_algs, ARRAY_SIZE(dcp_aes_algs));
 
        kthread_stop(sdcp->thread[DCP_CHAN_HASH_SHA]);
        kthread_stop(sdcp->thread[DCP_CHAN_CRYPTO]);