* submit the buffer_head for reading
         */
        set_buffer_new(bh);
-       clear_buffer_verified(bh);
        trace_ext4_read_block_bitmap_load(sb, block_group, ignore_locked);
-       bh->b_end_io = ext4_end_bitmap_read;
-       get_bh(bh);
-       submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO |
-                 (ignore_locked ? REQ_RAHEAD : 0), bh);
+       ext4_read_bh_nowait(bh, REQ_META | REQ_PRIO |
+                           (ignore_locked ? REQ_RAHEAD : 0),
+                           ext4_end_bitmap_read);
        return bh;
 verify:
        err = ext4_validate_block_bitmap(sb, desc, block_group, bh);
 
 
        if (!bh_uptodate_or_lock(bh)) {
                trace_ext4_ext_load_extent(inode, pblk, _RET_IP_);
-               clear_buffer_verified(bh);
-               err = bh_submit_read(bh);
+               err = ext4_read_bh(bh, 0, NULL);
                if (err < 0)
                        goto errout;
        }
 
        /*
         * submit the buffer_head for reading
         */
-       clear_buffer_verified(bh);
        trace_ext4_load_inode_bitmap(sb, block_group);
-       bh->b_end_io = ext4_end_bitmap_read;
-       get_bh(bh);
-       submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, bh);
-       wait_on_buffer(bh);
+       ext4_read_bh(bh, REQ_META | REQ_PRIO, ext4_end_bitmap_read);
        ext4_simulate_fail_bh(sb, bh, EXT4_SIM_IBITMAP_EIO);
        if (!buffer_uptodate(bh)) {
                put_bh(bh);
 
                }
 
                if (!bh_uptodate_or_lock(bh)) {
-                       if (bh_submit_read(bh) < 0) {
+                       if (ext4_read_bh(bh, 0, NULL) < 0) {
                                put_bh(bh);
                                goto failure;
                        }
 
                               ext4_lblk_t block, int map_flags)
 {
        struct buffer_head *bh;
+       int ret;
 
        bh = ext4_getblk(handle, inode, block, map_flags);
        if (IS_ERR(bh))
                return bh;
        if (!bh || ext4_buffer_uptodate(bh))
                return bh;
-       clear_buffer_verified(bh);
-       ll_rw_block(REQ_OP_READ, REQ_META | REQ_PRIO, 1, &bh);
-       wait_on_buffer(bh);
-       if (buffer_uptodate(bh))
-               return bh;
-       put_bh(bh);
-       return ERR_PTR(-EIO);
+
+       ret = ext4_read_bh_lock(bh, REQ_META | REQ_PRIO, true);
+       if (ret) {
+               put_bh(bh);
+               return ERR_PTR(ret);
+       }
+       return bh;
 }
 
 /* Read a contiguous batch of blocks. */
 
        for (i = 0; i < bh_count; i++)
                /* Note that NULL bhs[i] is valid because of holes. */
-               if (bhs[i] && !ext4_buffer_uptodate(bhs[i])) {
-                       clear_buffer_verified(bhs[i]);
-                       ll_rw_block(REQ_OP_READ, REQ_META | REQ_PRIO, 1,
-                                   &bhs[i]);
-               }
+               if (bhs[i] && !ext4_buffer_uptodate(bhs[i]))
+                       ext4_read_bh_lock(bhs[i], REQ_META | REQ_PRIO, false);
 
        if (!wait)
                return 0;
                if (!buffer_uptodate(bh) && !buffer_delay(bh) &&
                    !buffer_unwritten(bh) &&
                    (block_start < from || block_end > to)) {
-                       ll_rw_block(REQ_OP_READ, 0, 1, &bh);
+                       ext4_read_bh_lock(bh, 0, false);
                        wait[nr_wait++] = bh;
                }
        }
                set_buffer_uptodate(bh);
 
        if (!buffer_uptodate(bh)) {
-               err = -EIO;
-               ll_rw_block(REQ_OP_READ, 0, 1, &bh);
-               wait_on_buffer(bh);
-               /* Uhhuh. Read error. Complain and punt. */
-               if (!buffer_uptodate(bh))
+               err = ext4_read_bh_lock(bh, 0, true);
+               if (err)
                        goto unlock;
                if (fscrypt_inode_uses_fs_layer_crypto(inode)) {
                        /* We expect the key to be set. */
                 * Read the block from disk.
                 */
                trace_ext4_load_inode(inode);
-               get_bh(bh);
-               bh->b_end_io = end_buffer_read_sync;
-               submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, bh);
+               ext4_read_bh_nowait(bh, REQ_META | REQ_PRIO, NULL);
                blk_finish_plug(&plug);
                wait_on_buffer(bh);
                if (!buffer_uptodate(bh)) {
 
                }
        }
 
-       get_bh(*bh);
        lock_buffer(*bh);
-       (*bh)->b_end_io = end_buffer_read_sync;
-       submit_bh(REQ_OP_READ, REQ_META | REQ_PRIO, *bh);
-       wait_on_buffer(*bh);
-       if (!buffer_uptodate(*bh)) {
-               ret = -EIO;
+       ret = ext4_read_bh(*bh, REQ_META | REQ_PRIO, NULL);
+       if (ret)
                goto warn_exit;
-       }
+
        mmp = (struct mmp_struct *)((*bh)->b_data);
        if (le32_to_cpu(mmp->mmp_magic) != EXT4_MMP_MAGIC) {
                ret = -EFSCORRUPTED;
 
        for (i = 0; i < nr; i++) {
                bh = arr[i];
                if (!bh_uptodate_or_lock(bh)) {
-                       err = bh_submit_read(bh);
+                       err = ext4_read_bh(bh, 0, NULL);
                        if (err)
                                return err;
                }
 
        if (unlikely(!bh))
                return NULL;
        if (!bh_uptodate_or_lock(bh)) {
-               if (bh_submit_read(bh) < 0) {
+               if (ext4_read_bh(bh, 0, NULL) < 0) {
                        brelse(bh);
                        return NULL;
                }
 
 struct buffer_head *
 ext4_sb_bread(struct super_block *sb, sector_t block, int op_flags)
 {
-       struct buffer_head *bh = sb_getblk(sb, block);
+       struct buffer_head *bh;
+       int ret;
 
+       bh = sb_getblk(sb, block);
        if (bh == NULL)
                return ERR_PTR(-ENOMEM);
        if (ext4_buffer_uptodate(bh))
                return bh;
-       clear_buffer_verified(bh);
-       ll_rw_block(REQ_OP_READ, REQ_META | op_flags, 1, &bh);
-       wait_on_buffer(bh);
-       if (buffer_uptodate(bh))
-               return bh;
-       put_bh(bh);
-       return ERR_PTR(-EIO);
+
+       ret = ext4_read_bh_lock(bh, REQ_META | op_flags, true);
+       if (ret) {
+               put_bh(bh);
+               return ERR_PTR(ret);
+       }
+       return bh;
 }
 
 static int ext4_verify_csum_type(struct super_block *sb,
                goto out_bdev;
        }
        journal->j_private = sb;
-       ll_rw_block(REQ_OP_READ, REQ_META | REQ_PRIO, 1, &journal->j_sb_buffer);
-       wait_on_buffer(journal->j_sb_buffer);
-       if (!buffer_uptodate(journal->j_sb_buffer)) {
+       if (ext4_read_bh_lock(journal->j_sb_buffer, REQ_META | REQ_PRIO, true)) {
                ext4_msg(sb, KERN_ERR, "I/O error on journal device");
                goto out_journal;
        }