qsize_t i_reserved_quota;
 #endif
 
-       /* completed IOs that might need unwritten extents handling */
-       struct list_head i_completed_io_list;
+       /* Lock protecting lists below */
        spinlock_t i_completed_io_lock;
+       /*
+        * Completed IOs that need unwritten extents handling and have
+        * transaction reserved
+        */
+       struct list_head i_rsv_conversion_list;
+       /*
+        * Completed IOs that need unwritten extents handling and don't have
+        * transaction reserved
+        */
+       struct list_head i_unrsv_conversion_list;
        atomic_t i_ioend_count; /* Number of outstanding io_end structs */
        atomic_t i_unwritten; /* Nr. of inflight conversions pending */
-       struct work_struct i_unwritten_work;    /* deferred extent conversion */
+       struct work_struct i_rsv_conversion_work;
+       struct work_struct i_unrsv_conversion_work;
 
        spinlock_t i_block_reservation_lock;
 
        struct flex_groups *s_flex_groups;
        ext4_group_t s_flex_groups_allocated;
 
-       /* workqueue for dio unwritten */
-       struct workqueue_struct *dio_unwritten_wq;
+       /* workqueue for unreserved extent convertions (dio) */
+       struct workqueue_struct *unrsv_conversion_wq;
+       /* workqueue for reserved extent conversions (buffered io) */
+       struct workqueue_struct *rsv_conversion_wq;
 
        /* timer for periodic error stats printing */
        struct timer_list s_err_report;
 extern void ext4_put_io_end_defer(ext4_io_end_t *io_end);
 extern void ext4_io_submit_init(struct ext4_io_submit *io,
                                struct writeback_control *wbc);
-extern void ext4_end_io_work(struct work_struct *work);
+extern void ext4_end_io_rsv_work(struct work_struct *work);
+extern void ext4_end_io_unrsv_work(struct work_struct *work);
 extern void ext4_io_submit(struct ext4_io_submit *io);
 extern int ext4_bio_write_page(struct ext4_io_submit *io,
                               struct page *page,
 
         * We need to make sure the work structure is finished being
         * used before we let the inode get destroyed.
         */
-       if (work_pending(&EXT4_I(inode)->i_unwritten_work))
-               cancel_work_sync(&EXT4_I(inode)->i_unwritten_work);
+       if (work_pending(&EXT4_I(inode)->i_rsv_conversion_work))
+               cancel_work_sync(&EXT4_I(inode)->i_rsv_conversion_work);
+       if (work_pending(&EXT4_I(inode)->i_unrsv_conversion_work))
+               cancel_work_sync(&EXT4_I(inode)->i_unrsv_conversion_work);
 }
 
 static void ext4_release_io_end(ext4_io_end_t *io_end)
        return ret;
 }
 
-static void dump_completed_IO(struct inode *inode)
+static void dump_completed_IO(struct inode *inode, struct list_head *head)
 {
 #ifdef EXT4FS_DEBUG
        struct list_head *cur, *before, *after;
        ext4_io_end_t *io, *io0, *io1;
 
-       if (list_empty(&EXT4_I(inode)->i_completed_io_list)) {
-               ext4_debug("inode %lu completed_io list is empty\n",
-                          inode->i_ino);
+       if (list_empty(head))
                return;
-       }
 
-       ext4_debug("Dump inode %lu completed_io list\n", inode->i_ino);
-       list_for_each_entry(io, &EXT4_I(inode)->i_completed_io_list, list) {
+       ext4_debug("Dump inode %lu completed io list\n", inode->i_ino);
+       list_for_each_entry(io, head, list) {
                cur = &io->list;
                before = cur->prev;
                io0 = container_of(before, ext4_io_end_t, list);
        unsigned long flags;
 
        BUG_ON(!(io_end->flag & EXT4_IO_END_UNWRITTEN));
-       wq = EXT4_SB(io_end->inode->i_sb)->dio_unwritten_wq;
-
        spin_lock_irqsave(&ei->i_completed_io_lock, flags);
-       if (list_empty(&ei->i_completed_io_list))
-               queue_work(wq, &ei->i_unwritten_work);
-       list_add_tail(&io_end->list, &ei->i_completed_io_list);
+       if (io_end->handle) {
+               wq = EXT4_SB(io_end->inode->i_sb)->rsv_conversion_wq;
+               if (list_empty(&ei->i_rsv_conversion_list))
+                       queue_work(wq, &ei->i_rsv_conversion_work);
+               list_add_tail(&io_end->list, &ei->i_rsv_conversion_list);
+       } else {
+               wq = EXT4_SB(io_end->inode->i_sb)->unrsv_conversion_wq;
+               if (list_empty(&ei->i_unrsv_conversion_list))
+                       queue_work(wq, &ei->i_unrsv_conversion_work);
+               list_add_tail(&io_end->list, &ei->i_unrsv_conversion_list);
+       }
        spin_unlock_irqrestore(&ei->i_completed_io_lock, flags);
 }
 
-static int ext4_do_flush_completed_IO(struct inode *inode)
+static int ext4_do_flush_completed_IO(struct inode *inode,
+                                     struct list_head *head)
 {
        ext4_io_end_t *io;
        struct list_head unwritten;
        int err, ret = 0;
 
        spin_lock_irqsave(&ei->i_completed_io_lock, flags);
-       dump_completed_IO(inode);
-       list_replace_init(&ei->i_completed_io_list, &unwritten);
+       dump_completed_IO(inode, head);
+       list_replace_init(head, &unwritten);
        spin_unlock_irqrestore(&ei->i_completed_io_lock, flags);
 
        while (!list_empty(&unwritten)) {
 }
 
 /*
- * work on completed aio dio IO, to convert unwritten extents to extents
+ * work on completed IO, to convert unwritten extents to extents
  */
-void ext4_end_io_work(struct work_struct *work)
+void ext4_end_io_rsv_work(struct work_struct *work)
 {
        struct ext4_inode_info *ei = container_of(work, struct ext4_inode_info,
-                                                 i_unwritten_work);
-       ext4_do_flush_completed_IO(&ei->vfs_inode);
+                                                 i_rsv_conversion_work);
+       ext4_do_flush_completed_IO(&ei->vfs_inode, &ei->i_rsv_conversion_list);
+}
+
+void ext4_end_io_unrsv_work(struct work_struct *work)
+{
+       struct ext4_inode_info *ei = container_of(work, struct ext4_inode_info,
+                                                 i_unrsv_conversion_work);
+       ext4_do_flush_completed_IO(&ei->vfs_inode, &ei->i_unrsv_conversion_list);
 }
 
 int ext4_flush_unwritten_io(struct inode *inode)
 {
-       int ret;
+       int ret, err;
+
        WARN_ON_ONCE(!mutex_is_locked(&inode->i_mutex) &&
                     !(inode->i_state & I_FREEING));
-       ret = ext4_do_flush_completed_IO(inode);
+       ret = ext4_do_flush_completed_IO(inode,
+                                        &EXT4_I(inode)->i_rsv_conversion_list);
+       err = ext4_do_flush_completed_IO(inode,
+                                        &EXT4_I(inode)->i_unrsv_conversion_list);
+       if (!ret)
+               ret = err;
        ext4_unwritten_wait(inode);
        return ret;
 }
 
        ext4_unregister_li_request(sb);
        dquot_disable(sb, -1, DQUOT_USAGE_ENABLED | DQUOT_LIMITS_ENABLED);
 
-       flush_workqueue(sbi->dio_unwritten_wq);
-       destroy_workqueue(sbi->dio_unwritten_wq);
+       flush_workqueue(sbi->unrsv_conversion_wq);
+       flush_workqueue(sbi->rsv_conversion_wq);
+       destroy_workqueue(sbi->unrsv_conversion_wq);
+       destroy_workqueue(sbi->rsv_conversion_wq);
 
        if (sbi->s_journal) {
                err = jbd2_journal_destroy(sbi->s_journal);
        ei->i_reserved_quota = 0;
 #endif
        ei->jinode = NULL;
-       INIT_LIST_HEAD(&ei->i_completed_io_list);
+       INIT_LIST_HEAD(&ei->i_rsv_conversion_list);
+       INIT_LIST_HEAD(&ei->i_unrsv_conversion_list);
        spin_lock_init(&ei->i_completed_io_lock);
        ei->i_sync_tid = 0;
        ei->i_datasync_tid = 0;
        atomic_set(&ei->i_ioend_count, 0);
        atomic_set(&ei->i_unwritten, 0);
-       INIT_WORK(&ei->i_unwritten_work, ext4_end_io_work);
+       INIT_WORK(&ei->i_rsv_conversion_work, ext4_end_io_rsv_work);
+       INIT_WORK(&ei->i_unrsv_conversion_work, ext4_end_io_unrsv_work);
 
        return &ei->vfs_inode;
 }
         * The maximum number of concurrent works can be high and
         * concurrency isn't really necessary.  Limit it to 1.
         */
-       EXT4_SB(sb)->dio_unwritten_wq =
-               alloc_workqueue("ext4-dio-unwritten", WQ_MEM_RECLAIM | WQ_UNBOUND, 1);
-       if (!EXT4_SB(sb)->dio_unwritten_wq) {
-               printk(KERN_ERR "EXT4-fs: failed to create DIO workqueue\n");
+       EXT4_SB(sb)->rsv_conversion_wq =
+               alloc_workqueue("ext4-rsv-conversion", WQ_MEM_RECLAIM | WQ_UNBOUND, 1);
+       if (!EXT4_SB(sb)->rsv_conversion_wq) {
+               printk(KERN_ERR "EXT4-fs: failed to create workqueue\n");
                ret = -ENOMEM;
-               goto failed_mount_wq;
+               goto failed_mount4;
+       }
+
+       EXT4_SB(sb)->unrsv_conversion_wq =
+               alloc_workqueue("ext4-unrsv-conversion", WQ_MEM_RECLAIM | WQ_UNBOUND, 1);
+       if (!EXT4_SB(sb)->unrsv_conversion_wq) {
+               printk(KERN_ERR "EXT4-fs: failed to create workqueue\n");
+               ret = -ENOMEM;
+               goto failed_mount4;
        }
 
        /*
        sb->s_root = NULL;
 failed_mount4:
        ext4_msg(sb, KERN_ERR, "mount failed");
-       destroy_workqueue(EXT4_SB(sb)->dio_unwritten_wq);
+       if (EXT4_SB(sb)->rsv_conversion_wq)
+               destroy_workqueue(EXT4_SB(sb)->rsv_conversion_wq);
+       if (EXT4_SB(sb)->unrsv_conversion_wq)
+               destroy_workqueue(EXT4_SB(sb)->unrsv_conversion_wq);
 failed_mount_wq:
        if (sbi->s_journal) {
                jbd2_journal_destroy(sbi->s_journal);
        struct ext4_sb_info *sbi = EXT4_SB(sb);
 
        trace_ext4_sync_fs(sb, wait);
-       flush_workqueue(sbi->dio_unwritten_wq);
+       flush_workqueue(sbi->rsv_conversion_wq);
+       flush_workqueue(sbi->unrsv_conversion_wq);
        /*
         * Writeback quota in non-journalled quota case - journalled quota has
         * no dirty dquots