struct list_head i_completed_io_list;
        spinlock_t i_completed_io_lock;
        atomic_t i_ioend_count; /* Number of outstanding io_end structs */
-       atomic_t i_aiodio_unwritten; /* Nr. of inflight conversions pending */
+       atomic_t i_unwritten; /* Nr. of inflight conversions pending */
 
        spinlock_t i_block_reservation_lock;
 
 {
        if (!(io_end->flag & EXT4_IO_END_UNWRITTEN)) {
                io_end->flag |= EXT4_IO_END_UNWRITTEN;
-               atomic_inc(&EXT4_I(inode)->i_aiodio_unwritten);
+               atomic_inc(&EXT4_I(inode)->i_unwritten);
        }
 }
 
 
        return 0;
 }
 
-static void ext4_aiodio_wait(struct inode *inode)
+static void ext4_unwritten_wait(struct inode *inode)
 {
        wait_queue_head_t *wq = ext4_ioend_wq(inode);
 
-       wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_aiodio_unwritten) == 0));
+       wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_unwritten) == 0));
 }
 
 /*
                                 "performance will be poor.",
                                 inode->i_ino, current->comm);
                mutex_lock(ext4_aio_mutex(inode));
-               ext4_aiodio_wait(inode);
+               ext4_unwritten_wait(inode);
        }
 
        BUG_ON(iocb->ki_pos != pos);
 
        if (io->flag & EXT4_IO_END_DIRECT)
                inode_dio_done(inode);
        /* Wake up anyone waiting on unwritten extent conversion */
-       if (atomic_dec_and_test(&EXT4_I(inode)->i_aiodio_unwritten))
+       if (atomic_dec_and_test(&EXT4_I(inode)->i_unwritten))
                wake_up_all(ext4_ioend_wq(io->inode));
        return ret;
 }
 
        ei->i_sync_tid = 0;
        ei->i_datasync_tid = 0;
        atomic_set(&ei->i_ioend_count, 0);
-       atomic_set(&ei->i_aiodio_unwritten, 0);
+       atomic_set(&ei->i_unwritten, 0);
 
        return &ei->vfs_inode;
 }