if (buffer_new(bh)) {
                                /* blockdev mappings never come here */
                                clear_buffer_new(bh);
-                               unmap_underlying_metadata(bh->b_bdev,
-                                                       bh->b_blocknr);
+                               clean_bdev_bh_alias(bh);
                        }
                }
                bh = bh->b_this_page;
                        }
 
                        if (buffer_new(bh)) {
-                               unmap_underlying_metadata(bh->b_bdev,
-                                                       bh->b_blocknr);
+                               clean_bdev_bh_alias(bh);
                                if (PageUptodate(page)) {
                                        clear_buffer_new(bh);
                                        set_buffer_uptodate(bh);
                if (!buffer_mapped(bh))
                        is_mapped_to_disk = 0;
                if (buffer_new(bh))
-                       unmap_underlying_metadata(bh->b_bdev, bh->b_blocknr);
+                       clean_bdev_bh_alias(bh);
                if (PageUptodate(page)) {
                        set_buffer_uptodate(bh);
                        continue;
 
                        if (err)
                                break;
                        if (buffer_new(bh)) {
-                               unmap_underlying_metadata(bh->b_bdev,
-                                                         bh->b_blocknr);
+                               clean_bdev_bh_alias(bh);
                                if (PageUptodate(page)) {
                                        clear_buffer_new(bh);
                                        set_buffer_uptodate(bh);
 
                }
                if (buffer_new(bh)) {
                        clear_buffer_new(bh);
-                       unmap_underlying_metadata(bh->b_bdev, bh->b_blocknr);
+                       clean_bdev_bh_alias(bh);
                }
                set_buffer_async_write(bh);
                nr_to_submit++;
 
                if (mpd->get_block(inode, block_in_file, &map_bh, 1))
                        goto confused;
                if (buffer_new(&map_bh))
-                       unmap_underlying_metadata(map_bh.b_bdev,
-                                               map_bh.b_blocknr);
+                       clean_bdev_bh_alias(&map_bh);
                if (buffer_boundary(&map_bh)) {
                        boundary_block = map_bh.b_blocknr;
                        boundary_bdev = map_bh.b_bdev;
 
                        }
                        // TODO: Instantiate the hole.
                        // clear_buffer_new(bh);
-                       // unmap_underlying_metadata(bh->b_bdev, bh->b_blocknr);
+                       // clean_bdev_bh_alias(bh);
                        ntfs_error(vol->sb, "Writing into sparse regions is "
                                        "not supported yet. Sorry.");
                        err = -EOPNOTSUPP;
 
                                        set_buffer_uptodate(bh);
                                if (unlikely(was_hole)) {
                                        /* We allocated the buffer. */
-                                       unmap_underlying_metadata(bh->b_bdev,
-                                                       bh->b_blocknr);
+                                       clean_bdev_bh_alias(bh);
                                        if (bh_end <= pos || bh_pos >= end)
                                                mark_buffer_dirty(bh);
                                        else
                                continue;
                        }
                        /* We allocated the buffer. */
-                       unmap_underlying_metadata(bh->b_bdev, bh->b_blocknr);
+                       clean_bdev_bh_alias(bh);
                        /*
                         * If the buffer is fully outside the write, zero it,
                         * set it uptodate, and mark it dirty so it gets
 
 
                if (!buffer_mapped(bh)) {
                        map_bh(bh, inode->i_sb, *p_blkno);
-                       unmap_underlying_metadata(bh->b_bdev, bh->b_blocknr);
+                       clean_bdev_bh_alias(bh);
                }
 
                if (PageUptodate(page)) {
 
                             (unsigned long long)(pos + newb), pos);
 
                        bh->b_blocknr = newb + pos;
-                       unmap_underlying_metadata(bh->b_bdev,
-                                                 bh->b_blocknr);
+                       clean_bdev_bh_alias(bh);
                        mark_buffer_dirty(bh);
                        ++j;
                        bh = bh->b_this_page;
 
 
        if (buffer_new(bh)) {
               clear_buffer_new(bh);
-              unmap_underlying_metadata(bh->b_bdev,
-                                        bh->b_blocknr);
+              clean_bdev_bh_alias(bh);
               /*
                * we do not zeroize fragment, because of
                * if it maped to hole, it already contains zeroes
 
 void unmap_underlying_metadata(struct block_device *bdev, sector_t block);
 void clean_bdev_aliases(struct block_device *bdev, sector_t block,
                        sector_t len);
+static inline void clean_bdev_bh_alias(struct buffer_head *bh)
+{
+       clean_bdev_aliases(bh->b_bdev, bh->b_blocknr, 1);
+}
 
 void mark_buffer_async_write(struct buffer_head *bh);
 void __wait_on_buffer(struct buffer_head *);