.sector = to_sector(bh, mapping->host),
                .size = bh->b_size,
        };
-       int error;
        void *ret;
        void *entry = *entryp;
 
-       i_mmap_lock_read(mapping);
-
-       if (dax_map_atomic(bdev, &dax) < 0) {
-               error = PTR_ERR(dax.addr);
-               goto out;
-       }
+       if (dax_map_atomic(bdev, &dax) < 0)
+               return PTR_ERR(dax.addr);
        dax_unmap_atomic(bdev, &dax);
 
        ret = dax_insert_mapping_entry(mapping, vmf, entry, dax.sector);
-       if (IS_ERR(ret)) {
-               error = PTR_ERR(ret);
-               goto out;
-       }
+       if (IS_ERR(ret))
+               return PTR_ERR(ret);
        *entryp = ret;
 
-       error = vm_insert_mixed(vma, vaddr, dax.pfn);
- out:
-       i_mmap_unlock_read(mapping);
-       return error;
+       return vm_insert_mixed(vma, vaddr, dax.pfn);
 }
 
 /**
                truncate_pagecache_range(inode, lstart, lend);
        }
 
-       i_mmap_lock_read(mapping);
-
        if (!write && !buffer_mapped(&bh)) {
                spinlock_t *ptl;
                pmd_t entry;
        }
 
  out:
-       i_mmap_unlock_read(mapping);
-
        return result;
 
  fallback:
 
        if (details.last_index < details.first_index)
                details.last_index = ULONG_MAX;
 
-
-       /* DAX uses i_mmap_lock to serialise file truncate vs page fault */
        i_mmap_lock_write(mapping);
        if (unlikely(!RB_EMPTY_ROOT(&mapping->i_mmap)))
                unmap_mapping_range_tree(&mapping->i_mmap, &details);