if (error)
                return error;
 
-       kaddr = kmap_atomic(page, KM_USER0);
+       kaddr = kmap_atomic(page);
        if (dsize > (dibh->b_size - sizeof(struct gfs2_dinode)))
                dsize = (dibh->b_size - sizeof(struct gfs2_dinode));
        memcpy(kaddr, dibh->b_data + sizeof(struct gfs2_dinode), dsize);
        memset(kaddr + dsize, 0, PAGE_CACHE_SIZE - dsize);
-       kunmap_atomic(kaddr, KM_USER0);
+       kunmap_atomic(kaddr);
        flush_dcache_page(page);
        brelse(dibh);
        SetPageUptodate(page);
                page = read_cache_page(mapping, index, __gfs2_readpage, NULL);
                if (IS_ERR(page))
                        return PTR_ERR(page);
-               p = kmap_atomic(page, KM_USER0);
+               p = kmap_atomic(page);
                memcpy(buf + copied, p + offset, amt);
-               kunmap_atomic(p, KM_USER0);
+               kunmap_atomic(p);
                mark_page_accessed(page);
                page_cache_release(page);
                copied += amt;
        unsigned char *buf = dibh->b_data + sizeof(struct gfs2_dinode);
 
        BUG_ON((pos + len) > (dibh->b_size - sizeof(struct gfs2_dinode)));
-       kaddr = kmap_atomic(page, KM_USER0);
+       kaddr = kmap_atomic(page);
        memcpy(buf + pos, kaddr + pos, copied);
        memset(kaddr + pos + copied, 0, len - copied);
        flush_dcache_page(page);
-       kunmap_atomic(kaddr, KM_USER0);
+       kunmap_atomic(kaddr);
 
        if (!PageUptodate(page))
                SetPageUptodate(page);
 
        __be32 *ptr;
 
        clear_buffer_escaped(bh);
-       kaddr = kmap_atomic(bh->b_page, KM_USER0);
+       kaddr = kmap_atomic(bh->b_page);
        ptr = kaddr + bh_offset(bh);
        if (*ptr == cpu_to_be32(GFS2_MAGIC))
                set_buffer_escaped(bh);
-       kunmap_atomic(kaddr, KM_USER0);
+       kunmap_atomic(kaddr);
 }
 
 static void gfs2_write_blocks(struct gfs2_sbd *sdp, struct buffer_head *bh,
                if (buffer_escaped(bd->bd_bh)) {
                        void *kaddr;
                        bh1 = gfs2_log_get_buf(sdp);
-                       kaddr = kmap_atomic(bd->bd_bh->b_page, KM_USER0);
+                       kaddr = kmap_atomic(bd->bd_bh->b_page);
                        memcpy(bh1->b_data, kaddr + bh_offset(bd->bd_bh),
                               bh1->b_size);
-                       kunmap_atomic(kaddr, KM_USER0);
+                       kunmap_atomic(kaddr);
                        *(__be32 *)bh1->b_data = 0;
                        clear_buffer_escaped(bd->bd_bh);
                        unlock_buffer(bd->bd_bh);
 
 
        gfs2_trans_add_bh(ip->i_gl, bh, 0);
 
-       kaddr = kmap_atomic(page, KM_USER0);
+       kaddr = kmap_atomic(page);
        if (offset + sizeof(struct gfs2_quota) > PAGE_CACHE_SIZE)
                nbytes = PAGE_CACHE_SIZE - offset;
        memcpy(kaddr + offset, ptr, nbytes);
        flush_dcache_page(page);
-       kunmap_atomic(kaddr, KM_USER0);
+       kunmap_atomic(kaddr);
        unlock_page(page);
        page_cache_release(page);