struct vmap_area *first;
 
        BUG_ON(!size);
-       BUG_ON(size & ~PAGE_MASK);
+       BUG_ON(offset_in_page(size));
        BUG_ON(!is_power_of_2(align));
 
        va = kmalloc_node(sizeof(struct vmap_area),
        void *vaddr = NULL;
        unsigned int order;
 
-       BUG_ON(size & ~PAGE_MASK);
+       BUG_ON(offset_in_page(size));
        BUG_ON(size > PAGE_SIZE*VMAP_MAX_ALLOC);
        if (WARN_ON(size == 0)) {
                /*
        unsigned int order;
        struct vmap_block *vb;
 
-       BUG_ON(size & ~PAGE_MASK);
+       BUG_ON(offset_in_page(size));
        BUG_ON(size > PAGE_SIZE*VMAP_MAX_ALLOC);
 
        flush_cache_vunmap((unsigned long)addr, (unsigned long)addr + size);
        while (count) {
                unsigned long offset, length;
 
-               offset = (unsigned long)addr & ~PAGE_MASK;
+               offset = offset_in_page(addr);
                length = PAGE_SIZE - offset;
                if (length > count)
                        length = count;
        while (count) {
                unsigned long offset, length;
 
-               offset = (unsigned long)addr & ~PAGE_MASK;
+               offset = offset_in_page(addr);
                length = PAGE_SIZE - offset;
                if (length > count)
                        length = count;
        bool purged = false;
 
        /* verify parameters and allocate data structures */
-       BUG_ON(align & ~PAGE_MASK || !is_power_of_2(align));
+       BUG_ON(offset_in_page(align) || !is_power_of_2(align));
        for (last_area = 0, area = 0; area < nr_vms; area++) {
                start = offsets[area];
                end = start + sizes[area];