static int vdso_mremap(const struct vm_special_mapping *sm,
                struct vm_area_struct *new_vma)
 {
-       unsigned long new_size = new_vma->vm_end - new_vma->vm_start;
-       unsigned long vdso_size;
-
-       /* without VVAR page */
-       vdso_size = (vdso_total_pages - 1) << PAGE_SHIFT;
-
-       if (vdso_size != new_size)
-               return -EINVAL;
-
        current->mm->context.vdso = new_vma->vm_start;
 
        return 0;
 
 } vdso_data_store __page_aligned_data;
 struct vdso_data *vdso_data = vdso_data_store.data;
 
-static int __vdso_remap(enum vdso_abi abi,
-                       const struct vm_special_mapping *sm,
-                       struct vm_area_struct *new_vma)
+static int vdso_mremap(const struct vm_special_mapping *sm,
+               struct vm_area_struct *new_vma)
 {
-       unsigned long new_size = new_vma->vm_end - new_vma->vm_start;
-       unsigned long vdso_size = vdso_info[abi].vdso_code_end -
-                                 vdso_info[abi].vdso_code_start;
-
-       if (vdso_size != new_size)
-               return -EINVAL;
-
        current->mm->context.vdso = (void *)new_vma->vm_start;
 
        return 0;
        return vmf_insert_pfn(vma, vmf->address, pfn);
 }
 
-static int vvar_mremap(const struct vm_special_mapping *sm,
-                      struct vm_area_struct *new_vma)
-{
-       unsigned long new_size = new_vma->vm_end - new_vma->vm_start;
-
-       if (new_size != VVAR_NR_PAGES * PAGE_SIZE)
-               return -EINVAL;
-
-       return 0;
-}
-
 static int __setup_additional_pages(enum vdso_abi abi,
                                    struct mm_struct *mm,
                                    struct linux_binprm *bprm,
 /*
  * Create and map the vectors page for AArch32 tasks.
  */
-static int aarch32_vdso_mremap(const struct vm_special_mapping *sm,
-               struct vm_area_struct *new_vma)
-{
-       return __vdso_remap(VDSO_ABI_AA32, sm, new_vma);
-}
-
 enum aarch32_map {
        AA32_MAP_VECTORS, /* kuser helpers */
        AA32_MAP_SIGPAGE,
        [AA32_MAP_VVAR] = {
                .name = "[vvar]",
                .fault = vvar_fault,
-               .mremap = vvar_mremap,
        },
        [AA32_MAP_VDSO] = {
                .name = "[vdso]",
-               .mremap = aarch32_vdso_mremap,
+               .mremap = vdso_mremap,
        },
 };
 
 }
 #endif /* CONFIG_COMPAT */
 
-static int vdso_mremap(const struct vm_special_mapping *sm,
-               struct vm_area_struct *new_vma)
-{
-       return __vdso_remap(VDSO_ABI_AA64, sm, new_vma);
-}
-
 enum aarch64_map {
        AA64_MAP_VVAR,
        AA64_MAP_VDSO,
        [AA64_MAP_VVAR] = {
                .name   = "[vvar]",
                .fault = vvar_fault,
-               .mremap = vvar_mremap,
        },
        [AA64_MAP_VDSO] = {
                .name   = "[vdso]",
 
        fprintf(out_file, "     const struct vm_special_mapping *sm,\n");
        fprintf(out_file, "     struct vm_area_struct *new_vma)\n");
        fprintf(out_file, "{\n");
-       fprintf(out_file, "     unsigned long new_size =\n");
-       fprintf(out_file, "     new_vma->vm_end - new_vma->vm_start;\n");
-       fprintf(out_file, "     if (vdso_image.size != new_size)\n");
-       fprintf(out_file, "             return -EINVAL;\n");
        fprintf(out_file, "     current->mm->context.vdso =\n");
        fprintf(out_file, "     (void *)(new_vma->vm_start);\n");
        fprintf(out_file, "     return 0;\n");
 
 static int vdso_mremap(const struct vm_special_mapping *sm,
                       struct vm_area_struct *vma)
 {
-       unsigned long vdso_pages;
-
-       vdso_pages = vdso64_pages;
-
-       if ((vdso_pages << PAGE_SHIFT) != vma->vm_end - vma->vm_start)
-               return -EINVAL;
-
-       if (WARN_ON_ONCE(current->mm != vma->vm_mm))
-               return -EFAULT;
-
        current->mm->context.vdso_base = vma->vm_start;
+
        return 0;
 }
 
 
 static int vdso_mremap(const struct vm_special_mapping *sm,
                struct vm_area_struct *new_vma)
 {
-       unsigned long new_size = new_vma->vm_end - new_vma->vm_start;
        const struct vdso_image *image = current->mm->context.vdso_image;
 
-       if (image->size != new_size)
-               return -EINVAL;
-
        vdso_fix_landing(image, new_vma);
        current->mm->context.vdso = (void __user *)new_vma->vm_start;
 
        return 0;
 }
 
-static int vvar_mremap(const struct vm_special_mapping *sm,
-               struct vm_area_struct *new_vma)
-{
-       const struct vdso_image *image = new_vma->vm_mm->context.vdso_image;
-       unsigned long new_size = new_vma->vm_end - new_vma->vm_start;
-
-       if (new_size != -image->sym_vvar_start)
-               return -EINVAL;
-
-       return 0;
-}
-
 #ifdef CONFIG_TIME_NS
 static struct page *find_timens_vvar_page(struct vm_area_struct *vma)
 {
 static const struct vm_special_mapping vvar_mapping = {
        .name = "[vvar]",
        .fault = vvar_fault,
-       .mremap = vvar_mremap,
 };
 
 /*
 
        return 0;
 }
 
+static int special_mapping_split(struct vm_area_struct *vma, unsigned long addr)
+{
+       /*
+        * Forbid splitting special mappings - kernel has expectations over
+        * the number of pages in mapping. Together with VM_DONTEXPAND
+        * the size of vma should stay the same over the special mapping's
+        * lifetime.
+        */
+       return -EINVAL;
+}
+
 static const struct vm_operations_struct special_mapping_vmops = {
        .close = special_mapping_close,
        .fault = special_mapping_fault,
        .name = special_mapping_name,
        /* vDSO code relies that VVAR can't be accessed remotely */
        .access = NULL,
+       .may_split = special_mapping_split,
 };
 
 static const struct vm_operations_struct legacy_special_mapping_vmops = {