return next;
}
-static bool do_brk_munmap(struct vm_area_struct *vma, unsigned long newbrk,
- unsigned long oldbrk);
-static int do_brk_flags(unsigned long addr, unsigned long request,
- struct vm_area_struct *vma, unsigned long flags);
+static int do_brk_munmap(struct vm_area_struct *vma, unsigned long newbrk,
+ unsigned long oldbrk, struct list_head *uf);
+static int do_brk_flags(struct vm_area_struct **brkvma, unsigned long addr,
+ unsigned long request, unsigned long flags);
SYSCALL_DEFINE1(brk, unsigned long, brk)
{
unsigned long newbrk, oldbrk, origbrk;
unsigned long min_brk;
bool populate;
bool downgraded = false;
+ LIST_HEAD(uf);
if (mmap_write_lock_killable(mm))
return -EINTR;
goto success;
}
- brkvma = find_vma_intersection(mm, mm->start_brk, mm->brk);
+ brkvma = find_vma_intersection(mm, mm->start_brk, oldbrk);
if (brkvma) {
/*
* Always allow shrinking brk.
* do_brk_munmap() may downgrade mmap_lock to read.
*/
if (brk <= mm->brk) {
+ int ret;
/*
* mm->brk must to be protected by write mmap_lock.
* do_brk_munmap() may downgrade the lock, so update it
* before calling do_brk_munmap().
*/
mm->brk = brk;
- downgraded = do_brk_munmap(brkvma, newbrk, oldbrk);
- goto success;
+ ret = do_brk_munmap(brkvma, newbrk, oldbrk, &uf);
+ if (ret == 1) {
+ downgraded = true;
+ goto success;
+ } else if (!ret)
+ goto success;
+
+ mm->brk = origbrk;
+ goto out;
}
next = brkvma->vm_next;
} else {
goto out;
/* Ok, looks good - let it rip. */
- if (do_brk_flags(oldbrk, newbrk-oldbrk, brkvma, 0) < 0)
+ if (do_brk_flags(&brkvma, oldbrk, newbrk-oldbrk, 0) < 0)
goto out;
mm->brk = brk;
mmap_read_unlock(mm);
else
mmap_write_unlock(mm);
+ userfaultfd_unmap_complete(mm, &uf);
if (populate)
mm_populate_vma(brkvma, oldbrk, newbrk);
return brk;
* unmaps a partial VMA mapping. Does not handle alignment, downgrades lock if
* possible.
*/
-static bool do_brk_munmap(struct vm_area_struct *vma, unsigned long newbrk,
- unsigned long oldbrk)
+static int do_brk_munmap(struct vm_area_struct *vma, unsigned long newbrk,
+ unsigned long oldbrk, struct list_head *uf)
{
struct mm_struct *mm = vma->vm_mm;
struct vm_area_struct unmap;
unsigned long unmap_pages;
- int downgrade = true;
+ int ret = 1;
arch_unmap(mm, newbrk, oldbrk);
- if (vma->vm_start == newbrk) { // remove entire mapping.
+ if (vma->vm_start >= newbrk) { // remove entire mapping.
struct vm_area_struct *prev = vma->vm_prev;
+ ret = userfaultfd_unmap_prep(&unmap, newbrk, oldbrk, uf);
+ if (ret)
+ return ret;
+
if (mm->locked_vm)
unlock_range(vma, oldbrk);
if (!detach_vmas_to_be_unmapped(mm, vma, prev, oldbrk))
- downgrade = false;
+ ret = 0;
else
mmap_write_downgrade(mm);
vma_init(&unmap, mm);
unmap.vm_start = newbrk;
unmap.vm_end = oldbrk;
- unmap.vm_flags = vma->vm_flags;
-
- unmap_pages = vma_pages(&unmap);
-
- if (vma->vm_flags & VM_LOCKED) {
- mm->locked_vm -= unmap_pages;
- munlock_vma_pages_range(vma, newbrk, oldbrk);
- }
+ ret = userfaultfd_unmap_prep(&unmap, newbrk, oldbrk, uf);
+ if (ret)
+ return ret;
// Change the oldbrk of vma to the newbrk of the munmap area
+ vma_adjust_trans_huge(vma, vma->vm_start, newbrk, 0);
vma_mt_brk(vma, newbrk);
+ // Handle usertfaults here.
+ //
- if (vma->vm_next && (vma->vm_next->vm_flags & VM_GROWSDOWN))
- downgrade = false;
+ unmap_pages = vma_pages(&unmap);
+ if (unmap.vm_flags & VM_LOCKED) {
+ mm->locked_vm -= unmap_pages;
+ munlock_vma_pages_range(&unmap, newbrk, oldbrk);
+ }
- if (vma->vm_prev && (vma->vm_prev->vm_flags & VM_GROWSUP))
- downgrade = false;
+ // prev of unmap_pages is just vma.
+ if (vma->vm_flags & VM_GROWSUP)
+ ret = 0;
+ else if (vma->vm_next && (vma->vm_next->vm_flags & VM_GROWSDOWN))
+ ret = 0;
- if (downgrade)
+ if (ret)
mmap_write_downgrade(mm);
-
unmap_region(mm, &unmap, vma, newbrk, oldbrk);
/* Statistics */
vm_stat_account(mm, unmap.vm_flags, -unmap_pages);
munmap_full_vma:
validate_mm_mt(mm);
- return downgrade;
+ return ret;
}
/*
* do not match then create a new anonymous VMA. Eventually we may be able to
* do some brk-specific accounting here.
*/
-static int do_brk_flags(unsigned long addr, unsigned long len,
- struct vm_area_struct *vma, unsigned long flags)
+static int do_brk_flags(struct vm_area_struct **brkvma, unsigned long addr,
+ unsigned long len, unsigned long flags)
{
struct mm_struct *mm = current->mm;
- struct vm_area_struct *prev = NULL;
+ struct vm_area_struct *prev = NULL, *vma = NULL;
int error;
unsigned long mapped_addr;
validate_mm_mt(mm);
if (security_vm_enough_memory_mm(mm, len >> PAGE_SHIFT))
return -ENOMEM;
- /* Fast path, expand the existing vma if possible */
- if (vma && ((vma->vm_flags & ~VM_SOFTDIRTY) == flags)) {
- vma_mt_brk(vma, addr + len);
- goto out;
+ if (brkvma) {
+ vma = &brkvma;
+ /* Fast path, expand the existing vma if possible */
+ if (vma && ((vma->vm_flags & ~VM_SOFTDIRTY) == flags)){
+ vma_mt_brk(vma, addr + len);
+ goto out;
+ }
}
/* create a vma struct for an anonymous mapping */
if (!prev)
find_vma_prev(mm, addr, &prev);
vma_link(mm, vma, prev);
+ *brkvma = vma;
out:
perf_event_mmap(vma);
mm->total_vm += len >> PAGE_SHIFT;
if (mmap_write_lock_killable(mm))
return -EINTR;
- ret = do_brk_flags(addr, len, NULL, flags);
+ ret = do_brk_flags(NULL, addr, len, flags);
populate = ((mm->def_flags & VM_LOCKED) != 0);
mmap_write_unlock(mm);
if (populate && !ret)