}
 
 static unsigned long
-elf32_map (struct file *filep, unsigned long addr, struct elf_phdr *eppnt, int prot, int type, unsigned long unused)
+elf32_map(struct file *filep, unsigned long addr, struct elf_phdr *eppnt,
+               int prot, int type, unsigned long unused)
 {
        unsigned long pgoff = (eppnt->p_vaddr) & ~IA32_PAGE_MASK;
 
 
 
        /* check if free_area_cache is useful for us */
        if (len <= mm->cached_hole_size) {
-               mm->cached_hole_size = 0;
-               mm->free_area_cache = mm->mmap_base;
-       }
+               mm->cached_hole_size = 0;
+               mm->free_area_cache = mm->mmap_base;
+       }
 
        /* either no address requested or can't fit in requested address hole */
        addr = mm->free_area_cache;
                        /* remember the address as a hint for next time */
                        return (mm->free_area_cache = addr);
 
-               /* remember the largest hole we saw so far */
-               if (addr + mm->cached_hole_size < vma->vm_start)
-                       mm->cached_hole_size = vma->vm_start - addr;
+               /* remember the largest hole we saw so far */
+               if (addr + mm->cached_hole_size < vma->vm_start)
+                       mm->cached_hole_size = vma->vm_start - addr;
 
                /* try just below the current vma->vm_start */
                addr = vma->vm_start-len;
         * allocations.
         */
        mm->cached_hole_size = ~0UL;
-       mm->free_area_cache = TASK_UNMAPPED_BASE;
+       mm->free_area_cache = TASK_UNMAPPED_BASE;
        addr = arch_get_unmapped_area(filp, addr0, len, pgoff, flags);
        /*
         * Restore the topdown base:
 
                /* ia32_pick_mmap_layout has its own. */
                return ia32_pick_mmap_layout(mm);
 #endif
-       } else if(mmap_is_legacy()) {
+       } else if (mmap_is_legacy()) {
                mm->mmap_base = TASK_UNMAPPED_BASE;
                mm->get_unmapped_area = arch_get_unmapped_area;
                mm->unmap_area = arch_unmap_area;
                if (current->flags & PF_RANDOMIZE)
                        rnd = -rnd;
        }
-       if (current->flags & PF_RANDOMIZE) {
+       if (current->flags & PF_RANDOMIZE)
                mm->mmap_base += ((long)rnd) << PAGE_SHIFT;
-       }
 }
 
 
 static int load_elf_binary(struct linux_binprm *bprm, struct pt_regs *regs);
 static int load_elf_library(struct file *);
-static unsigned long elf_map (struct file *, unsigned long, struct elf_phdr *, int, int, unsigned long);
+static unsigned long elf_map(struct file *, unsigned long, struct elf_phdr *,
+                               int, int, unsigned long);
 
 /*
  * If we don't support core dumping, then supply a NULL so we
                                load_addr = -vaddr;
 
                        map_addr = elf_map(interpreter, load_addr + vaddr,
-                                          eppnt, elf_prot, elf_type, total_size);
+                                       eppnt, elf_prot, elf_type, total_size);
                        total_size = 0;
                        if (!*interp_map_addr)
                                *interp_map_addr = map_addr;
                }
 
                error = elf_map(bprm->file, load_bias + vaddr, elf_ppnt,
-                               elf_prot, elf_flags,0);
+                               elf_prot, elf_flags, 0);
                if (BAD_ADDR(error)) {
                        send_sig(SIGKILL, current, 0);
                        retval = IS_ERR((void *)error) ?