Most of the mm subsystem uses pr_<level> so make it consistent.
Miscellanea:
 - Realign arguments
 - Add missing newline to format
 - kmemleak-test.c has a "kmemleak: " prefix added to the
   "Kmemleak testing" logging message via pr_fmt
Signed-off-by: Joe Perches <joe@perches.com>
Acked-by: Tejun Heo <tj@kernel.org>	[percpu]
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
 
        if (copy_to_user(buffer, kbuf, sizeof(kbuf)))
                return -EFAULT;
-       printk_once(KERN_WARNING "%s exported in /proc is scheduled for removal\n",
-                       table->procname);
+       pr_warn_once("%s exported in /proc is scheduled for removal\n",
+                    table->procname);
 
        *lenp = 2;
        *ppos += *lenp;
 
 
 #define bdebug(fmt, args...) ({                                \
        if (unlikely(bootmem_debug))                    \
-               printk(KERN_INFO                        \
-                       "bootmem::%s " fmt,             \
+               pr_info("bootmem::%s " fmt,             \
                        __func__, ## args);             \
 })
 
        /*
         * Whoops, we cannot satisfy the allocation request.
         */
-       printk(KERN_ALERT "bootmem alloc of %lu bytes failed!\n", size);
+       pr_alert("bootmem alloc of %lu bytes failed!\n", size);
        panic("Out of memory");
        return NULL;
 }
        if (ptr)
                return ptr;
 
-       printk(KERN_ALERT "bootmem alloc of %lu bytes failed!\n", size);
+       pr_alert("bootmem alloc of %lu bytes failed!\n", size);
        panic("Out of memory");
        return NULL;
 }
 
                                        "dma_pool_destroy %s, %p busy\n",
                                        pool->name, page->vaddr);
                        else
-                               printk(KERN_ERR
-                                      "dma_pool_destroy %s, %p busy\n",
+                               pr_err("dma_pool_destroy %s, %p busy\n",
                                       pool->name, page->vaddr);
                        /* leak the still-in-use consistent memory */
                        list_del(&page->page_list);
                                "dma_pool_free %s, %p/%lx (bad dma)\n",
                                pool->name, vaddr, (unsigned long)dma);
                else
-                       printk(KERN_ERR "dma_pool_free %s, %p/%lx (bad dma)\n",
+                       pr_err("dma_pool_free %s, %p/%lx (bad dma)\n",
                               pool->name, vaddr, (unsigned long)dma);
                return;
        }
                                "dma_pool_free %s, %p (bad vaddr)/%Lx\n",
                                pool->name, vaddr, (unsigned long long)dma);
                else
-                       printk(KERN_ERR
-                              "dma_pool_free %s, %p (bad vaddr)/%Lx\n",
+                       pr_err("dma_pool_free %s, %p (bad vaddr)/%Lx\n",
                               pool->name, vaddr, (unsigned long long)dma);
                return;
        }
                                dev_err(pool->dev, "dma_pool_free %s, dma %Lx already free\n",
                                        pool->name, (unsigned long long)dma);
                        else
-                               printk(KERN_ERR "dma_pool_free %s, dma %Lx already free\n",
-                                       pool->name, (unsigned long long)dma);
+                               pr_err("dma_pool_free %s, dma %Lx already free\n",
+                                      pool->name, (unsigned long long)dma);
                        return;
                }
        }
 
 do { \
        if (level < mminit_loglevel) { \
                if (level <= MMINIT_WARNING) \
-                       printk(KERN_WARNING "mminit::" prefix " " fmt, ##arg); \
+                       pr_warn("mminit::" prefix " " fmt, ##arg);      \
                else \
                        printk(KERN_DEBUG "mminit::" prefix " " fmt, ##arg); \
        } \
 
        shadow = alloc_pages_node(node, flags | __GFP_NOTRACK, order);
        if (!shadow) {
                if (printk_ratelimit())
-                       printk(KERN_ERR "kmemcheck: failed to allocate shadow bitmap\n");
+                       pr_err("kmemcheck: failed to allocate shadow bitmap\n");
                return;
        }
 
 
        struct test_node *elem;
        int i;
 
-       printk(KERN_INFO "Kmemleak testing\n");
+       pr_info("Kmemleak testing\n");
 
        /* make some orphan objects */
        pr_info("kmalloc(32) = %p\n", kmalloc(32, GFP_KERNEL));
 
        struct siginfo si;
        int ret;
 
-       printk(KERN_ERR
-               "MCE %#lx: Killing %s:%d due to hardware memory corruption\n",
-               pfn, t->comm, t->pid);
+       pr_err("MCE %#lx: Killing %s:%d due to hardware memory corruption\n",
+              pfn, t->comm, t->pid);
        si.si_signo = SIGBUS;
        si.si_errno = 0;
        si.si_addr = (void *)addr;
                ret = send_sig_info(SIGBUS, &si, t);  /* synchronous? */
        }
        if (ret < 0)
-               printk(KERN_INFO "MCE: Error sending signal to %s:%d: %d\n",
-                      t->comm, t->pid, ret);
+               pr_info("MCE: Error sending signal to %s:%d: %d\n",
+                       t->comm, t->pid, ret);
        return ret;
 }
 
        } else {
                tk = kmalloc(sizeof(struct to_kill), GFP_ATOMIC);
                if (!tk) {
-                       printk(KERN_ERR
-               "MCE: Out of memory while machine check handling\n");
+                       pr_err("MCE: Out of memory while machine check handling\n");
                        return;
                }
        }
                         * signal and then access the memory. Just kill it.
                         */
                        if (fail || tk->addr_valid == 0) {
-                               printk(KERN_ERR
-               "MCE %#lx: forcibly killing %s:%d because of failure to unmap corrupted page\n",
-                                       pfn, tk->tsk->comm, tk->tsk->pid);
+                               pr_err("MCE %#lx: forcibly killing %s:%d because of failure to unmap corrupted page\n",
+                                      pfn, tk->tsk->comm, tk->tsk->pid);
                                force_sig(SIGKILL, tk->tsk);
                        }
 
                         */
                        else if (kill_proc(tk->tsk, tk->addr, trapno,
                                              pfn, page, flags) < 0)
-                               printk(KERN_ERR
-               "MCE %#lx: Cannot send advisory machine check signal to %s:%d\n",
-                                       pfn, tk->tsk->comm, tk->tsk->pid);
+                               pr_err("MCE %#lx: Cannot send advisory machine check signal to %s:%d\n",
+                                      pfn, tk->tsk->comm, tk->tsk->pid);
                }
                put_task_struct(tk->tsk);
                kfree(tk);
  */
 static int me_unknown(struct page *p, unsigned long pfn)
 {
-       printk(KERN_ERR "MCE %#lx: Unknown page state\n", pfn);
+       pr_err("MCE %#lx: Unknown page state\n", pfn);
        return MF_FAILED;
 }
 
        if (mapping->a_ops->error_remove_page) {
                err = mapping->a_ops->error_remove_page(mapping, p);
                if (err != 0) {
-                       printk(KERN_INFO "MCE %#lx: Failed to punch page: %d\n",
-                                       pfn, err);
+                       pr_info("MCE %#lx: Failed to punch page: %d\n",
+                               pfn, err);
                } else if (page_has_private(p) &&
                                !try_to_release_page(p, GFP_NOIO)) {
                        pr_info("MCE %#lx: failed to release buffers\n", pfn);
                if (invalidate_inode_page(p))
                        ret = MF_RECOVERED;
                else
-                       printk(KERN_INFO "MCE %#lx: Failed to invalidate\n",
-                               pfn);
+                       pr_info("MCE %#lx: Failed to invalidate\n", pfn);
        }
        return ret;
 }
        if (ps->action == me_swapcache_dirty && result == MF_DELAYED)
                count--;
        if (count != 0) {
-               printk(KERN_ERR
-                      "MCE %#lx: %s still referenced by %d users\n",
+               pr_err("MCE %#lx: %s still referenced by %d users\n",
                       pfn, action_page_types[ps->type], count);
                result = MF_FAILED;
        }
        }
 
        if (PageSwapCache(p)) {
-               printk(KERN_ERR
-                      "MCE %#lx: keeping poisoned page in swap cache\n", pfn);
+               pr_err("MCE %#lx: keeping poisoned page in swap cache\n", pfn);
                ttu |= TTU_IGNORE_HWPOISON;
        }
 
                } else {
                        kill = 0;
                        ttu |= TTU_IGNORE_HWPOISON;
-                       printk(KERN_INFO
-       "MCE %#lx: corrupted page was clean: dropped without side effects\n",
+                       pr_info("MCE %#lx: corrupted page was clean: dropped without side effects\n",
                                pfn);
                }
        }
 
        ret = try_to_unmap(hpage, ttu);
        if (ret != SWAP_SUCCESS)
-               printk(KERN_ERR "MCE %#lx: failed to unmap page (mapcount=%d)\n",
-                               pfn, page_mapcount(hpage));
+               pr_err("MCE %#lx: failed to unmap page (mapcount=%d)\n",
+                      pfn, page_mapcount(hpage));
 
        /*
         * Now that the dirty bit has been propagated to the
                panic("Memory failure from trap %d on page %lx", trapno, pfn);
 
        if (!pfn_valid(pfn)) {
-               printk(KERN_ERR
-                      "MCE %#lx: memory outside kernel control\n",
-                      pfn);
+               pr_err("MCE %#lx: memory outside kernel control\n", pfn);
                return -ENXIO;
        }
 
        p = pfn_to_page(pfn);
        orig_head = hpage = compound_head(p);
        if (TestSetPageHWPoison(p)) {
-               printk(KERN_ERR "MCE %#lx: already hardware poisoned\n", pfn);
+               pr_err("MCE %#lx: already hardware poisoned\n", pfn);
                return 0;
        }
 
         * unpoison always clear PG_hwpoison inside page lock
         */
        if (!PageHWPoison(p)) {
-               printk(KERN_ERR "MCE %#lx: just unpoisoned\n", pfn);
+               pr_err("MCE %#lx: just unpoisoned\n", pfn);
                num_poisoned_pages_sub(nr_pages);
                unlock_page(hpage);
                put_hwpoison_page(hpage);
 
                        return;
                }
                if (nr_unshown) {
-                       printk(KERN_ALERT
-                               "BUG: Bad page map: %lu messages suppressed\n",
-                               nr_unshown);
+                       pr_alert("BUG: Bad page map: %lu messages suppressed\n",
+                                nr_unshown);
                        nr_unshown = 0;
                }
                nr_shown = 0;
        mapping = vma->vm_file ? vma->vm_file->f_mapping : NULL;
        index = linear_page_index(vma, addr);
 
-       printk(KERN_ALERT
-               "BUG: Bad page map in process %s  pte:%08llx pmd:%08llx\n",
-               current->comm,
-               (long long)pte_val(pte), (long long)pmd_val(*pmd));
+       pr_alert("BUG: Bad page map in process %s  pte:%08llx pmd:%08llx\n",
+                current->comm,
+                (long long)pte_val(pte), (long long)pmd_val(*pmd));
        if (page)
                dump_page(page, "bad pte");
-       printk(KERN_ALERT
-               "addr:%p vm_flags:%08lx anon_vma:%p mapping:%p index:%lx\n",
-               (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index);
+       pr_alert("addr:%p vm_flags:%08lx anon_vma:%p mapping:%p index:%lx\n",
+                (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index);
        /*
         * Choose text because data symbols depend on CONFIG_KALLSYMS_ALL=y
         */
 
                        /* Iterate the zonelist */
                        for_each_zone_zonelist(zone, z, zonelist, zoneid) {
 #ifdef CONFIG_NUMA
-                               printk(KERN_CONT "%d:%s ",
-                                       zone->node, zone->name);
+                               pr_cont("%d:%s ", zone->node, zone->name);
 #else
-                               printk(KERN_CONT "0:%s ", zone->name);
+                               pr_cont("0:%s ", zone->name);
 #endif /* CONFIG_NUMA */
                        }
-                       printk(KERN_CONT "\n");
+                       pr_cont("\n");
                }
        }
 }
 
        /*
         * Whoops, we cannot satisfy the allocation request.
         */
-       printk(KERN_ALERT "bootmem alloc of %lu bytes failed!\n", size);
+       pr_alert("bootmem alloc of %lu bytes failed!\n", size);
        panic("Out of memory");
        return NULL;
 }
        if (ptr)
                return ptr;
 
-       printk(KERN_ALERT "bootmem alloc of %lu bytes failed!\n", size);
+       pr_alert("bootmem alloc of %lu bytes failed!\n", size);
        panic("Out of memory");
        return NULL;
 }
 
        unsigned long res;
 
        if (kstrtoul(buf, 10, &res) < 0 ||  res > MAX_ORDER / 2) {
-               printk(KERN_ERR "Bad debug_guardpage_minorder value\n");
+               pr_err("Bad debug_guardpage_minorder value\n");
                return 0;
        }
        _debug_guardpage_minorder = res;
-       printk(KERN_INFO "Setting debug_guardpage_minorder to %lu\n", res);
+       pr_info("Setting debug_guardpage_minorder to %lu\n", res);
        return 0;
 }
 __setup("debug_guardpage_minorder=", debug_guardpage_minorder_setup);
        } else if (*s == 'z' || *s == 'Z') {
                user_zonelist_order = ZONELIST_ORDER_ZONE;
        } else {
-               printk(KERN_WARNING
-                      "Ignoring invalid numa_zonelist_order value:  %s\n", s);
+               pr_warn("Ignoring invalid numa_zonelist_order value:  %s\n", s);
                return -EINVAL;
        }
        return 0;
                                               "  %s zone: %lu pages used for memmap\n",
                                               zone_names[j], memmap_pages);
                        } else
-                               printk(KERN_WARNING
-                                       "  %s zone: %lu pages exceeds freesize %lu\n",
+                               pr_warn("  %s zone: %lu pages exceeds freesize %lu\n",
                                        zone_names[j], memmap_pages, freesize);
                }
 
                min_pfn = min(min_pfn, start_pfn);
 
        if (min_pfn == ULONG_MAX) {
-               printk(KERN_WARNING
-                       "Could not find start_pfn for node %d\n", nid);
+               pr_warn("Could not find start_pfn for node %d\n", nid);
                return 0;
        }
 
        if (!table)
                panic("Failed to allocate %s hash table\n", tablename);
 
-       printk(KERN_INFO "%s hash table entries: %ld (order: %d, %lu bytes)\n",
-              tablename,
-              (1UL << log2qty),
-              ilog2(size) - PAGE_SHIFT,
-              size);
+       pr_info("%s hash table entries: %ld (order: %d, %lu bytes)\n",
+               tablename, 1UL << log2qty, ilog2(size) - PAGE_SHIFT, size);
 
        if (_hash_shift)
                *_hash_shift = log2qty;
                BUG_ON(!PageBuddy(page));
                order = page_order(page);
 #ifdef CONFIG_DEBUG_VM
-               printk(KERN_INFO "remove from free list %lx %d %lx\n",
-                      pfn, 1 << order, end_pfn);
+               pr_info("remove from free list %lx %d %lx\n",
+                       pfn, 1 << order, end_pfn);
 #endif
                list_del(&page->lru);
                rmv_page_order(page);
 
                 * Also clear PG_reclaim to avoid rotate_reclaimable_page()
                 */
                set_page_dirty(page);
-               printk(KERN_ALERT "Write-error on swap-device (%u:%u:%Lu)\n",
-                               imajor(bio->bi_bdev->bd_inode),
-                               iminor(bio->bi_bdev->bd_inode),
-                               (unsigned long long)bio->bi_iter.bi_sector);
+               pr_alert("Write-error on swap-device (%u:%u:%llu)\n",
+                        imajor(bio->bi_bdev->bd_inode),
+                        iminor(bio->bi_bdev->bd_inode),
+                        (unsigned long long)bio->bi_iter.bi_sector);
                ClearPageReclaim(page);
        }
        end_page_writeback(page);
        if (bio->bi_error) {
                SetPageError(page);
                ClearPageUptodate(page);
-               printk(KERN_ALERT "Read-error on swap-device (%u:%u:%Lu)\n",
-                               imajor(bio->bi_bdev->bd_inode),
-                               iminor(bio->bi_bdev->bd_inode),
-                               (unsigned long long)bio->bi_iter.bi_sector);
+               pr_alert("Read-error on swap-device (%u:%u:%llu)\n",
+                        imajor(bio->bi_bdev->bd_inode),
+                        iminor(bio->bi_bdev->bd_inode),
+                        (unsigned long long)bio->bi_iter.bi_sector);
                goto out;
        }
 
 out:
        return ret;
 bad_bmap:
-       printk(KERN_ERR "swapon: swapfile has holes\n");
+       pr_err("swapon: swapfile has holes\n");
        ret = -EINVAL;
        goto out;
 }
                         */
                        set_page_dirty(page);
                        ClearPageReclaim(page);
-                       pr_err_ratelimited("Write error on dio swapfile (%Lu)\n",
-                               page_file_offset(page));
+                       pr_err_ratelimited("Write error on dio swapfile (%llu)\n",
+                                          page_file_offset(page));
                }
                end_page_writeback(page);
                return ret;
 
 
        /* all units must be in a single group */
        if (ai->nr_groups != 1) {
-               printk(KERN_CRIT "percpu: can't handle more than one groups\n");
+               pr_crit("percpu: can't handle more than one groups\n");
                return -EINVAL;
        }
 
        alloc_pages = roundup_pow_of_two(nr_pages);
 
        if (alloc_pages > nr_pages)
-               printk(KERN_WARNING "percpu: wasting %zu pages per chunk\n",
-                      alloc_pages - nr_pages);
+               pr_warn("percpu: wasting %zu pages per chunk\n",
+                       alloc_pages - nr_pages);
 
        return 0;
 }
 
                for (alloc_end += gi->nr_units / upa;
                     alloc < alloc_end; alloc++) {
                        if (!(alloc % apl)) {
-                               printk(KERN_CONT "\n");
+                               pr_cont("\n");
                                printk("%spcpu-alloc: ", lvl);
                        }
-                       printk(KERN_CONT "[%0*d] ", group_width, group);
+                       pr_cont("[%0*d] ", group_width, group);
 
                        for (unit_end += upa; unit < unit_end; unit++)
                                if (gi->cpu_map[unit] != NR_CPUS)
-                                       printk(KERN_CONT "%0*d ", cpu_width,
-                                              gi->cpu_map[unit]);
+                                       pr_cont("%0*d ",
+                                               cpu_width, gi->cpu_map[unit]);
                                else
-                                       printk(KERN_CONT "%s ", empty_str);
+                                       pr_cont("%s ", empty_str);
                }
        }
-       printk(KERN_CONT "\n");
+       pr_cont("\n");
 }
 
 /**
 
                if ((value = strchr(this_char,'=')) != NULL) {
                        *value++ = 0;
                } else {
-                       printk(KERN_ERR
-                           "tmpfs: No value for mount option '%s'\n",
-                           this_char);
+                       pr_err("tmpfs: No value for mount option '%s'\n",
+                              this_char);
                        goto error;
                }
 
                        if (mpol_parse_str(value, &mpol))
                                goto bad_val;
                } else {
-                       printk(KERN_ERR "tmpfs: Bad mount option %s\n",
-                              this_char);
+                       pr_err("tmpfs: Bad mount option %s\n", this_char);
                        goto error;
                }
        }
        return 0;
 
 bad_val:
-       printk(KERN_ERR "tmpfs: Bad value '%s' for mount option '%s'\n",
+       pr_err("tmpfs: Bad value '%s' for mount option '%s'\n",
               value, this_char);
 error:
        mpol_put(mpol);
 
        error = register_filesystem(&shmem_fs_type);
        if (error) {
-               printk(KERN_ERR "Could not register tmpfs\n");
+               pr_err("Could not register tmpfs\n");
                goto out2;
        }
 
        shm_mnt = kern_mount(&shmem_fs_type);
        if (IS_ERR(shm_mnt)) {
                error = PTR_ERR(shm_mnt);
-               printk(KERN_ERR "Could not kern_mount tmpfs\n");
+               pr_err("Could not kern_mount tmpfs\n");
                goto out1;
        }
        return 0;
 
 static void __slab_error(const char *function, struct kmem_cache *cachep,
                        char *msg)
 {
-       printk(KERN_ERR "slab error in %s(): cache `%s': %s\n",
+       pr_err("slab error in %s(): cache `%s': %s\n",
               function, cachep->name, msg);
        dump_stack();
        add_taint(TAINT_BAD_PAGE, LOCKDEP_NOW_UNRELIABLE);
        unsigned char error = 0;
        int bad_count = 0;
 
-       printk(KERN_ERR "%03x: ", offset);
+       pr_err("%03x: ", offset);
        for (i = 0; i < limit; i++) {
                if (data[offset + i] != POISON_FREE) {
                        error = data[offset + i];
        if (bad_count == 1) {
                error ^= POISON_FREE;
                if (!(error & (error - 1))) {
-                       printk(KERN_ERR "Single bit error detected. Probably bad RAM.\n");
+                       pr_err("Single bit error detected. Probably bad RAM.\n");
 #ifdef CONFIG_X86
-                       printk(KERN_ERR "Run memtest86+ or a similar memory test tool.\n");
+                       pr_err("Run memtest86+ or a similar memory test tool.\n");
 #else
-                       printk(KERN_ERR "Run a memory test tool.\n");
+                       pr_err("Run a memory test tool.\n");
 #endif
                }
        }
        char *realobj;
 
        if (cachep->flags & SLAB_RED_ZONE) {
-               printk(KERN_ERR "Redzone: 0x%llx/0x%llx.\n",
-                       *dbg_redzone1(cachep, objp),
-                       *dbg_redzone2(cachep, objp));
+               pr_err("Redzone: 0x%llx/0x%llx\n",
+                      *dbg_redzone1(cachep, objp),
+                      *dbg_redzone2(cachep, objp));
        }
 
        if (cachep->flags & SLAB_STORE_USER) {
-               printk(KERN_ERR "Last user: [<%p>](%pSR)\n",
+               pr_err("Last user: [<%p>](%pSR)\n",
                       *dbg_userword(cachep, objp),
                       *dbg_userword(cachep, objp));
        }
                        /* Mismatch ! */
                        /* Print header */
                        if (lines == 0) {
-                               printk(KERN_ERR
-                                       "Slab corruption (%s): %s start=%p, len=%d\n",
-                                       print_tainted(), cachep->name, realobj, size);
+                               pr_err("Slab corruption (%s): %s start=%p, len=%d\n",
+                                      print_tainted(), cachep->name,
+                                      realobj, size);
                                print_objinfo(cachep, objp, 0);
                        }
                        /* Hexdump the affected line */
                if (objnr) {
                        objp = index_to_obj(cachep, page, objnr - 1);
                        realobj = (char *)objp + obj_offset(cachep);
-                       printk(KERN_ERR "Prev obj: start=%p, len=%d\n",
-                              realobj, size);
+                       pr_err("Prev obj: start=%p, len=%d\n", realobj, size);
                        print_objinfo(cachep, objp, 2);
                }
                if (objnr + 1 < cachep->num) {
                        objp = index_to_obj(cachep, page, objnr + 1);
                        realobj = (char *)objp + obj_offset(cachep);
-                       printk(KERN_ERR "Next obj: start=%p, len=%d\n",
-                              realobj, size);
+                       pr_err("Next obj: start=%p, len=%d\n", realobj, size);
                        print_objinfo(cachep, objp, 2);
                }
        }
        /* Verify double free bug */
        for (i = page->active; i < cachep->num; i++) {
                if (get_free_obj(page, i) == objnr) {
-                       printk(KERN_ERR "slab: double free detected in cache '%s', objp %p\n",
+                       pr_err("slab: double free detected in cache '%s', objp %p\n",
                               cachep->name, objp);
                        BUG();
                }
 static void kfree_debugcheck(const void *objp)
 {
        if (!virt_addr_valid(objp)) {
-               printk(KERN_ERR "kfree_debugcheck: out of range ptr %lxh.\n",
+               pr_err("kfree_debugcheck: out of range ptr %lxh\n",
                       (unsigned long)objp);
                BUG();
        }
        else
                slab_error(cache, "memory outside object was overwritten");
 
-       printk(KERN_ERR "%p: redzone 1:0x%llx, redzone 2:0x%llx.\n",
-                       obj, redzone1, redzone2);
+       pr_err("%p: redzone 1:0x%llx, redzone 2:0x%llx\n",
+              obj, redzone1, redzone2);
 }
 
 static void *cache_free_debugcheck(struct kmem_cache *cachep, void *objp,
                if (*dbg_redzone1(cachep, objp) != RED_INACTIVE ||
                                *dbg_redzone2(cachep, objp) != RED_INACTIVE) {
                        slab_error(cachep, "double free, or memory outside object was overwritten");
-                       printk(KERN_ERR
-                               "%p: redzone 1:0x%llx, redzone 2:0x%llx\n",
-                               objp, *dbg_redzone1(cachep, objp),
-                               *dbg_redzone2(cachep, objp));
+                       pr_err("%p: redzone 1:0x%llx, redzone 2:0x%llx\n",
+                              objp, *dbg_redzone1(cachep, objp),
+                              *dbg_redzone2(cachep, objp));
                }
                *dbg_redzone1(cachep, objp) = RED_ACTIVE;
                *dbg_redzone2(cachep, objp) = RED_ACTIVE;
                cachep->ctor(objp);
        if (ARCH_SLAB_MINALIGN &&
            ((unsigned long)objp & (ARCH_SLAB_MINALIGN-1))) {
-               printk(KERN_ERR "0x%p: not aligned to ARCH_SLAB_MINALIGN=%d\n",
+               pr_err("0x%p: not aligned to ARCH_SLAB_MINALIGN=%d\n",
                       objp, (int)ARCH_SLAB_MINALIGN);
        }
        return objp;
 skip_setup:
        err = do_tune_cpucache(cachep, limit, batchcount, shared, gfp);
        if (err)
-               printk(KERN_ERR "enable_cpucache failed for %s, error %d.\n",
+               pr_err("enable_cpucache failed for %s, error %d\n",
                       cachep->name, -err);
        return err;
 }
 
        name = cachep->name;
        if (error)
-               printk(KERN_ERR "slab: cache %s error: %s\n", name, error);
+               pr_err("slab: cache %s error: %s\n", name, error);
 
        sinfo->active_objs = active_objs;
        sinfo->num_objs = num_objs;
 
                        panic("kmem_cache_create: Failed to create slab '%s'. Error %d\n",
                                name, err);
                else {
-                       printk(KERN_WARNING "kmem_cache_create(%s) failed with error %d",
+                       pr_warn("kmem_cache_create(%s) failed with error %d\n",
                                name, err);
                        dump_stack();
                }
 
        int actual_node = early_pfn_to_nid(pfn);
 
        if (node_distance(actual_node, node) > LOCAL_DISTANCE)
-               printk(KERN_WARNING "[%lx-%lx] potential offnode page_structs\n",
-                      start, end - 1);
+               pr_warn("[%lx-%lx] potential offnode page_structs\n",
+                       start, end - 1);
 }
 
 pte_t * __meminit vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node)
                if (map_map[pnum])
                        continue;
                ms = __nr_to_section(pnum);
-               printk(KERN_ERR "%s: sparsemem memory map backing failed some memory will not be available.\n",
+               pr_err("%s: sparsemem memory map backing failed some memory will not be available\n",
                       __func__);
                ms->section_mem_map = 0;
        }
 
 
        usemap_nid = sparse_early_nid(__nr_to_section(usemap_snr));
        if (usemap_nid != nid) {
-               printk(KERN_INFO
-                      "node %d must be removed before remove section %ld\n",
-                      nid, usemap_snr);
+               pr_info("node %d must be removed before remove section %ld\n",
+                       nid, usemap_snr);
                return;
        }
        /*
         * gather other removable sections for dynamic partitioning.
         * Just notify un-removable section's number here.
         */
-       printk(KERN_INFO "Section %ld and %ld (node %d)", usemap_snr,
-              pgdat_snr, nid);
-       printk(KERN_CONT
-              " have a circular dependency on usemap and pgdat allocations\n");
+       pr_info("Section %ld and %ld (node %d) have a circular dependency on usemap and pgdat allocations\n",
+               usemap_snr, pgdat_snr, nid);
 }
 #else
 static unsigned long * __init
        usemap = sparse_early_usemaps_alloc_pgdat_section(NODE_DATA(nodeid),
                                                          size * usemap_count);
        if (!usemap) {
-               printk(KERN_WARNING "%s: allocation failed\n", __func__);
+               pr_warn("%s: allocation failed\n", __func__);
                return;
        }
 
                if (map_map[pnum])
                        continue;
                ms = __nr_to_section(pnum);
-               printk(KERN_ERR "%s: sparsemem memory map backing failed some memory will not be available.\n",
+               pr_err("%s: sparsemem memory map backing failed some memory will not be available\n",
                       __func__);
                ms->section_mem_map = 0;
        }
        if (map)
                return map;
 
-       printk(KERN_ERR "%s: sparsemem memory map backing failed some memory will not be available.\n",
+       pr_err("%s: sparsemem memory map backing failed some memory will not be available\n",
               __func__);
        ms->section_mem_map = 0;
        return NULL;
 
 
        return 0;
 nomem:
-       printk(KERN_INFO "couldn't allocate enough memory for swap_cgroup.\n");
-       printk(KERN_INFO
-               "swap_cgroup can be disabled by swapaccount=0 boot option\n");
+       pr_info("couldn't allocate enough memory for swap_cgroup\n");
+       pr_info("swap_cgroup can be disabled by swapaccount=0 boot option\n");
        return -ENOMEM;
 }