int node)
 {
        const gfp_t nested_gfp = (gfp_mask & GFP_RECLAIM_MASK) | __GFP_ZERO;
-       const gfp_t orig_gfp_mask = gfp_mask;
        bool nofail = gfp_mask & __GFP_NOFAIL;
        unsigned long addr = (unsigned long)area->addr;
        unsigned long size = get_vm_area_size(area);
        }
 
        if (!area->pages) {
-               warn_alloc(orig_gfp_mask, NULL,
+               warn_alloc(gfp_mask, NULL,
                        "vmalloc error: size %lu, failed to allocated page array size %lu",
                        nr_small_pages * PAGE_SIZE, array_size);
                free_vm_area(area);
        set_vm_area_page_order(area, page_shift - PAGE_SHIFT);
        page_order = vm_area_page_order(area);
 
-       area->nr_pages = vm_area_alloc_pages(gfp_mask, node,
-               page_order, nr_small_pages, area->pages);
+       area->nr_pages = vm_area_alloc_pages(gfp_mask | __GFP_NOWARN,
+               node, page_order, nr_small_pages, area->pages);
 
        atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
        if (gfp_mask & __GFP_ACCOUNT) {
         * allocation request, free them via __vfree() if any.
         */
        if (area->nr_pages != nr_small_pages) {
-               warn_alloc(orig_gfp_mask, NULL,
+               warn_alloc(gfp_mask, NULL,
                        "vmalloc error: size %lu, page order %u, failed to allocate pages",
                        area->nr_pages * PAGE_SIZE, page_order);
                goto fail;
                memalloc_noio_restore(flags);
 
        if (ret < 0) {
-               warn_alloc(orig_gfp_mask, NULL,
+               warn_alloc(gfp_mask, NULL,
                        "vmalloc error: size %lu, failed to map pages",
                        area->nr_pages * PAGE_SIZE);
                goto fail;