extern void putback_lru_page(struct page *page);
 extern bool zone_reclaimable(struct zone *zone);
 
+#define lru_to_page(_head) (list_entry((_head)->prev, struct page, lru))
+
 /*
  * in mm/rmap.c:
  */
 
 }
 EXPORT_SYMBOL_GPL(file_ra_state_init);
 
-#define list_to_page(head) (list_entry((head)->prev, struct page, lru))
-
 /*
  * see if a page needs releasing upon read_cache_pages() failure
  * - the caller of read_cache_pages() may have set PG_private or PG_fscache
        struct page *victim;
 
        while (!list_empty(pages)) {
-               victim = list_to_page(pages);
+               victim = lru_to_page(pages);
                list_del(&victim->lru);
                read_cache_pages_invalidate_page(mapping, victim);
        }
        int ret = 0;
 
        while (!list_empty(pages)) {
-               page = list_to_page(pages);
+               page = lru_to_page(pages);
                list_del(&page->lru);
                if (add_to_page_cache_lru(page, mapping, page->index,
                                mapping_gfp_constraint(mapping, GFP_KERNEL))) {
        }
 
        for (page_idx = 0; page_idx < nr_pages; page_idx++) {
-               struct page *page = list_to_page(pages);
+               struct page *page = lru_to_page(pages);
                list_del(&page->lru);
                if (!add_to_page_cache_lru(page, mapping, page->index,
                                mapping_gfp_constraint(mapping, GFP_KERNEL))) {
 
        unsigned long nr_reclaimed;
 };
 
-#define lru_to_page(_head) (list_entry((_head)->prev, struct page, lru))
-
 #ifdef ARCH_HAS_PREFETCH
 #define prefetch_prev_lru_page(_page, _base, _field)                   \
        do {                                                            \