* the zone and PFN linkages are still valid. This is expensive, but walkers
  * of the full memmap are extremely rare.
  */
-int memmap_valid_within(unsigned long pfn,
+bool memmap_valid_within(unsigned long pfn,
                                        struct page *page, struct zone *zone);
 #else
-static inline int memmap_valid_within(unsigned long pfn,
+static inline bool memmap_valid_within(unsigned long pfn,
                                        struct page *page, struct zone *zone)
 {
-       return 1;
+       return true;
 }
 #endif /* CONFIG_ARCH_HAS_HOLES_MEMORYMODEL */
 
 
 }
 
 #ifdef CONFIG_ARCH_HAS_HOLES_MEMORYMODEL
-int memmap_valid_within(unsigned long pfn,
+bool memmap_valid_within(unsigned long pfn,
                                        struct page *page, struct zone *zone)
 {
        if (page_to_pfn(page) != pfn)
-               return 0;
+               return false;
 
        if (page_zone(page) != zone)
-               return 0;
+               return false;
 
-       return 1;
+       return true;
 }
 #endif /* CONFIG_ARCH_HAS_HOLES_MEMORYMODEL */