int hugetlb_report_meminfo(char *);
 int hugetlb_report_node_meminfo(int, char *);
 unsigned long hugetlb_total_pages(void);
-struct page *alloc_huge_page(struct vm_area_struct *, unsigned long);
-void free_huge_page(struct page *);
 int hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
                        unsigned long address, int write_access);
 
 #define is_hugepage_only_range(mm, addr, len)  0
 #define hugetlb_free_pgd_range(tlb, addr, end, floor, ceiling) \
                                                do { } while (0)
-#define alloc_huge_page(vma, addr)             ({ NULL; })
-#define free_huge_page(p)                      ({ (void)(p); BUG(); })
 #define hugetlb_fault(mm, vma, addr, write)    ({ BUG(); 0; })
 
 #define hugetlb_change_protection(vma, address, end, newprot)
 
        return page;
 }
 
+static void free_huge_page(struct page *page)
+{
+       BUG_ON(page_count(page));
+
+       INIT_LIST_HEAD(&page->lru);
+
+       spin_lock(&hugetlb_lock);
+       enqueue_huge_page(page);
+       spin_unlock(&hugetlb_lock);
+}
+
 static int alloc_fresh_huge_page(void)
 {
        static int nid = 0;
        return 0;
 }
 
-void free_huge_page(struct page *page)
-{
-       BUG_ON(page_count(page));
-
-       INIT_LIST_HEAD(&page->lru);
-
-       spin_lock(&hugetlb_lock);
-       enqueue_huge_page(page);
-       spin_unlock(&hugetlb_lock);
-}
-
-struct page *alloc_huge_page(struct vm_area_struct *vma, unsigned long addr)
+static struct page *alloc_huge_page(struct vm_area_struct *vma,
+                                   unsigned long addr)
 {
        struct inode *inode = vma->vm_file->f_dentry->d_inode;
        struct page *page;