pgd_val(*pgdp) = virt_to_phys(pmdp);
 }
 
-#define __pte_page(pte)        ((unsigned long) (pte_val(pte) & PAGE_MASK))
+#define __pte_page(pte)        ((void *) (pte_val(pte) & PAGE_MASK))
 #define pmd_page_vaddr(pmd)    ((unsigned long) (pmd_val(pmd)))
 
 static inline int pte_none(pte_t pte)
        pte_val(*ptep) = 0;
 }
 
-#define pte_pagenr(pte)        ((__pte_page(pte) - PAGE_OFFSET) >> PAGE_SHIFT)
 #define pte_page(pte)  virt_to_page(__pte_page(pte))
 
 static inline int pmd_none2(pmd_t *pmd) { return !pmd_val(*pmd); }
 
 #define pmd_set(pmdp,ptep) do {} while (0)
 
 #define __pte_page(pte) \
-((unsigned long) __va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT))
+(__va ((pte_val (pte) & SUN3_PAGE_PGNUM_MASK) << PAGE_SHIFT))
 
 static inline unsigned long pmd_page_vaddr(pmd_t pmd)
 {
 
 #define pte_page(pte)          virt_to_page(__pte_page(pte))
 #define pmd_pfn(pmd)           (pmd_val(pmd) >> PAGE_SHIFT)
-#define pmd_page(pmd)          virt_to_page(pmd_page_vaddr(pmd))
+#define pmd_page(pmd)          virt_to_page((void *)pmd_page_vaddr(pmd))
 
 
 static inline int pmd_none2 (pmd_t *pmd) { return !pmd_val (*pmd); }
 
 
                /* now change pg_table to kernel virtual addresses */
                for (i = 0; i < PTRS_PER_PTE; ++i, ++pg_table) {
-                       pte_t pte = pfn_pte(virt_to_pfn(address), PAGE_INIT);
+                       pte_t pte = pfn_pte(virt_to_pfn((void *)address),
+                                           PAGE_INIT);
                        if (address >= (unsigned long) high_memory)
                                pte_val(pte) = 0;
 
 
        LIST_HEAD_INIT(ptable_list[1]),
 };
 
-#define PD_PTABLE(page) ((ptable_desc *)&(virt_to_page(page)->lru))
+#define PD_PTABLE(page) ((ptable_desc *)&(virt_to_page((void *)(page))->lru))
 #define PD_PAGE(ptable) (list_entry(ptable, struct page, lru))
 #define PD_MARKBITS(dp) (*(unsigned int *)&PD_PAGE(dp)->index)
 
                list_del(dp);
                mmu_page_dtor((void *)page);
                if (type == TABLE_PTE)
-                       pgtable_pte_page_dtor(virt_to_page(page));
+                       pgtable_pte_page_dtor(virt_to_page((void *)page));
                free_page (page);
                return 1;
        } else if (ptable_list[type].next != dp) {
 
                /* now change pg_table to kernel virtual addresses */
                pg_table = (pte_t *) __va ((unsigned long) pg_table);
                for (i=0; i<PTRS_PER_PTE; ++i, ++pg_table) {
-                       pte_t pte = pfn_pte(virt_to_pfn(address), PAGE_INIT);
+                       pte_t pte = pfn_pte(virt_to_pfn((void *)address), PAGE_INIT);
                        if (address >= (unsigned long)high_memory)
                                pte_val (pte) = 0;
                        set_pte (pg_table, pte);
 
        j = *(volatile unsigned long *)kaddr;
        *(volatile unsigned long *)kaddr = j;
 
-       ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL);
+       ptep = pfn_pte(virt_to_pfn((void *)kaddr), PAGE_KERNEL);
        pte = pte_val(ptep);
 //     pr_info("dvma_remap: addr %lx -> %lx pte %08lx\n", kaddr, vaddr, pte);
        if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) {
 
                        do {
                                pr_debug("mapping %08lx phys to %08lx\n",
                                         __pa(kaddr), vaddr);
-                               set_pte(pte, pfn_pte(virt_to_pfn(kaddr),
+                               set_pte(pte, pfn_pte(virt_to_pfn((void *)kaddr),
                                                     PAGE_KERNEL));
                                pte++;
                                kaddr += PAGE_SIZE;