#define radix__flush_tlb_page_psize(mm,addr,p) radix__local_flush_tlb_page_psize(mm,addr,p)
 #endif
 extern void radix__flush_tlb_pwc(struct mmu_gather *tlb, unsigned long addr);
+extern void radix__flush_tlb_collapsed_pmd(struct mm_struct *mm, unsigned long addr);
 extern void radix__flush_tlb_lpid_va(unsigned long lpid, unsigned long gpa,
                                     unsigned long page_size);
 extern void radix__flush_tlb_lpid(unsigned long lpid);
 
 {
        struct mm_struct *mm = vma->vm_mm;
 
-       /*
-        * This is currently used when collapsing THPs so we need to
-        * flush the PWC. We should fix this.
-        */
-       radix__flush_all_mm(mm);
+       radix__flush_tlb_mm(mm);
 }
 EXPORT_SYMBOL(radix__flush_tlb_range);
 
        preempt_enable();
 }
 
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+void radix__flush_tlb_collapsed_pmd(struct mm_struct *mm, unsigned long addr)
+{
+       int local = mm_is_thread_local(mm);
+       unsigned long ap = mmu_get_ap(mmu_virtual_psize);
+       unsigned long pid, end;
+
+
+       pid = mm ? mm->context.id : 0;
+       if (unlikely(pid == MMU_NO_CONTEXT))
+               goto no_context;
+
+       /* 4k page size, just blow the world */
+       if (PAGE_SIZE == 0x1000) {
+               radix__flush_all_mm(mm);
+               return;
+       }
+
+       /* Otherwise first do the PWC */
+       if (local)
+               _tlbiel_pid(pid, RIC_FLUSH_PWC);
+       else
+               _tlbie_pid(pid, RIC_FLUSH_PWC);
+
+       /* Then iterate the pages */
+       end = addr + HPAGE_PMD_SIZE;
+       for (; addr < end; addr += PAGE_SIZE) {
+               if (local)
+                       _tlbiel_va(addr, pid, ap, RIC_FLUSH_TLB);
+               else
+                       _tlbie_va(addr, pid, ap, RIC_FLUSH_TLB);
+       }
+no_context:
+       preempt_enable();
+}
+#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
+
 void radix__flush_tlb_lpid_va(unsigned long lpid, unsigned long gpa,
                              unsigned long page_size)
 {