void *addr = page_address(page) + offset;
        BUG_ON(direction == DMA_NONE);
 
-       flush_kernel_dcache_range((unsigned long) addr, size);
+       if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
+               flush_kernel_dcache_range((unsigned long) addr, size);
+
        return virt_to_phys(addr);
 }
 
 {
        BUG_ON(direction == DMA_NONE);
 
+       if (attrs & DMA_ATTR_SKIP_CPU_SYNC)
+               return;
+
        if (direction == DMA_TO_DEVICE)
-           return;
+               return;
 
        /*
         * For PCI_DMA_FROMDEVICE this flush is not necessary for the
         */
 
        flush_kernel_dcache_range((unsigned long) phys_to_virt(dma_handle), size);
-       return;
 }
 
 static int pa11_dma_map_sg(struct device *dev, struct scatterlist *sglist,
 
                sg_dma_address(sg) = (dma_addr_t) virt_to_phys(vaddr);
                sg_dma_len(sg) = sg->length;
+
+               if (attrs & DMA_ATTR_SKIP_CPU_SYNC)
+                       continue;
+
                flush_kernel_dcache_range(vaddr, sg->length);
        }
        return nents;
 
        BUG_ON(direction == DMA_NONE);
 
+       if (attrs & DMA_ATTR_SKIP_CPU_SYNC)
+               return;
+
        if (direction == DMA_TO_DEVICE)
-           return;
+               return;
 
        /* once we do combining we'll need to use phys_to_virt(sg_dma_address(sglist)) */
 
        for_each_sg(sglist, sg, nents, i)
                flush_kernel_vmap_range(sg_virt(sg), sg->length);
-       return;
 }
 
 static void pa11_dma_sync_single_for_cpu(struct device *dev,