if (dma_capable(dev, dev_addr, size))
                return dev_addr;
 
-       swiotlb_tbl_unmap_single(dev, map, size, dir,
-                                attrs | DMA_ATTR_SKIP_CPU_SYNC);
+       attrs |= DMA_ATTR_SKIP_CPU_SYNC;
+       swiotlb_tbl_unmap_single(dev, map, size, dir, attrs);
 
        return DMA_ERROR_CODE;
 }
 
        if (dma_capable(dev, dev_addr, size))
                return dev_addr;
 
-       swiotlb_tbl_unmap_single(dev, map, size, dir,
-                                attrs | DMA_ATTR_SKIP_CPU_SYNC);
+       attrs |= DMA_ATTR_SKIP_CPU_SYNC;
+       swiotlb_tbl_unmap_single(dev, map, size, dir, attrs);
 
        return phys_to_dma(dev, io_tlb_overflow_buffer);
 }
                                /* Don't panic here, we expect map_sg users
                                   to do proper error handling. */
                                swiotlb_full(hwdev, sg->length, dir, 0);
+                               attrs |= DMA_ATTR_SKIP_CPU_SYNC;
                                swiotlb_unmap_sg_attrs(hwdev, sgl, i, dir,
                                                       attrs);
                                sg_dma_len(sgl) = 0;
        for_each_sg(sgl, sg, nelems, i)
                unmap_single(hwdev, sg->dma_address, sg_dma_len(sg), dir,
                             attrs);
-
 }
 EXPORT_SYMBOL(swiotlb_unmap_sg_attrs);