#define __read_mostly __attribute__((__section__(".data..read_mostly")))
 
+static inline int cache_line_size_of_cpu(void)
+{
+       u32 cwg = cache_type_cwg();
+
+       return cwg ? 4 << cwg : ARCH_DMA_MINALIGN;
+}
+
 int cache_line_size(void);
 
 /*
 
 
 int cache_line_size(void)
 {
-       u32 cwg = cache_type_cwg();
-
        if (coherency_max_size != 0)
                return coherency_max_size;
 
-       return cwg ? 4 << cwg : ARCH_DMA_MINALIGN;
+       return cache_line_size_of_cpu();
 }
 EXPORT_SYMBOL_GPL(cache_line_size);
 
 
 
 static int __init arm64_dma_init(void)
 {
-       WARN_TAINT(ARCH_DMA_MINALIGN < cache_line_size(),
-                  TAINT_CPU_OUT_OF_SPEC,
-                  "ARCH_DMA_MINALIGN smaller than CTR_EL0.CWG (%d < %d)",
-                  ARCH_DMA_MINALIGN, cache_line_size());
        return dma_atomic_pool_init(GFP_DMA32, __pgprot(PROT_NORMAL_NC));
 }
 arch_initcall(arm64_dma_init);
 void arch_setup_dma_ops(struct device *dev, u64 dma_base, u64 size,
                        const struct iommu_ops *iommu, bool coherent)
 {
+       int cls = cache_line_size_of_cpu();
+
+       WARN_TAINT(!coherent && cls > ARCH_DMA_MINALIGN,
+                  TAINT_CPU_OUT_OF_SPEC,
+                  "%s %s: ARCH_DMA_MINALIGN smaller than CTR_EL0.CWG (%d < %d)",
+                  dev_driver_string(dev), dev_name(dev),
+                  ARCH_DMA_MINALIGN, cls);
+
        dev->dma_coherent = coherent;
        __iommu_setup_dma_ops(dev, dma_base, size, iommu);