#include <linux/dmar.h>
 #include <linux/iommu.h>
 #include <linux/memory.h>
+#include <linux/pci.h>
 #include <linux/spinlock.h>
 
 #include "iommu.h"
        if (domain->domain.type == IOMMU_DOMAIN_NESTED)
                __cache_tag_unassign_parent_domain(domain->s2_domain, did, dev, pasid);
 }
+
+static unsigned long calculate_psi_aligned_address(unsigned long start,
+                                                  unsigned long end,
+                                                  unsigned long *_pages,
+                                                  unsigned long *_mask)
+{
+       unsigned long pages = aligned_nrpages(start, end - start + 1);
+       unsigned long aligned_pages = __roundup_pow_of_two(pages);
+       unsigned long bitmask = aligned_pages - 1;
+       unsigned long mask = ilog2(aligned_pages);
+       unsigned long pfn = IOVA_PFN(start);
+
+       /*
+        * PSI masks the low order bits of the base address. If the
+        * address isn't aligned to the mask, then compute a mask value
+        * needed to ensure the target range is flushed.
+        */
+       if (unlikely(bitmask & pfn)) {
+               unsigned long end_pfn = pfn + pages - 1, shared_bits;
+
+               /*
+                * Since end_pfn <= pfn + bitmask, the only way bits
+                * higher than bitmask can differ in pfn and end_pfn is
+                * by carrying. This means after masking out bitmask,
+                * high bits starting with the first set bit in
+                * shared_bits are all equal in both pfn and end_pfn.
+                */
+               shared_bits = ~(pfn ^ end_pfn) & ~bitmask;
+               mask = shared_bits ? __ffs(shared_bits) : BITS_PER_LONG;
+       }
+
+       *_pages = aligned_pages;
+       *_mask = mask;
+
+       return ALIGN_DOWN(start, VTD_PAGE_SIZE << mask);
+}
+
+/*
+ * Invalidates a range of IOVA from @start (inclusive) to @end (inclusive)
+ * when the memory mappings in the target domain have been modified.
+ */
+void cache_tag_flush_range(struct dmar_domain *domain, unsigned long start,
+                          unsigned long end, int ih)
+{
+       unsigned long pages, mask, addr;
+       struct cache_tag *tag;
+       unsigned long flags;
+
+       addr = calculate_psi_aligned_address(start, end, &pages, &mask);
+
+       spin_lock_irqsave(&domain->cache_lock, flags);
+       list_for_each_entry(tag, &domain->cache_tags, node) {
+               struct intel_iommu *iommu = tag->iommu;
+               struct device_domain_info *info;
+               u16 sid;
+
+               switch (tag->type) {
+               case CACHE_TAG_IOTLB:
+               case CACHE_TAG_NESTING_IOTLB:
+                       if (domain->use_first_level) {
+                               qi_flush_piotlb(iommu, tag->domain_id,
+                                               tag->pasid, addr, pages, ih);
+                       } else {
+                               /*
+                                * Fallback to domain selective flush if no
+                                * PSI support or the size is too big.
+                                */
+                               if (!cap_pgsel_inv(iommu->cap) ||
+                                   mask > cap_max_amask_val(iommu->cap))
+                                       iommu->flush.flush_iotlb(iommu, tag->domain_id,
+                                                                0, 0, DMA_TLB_DSI_FLUSH);
+                               else
+                                       iommu->flush.flush_iotlb(iommu, tag->domain_id,
+                                                                addr | ih, mask,
+                                                                DMA_TLB_PSI_FLUSH);
+                       }
+                       break;
+               case CACHE_TAG_NESTING_DEVTLB:
+                       /*
+                        * Address translation cache in device side caches the
+                        * result of nested translation. There is no easy way
+                        * to identify the exact set of nested translations
+                        * affected by a change in S2. So just flush the entire
+                        * device cache.
+                        */
+                       addr = 0;
+                       mask = MAX_AGAW_PFN_WIDTH;
+                       fallthrough;
+               case CACHE_TAG_DEVTLB:
+                       info = dev_iommu_priv_get(tag->dev);
+                       sid = PCI_DEVID(info->bus, info->devfn);
+
+                       if (tag->pasid == IOMMU_NO_PASID)
+                               qi_flush_dev_iotlb(iommu, sid, info->pfsid,
+                                                  info->ats_qdep, addr, mask);
+                       else
+                               qi_flush_dev_iotlb_pasid(iommu, sid, info->pfsid,
+                                                        tag->pasid, info->ats_qdep,
+                                                        addr, mask);
+
+                       quirk_extra_dev_tlb_flush(info, addr, mask, tag->pasid, info->ats_qdep);
+                       break;
+               }
+       }
+       spin_unlock_irqrestore(&domain->cache_lock, flags);
+}
+
+/*
+ * Invalidates all ranges of IOVA when the memory mappings in the target
+ * domain have been modified.
+ */
+void cache_tag_flush_all(struct dmar_domain *domain)
+{
+       struct cache_tag *tag;
+       unsigned long flags;
+
+       spin_lock_irqsave(&domain->cache_lock, flags);
+       list_for_each_entry(tag, &domain->cache_tags, node) {
+               struct intel_iommu *iommu = tag->iommu;
+               struct device_domain_info *info;
+               u16 sid;
+
+               switch (tag->type) {
+               case CACHE_TAG_IOTLB:
+               case CACHE_TAG_NESTING_IOTLB:
+                       if (domain->use_first_level)
+                               qi_flush_piotlb(iommu, tag->domain_id,
+                                               tag->pasid, 0, -1, 0);
+                       else
+                               iommu->flush.flush_iotlb(iommu, tag->domain_id,
+                                                        0, 0, DMA_TLB_DSI_FLUSH);
+                       break;
+               case CACHE_TAG_DEVTLB:
+               case CACHE_TAG_NESTING_DEVTLB:
+                       info = dev_iommu_priv_get(tag->dev);
+                       sid = PCI_DEVID(info->bus, info->devfn);
+
+                       qi_flush_dev_iotlb(iommu, sid, info->pfsid, info->ats_qdep,
+                                          0, MAX_AGAW_PFN_WIDTH);
+                       quirk_extra_dev_tlb_flush(info, 0, MAX_AGAW_PFN_WIDTH,
+                                                 IOMMU_NO_PASID, info->ats_qdep);
+                       break;
+               }
+       }
+       spin_unlock_irqrestore(&domain->cache_lock, flags);
+}
+
+/*
+ * Invalidate a range of IOVA when new mappings are created in the target
+ * domain.
+ *
+ * - VT-d spec, Section 6.1 Caching Mode: When the CM field is reported as
+ *   Set, any software updates to remapping structures other than first-
+ *   stage mapping requires explicit invalidation of the caches.
+ * - VT-d spec, Section 6.8 Write Buffer Flushing: For hardware that requires
+ *   write buffer flushing, software must explicitly perform write-buffer
+ *   flushing, if cache invalidation is not required.
+ */
+void cache_tag_flush_range_np(struct dmar_domain *domain, unsigned long start,
+                             unsigned long end)
+{
+       unsigned long pages, mask, addr;
+       struct cache_tag *tag;
+       unsigned long flags;
+
+       addr = calculate_psi_aligned_address(start, end, &pages, &mask);
+
+       spin_lock_irqsave(&domain->cache_lock, flags);
+       list_for_each_entry(tag, &domain->cache_tags, node) {
+               struct intel_iommu *iommu = tag->iommu;
+
+               if (!cap_caching_mode(iommu->cap) || domain->use_first_level) {
+                       iommu_flush_write_buffer(iommu);
+                       continue;
+               }
+
+               if (tag->type == CACHE_TAG_IOTLB ||
+                   tag->type == CACHE_TAG_NESTING_IOTLB) {
+                       /*
+                        * Fallback to domain selective flush if no
+                        * PSI support or the size is too big.
+                        */
+                       if (!cap_pgsel_inv(iommu->cap) ||
+                           mask > cap_max_amask_val(iommu->cap))
+                               iommu->flush.flush_iotlb(iommu, tag->domain_id,
+                                                        0, 0, DMA_TLB_DSI_FLUSH);
+                       else
+                               iommu->flush.flush_iotlb(iommu, tag->domain_id,
+                                                        addr, mask,
+                                                        DMA_TLB_PSI_FLUSH);
+               }
+       }
+       spin_unlock_irqrestore(&domain->cache_lock, flags);
+}