int swiotlb __read_mostly;
 
-void *x86_swiotlb_alloc_coherent(struct device *hwdev, size_t size,
-                                       dma_addr_t *dma_handle, gfp_t flags,
-                                       unsigned long attrs)
-{
-       void *vaddr;
-
-       /*
-        * Don't print a warning when the first allocation attempt fails.
-        * swiotlb_alloc_coherent() will print a warning when the DMA
-        * memory allocation ultimately failed.
-        */
-       flags |= __GFP_NOWARN;
-
-       vaddr = dma_direct_alloc(hwdev, size, dma_handle, flags, attrs);
-       if (vaddr)
-               return vaddr;
-
-       return swiotlb_alloc_coherent(hwdev, size, dma_handle, flags);
-}
-
-void x86_swiotlb_free_coherent(struct device *dev, size_t size,
-                                     void *vaddr, dma_addr_t dma_addr,
-                                     unsigned long attrs)
-{
-       if (is_swiotlb_buffer(dma_to_phys(dev, dma_addr)))
-               swiotlb_free_coherent(dev, size, vaddr, dma_addr);
-       else
-               dma_direct_free(dev, size, vaddr, dma_addr, attrs);
-}
-
-static const struct dma_map_ops x86_swiotlb_dma_ops = {
-       .mapping_error = swiotlb_dma_mapping_error,
-       .alloc = x86_swiotlb_alloc_coherent,
-       .free = x86_swiotlb_free_coherent,
-       .sync_single_for_cpu = swiotlb_sync_single_for_cpu,
-       .sync_single_for_device = swiotlb_sync_single_for_device,
-       .sync_sg_for_cpu = swiotlb_sync_sg_for_cpu,
-       .sync_sg_for_device = swiotlb_sync_sg_for_device,
-       .map_sg = swiotlb_map_sg_attrs,
-       .unmap_sg = swiotlb_unmap_sg_attrs,
-       .map_page = swiotlb_map_page,
-       .unmap_page = swiotlb_unmap_page,
-       .dma_supported = NULL,
-};
-
 /*
  * pci_swiotlb_detect_override - set swiotlb to 1 if necessary
  *
 {
        if (swiotlb) {
                swiotlb_init(0);
-               dma_ops = &x86_swiotlb_dma_ops;
+               dma_ops = &swiotlb_dma_ops;
        }
 }
 
 
        return p;
 }
 
-/**
- * sta2x11_swiotlb_alloc_coherent - Allocate swiotlb bounce buffers
- *     returns virtual address. This is the only "special" function here.
- * @dev: PCI device
- * @size: Size of the buffer
- * @dma_handle: DMA address
- * @flags: memory flags
- */
-static void *sta2x11_swiotlb_alloc_coherent(struct device *dev,
-                                           size_t size,
-                                           dma_addr_t *dma_handle,
-                                           gfp_t flags,
-                                           unsigned long attrs)
-{
-       void *vaddr;
-
-       vaddr = x86_swiotlb_alloc_coherent(dev, size, dma_handle, flags, attrs);
-       *dma_handle = p2a(*dma_handle, to_pci_dev(dev));
-       return vaddr;
-}
-
-/* We have our own dma_ops: the same as swiotlb but from alloc (above) */
-static const struct dma_map_ops sta2x11_dma_ops = {
-       .alloc = sta2x11_swiotlb_alloc_coherent,
-       .free = x86_swiotlb_free_coherent,
-       .map_page = swiotlb_map_page,
-       .unmap_page = swiotlb_unmap_page,
-       .map_sg = swiotlb_map_sg_attrs,
-       .unmap_sg = swiotlb_unmap_sg_attrs,
-       .sync_single_for_cpu = swiotlb_sync_single_for_cpu,
-       .sync_single_for_device = swiotlb_sync_single_for_device,
-       .sync_sg_for_cpu = swiotlb_sync_sg_for_cpu,
-       .sync_sg_for_device = swiotlb_sync_sg_for_device,
-       .mapping_error = swiotlb_dma_mapping_error,
-       .dma_supported = dma_direct_supported,
-};
-
 /* At setup time, we use our own ops if the device is a ConneXt one */
 static void sta2x11_setup_pdev(struct pci_dev *pdev)
 {
                return;
        pci_set_consistent_dma_mask(pdev, STA2X11_AMBA_SIZE - 1);
        pci_set_dma_mask(pdev, STA2X11_AMBA_SIZE - 1);
-       pdev->dev.dma_ops = &sta2x11_dma_ops;
+       pdev->dev.dma_ops = &swiotlb_dma_ops;
+       pdev->dev.archdata.is_sta2x11 = true;
 
        /* We must enable all devices as master, for audio DMA to work */
        pci_set_master(pdev);
 {
        struct sta2x11_mapping *map;
 
-       if (dev->dma_ops != &sta2x11_dma_ops) {
+       if (!dev->archdata.is_sta2x11) {
                if (!dev->dma_mask)
                        return false;
                return addr + size - 1 <= *dev->dma_mask;
  */
 dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
 {
-       if (dev->dma_ops != &sta2x11_dma_ops)
+       if (!dev->archdata.is_sta2x11)
                return paddr;
        return p2a(paddr, to_pci_dev(dev));
 }
  */
 phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr)
 {
-       if (dev->dma_ops != &sta2x11_dma_ops)
+       if (!dev->archdata.is_sta2x11)
                return daddr;
        return a2p(daddr, to_pci_dev(dev));
 }