goto out;
        }
 
-       tmpptr = pci_alloc_consistent(pdev, TX_TOTAL_SIZE, &ring_dma);
+       tmpptr = dma_alloc_coherent(&pdev->dev, TX_TOTAL_SIZE, &ring_dma,
+                                   GFP_KERNEL);
        rrpriv->tx_ring = tmpptr;
        rrpriv->tx_ring_dma = ring_dma;
 
                goto out;
        }
 
-       tmpptr = pci_alloc_consistent(pdev, RX_TOTAL_SIZE, &ring_dma);
+       tmpptr = dma_alloc_coherent(&pdev->dev, RX_TOTAL_SIZE, &ring_dma,
+                                   GFP_KERNEL);
        rrpriv->rx_ring = tmpptr;
        rrpriv->rx_ring_dma = ring_dma;
 
                goto out;
        }
 
-       tmpptr = pci_alloc_consistent(pdev, EVT_RING_SIZE, &ring_dma);
+       tmpptr = dma_alloc_coherent(&pdev->dev, EVT_RING_SIZE, &ring_dma,
+                                   GFP_KERNEL);
        rrpriv->evt_ring = tmpptr;
        rrpriv->evt_ring_dma = ring_dma;
 
 
  out:
        if (rrpriv->evt_ring)
-               pci_free_consistent(pdev, EVT_RING_SIZE, rrpriv->evt_ring,
-                                   rrpriv->evt_ring_dma);
+               dma_free_coherent(&pdev->dev, EVT_RING_SIZE, rrpriv->evt_ring,
+                                 rrpriv->evt_ring_dma);
        if (rrpriv->rx_ring)
-               pci_free_consistent(pdev, RX_TOTAL_SIZE, rrpriv->rx_ring,
-                                   rrpriv->rx_ring_dma);
+               dma_free_coherent(&pdev->dev, RX_TOTAL_SIZE, rrpriv->rx_ring,
+                                 rrpriv->rx_ring_dma);
        if (rrpriv->tx_ring)
-               pci_free_consistent(pdev, TX_TOTAL_SIZE, rrpriv->tx_ring,
-                                   rrpriv->tx_ring_dma);
+               dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, rrpriv->tx_ring,
+                                 rrpriv->tx_ring_dma);
        if (rrpriv->regs)
                pci_iounmap(pdev, rrpriv->regs);
        if (pdev)
        }
 
        unregister_netdev(dev);
-       pci_free_consistent(pdev, EVT_RING_SIZE, rr->evt_ring,
-                           rr->evt_ring_dma);
-       pci_free_consistent(pdev, RX_TOTAL_SIZE, rr->rx_ring,
-                           rr->rx_ring_dma);
-       pci_free_consistent(pdev, TX_TOTAL_SIZE, rr->tx_ring,
-                           rr->tx_ring_dma);
+       dma_free_coherent(&pdev->dev, EVT_RING_SIZE, rr->evt_ring,
+                         rr->evt_ring_dma);
+       dma_free_coherent(&pdev->dev, RX_TOTAL_SIZE, rr->rx_ring,
+                         rr->rx_ring_dma);
+       dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, rr->tx_ring,
+                         rr->tx_ring_dma);
        pci_iounmap(pdev, rr->regs);
        pci_release_regions(pdev);
        pci_disable_device(pdev);
                        goto error;
                }
                rrpriv->rx_skbuff[i] = skb;
-               addr = pci_map_single(rrpriv->pci_dev, skb->data,
-                       dev->mtu + HIPPI_HLEN, PCI_DMA_FROMDEVICE);
+               addr = dma_map_single(&rrpriv->pci_dev->dev, skb->data,
+                                     dev->mtu + HIPPI_HLEN, DMA_FROM_DEVICE);
                /*
                 * Sanity test to see if we conflict with the DMA
                 * limitations of the Roadrunner.
                struct sk_buff *skb = rrpriv->rx_skbuff[i];
 
                if (skb) {
-                       pci_unmap_single(rrpriv->pci_dev,
+                       dma_unmap_single(&rrpriv->pci_dev->dev,
                                         rrpriv->rx_ring[i].addr.addrlo,
                                         dev->mtu + HIPPI_HLEN,
-                                        PCI_DMA_FROMDEVICE);
+                                        DMA_FROM_DEVICE);
                        rrpriv->rx_ring[i].size = 0;
                        set_rraddr(&rrpriv->rx_ring[i].addr, 0);
                        dev_kfree_skb(skb);
                                        dev->stats.rx_dropped++;
                                        goto defer;
                                } else {
-                                       pci_dma_sync_single_for_cpu(rrpriv->pci_dev,
-                                                                   desc->addr.addrlo,
-                                                                   pkt_len,
-                                                                   PCI_DMA_FROMDEVICE);
+                                       dma_sync_single_for_cpu(&rrpriv->pci_dev->dev,
+                                                               desc->addr.addrlo,
+                                                               pkt_len,
+                                                               DMA_FROM_DEVICE);
 
                                        skb_put_data(skb, rx_skb->data,
                                                     pkt_len);
 
-                                       pci_dma_sync_single_for_device(rrpriv->pci_dev,
-                                                                      desc->addr.addrlo,
-                                                                      pkt_len,
-                                                                      PCI_DMA_FROMDEVICE);
+                                       dma_sync_single_for_device(&rrpriv->pci_dev->dev,
+                                                                  desc->addr.addrlo,
+                                                                  pkt_len,
+                                                                  DMA_FROM_DEVICE);
                                }
                        }else{
                                struct sk_buff *newskb;
                                if (newskb){
                                        dma_addr_t addr;
 
-                                       pci_unmap_single(rrpriv->pci_dev,
-                                               desc->addr.addrlo, dev->mtu +
-                                               HIPPI_HLEN, PCI_DMA_FROMDEVICE);
+                                       dma_unmap_single(&rrpriv->pci_dev->dev,
+                                                        desc->addr.addrlo,
+                                                        dev->mtu + HIPPI_HLEN,
+                                                        DMA_FROM_DEVICE);
                                        skb = rx_skb;
                                        skb_put(skb, pkt_len);
                                        rrpriv->rx_skbuff[index] = newskb;
-                                       addr = pci_map_single(rrpriv->pci_dev,
-                                               newskb->data,
-                                               dev->mtu + HIPPI_HLEN,
-                                               PCI_DMA_FROMDEVICE);
+                                       addr = dma_map_single(&rrpriv->pci_dev->dev,
+                                                             newskb->data,
+                                                             dev->mtu + HIPPI_HLEN,
+                                                             DMA_FROM_DEVICE);
                                        set_rraddr(&desc->addr, addr);
                                } else {
                                        printk("%s: Out of memory, deferring "
                                dev->stats.tx_packets++;
                                dev->stats.tx_bytes += skb->len;
 
-                               pci_unmap_single(rrpriv->pci_dev,
+                               dma_unmap_single(&rrpriv->pci_dev->dev,
                                                 desc->addr.addrlo, skb->len,
-                                                PCI_DMA_TODEVICE);
+                                                DMA_TO_DEVICE);
                                dev_kfree_skb_irq(skb);
 
                                rrpriv->tx_skbuff[txcon] = NULL;
                if (skb) {
                        struct tx_desc *desc = &(rrpriv->tx_ring[i]);
 
-                       pci_unmap_single(rrpriv->pci_dev, desc->addr.addrlo,
-                               skb->len, PCI_DMA_TODEVICE);
+                       dma_unmap_single(&rrpriv->pci_dev->dev,
+                                        desc->addr.addrlo, skb->len,
+                                        DMA_TO_DEVICE);
                        desc->size = 0;
                        set_rraddr(&desc->addr, 0);
                        dev_kfree_skb(skb);
                if (skb) {
                        struct rx_desc *desc = &(rrpriv->rx_ring[i]);
 
-                       pci_unmap_single(rrpriv->pci_dev, desc->addr.addrlo,
-                               dev->mtu + HIPPI_HLEN, PCI_DMA_FROMDEVICE);
+                       dma_unmap_single(&rrpriv->pci_dev->dev,
+                                        desc->addr.addrlo,
+                                        dev->mtu + HIPPI_HLEN,
+                                        DMA_FROM_DEVICE);
                        desc->size = 0;
                        set_rraddr(&desc->addr, 0);
                        dev_kfree_skb(skb);
                goto error;
        }
 
-       rrpriv->rx_ctrl = pci_alloc_consistent(pdev,
-                                              256 * sizeof(struct ring_ctrl),
-                                              &dma_addr);
+       rrpriv->rx_ctrl = dma_alloc_coherent(&pdev->dev,
+                                            256 * sizeof(struct ring_ctrl),
+                                            &dma_addr, GFP_KERNEL);
        if (!rrpriv->rx_ctrl) {
                ecode = -ENOMEM;
                goto error;
        }
        rrpriv->rx_ctrl_dma = dma_addr;
 
-       rrpriv->info = pci_alloc_consistent(pdev, sizeof(struct rr_info),
-                                           &dma_addr);
+       rrpriv->info = dma_alloc_coherent(&pdev->dev, sizeof(struct rr_info),
+                                         &dma_addr, GFP_KERNEL);
        if (!rrpriv->info) {
                ecode = -ENOMEM;
                goto error;
        spin_unlock_irqrestore(&rrpriv->lock, flags);
 
        if (rrpriv->info) {
-               pci_free_consistent(pdev, sizeof(struct rr_info), rrpriv->info,
-                                   rrpriv->info_dma);
+               dma_free_coherent(&pdev->dev, sizeof(struct rr_info),
+                                 rrpriv->info, rrpriv->info_dma);
                rrpriv->info = NULL;
        }
        if (rrpriv->rx_ctrl) {
-               pci_free_consistent(pdev, 256 * sizeof(struct ring_ctrl),
-                                   rrpriv->rx_ctrl, rrpriv->rx_ctrl_dma);
+               dma_free_coherent(&pdev->dev, 256 * sizeof(struct ring_ctrl),
+                                 rrpriv->rx_ctrl, rrpriv->rx_ctrl_dma);
                rrpriv->rx_ctrl = NULL;
        }
 
        rr_raz_tx(rrpriv, dev);
        rr_raz_rx(rrpriv, dev);
 
-       pci_free_consistent(pdev, 256 * sizeof(struct ring_ctrl),
-                           rrpriv->rx_ctrl, rrpriv->rx_ctrl_dma);
+       dma_free_coherent(&pdev->dev, 256 * sizeof(struct ring_ctrl),
+                         rrpriv->rx_ctrl, rrpriv->rx_ctrl_dma);
        rrpriv->rx_ctrl = NULL;
 
-       pci_free_consistent(pdev, sizeof(struct rr_info), rrpriv->info,
-                           rrpriv->info_dma);
+       dma_free_coherent(&pdev->dev, sizeof(struct rr_info), rrpriv->info,
+                         rrpriv->info_dma);
        rrpriv->info = NULL;
 
        spin_unlock_irqrestore(&rrpriv->lock, flags);
        index = txctrl->pi;
 
        rrpriv->tx_skbuff[index] = skb;
-       set_rraddr(&rrpriv->tx_ring[index].addr, pci_map_single(
-               rrpriv->pci_dev, skb->data, len + 8, PCI_DMA_TODEVICE));
+       set_rraddr(&rrpriv->tx_ring[index].addr,
+                  dma_map_single(&rrpriv->pci_dev->dev, skb->data, len + 8, DMA_TO_DEVICE));
        rrpriv->tx_ring[index].size = len + 8; /* include IFIELD */
        rrpriv->tx_ring[index].mode = PACKET_START | PACKET_END;
        txctrl->pi = (index + 1) % TX_RING_ENTRIES;