}
 }
 
-int aq_ring_tx_clean(struct aq_ring_s *self)
+void aq_ring_tx_clean(struct aq_ring_s *self)
 {
        struct device *dev = aq_nic_get_dev(self->aq_nic);
 
                if (unlikely(buff->is_eop))
                        dev_kfree_skb_any(buff->skb);
        }
-
-       if (aq_ring_avail_dx(self) > AQ_CFG_SKB_FRAGS_MAX)
-               aq_nic_ndev_queue_start(self->aq_nic, self->idx);
-
-       return 0;
 }
 
 static inline unsigned int aq_ring_dx_in_range(unsigned int h, unsigned int i,
 err_exit:;
 }
 
-void aq_ring_tx_deinit(struct aq_ring_s *self)
-{
-       if (!self)
-               goto err_exit;
-
-       for (; self->sw_head != self->sw_tail;
-               self->sw_head = aq_ring_next_dx(self, self->sw_head)) {
-               struct aq_ring_buff_s *buff = &self->buff_ring[self->sw_head];
-               struct device *ndev = aq_nic_get_dev(self->aq_nic);
-
-               if (likely(buff->is_mapped)) {
-                       if (unlikely(buff->is_sop)) {
-                               dma_unmap_single(ndev, buff->pa, buff->len,
-                                                DMA_TO_DEVICE);
-                       } else {
-                               dma_unmap_page(ndev, buff->pa, buff->len,
-                                              DMA_TO_DEVICE);
-                       }
-               }
-
-               if (unlikely(buff->is_eop))
-                       dev_kfree_skb_any(buff->skb);
-       }
-err_exit:;
-}
-
 void aq_ring_free(struct aq_ring_s *self)
 {
        if (!self)
 
                                   unsigned int idx,
                                   struct aq_nic_cfg_s *aq_nic_cfg);
 int aq_ring_init(struct aq_ring_s *self);
-void aq_ring_tx_deinit(struct aq_ring_s *self);
 void aq_ring_rx_deinit(struct aq_ring_s *self);
 void aq_ring_free(struct aq_ring_s *self);
 void aq_ring_tx_append_buffs(struct aq_ring_s *ring,
                             struct aq_ring_buff_s *buffer,
                             unsigned int buffers);
-int aq_ring_tx_clean(struct aq_ring_s *self);
+void aq_ring_tx_clean(struct aq_ring_s *self);
 int aq_ring_rx_clean(struct aq_ring_s *self, int *work_done, int budget);
 int aq_ring_rx_fill(struct aq_ring_s *self);
 
 
                        }
 
                        if (ring[AQ_VEC_TX_ID].sw_head !=
-                               ring[AQ_VEC_TX_ID].hw_head) {
-                               err = aq_ring_tx_clean(&ring[AQ_VEC_TX_ID]);
-                               if (err < 0)
-                                       goto err_exit;
+                           ring[AQ_VEC_TX_ID].hw_head) {
+                               aq_ring_tx_clean(&ring[AQ_VEC_TX_ID]);
+
+                               if (aq_ring_avail_dx(&ring[AQ_VEC_TX_ID]) >
+                                   AQ_CFG_SKB_FRAGS_MAX) {
+                                       aq_nic_ndev_queue_start(self->aq_nic,
+                                               ring[AQ_VEC_TX_ID].idx);
+                               }
                                was_tx_cleaned = true;
                        }
 
 
        for (i = 0U, ring = self->ring[0];
                self->tx_rings > i; ++i, ring = self->ring[i]) {
-               aq_ring_tx_deinit(&ring[AQ_VEC_TX_ID]);
+               aq_ring_tx_clean(&ring[AQ_VEC_TX_ID]);
                aq_ring_rx_deinit(&ring[AQ_VEC_RX_ID]);
        }
 err_exit:;