cppi5_hdesc_init(desc_rx, CPPI5_INFO0_HDESC_EPIB_PRESENT,
                         AM65_CPSW_NAV_PS_DATA_SIZE);
-       cppi5_hdesc_attach_buf(desc_rx, 0, 0, buf_dma, skb_tailroom(skb));
+       cppi5_hdesc_attach_buf(desc_rx, buf_dma, skb_tailroom(skb), buf_dma, skb_tailroom(skb));
        swdata = cppi5_hdesc_get_swdata(desc_rx);
        *((void **)swdata) = skb;
 
        dev_kfree_skb_any(skb);
 }
 
+static struct sk_buff *
+am65_cpsw_nuss_tx_compl_packet(struct am65_cpsw_tx_chn *tx_chn,
+                              dma_addr_t desc_dma)
+{
+       struct am65_cpsw_ndev_priv *ndev_priv;
+       struct am65_cpsw_ndev_stats *stats;
+       struct cppi5_host_desc_t *desc_tx;
+       struct net_device *ndev;
+       struct sk_buff *skb;
+       void **swdata;
+
+       desc_tx = k3_cppi_desc_pool_dma2virt(tx_chn->desc_pool,
+                                            desc_dma);
+       swdata = cppi5_hdesc_get_swdata(desc_tx);
+       skb = *(swdata);
+       am65_cpsw_nuss_xmit_free(tx_chn, tx_chn->common->dev, desc_tx);
+
+       ndev = skb->dev;
+
+       am65_cpts_tx_timestamp(tx_chn->common->cpts, skb);
+
+       ndev_priv = netdev_priv(ndev);
+       stats = this_cpu_ptr(ndev_priv->stats);
+       u64_stats_update_begin(&stats->syncp);
+       stats->tx_packets++;
+       stats->tx_bytes += skb->len;
+       u64_stats_update_end(&stats->syncp);
+
+       return skb;
+}
+
+static void am65_cpsw_nuss_tx_wake(struct am65_cpsw_tx_chn *tx_chn, struct net_device *ndev,
+                                  struct netdev_queue *netif_txq)
+{
+       if (netif_tx_queue_stopped(netif_txq)) {
+               /* Check whether the queue is stopped due to stalled
+                * tx dma, if the queue is stopped then wake the queue
+                * as we have free desc for tx
+                */
+               __netif_tx_lock(netif_txq, smp_processor_id());
+               if (netif_running(ndev) &&
+                   (k3_cppi_desc_pool_avail(tx_chn->desc_pool) >= MAX_SKB_FRAGS))
+                       netif_tx_wake_queue(netif_txq);
+
+               __netif_tx_unlock(netif_txq);
+       }
+}
+
 static int am65_cpsw_nuss_tx_compl_packets(struct am65_cpsw_common *common,
                                           int chn, unsigned int budget)
 {
-       struct cppi5_host_desc_t *desc_tx;
        struct device *dev = common->dev;
        struct am65_cpsw_tx_chn *tx_chn;
        struct netdev_queue *netif_txq;
        struct sk_buff *skb;
        dma_addr_t desc_dma;
        int res, num_tx = 0;
-       void **swdata;
 
        tx_chn = &common->tx_chns[chn];
 
        while (true) {
-               struct am65_cpsw_ndev_priv *ndev_priv;
-               struct am65_cpsw_ndev_stats *stats;
-
+               spin_lock(&tx_chn->lock);
                res = k3_udma_glue_pop_tx_chn(tx_chn->tx_chn, &desc_dma);
+               spin_unlock(&tx_chn->lock);
                if (res == -ENODATA)
                        break;
 
                        break;
                }
 
-               desc_tx = k3_cppi_desc_pool_dma2virt(tx_chn->desc_pool,
-                                                    desc_dma);
-               swdata = cppi5_hdesc_get_swdata(desc_tx);
-               skb = *(swdata);
-               am65_cpsw_nuss_xmit_free(tx_chn, dev, desc_tx);
-
+               skb = am65_cpsw_nuss_tx_compl_packet(tx_chn, desc_dma);
+               total_bytes = skb->len;
                ndev = skb->dev;
+               napi_consume_skb(skb, budget);
+               num_tx++;
 
-               am65_cpts_tx_timestamp(common->cpts, skb);
+               netif_txq = netdev_get_tx_queue(ndev, chn);
 
-               ndev_priv = netdev_priv(ndev);
-               stats = this_cpu_ptr(ndev_priv->stats);
-               u64_stats_update_begin(&stats->syncp);
-               stats->tx_packets++;
-               stats->tx_bytes += skb->len;
-               u64_stats_update_end(&stats->syncp);
+               netdev_tx_completed_queue(netif_txq, num_tx, total_bytes);
+
+               am65_cpsw_nuss_tx_wake(tx_chn, ndev, netif_txq);
+       }
 
+       dev_dbg(dev, "%s:%u pkt:%d\n", __func__, chn, num_tx);
+
+       return num_tx;
+}
+
+static int am65_cpsw_nuss_tx_compl_packets_2g(struct am65_cpsw_common *common,
+                                             int chn, unsigned int budget)
+{
+       struct device *dev = common->dev;
+       struct am65_cpsw_tx_chn *tx_chn;
+       struct netdev_queue *netif_txq;
+       unsigned int total_bytes = 0;
+       struct net_device *ndev;
+       struct sk_buff *skb;
+       dma_addr_t desc_dma;
+       int res, num_tx = 0;
+
+       tx_chn = &common->tx_chns[chn];
+
+       while (true) {
+               res = k3_udma_glue_pop_tx_chn(tx_chn->tx_chn, &desc_dma);
+               if (res == -ENODATA)
+                       break;
+
+               if (cppi5_desc_is_tdcm(desc_dma)) {
+                       if (atomic_dec_and_test(&common->tdown_cnt))
+                               complete(&common->tdown_complete);
+                       break;
+               }
+
+               skb = am65_cpsw_nuss_tx_compl_packet(tx_chn, desc_dma);
+
+               ndev = skb->dev;
                total_bytes += skb->len;
                napi_consume_skb(skb, budget);
                num_tx++;
 
        netdev_tx_completed_queue(netif_txq, num_tx, total_bytes);
 
-       if (netif_tx_queue_stopped(netif_txq)) {
-               /* Check whether the queue is stopped due to stalled tx dma,
-                * if the queue is stopped then wake the queue as
-                * we have free desc for tx
-                */
-               __netif_tx_lock(netif_txq, smp_processor_id());
-               if (netif_running(ndev) &&
-                   (k3_cppi_desc_pool_avail(tx_chn->desc_pool) >=
-                    MAX_SKB_FRAGS))
-                       netif_tx_wake_queue(netif_txq);
+       am65_cpsw_nuss_tx_wake(tx_chn, ndev, netif_txq);
 
-               __netif_tx_unlock(netif_txq);
-       }
        dev_dbg(dev, "%s:%u pkt:%d\n", __func__, chn, num_tx);
 
        return num_tx;
        struct am65_cpsw_tx_chn *tx_chn = am65_cpsw_napi_to_tx_chn(napi_tx);
        int num_tx;
 
-       num_tx = am65_cpsw_nuss_tx_compl_packets(tx_chn->common, tx_chn->id,
-                                                budget);
+       if (AM65_CPSW_IS_CPSW2G(tx_chn->common))
+               num_tx = am65_cpsw_nuss_tx_compl_packets_2g(tx_chn->common, tx_chn->id, budget);
+       else
+               num_tx = am65_cpsw_nuss_tx_compl_packets(tx_chn->common, tx_chn->id, budget);
+
        num_tx = min(num_tx, budget);
        if (num_tx < budget) {
                napi_complete(napi_tx);
 
        cppi5_hdesc_set_pktlen(first_desc, pkt_len);
        desc_dma = k3_cppi_desc_pool_virt2dma(tx_chn->desc_pool, first_desc);
-       ret = k3_udma_glue_push_tx_chn(tx_chn->tx_chn, first_desc, desc_dma);
+       if (AM65_CPSW_IS_CPSW2G(common)) {
+               ret = k3_udma_glue_push_tx_chn(tx_chn->tx_chn, first_desc, desc_dma);
+       } else {
+               spin_lock_bh(&tx_chn->lock);
+               ret = k3_udma_glue_push_tx_chn(tx_chn->tx_chn, first_desc, desc_dma);
+               spin_unlock_bh(&tx_chn->lock);
+       }
        if (ret) {
                dev_err(dev, "can't push desc %d\n", ret);
                /* inform bql */
                snprintf(tx_chn->tx_chn_name,
                         sizeof(tx_chn->tx_chn_name), "tx%d", i);
 
+               spin_lock_init(&tx_chn->lock);
                tx_chn->common = common;
                tx_chn->id = i;
                tx_chn->descs_num = max_desc_num;