dev_kfree_skb(skb);
                return -ENOMEM;
        }
+       netdev_sent_queue(dev, skb->len);
        spin_lock_irqsave(&priv->lock, flags);
 
        /* Start from the next BD that should be filled */
 {
        /* Start from the next BD that should be filled */
        struct net_device *dev = priv->ndev;
+       unsigned int bytes_sent = 0;
+       int howmany = 0;
        struct qe_bd *bd;               /* BD pointer */
        u16 bd_status;
        int tx_restart = 0;
                skb = priv->tx_skbuff[priv->skb_dirtytx];
                if (!skb)
                        break;
+               howmany++;
+               bytes_sent += skb->len;
                dev->stats.tx_packets++;
                memset(priv->tx_buffer +
                       (be32_to_cpu(bd->buf) - priv->dma_tx_addr),
        if (tx_restart)
                hdlc_tx_restart(priv);
 
+       netdev_completed_queue(dev, howmany, bytes_sent);
        return 0;
 }
 
                priv->hdlc_busy = 1;
                netif_device_attach(priv->ndev);
                napi_enable(&priv->napi);
+               netdev_reset_queue(dev);
                netif_start_queue(dev);
                hdlc_open(dev);
        }
 
        free_irq(priv->ut_info->uf_info.irq, priv);
        netif_stop_queue(dev);
+       netdev_reset_queue(dev);
        priv->hdlc_busy = 0;
 
        return 0;