if (unlikely(ib_req_notify_cq(priv->recv_cq,
                                              IB_CQ_NEXT_COMP |
                                              IB_CQ_REPORT_MISSED_EVENTS)) &&
-                   napi_reschedule(napi))
+                   napi_schedule(napi))
                        goto poll_more;
        }
 
                napi_complete(napi);
                if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP |
                                              IB_CQ_REPORT_MISSED_EVENTS)) &&
-                   napi_reschedule(napi))
+                   napi_schedule(napi))
                        goto poll_more;
        }
        return n < 0 ? 0 : n;
 
 
                /* Check if there was another interrupt */
                if (!skb_queue_empty(&offload->skb_queue))
-                       napi_reschedule(&offload->napi);
+                       napi_schedule(&offload->napi);
        }
 
        return work_done;
 
 
                        if (fl_starving(adap, fl)) {
                                rxq = container_of(fl, struct sge_eth_rxq, fl);
-                               if (napi_reschedule(&rxq->rspq.napi))
+                               if (napi_schedule(&rxq->rspq.napi))
                                        fl->starving++;
                                else
                                        set_bit(id, s->starving_fl);
 
                                struct sge_eth_rxq *rxq;
 
                                rxq = container_of(fl, struct sge_eth_rxq, fl);
-                               if (napi_reschedule(&rxq->rspq.napi))
+                               if (napi_schedule(&rxq->rspq.napi))
                                        fl->starving++;
                                else
                                        set_bit(id, s->starving_fl);
 
                 */
                if (nps_enet_is_tx_pending(priv)) {
                        nps_enet_reg_set(priv, NPS_ENET_REG_BUF_INT_ENABLE, 0);
-                       napi_reschedule(napi);
+                       napi_schedule(napi);
                }
        }
 
 
                if (block->rx)
                        reschedule |= gve_rx_work_pending(block->rx);
 
-               if (reschedule && napi_reschedule(napi))
+               if (reschedule && napi_schedule(napi))
                        iowrite32be(GVE_IRQ_MASK, irq_doorbell);
        }
        return work_done;
 
                if (!cqe && !cqe_skb)
                        return rx;
 
-               if (!napi_reschedule(napi))
+               if (!napi_schedule(napi))
                        return rx;
 
                cqe_skb = ehea_proc_cqes(pr, EHEA_POLL_MAX_CQES);
 
                if (unlikely(mc->ops->peek_rx(mc->dev) ||
                             test_bit(MAL_COMMAC_RX_STOPPED, &mc->flags))) {
                        MAL_DBG2(mal, "rotting packet" NL);
-                       if (!napi_reschedule(napi))
+                       if (!napi_schedule(napi))
                                goto more_work;
 
                        spin_lock_irqsave(&mal->lock, flags);
 
                BUG_ON(lpar_rc != H_SUCCESS);
 
                if (ibmveth_rxq_pending_buffer(adapter) &&
-                   napi_reschedule(napi)) {
+                   napi_schedule(napi)) {
                        lpar_rc = h_vio_signal(adapter->vdev->unit_address,
                                               VIO_IRQ_DISABLE);
                }
 
                if (napi_complete_done(napi, frames_processed)) {
                        enable_scrq_irq(adapter, rx_scrq);
                        if (pending_scrq(adapter, rx_scrq)) {
-                               if (napi_reschedule(napi)) {
+                               if (napi_schedule(napi)) {
                                        disable_scrq_irq(adapter, rx_scrq);
                                        goto restart_poll;
                                }
 
        for (ring = 0; ring < priv->rx_ring_num; ring++) {
                if (mlx4_en_is_ring_empty(priv->rx_ring[ring])) {
                        local_bh_disable();
-                       napi_reschedule(&priv->rx_cq[ring]->napi);
+                       napi_schedule(&priv->rx_cq[ring]->napi);
                        local_bh_enable();
                }
        }
 
                if (status & (XAXIDMA_IRQ_IOC_MASK | XAXIDMA_IRQ_DELAY_MASK)) {
                        /* If there's more, reschedule, but clear */
                        nixge_dma_write_reg(priv, XAXIDMA_RX_SR_OFFSET, status);
-                       napi_reschedule(napi);
+                       napi_schedule(napi);
                } else {
                        /* if not, turn on RX IRQs again ... */
                        cr = nixge_dma_read_reg(priv, XAXIDMA_RX_CR_OFFSET);
 
                stmmac_start_rx(priv, priv->ioaddr, i);
 
                local_bh_disable();
-               napi_reschedule(&ch->rx_napi);
+               napi_schedule(&ch->rx_napi);
                local_bh_enable();
        }
 
 
                        napi_complete(napi);
                        qmgr_enable_irq(rxq);
                        if (!qmgr_stat_below_low_watermark(rxq) &&
-                           napi_reschedule(napi)) { /* not empty again */
+                           napi_schedule(napi)) { /* not empty again */
 #if DEBUG_RX
-                               netdev_debug(dev, "eth_poll napi_reschedule succeeded\n");
+                               netdev_debug(dev, "eth_poll napi_schedule succeeded\n");
 #endif
                                qmgr_disable_irq(rxq);
                                continue;
 
                }
 
                if (((long)jiffies - (long)adapter->rx_last_jiffies) < 3) {
-                       napi_reschedule(napi);
+                       napi_schedule(napi);
                } else {
                        spin_lock(&hw->rx_status_lock);
                        for (epidx = 0; epidx < hw->max_epid; epidx++) {
 
                        napi_complete(napi);
                        qmgr_enable_irq(rxq);
                        if (!qmgr_stat_empty(rxq) &&
-                           napi_reschedule(napi)) {
+                           napi_schedule(napi)) {
 #if DEBUG_RX
                                printk(KERN_DEBUG "%s: hss_hdlc_poll"
-                                      " napi_reschedule succeeded\n",
+                                      " napi_schedule succeeded\n",
                                       dev->name);
 #endif
                                qmgr_disable_irq(rxq);
 
                 * immediate servicing.
                 */
                if (ath10k_ce_interrupt_summary(ar)) {
-                       napi_reschedule(ctx);
+                       napi_schedule(ctx);
                        goto out;
                }
                ath10k_pci_enable_legacy_irq(ar);
 
        if (!ret) {
                napi_complete_done(napi, work_done);
                rxq->sleep_lock_pending = true;
-               napi_reschedule(napi);
+               napi_schedule(napi);
                return work_done;
        }
 
 
                __napi_schedule_irqoff(n);
 }
 
-/* Try to reschedule poll. Called by dev->poll() after napi_complete().  */
-static inline bool napi_reschedule(struct napi_struct *napi)
-{
-       if (napi_schedule_prep(napi)) {
-               __napi_schedule(napi);
-               return true;
-       }
-       return false;
-}
-
 /**
  * napi_complete_done - NAPI processing complete
  * @n: NAPI context