}
 
 static void brcmf_proto_bcdc_rxreorder(struct brcmf_if *ifp,
-                                      struct sk_buff *skb, bool inirq)
+                                      struct sk_buff *skb)
 {
-       brcmf_fws_rxreorder(ifp, skb, inirq);
+       brcmf_fws_rxreorder(ifp, skb);
 }
 
 static void
 
        spin_unlock_irqrestore(&ifp->netif_stop_lock, flags);
 }
 
-void brcmf_netif_rx(struct brcmf_if *ifp, struct sk_buff *skb, bool inirq)
+void brcmf_netif_rx(struct brcmf_if *ifp, struct sk_buff *skb)
 {
        /* Most of Broadcom's firmwares send 802.11f ADD frame every time a new
         * STA connects to the AP interface. This is an obsoleted standard most
        ifp->ndev->stats.rx_packets++;
 
        brcmf_dbg(DATA, "rx proto=0x%X\n", ntohs(skb->protocol));
-       if (inirq) {
-               netif_rx(skb);
-       } else {
-               /* If the receive is not processed inside an ISR,
-                * the softirqd must be woken explicitly to service
-                * the NET_RX_SOFTIRQ.  This is handled by netif_rx_ni().
-                */
-               netif_rx_ni(skb);
-       }
+       netif_rx(skb);
 }
 
 void brcmf_netif_mon_rx(struct brcmf_if *ifp, struct sk_buff *skb)
        skb->pkt_type = PACKET_OTHERHOST;
        skb->protocol = htons(ETH_P_802_2);
 
-       brcmf_netif_rx(ifp, skb, false);
+       brcmf_netif_rx(ifp, skb);
 }
 
 static int brcmf_rx_hdrpull(struct brcmf_pub *drvr, struct sk_buff *skb,
                return;
 
        if (brcmf_proto_is_reorder_skb(skb)) {
-               brcmf_proto_rxreorder(ifp, skb, inirq);
+               brcmf_proto_rxreorder(ifp, skb);
        } else {
                /* Process special event packets */
                if (handle_event) {
                        brcmf_fweh_process_skb(ifp->drvr, skb,
                                               BCMILCP_SUBTYPE_VENDOR_LONG, gfp);
                }
-               brcmf_netif_rx(ifp, skb, inirq);
+               brcmf_netif_rx(ifp, skb);
        }
 }
 
 
 void brcmf_txflowblock_if(struct brcmf_if *ifp,
                          enum brcmf_netif_stop_reason reason, bool state);
 void brcmf_txfinalize(struct brcmf_if *ifp, struct sk_buff *txp, bool success);
-void brcmf_netif_rx(struct brcmf_if *ifp, struct sk_buff *skb, bool inirq);
+void brcmf_netif_rx(struct brcmf_if *ifp, struct sk_buff *skb);
 void brcmf_netif_mon_rx(struct brcmf_if *ifp, struct sk_buff *skb);
 void brcmf_net_detach(struct net_device *ndev, bool locked);
 int brcmf_net_mon_attach(struct brcmf_if *ifp);
 
        rfi->pend_pkts -= skb_queue_len(skb_list);
 }
 
-void brcmf_fws_rxreorder(struct brcmf_if *ifp, struct sk_buff *pkt, bool inirq)
+void brcmf_fws_rxreorder(struct brcmf_if *ifp, struct sk_buff *pkt)
 {
        struct brcmf_pub *drvr = ifp->drvr;
        u8 *reorder_data;
        /* validate flags and flow id */
        if (flags == 0xFF) {
                bphy_err(drvr, "invalid flags...so ignore this packet\n");
-               brcmf_netif_rx(ifp, pkt, inirq);
+               brcmf_netif_rx(ifp, pkt);
                return;
        }
 
                if (rfi == NULL) {
                        brcmf_dbg(INFO, "received flags to cleanup, but no flow (%d) yet\n",
                                  flow_id);
-                       brcmf_netif_rx(ifp, pkt, inirq);
+                       brcmf_netif_rx(ifp, pkt);
                        return;
                }
 
                rfi = kzalloc(buf_size, GFP_ATOMIC);
                if (rfi == NULL) {
                        bphy_err(drvr, "failed to alloc buffer\n");
-                       brcmf_netif_rx(ifp, pkt, inirq);
+                       brcmf_netif_rx(ifp, pkt);
                        return;
                }
 
 netif_rx:
        skb_queue_walk_safe(&reorder_list, pkt, pnext) {
                __skb_unlink(pkt, &reorder_list);
-               brcmf_netif_rx(ifp, pkt, inirq);
+               brcmf_netif_rx(ifp, pkt);
        }
 }
 
 
 void brcmf_fws_del_interface(struct brcmf_if *ifp);
 void brcmf_fws_bustxfail(struct brcmf_fws_info *fws, struct sk_buff *skb);
 void brcmf_fws_bus_blocked(struct brcmf_pub *drvr, bool flow_blocked);
-void brcmf_fws_rxreorder(struct brcmf_if *ifp, struct sk_buff *skb, bool inirq);
+void brcmf_fws_rxreorder(struct brcmf_if *ifp, struct sk_buff *skb);
 
 #endif /* FWSIGNAL_H_ */
 
        return -ENODEV;
 }
 
-static void brcmf_msgbuf_rxreorder(struct brcmf_if *ifp, struct sk_buff *skb,
-                                  bool inirq)
+static void brcmf_msgbuf_rxreorder(struct brcmf_if *ifp, struct sk_buff *skb)
 {
 }
 
        }
 
        skb->protocol = eth_type_trans(skb, ifp->ndev);
-       brcmf_netif_rx(ifp, skb, false);
+       brcmf_netif_rx(ifp, skb);
 }
 
 static void brcmf_msgbuf_process_gen_status(struct brcmf_msgbuf *msgbuf,
 
                            u8 peer[ETH_ALEN]);
        void (*add_tdls_peer)(struct brcmf_pub *drvr, int ifidx,
                              u8 peer[ETH_ALEN]);
-       void (*rxreorder)(struct brcmf_if *ifp, struct sk_buff *skb, bool inirq);
+       void (*rxreorder)(struct brcmf_if *ifp, struct sk_buff *skb);
        void (*add_if)(struct brcmf_if *ifp);
        void (*del_if)(struct brcmf_if *ifp);
        void (*reset_if)(struct brcmf_if *ifp);
 }
 
 static inline void
-brcmf_proto_rxreorder(struct brcmf_if *ifp, struct sk_buff *skb, bool inirq)
+brcmf_proto_rxreorder(struct brcmf_if *ifp, struct sk_buff *skb)
 {
-       ifp->drvr->proto->rxreorder(ifp, skb, inirq);
+       ifp->drvr->proto->rxreorder(ifp, skb);
 }
 
 static inline void