struct bpf_prog *xdp_prog;
        int cq_ring = cq->ring;
        bool doorbell_pending;
+       bool xdp_redir_flush;
        struct mlx4_cqe *cqe;
        struct xdp_buff xdp;
        int polled = 0;
        xdp_prog = rcu_dereference_bh(ring->xdp_prog);
        xdp_init_buff(&xdp, priv->frag_info[0].frag_stride, &ring->xdp_rxq);
        doorbell_pending = false;
+       xdp_redir_flush = false;
 
        /* We assume a 1:1 mapping between CQEs and Rx descriptors, so Rx
         * descriptor offset can be deduced from the CQE index instead of
                        switch (act) {
                        case XDP_PASS:
                                break;
+                       case XDP_REDIRECT:
+                               if (xdp_do_redirect(dev, &xdp, xdp_prog) >= 0) {
+                                       xdp_redir_flush = true;
+                                       frags[0].page = NULL;
+                                       goto next;
+                               }
+                               trace_xdp_exception(dev, xdp_prog, act);
+                               goto xdp_drop_no_cnt;
                        case XDP_TX:
                                if (likely(!mlx4_en_xmit_frame(ring, frags, priv,
                                                        length, cq_ring,
                        break;
        }
 
+       if (xdp_redir_flush)
+               xdp_do_flush();
+
        if (likely(polled)) {
                if (doorbell_pending) {
                        priv->tx_cq[TX_XDP][cq_ring]->xdp_busy = true;