struct tun_struct *tun;
 
        u32 rxhash;
+       u32 rps_rxhash;
        int queue_index;
        unsigned long updated;
 };
                          rxhash, queue_index);
                e->updated = jiffies;
                e->rxhash = rxhash;
+               e->rps_rxhash = 0;
                e->queue_index = queue_index;
                e->tun = tun;
                hlist_add_head_rcu(&e->hash_link, head);
 {
        tun_debug(KERN_INFO, tun, "delete flow: hash %u index %u\n",
                  e->rxhash, e->queue_index);
+       sock_rps_reset_flow_hash(e->rps_rxhash);
        hlist_del_rcu(&e->hash_link);
        kfree_rcu(e, rcu);
        --tun->flow_count;
                /* TODO: keep queueing to old queue until it's empty? */
                e->queue_index = queue_index;
                e->updated = jiffies;
+               sock_rps_record_flow_hash(e->rps_rxhash);
        } else {
                spin_lock_bh(&tun->lock);
                if (!tun_flow_find(head, rxhash) &&
        rcu_read_unlock();
 }
 
+/**
+ * Save the hash received in the stack receive path and update the
+ * flow_hash table accordingly.
+ */
+static inline void tun_flow_save_rps_rxhash(struct tun_flow_entry *e, u32 hash)
+{
+       if (unlikely(e->rps_rxhash != hash)) {
+               sock_rps_reset_flow_hash(e->rps_rxhash);
+               e->rps_rxhash = hash;
+       }
+}
+
 /* We try to identify a flow through its rxhash first. The reason that
  * we do not check rxq no. is because some cards(e.g 82599), chooses
  * the rxq based on the txq where the last packet of the flow comes. As
        txq = skb_get_hash(skb);
        if (txq) {
                e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq);
-               if (e)
+               if (e) {
                        txq = e->queue_index;
-               else
+                       tun_flow_save_rps_rxhash(e, txq);
+               } else
                        /* use multiply and shift instead of expensive divide */
                        txq = ((u64)txq * numqueues) >> 32;
        } else if (likely(skb_rx_queue_recorded(skb))) {
        if (txq >= tun->numqueues)
                goto drop;
 
+       if (tun->numqueues == 1) {
+               /* Select queue was not called for the skbuff, so we extract the
+                * RPS hash and save it into the flow_table here.
+                */
+               __u32 rxhash;
+
+               rxhash = skb_get_hash(skb);
+               if (rxhash) {
+                       struct tun_flow_entry *e;
+                       e = tun_flow_find(&tun->flows[tun_hashfn(rxhash)],
+                                       rxhash);
+                       if (e)
+                               tun_flow_save_rps_rxhash(e, rxhash);
+               }
+       }
+
        tun_debug(KERN_INFO, tun, "tun_net_xmit %d\n", skb->len);
 
        BUG_ON(!tfile);