#define FL_HASH(l)     (ntohl(l)&FL_HASH_MASK)
 
 static atomic_t fl_size = ATOMIC_INIT(0);
-static struct ip6_flowlabel *fl_ht[FL_HASH_MASK+1];
+static struct ip6_flowlabel __rcu *fl_ht[FL_HASH_MASK+1];
 
 static void ip6_fl_gc(unsigned long dummy);
 static DEFINE_TIMER(ip6_fl_gc_timer, ip6_fl_gc, 0, 0);
 
 /* FL hash table lock: it protects only of GC */
 
-static DEFINE_RWLOCK(ip6_fl_lock);
+static DEFINE_SPINLOCK(ip6_fl_lock);
 
 /* Big socket sock */
 
 static DEFINE_RWLOCK(ip6_sk_fl_lock);
 
+#define for_each_fl_rcu(hash, fl)                              \
+       for (fl = rcu_dereference(fl_ht[(hash)]);               \
+            fl != NULL;                                        \
+            fl = rcu_dereference(fl->next))
+#define for_each_fl_continue_rcu(fl)                           \
+       for (fl = rcu_dereference(fl->next);                    \
+            fl != NULL;                                        \
+            fl = rcu_dereference(fl->next))
 
 static inline struct ip6_flowlabel *__fl_lookup(struct net *net, __be32 label)
 {
        struct ip6_flowlabel *fl;
 
-       for (fl=fl_ht[FL_HASH(label)]; fl; fl = fl->next) {
+       for_each_fl_rcu(FL_HASH(label), fl) {
                if (fl->label == label && net_eq(fl->fl_net, net))
                        return fl;
        }
 {
        struct ip6_flowlabel *fl;
 
-       read_lock_bh(&ip6_fl_lock);
+       rcu_read_lock_bh();
        fl = __fl_lookup(net, label);
-       if (fl)
-               atomic_inc(&fl->users);
-       read_unlock_bh(&ip6_fl_lock);
+       if (fl && !atomic_inc_not_zero(&fl->users))
+               fl = NULL;
+       rcu_read_unlock_bh();
        return fl;
 }
 
                        put_pid(fl->owner.pid);
                release_net(fl->fl_net);
                kfree(fl->opt);
+               kfree_rcu(fl, rcu);
        }
-       kfree(fl);
 }
 
 static void fl_release(struct ip6_flowlabel *fl)
 {
-       write_lock_bh(&ip6_fl_lock);
+       spin_lock_bh(&ip6_fl_lock);
 
        fl->lastuse = jiffies;
        if (atomic_dec_and_test(&fl->users)) {
                    time_after(ip6_fl_gc_timer.expires, ttd))
                        mod_timer(&ip6_fl_gc_timer, ttd);
        }
-       write_unlock_bh(&ip6_fl_lock);
+       spin_unlock_bh(&ip6_fl_lock);
 }
 
 static void ip6_fl_gc(unsigned long dummy)
        unsigned long now = jiffies;
        unsigned long sched = 0;
 
-       write_lock(&ip6_fl_lock);
+       spin_lock(&ip6_fl_lock);
 
        for (i=0; i<=FL_HASH_MASK; i++) {
                struct ip6_flowlabel *fl, **flp;
                flp = &fl_ht[i];
-               while ((fl=*flp) != NULL) {
+               while ((fl = rcu_dereference_protected(*flp,
+                                                      lockdep_is_held(&ip6_fl_lock))) != NULL) {
                        if (atomic_read(&fl->users) == 0) {
                                unsigned long ttd = fl->lastuse + fl->linger;
                                if (time_after(ttd, fl->expires))
        if (sched) {
                mod_timer(&ip6_fl_gc_timer, sched);
        }
-       write_unlock(&ip6_fl_lock);
+       spin_unlock(&ip6_fl_lock);
 }
 
 static void __net_exit ip6_fl_purge(struct net *net)
 {
        int i;
 
-       write_lock(&ip6_fl_lock);
+       spin_lock(&ip6_fl_lock);
        for (i = 0; i <= FL_HASH_MASK; i++) {
                struct ip6_flowlabel *fl, **flp;
                flp = &fl_ht[i];
-               while ((fl = *flp) != NULL) {
+               while ((fl = rcu_dereference_protected(*flp,
+                                                      lockdep_is_held(&ip6_fl_lock))) != NULL) {
                        if (net_eq(fl->fl_net, net) &&
                            atomic_read(&fl->users) == 0) {
                                *flp = fl->next;
                        flp = &fl->next;
                }
        }
-       write_unlock(&ip6_fl_lock);
+       spin_unlock(&ip6_fl_lock);
 }
 
 static struct ip6_flowlabel *fl_intern(struct net *net,
 
        fl->label = label & IPV6_FLOWLABEL_MASK;
 
-       write_lock_bh(&ip6_fl_lock);
+       spin_lock_bh(&ip6_fl_lock);
        if (label == 0) {
                for (;;) {
                        fl->label = htonl(net_random())&IPV6_FLOWLABEL_MASK;
                lfl = __fl_lookup(net, fl->label);
                if (lfl != NULL) {
                        atomic_inc(&lfl->users);
-                       write_unlock_bh(&ip6_fl_lock);
+                       spin_unlock_bh(&ip6_fl_lock);
                        return lfl;
                }
        }
 
        fl->lastuse = jiffies;
        fl->next = fl_ht[FL_HASH(fl->label)];
-       fl_ht[FL_HASH(fl->label)] = fl;
+       rcu_assign_pointer(fl_ht[FL_HASH(fl->label)], fl);
        atomic_inc(&fl_size);
-       write_unlock_bh(&ip6_fl_lock);
+       spin_unlock_bh(&ip6_fl_lock);
        return NULL;
 }
 
        struct net *net = seq_file_net(seq);
 
        for (state->bucket = 0; state->bucket <= FL_HASH_MASK; ++state->bucket) {
-               fl = fl_ht[state->bucket];
-
-               while (fl && !net_eq(fl->fl_net, net))
-                       fl = fl->next;
-               if (fl)
-                       break;
+               for_each_fl_rcu(state->bucket, fl) {
+                       if (net_eq(fl->fl_net, net))
+                               goto out;
+               }
        }
+       fl = NULL;
+out:
        return fl;
 }
 
        struct ip6fl_iter_state *state = ip6fl_seq_private(seq);
        struct net *net = seq_file_net(seq);
 
-       fl = fl->next;
+       for_each_fl_continue_rcu(fl) {
+               if (net_eq(fl->fl_net, net))
+                       goto out;
+       }
+
 try_again:
-       while (fl && !net_eq(fl->fl_net, net))
-               fl = fl->next;
-
-       while (!fl) {
-               if (++state->bucket <= FL_HASH_MASK) {
-                       fl = fl_ht[state->bucket];
-                       goto try_again;
-               } else
-                       break;
+       if (++state->bucket <= FL_HASH_MASK) {
+               for_each_fl_rcu(state->bucket, fl) {
+                       if (net_eq(fl->fl_net, net))
+                               goto out;
+               }
+               goto try_again;
        }
+       fl = NULL;
+
+out:
        return fl;
 }
 
 }
 
 static void *ip6fl_seq_start(struct seq_file *seq, loff_t *pos)
-       __acquires(ip6_fl_lock)
+       __acquires(RCU)
 {
-       read_lock_bh(&ip6_fl_lock);
+       rcu_read_lock_bh();
        return *pos ? ip6fl_get_idx(seq, *pos - 1) : SEQ_START_TOKEN;
 }
 
 }
 
 static void ip6fl_seq_stop(struct seq_file *seq, void *v)
-       __releases(ip6_fl_lock)
+       __releases(RCU)
 {
-       read_unlock_bh(&ip6_fl_lock);
+       rcu_read_unlock_bh();
 }
 
 static int ip6fl_seq_show(struct seq_file *seq, void *v)