|
@@ -51,25 +51,33 @@
|
|
|
#define FL_HASH(l) (ntohl(l)&FL_HASH_MASK)
|
|
|
|
|
|
static atomic_t fl_size = ATOMIC_INIT(0);
|
|
|
-static struct ip6_flowlabel *fl_ht[FL_HASH_MASK+1];
|
|
|
+static struct ip6_flowlabel __rcu *fl_ht[FL_HASH_MASK+1];
|
|
|
|
|
|
static void ip6_fl_gc(unsigned long dummy);
|
|
|
static DEFINE_TIMER(ip6_fl_gc_timer, ip6_fl_gc, 0, 0);
|
|
|
|
|
|
/* FL hash table lock: it protects only of GC */
|
|
|
|
|
|
-static DEFINE_RWLOCK(ip6_fl_lock);
|
|
|
+static DEFINE_SPINLOCK(ip6_fl_lock);
|
|
|
|
|
|
/* Big socket sock */
|
|
|
|
|
|
static DEFINE_RWLOCK(ip6_sk_fl_lock);
|
|
|
|
|
|
+#define for_each_fl_rcu(hash, fl) \
|
|
|
+ for (fl = rcu_dereference(fl_ht[(hash)]); \
|
|
|
+ fl != NULL; \
|
|
|
+ fl = rcu_dereference(fl->next))
|
|
|
+#define for_each_fl_continue_rcu(fl) \
|
|
|
+ for (fl = rcu_dereference(fl->next); \
|
|
|
+ fl != NULL; \
|
|
|
+ fl = rcu_dereference(fl->next))
|
|
|
|
|
|
static inline struct ip6_flowlabel *__fl_lookup(struct net *net, __be32 label)
|
|
|
{
|
|
|
struct ip6_flowlabel *fl;
|
|
|
|
|
|
- for (fl=fl_ht[FL_HASH(label)]; fl; fl = fl->next) {
|
|
|
+ for_each_fl_rcu(FL_HASH(label), fl) {
|
|
|
if (fl->label == label && net_eq(fl->fl_net, net))
|
|
|
return fl;
|
|
|
}
|
|
@@ -80,11 +88,11 @@ static struct ip6_flowlabel *fl_lookup(struct net *net, __be32 label)
|
|
|
{
|
|
|
struct ip6_flowlabel *fl;
|
|
|
|
|
|
- read_lock_bh(&ip6_fl_lock);
|
|
|
+ rcu_read_lock_bh();
|
|
|
fl = __fl_lookup(net, label);
|
|
|
- if (fl)
|
|
|
- atomic_inc(&fl->users);
|
|
|
- read_unlock_bh(&ip6_fl_lock);
|
|
|
+ if (fl && !atomic_inc_not_zero(&fl->users))
|
|
|
+ fl = NULL;
|
|
|
+ rcu_read_unlock_bh();
|
|
|
return fl;
|
|
|
}
|
|
|
|
|
@@ -96,13 +104,13 @@ static void fl_free(struct ip6_flowlabel *fl)
|
|
|
put_pid(fl->owner.pid);
|
|
|
release_net(fl->fl_net);
|
|
|
kfree(fl->opt);
|
|
|
+ kfree_rcu(fl, rcu);
|
|
|
}
|
|
|
- kfree(fl);
|
|
|
}
|
|
|
|
|
|
static void fl_release(struct ip6_flowlabel *fl)
|
|
|
{
|
|
|
- write_lock_bh(&ip6_fl_lock);
|
|
|
+ spin_lock_bh(&ip6_fl_lock);
|
|
|
|
|
|
fl->lastuse = jiffies;
|
|
|
if (atomic_dec_and_test(&fl->users)) {
|
|
@@ -119,7 +127,7 @@ static void fl_release(struct ip6_flowlabel *fl)
|
|
|
time_after(ip6_fl_gc_timer.expires, ttd))
|
|
|
mod_timer(&ip6_fl_gc_timer, ttd);
|
|
|
}
|
|
|
- write_unlock_bh(&ip6_fl_lock);
|
|
|
+ spin_unlock_bh(&ip6_fl_lock);
|
|
|
}
|
|
|
|
|
|
static void ip6_fl_gc(unsigned long dummy)
|
|
@@ -128,12 +136,13 @@ static void ip6_fl_gc(unsigned long dummy)
|
|
|
unsigned long now = jiffies;
|
|
|
unsigned long sched = 0;
|
|
|
|
|
|
- write_lock(&ip6_fl_lock);
|
|
|
+ spin_lock(&ip6_fl_lock);
|
|
|
|
|
|
for (i=0; i<=FL_HASH_MASK; i++) {
|
|
|
struct ip6_flowlabel *fl, **flp;
|
|
|
flp = &fl_ht[i];
|
|
|
- while ((fl=*flp) != NULL) {
|
|
|
+ while ((fl = rcu_dereference_protected(*flp,
|
|
|
+ lockdep_is_held(&ip6_fl_lock))) != NULL) {
|
|
|
if (atomic_read(&fl->users) == 0) {
|
|
|
unsigned long ttd = fl->lastuse + fl->linger;
|
|
|
if (time_after(ttd, fl->expires))
|
|
@@ -156,18 +165,19 @@ static void ip6_fl_gc(unsigned long dummy)
|
|
|
if (sched) {
|
|
|
mod_timer(&ip6_fl_gc_timer, sched);
|
|
|
}
|
|
|
- write_unlock(&ip6_fl_lock);
|
|
|
+ spin_unlock(&ip6_fl_lock);
|
|
|
}
|
|
|
|
|
|
static void __net_exit ip6_fl_purge(struct net *net)
|
|
|
{
|
|
|
int i;
|
|
|
|
|
|
- write_lock(&ip6_fl_lock);
|
|
|
+ spin_lock(&ip6_fl_lock);
|
|
|
for (i = 0; i <= FL_HASH_MASK; i++) {
|
|
|
struct ip6_flowlabel *fl, **flp;
|
|
|
flp = &fl_ht[i];
|
|
|
- while ((fl = *flp) != NULL) {
|
|
|
+ while ((fl = rcu_dereference_protected(*flp,
|
|
|
+ lockdep_is_held(&ip6_fl_lock))) != NULL) {
|
|
|
if (net_eq(fl->fl_net, net) &&
|
|
|
atomic_read(&fl->users) == 0) {
|
|
|
*flp = fl->next;
|
|
@@ -178,7 +188,7 @@ static void __net_exit ip6_fl_purge(struct net *net)
|
|
|
flp = &fl->next;
|
|
|
}
|
|
|
}
|
|
|
- write_unlock(&ip6_fl_lock);
|
|
|
+ spin_unlock(&ip6_fl_lock);
|
|
|
}
|
|
|
|
|
|
static struct ip6_flowlabel *fl_intern(struct net *net,
|
|
@@ -188,7 +198,7 @@ static struct ip6_flowlabel *fl_intern(struct net *net,
|
|
|
|
|
|
fl->label = label & IPV6_FLOWLABEL_MASK;
|
|
|
|
|
|
- write_lock_bh(&ip6_fl_lock);
|
|
|
+ spin_lock_bh(&ip6_fl_lock);
|
|
|
if (label == 0) {
|
|
|
for (;;) {
|
|
|
fl->label = htonl(net_random())&IPV6_FLOWLABEL_MASK;
|
|
@@ -210,16 +220,16 @@ static struct ip6_flowlabel *fl_intern(struct net *net,
|
|
|
lfl = __fl_lookup(net, fl->label);
|
|
|
if (lfl != NULL) {
|
|
|
atomic_inc(&lfl->users);
|
|
|
- write_unlock_bh(&ip6_fl_lock);
|
|
|
+ spin_unlock_bh(&ip6_fl_lock);
|
|
|
return lfl;
|
|
|
}
|
|
|
}
|
|
|
|
|
|
fl->lastuse = jiffies;
|
|
|
fl->next = fl_ht[FL_HASH(fl->label)];
|
|
|
- fl_ht[FL_HASH(fl->label)] = fl;
|
|
|
+ rcu_assign_pointer(fl_ht[FL_HASH(fl->label)], fl);
|
|
|
atomic_inc(&fl_size);
|
|
|
- write_unlock_bh(&ip6_fl_lock);
|
|
|
+ spin_unlock_bh(&ip6_fl_lock);
|
|
|
return NULL;
|
|
|
}
|
|
|
|
|
@@ -650,13 +660,13 @@ static struct ip6_flowlabel *ip6fl_get_first(struct seq_file *seq)
|
|
|
struct net *net = seq_file_net(seq);
|
|
|
|
|
|
for (state->bucket = 0; state->bucket <= FL_HASH_MASK; ++state->bucket) {
|
|
|
- fl = fl_ht[state->bucket];
|
|
|
-
|
|
|
- while (fl && !net_eq(fl->fl_net, net))
|
|
|
- fl = fl->next;
|
|
|
- if (fl)
|
|
|
- break;
|
|
|
+ for_each_fl_rcu(state->bucket, fl) {
|
|
|
+ if (net_eq(fl->fl_net, net))
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
}
|
|
|
+ fl = NULL;
|
|
|
+out:
|
|
|
return fl;
|
|
|
}
|
|
|
|
|
@@ -665,18 +675,22 @@ static struct ip6_flowlabel *ip6fl_get_next(struct seq_file *seq, struct ip6_flo
|
|
|
struct ip6fl_iter_state *state = ip6fl_seq_private(seq);
|
|
|
struct net *net = seq_file_net(seq);
|
|
|
|
|
|
- fl = fl->next;
|
|
|
+ for_each_fl_continue_rcu(fl) {
|
|
|
+ if (net_eq(fl->fl_net, net))
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+
|
|
|
try_again:
|
|
|
- while (fl && !net_eq(fl->fl_net, net))
|
|
|
- fl = fl->next;
|
|
|
-
|
|
|
- while (!fl) {
|
|
|
- if (++state->bucket <= FL_HASH_MASK) {
|
|
|
- fl = fl_ht[state->bucket];
|
|
|
- goto try_again;
|
|
|
- } else
|
|
|
- break;
|
|
|
+ if (++state->bucket <= FL_HASH_MASK) {
|
|
|
+ for_each_fl_rcu(state->bucket, fl) {
|
|
|
+ if (net_eq(fl->fl_net, net))
|
|
|
+ goto out;
|
|
|
+ }
|
|
|
+ goto try_again;
|
|
|
}
|
|
|
+ fl = NULL;
|
|
|
+
|
|
|
+out:
|
|
|
return fl;
|
|
|
}
|
|
|
|
|
@@ -690,9 +704,9 @@ static struct ip6_flowlabel *ip6fl_get_idx(struct seq_file *seq, loff_t pos)
|
|
|
}
|
|
|
|
|
|
static void *ip6fl_seq_start(struct seq_file *seq, loff_t *pos)
|
|
|
- __acquires(ip6_fl_lock)
|
|
|
+ __acquires(RCU)
|
|
|
{
|
|
|
- read_lock_bh(&ip6_fl_lock);
|
|
|
+ rcu_read_lock_bh();
|
|
|
return *pos ? ip6fl_get_idx(seq, *pos - 1) : SEQ_START_TOKEN;
|
|
|
}
|
|
|
|
|
@@ -709,9 +723,9 @@ static void *ip6fl_seq_next(struct seq_file *seq, void *v, loff_t *pos)
|
|
|
}
|
|
|
|
|
|
static void ip6fl_seq_stop(struct seq_file *seq, void *v)
|
|
|
- __releases(ip6_fl_lock)
|
|
|
+ __releases(RCU)
|
|
|
{
|
|
|
- read_unlock_bh(&ip6_fl_lock);
|
|
|
+ rcu_read_unlock_bh();
|
|
|
}
|
|
|
|
|
|
static int ip6fl_seq_show(struct seq_file *seq, void *v)
|