|
@@ -82,7 +82,7 @@ static void *__mld2_query_bugs[] __attribute__((__unused__)) = {
|
|
|
static struct in6_addr mld2_all_mcr = MLD2_ALL_MCR_INIT;
|
|
|
|
|
|
/* Big mc list lock for all the sockets */
|
|
|
-static DEFINE_RWLOCK(ipv6_sk_mc_lock);
|
|
|
+static DEFINE_SPINLOCK(ipv6_sk_mc_lock);
|
|
|
|
|
|
static void igmp6_join_group(struct ifmcaddr6 *ma);
|
|
|
static void igmp6_leave_group(struct ifmcaddr6 *ma);
|
|
@@ -123,6 +123,11 @@ int sysctl_mld_max_msf __read_mostly = IPV6_MLD_MAX_MSF;
|
|
|
* socket join on multicast group
|
|
|
*/
|
|
|
|
|
|
+#define for_each_pmc_rcu(np, pmc) \
|
|
|
+ for (pmc = rcu_dereference(np->ipv6_mc_list); \
|
|
|
+ pmc != NULL; \
|
|
|
+ pmc = rcu_dereference(pmc->next))
|
|
|
+
|
|
|
int ipv6_sock_mc_join(struct sock *sk, int ifindex, const struct in6_addr *addr)
|
|
|
{
|
|
|
struct net_device *dev = NULL;
|
|
@@ -134,15 +139,15 @@ int ipv6_sock_mc_join(struct sock *sk, int ifindex, const struct in6_addr *addr)
|
|
|
if (!ipv6_addr_is_multicast(addr))
|
|
|
return -EINVAL;
|
|
|
|
|
|
- read_lock_bh(&ipv6_sk_mc_lock);
|
|
|
- for (mc_lst=np->ipv6_mc_list; mc_lst; mc_lst=mc_lst->next) {
|
|
|
+ rcu_read_lock();
|
|
|
+ for_each_pmc_rcu(np, mc_lst) {
|
|
|
if ((ifindex == 0 || mc_lst->ifindex == ifindex) &&
|
|
|
ipv6_addr_equal(&mc_lst->addr, addr)) {
|
|
|
- read_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ rcu_read_unlock();
|
|
|
return -EADDRINUSE;
|
|
|
}
|
|
|
}
|
|
|
- read_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ rcu_read_unlock();
|
|
|
|
|
|
mc_lst = sock_kmalloc(sk, sizeof(struct ipv6_mc_socklist), GFP_KERNEL);
|
|
|
|
|
@@ -186,33 +191,41 @@ int ipv6_sock_mc_join(struct sock *sk, int ifindex, const struct in6_addr *addr)
|
|
|
return err;
|
|
|
}
|
|
|
|
|
|
- write_lock_bh(&ipv6_sk_mc_lock);
|
|
|
+ spin_lock(&ipv6_sk_mc_lock);
|
|
|
mc_lst->next = np->ipv6_mc_list;
|
|
|
- np->ipv6_mc_list = mc_lst;
|
|
|
- write_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ rcu_assign_pointer(np->ipv6_mc_list, mc_lst);
|
|
|
+ spin_unlock(&ipv6_sk_mc_lock);
|
|
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
+static void ipv6_mc_socklist_reclaim(struct rcu_head *head)
|
|
|
+{
|
|
|
+ kfree(container_of(head, struct ipv6_mc_socklist, rcu));
|
|
|
+}
|
|
|
/*
|
|
|
* socket leave on multicast group
|
|
|
*/
|
|
|
int ipv6_sock_mc_drop(struct sock *sk, int ifindex, const struct in6_addr *addr)
|
|
|
{
|
|
|
struct ipv6_pinfo *np = inet6_sk(sk);
|
|
|
- struct ipv6_mc_socklist *mc_lst, **lnk;
|
|
|
+ struct ipv6_mc_socklist *mc_lst;
|
|
|
+ struct ipv6_mc_socklist __rcu **lnk;
|
|
|
struct net *net = sock_net(sk);
|
|
|
|
|
|
- write_lock_bh(&ipv6_sk_mc_lock);
|
|
|
- for (lnk = &np->ipv6_mc_list; (mc_lst = *lnk) !=NULL ; lnk = &mc_lst->next) {
|
|
|
+ spin_lock(&ipv6_sk_mc_lock);
|
|
|
+ for (lnk = &np->ipv6_mc_list;
|
|
|
+ (mc_lst = rcu_dereference_protected(*lnk,
|
|
|
+ lockdep_is_held(&ipv6_sk_mc_lock))) !=NULL ;
|
|
|
+ lnk = &mc_lst->next) {
|
|
|
if ((ifindex == 0 || mc_lst->ifindex == ifindex) &&
|
|
|
ipv6_addr_equal(&mc_lst->addr, addr)) {
|
|
|
struct net_device *dev;
|
|
|
|
|
|
*lnk = mc_lst->next;
|
|
|
- write_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ spin_unlock(&ipv6_sk_mc_lock);
|
|
|
|
|
|
rcu_read_lock();
|
|
|
dev = dev_get_by_index_rcu(net, mc_lst->ifindex);
|
|
@@ -225,11 +238,12 @@ int ipv6_sock_mc_drop(struct sock *sk, int ifindex, const struct in6_addr *addr)
|
|
|
} else
|
|
|
(void) ip6_mc_leave_src(sk, mc_lst, NULL);
|
|
|
rcu_read_unlock();
|
|
|
- sock_kfree_s(sk, mc_lst, sizeof(*mc_lst));
|
|
|
+ atomic_sub(sizeof(*mc_lst), &sk->sk_omem_alloc);
|
|
|
+ call_rcu(&mc_lst->rcu, ipv6_mc_socklist_reclaim);
|
|
|
return 0;
|
|
|
}
|
|
|
}
|
|
|
- write_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ spin_unlock(&ipv6_sk_mc_lock);
|
|
|
|
|
|
return -EADDRNOTAVAIL;
|
|
|
}
|
|
@@ -272,12 +286,13 @@ void ipv6_sock_mc_close(struct sock *sk)
|
|
|
struct ipv6_mc_socklist *mc_lst;
|
|
|
struct net *net = sock_net(sk);
|
|
|
|
|
|
- write_lock_bh(&ipv6_sk_mc_lock);
|
|
|
- while ((mc_lst = np->ipv6_mc_list) != NULL) {
|
|
|
+ spin_lock(&ipv6_sk_mc_lock);
|
|
|
+ while ((mc_lst = rcu_dereference_protected(np->ipv6_mc_list,
|
|
|
+ lockdep_is_held(&ipv6_sk_mc_lock))) != NULL) {
|
|
|
struct net_device *dev;
|
|
|
|
|
|
np->ipv6_mc_list = mc_lst->next;
|
|
|
- write_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ spin_unlock(&ipv6_sk_mc_lock);
|
|
|
|
|
|
rcu_read_lock();
|
|
|
dev = dev_get_by_index_rcu(net, mc_lst->ifindex);
|
|
@@ -290,11 +305,13 @@ void ipv6_sock_mc_close(struct sock *sk)
|
|
|
} else
|
|
|
(void) ip6_mc_leave_src(sk, mc_lst, NULL);
|
|
|
rcu_read_unlock();
|
|
|
- sock_kfree_s(sk, mc_lst, sizeof(*mc_lst));
|
|
|
|
|
|
- write_lock_bh(&ipv6_sk_mc_lock);
|
|
|
+ atomic_sub(sizeof(*mc_lst), &sk->sk_omem_alloc);
|
|
|
+ call_rcu(&mc_lst->rcu, ipv6_mc_socklist_reclaim);
|
|
|
+
|
|
|
+ spin_lock(&ipv6_sk_mc_lock);
|
|
|
}
|
|
|
- write_unlock_bh(&ipv6_sk_mc_lock);
|
|
|
+ spin_unlock(&ipv6_sk_mc_lock);
|
|
|
}
|
|
|
|
|
|
int ip6_mc_source(int add, int omode, struct sock *sk,
|
|
@@ -328,8 +345,7 @@ int ip6_mc_source(int add, int omode, struct sock *sk,
|
|
|
|
|
|
err = -EADDRNOTAVAIL;
|
|
|
|
|
|
- read_lock(&ipv6_sk_mc_lock);
|
|
|
- for (pmc=inet6->ipv6_mc_list; pmc; pmc=pmc->next) {
|
|
|
+ for_each_pmc_rcu(inet6, pmc) {
|
|
|
if (pgsr->gsr_interface && pmc->ifindex != pgsr->gsr_interface)
|
|
|
continue;
|
|
|
if (ipv6_addr_equal(&pmc->addr, group))
|
|
@@ -428,7 +444,6 @@ int ip6_mc_source(int add, int omode, struct sock *sk,
|
|
|
done:
|
|
|
if (pmclocked)
|
|
|
write_unlock(&pmc->sflock);
|
|
|
- read_unlock(&ipv6_sk_mc_lock);
|
|
|
read_unlock_bh(&idev->lock);
|
|
|
rcu_read_unlock();
|
|
|
if (leavegroup)
|
|
@@ -466,14 +481,13 @@ int ip6_mc_msfilter(struct sock *sk, struct group_filter *gsf)
|
|
|
dev = idev->dev;
|
|
|
|
|
|
err = 0;
|
|
|
- read_lock(&ipv6_sk_mc_lock);
|
|
|
|
|
|
if (gsf->gf_fmode == MCAST_INCLUDE && gsf->gf_numsrc == 0) {
|
|
|
leavegroup = 1;
|
|
|
goto done;
|
|
|
}
|
|
|
|
|
|
- for (pmc=inet6->ipv6_mc_list; pmc; pmc=pmc->next) {
|
|
|
+ for_each_pmc_rcu(inet6, pmc) {
|
|
|
if (pmc->ifindex != gsf->gf_interface)
|
|
|
continue;
|
|
|
if (ipv6_addr_equal(&pmc->addr, group))
|
|
@@ -521,7 +535,6 @@ int ip6_mc_msfilter(struct sock *sk, struct group_filter *gsf)
|
|
|
write_unlock(&pmc->sflock);
|
|
|
err = 0;
|
|
|
done:
|
|
|
- read_unlock(&ipv6_sk_mc_lock);
|
|
|
read_unlock_bh(&idev->lock);
|
|
|
rcu_read_unlock();
|
|
|
if (leavegroup)
|
|
@@ -562,7 +575,7 @@ int ip6_mc_msfget(struct sock *sk, struct group_filter *gsf,
|
|
|
* so reading the list is safe.
|
|
|
*/
|
|
|
|
|
|
- for (pmc=inet6->ipv6_mc_list; pmc; pmc=pmc->next) {
|
|
|
+ for_each_pmc_rcu(inet6, pmc) {
|
|
|
if (pmc->ifindex != gsf->gf_interface)
|
|
|
continue;
|
|
|
if (ipv6_addr_equal(group, &pmc->addr))
|
|
@@ -612,13 +625,13 @@ int inet6_mc_check(struct sock *sk, const struct in6_addr *mc_addr,
|
|
|
struct ip6_sf_socklist *psl;
|
|
|
int rv = 1;
|
|
|
|
|
|
- read_lock(&ipv6_sk_mc_lock);
|
|
|
- for (mc = np->ipv6_mc_list; mc; mc = mc->next) {
|
|
|
+ rcu_read_lock();
|
|
|
+ for_each_pmc_rcu(np, mc) {
|
|
|
if (ipv6_addr_equal(&mc->addr, mc_addr))
|
|
|
break;
|
|
|
}
|
|
|
if (!mc) {
|
|
|
- read_unlock(&ipv6_sk_mc_lock);
|
|
|
+ rcu_read_unlock();
|
|
|
return 1;
|
|
|
}
|
|
|
read_lock(&mc->sflock);
|
|
@@ -638,7 +651,7 @@ int inet6_mc_check(struct sock *sk, const struct in6_addr *mc_addr,
|
|
|
rv = 0;
|
|
|
}
|
|
|
read_unlock(&mc->sflock);
|
|
|
- read_unlock(&ipv6_sk_mc_lock);
|
|
|
+ rcu_read_unlock();
|
|
|
|
|
|
return rv;
|
|
|
}
|