On Fri, Jun 05, 2015 at 02:08:05PM -0300, mleit...@redhat.com wrote: > From: Marcelo Ricardo Leitner <marcelo.leit...@gmail.com> > > That's needed for the next patch, so we break the lock inversion between > netns_sctp->addr_wq_lock and socket lock on > sctp_addr_wq_timeout_handler(). With this, we can traverse addr_waitq > without taking addr_wq_lock, taking it just for the write operations. > > Signed-off-by: Marcelo Ricardo Leitner <marcelo.leit...@gmail.com> > --- > > Notes: > v2->v3: > placed break statement on sctp_free_addr_wq_entry() > removed unnecessary spin_lock noticed by Neil > > include/net/netns/sctp.h | 2 +- > net/sctp/protocol.c | 80 > +++++++++++++++++++++++++++++------------------- > 2 files changed, 49 insertions(+), 33 deletions(-) > > diff --git a/include/net/netns/sctp.h b/include/net/netns/sctp.h > index > 3573a81815ad9e0efb6ceb721eb066d3726419f0..9e53412c4ed829e8e45777a6d95406d490dbaa75 > 100644 > --- a/include/net/netns/sctp.h > +++ b/include/net/netns/sctp.h > @@ -28,7 +28,7 @@ struct netns_sctp { > * It is a list of sctp_sockaddr_entry. > */ > struct list_head local_addr_list; > - struct list_head addr_waitq; > + struct list_head __rcu addr_waitq; > struct timer_list addr_wq_timer; > struct list_head auto_asconf_splist; > spinlock_t addr_wq_lock; > diff --git a/net/sctp/protocol.c b/net/sctp/protocol.c > index > 53b7acde9aa37bf3d4029c459421564d5270f4c0..9954fb8c9a9455d5ad7a627e2d7f9a1fef861fc2 > 100644 > --- a/net/sctp/protocol.c > +++ b/net/sctp/protocol.c > @@ -593,15 +593,47 @@ static void sctp_v4_ecn_capable(struct sock *sk) > INET_ECN_xmit(sk); > } > > +static void sctp_free_addr_wq(struct net *net) > +{ > + struct sctp_sockaddr_entry *addrw; > + > + spin_lock_bh(&net->sctp.addr_wq_lock); > + del_timer(&net->sctp.addr_wq_timer); > + list_for_each_entry_rcu(addrw, &net->sctp.addr_waitq, list) { > + list_del_rcu(&addrw->list); > + kfree_rcu(addrw, rcu); > + } > + spin_unlock_bh(&net->sctp.addr_wq_lock); > +} > + > +/* As there is no refcnt on sctp_sockaddr_entry, we must check inside > + * the lock if it wasn't removed from addr_waitq already, otherwise we > + * could double-free it. > + */ > +static void sctp_free_addr_wq_entry(struct net *net, > + struct sctp_sockaddr_entry *addrw) > +{ > + struct sctp_sockaddr_entry *temp; > + > + spin_lock_bh(&net->sctp.addr_wq_lock); > + list_for_each_entry_rcu(temp, &net->sctp.addr_waitq, list) { > + if (temp == addrw) { > + list_del_rcu(&addrw->list); > + kfree_rcu(addrw, rcu); > + break; > + } > + } > + spin_unlock_bh(&net->sctp.addr_wq_lock); > +} > + > static void sctp_addr_wq_timeout_handler(unsigned long arg) > { > struct net *net = (struct net *)arg; > - struct sctp_sockaddr_entry *addrw, *temp; > + struct sctp_sockaddr_entry *addrw; > struct sctp_sock *sp; > > - spin_lock_bh(&net->sctp.addr_wq_lock); > - > - list_for_each_entry_safe(addrw, temp, &net->sctp.addr_waitq, list) { > + rcu_read_lock_bh(); > + list_for_each_entry_rcu(addrw, &net->sctp.addr_waitq, list) { > pr_debug("%s: the first ent in wq:%p is addr:%pISc for cmd:%d > at " > "entry:%p\n", __func__, &net->sctp.addr_waitq, > &addrw->a.sa, > addrw->state, addrw); > @@ -647,35 +679,20 @@ static void sctp_addr_wq_timeout_handler(unsigned long > arg) > #if IS_ENABLED(CONFIG_IPV6) > free_next: > #endif > - list_del(&addrw->list); > - kfree(addrw); > - } > - spin_unlock_bh(&net->sctp.addr_wq_lock); > -} > - > -static void sctp_free_addr_wq(struct net *net) > -{ > - struct sctp_sockaddr_entry *addrw; > - struct sctp_sockaddr_entry *temp; > - > - spin_lock_bh(&net->sctp.addr_wq_lock); > - del_timer(&net->sctp.addr_wq_timer); > - list_for_each_entry_safe(addrw, temp, &net->sctp.addr_waitq, list) { > - list_del(&addrw->list); > - kfree(addrw); > + sctp_free_addr_wq_entry(net, addrw); > } > - spin_unlock_bh(&net->sctp.addr_wq_lock); > + rcu_read_unlock_bh(); > } > > /* lookup the entry for the same address in the addr_waitq > - * sctp_addr_wq MUST be locked > + * rcu read MUST be locked > */ > static struct sctp_sockaddr_entry *sctp_addr_wq_lookup(struct net *net, > struct sctp_sockaddr_entry *addr) > { > struct sctp_sockaddr_entry *addrw; > > - list_for_each_entry(addrw, &net->sctp.addr_waitq, list) { > + list_for_each_entry_rcu(addrw, &net->sctp.addr_waitq, list) { > if (addrw->a.sa.sa_family != addr->a.sa.sa_family) > continue; > if (addrw->a.sa.sa_family == AF_INET) { > @@ -702,7 +719,7 @@ void sctp_addr_wq_mgmt(struct net *net, struct > sctp_sockaddr_entry *addr, int cm > * new address after a couple of addition and deletion of that address > */ > > - spin_lock_bh(&net->sctp.addr_wq_lock); > + rcu_read_lock_bh(); > /* Offsets existing events in addr_wq */ > addrw = sctp_addr_wq_lookup(net, addr); > if (addrw) { > @@ -710,22 +727,21 @@ void sctp_addr_wq_mgmt(struct net *net, struct > sctp_sockaddr_entry *addr, int cm > pr_debug("%s: offsets existing entry for %d, addr:%pISc > " > "in wq:%p\n", __func__, addrw->state, > &addrw->a.sa, > &net->sctp.addr_waitq); > - > - list_del(&addrw->list); > - kfree(addrw); > + sctp_free_addr_wq_entry(net, addrw); > } > - spin_unlock_bh(&net->sctp.addr_wq_lock); > + rcu_read_unlock_bh(); > return; > } > + rcu_read_unlock_bh(); > > /* OK, we have to add the new address to the wait queue */ > addrw = kmemdup(addr, sizeof(struct sctp_sockaddr_entry), GFP_ATOMIC); > - if (addrw == NULL) { > - spin_unlock_bh(&net->sctp.addr_wq_lock); > + if (!addrw) > return; > - } > addrw->state = cmd; > - list_add_tail(&addrw->list, &net->sctp.addr_waitq); > + > + spin_lock_bh(&net->sctp.addr_wq_lock); > + list_add_tail_rcu(&addrw->list, &net->sctp.addr_waitq); > > pr_debug("%s: add new entry for cmd:%d, addr:%pISc in wq:%p\n", > __func__, addrw->state, &addrw->a.sa, &net->sctp.addr_waitq); > -- > 2.4.1 > > Acked-by: Neil Horman <nhor...@tuxdriver.com>
-- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html