struct net_device *dev = pmc->interface->dev;
struct igmpv3_report *pih;
struct igmpv3_grec *pgr = 0;
- struct ip_sf_list *psf, *psf_next, *psf_prev, *psf_list;
+ struct ip_sf_list *psf, *psf_next, *psf_prev, **psf_list;
int scount, first, isquery, truncate;
if (pmc->multiaddr == IGMP_ALL_HOSTS)
truncate = type == IGMPV3_MODE_IS_EXCLUDE ||
type == IGMPV3_CHANGE_TO_EXCLUDE;
- psf_list = sdeleted ? pmc->tomb : pmc->sources;
+ psf_list = sdeleted ? &pmc->tomb : &pmc->sources;
- if (!psf_list) {
+ if (!*psf_list) {
if (type == IGMPV3_ALLOW_NEW_SOURCES ||
type == IGMPV3_BLOCK_OLD_SOURCES)
return skb;
first = 1;
scount = 0;
psf_prev = 0;
- for (psf=psf_list; psf; psf=psf_next) {
+ for (psf=*psf_list; psf; psf=psf_next) {
u32 *psrc;
psf_next = psf->sf_next;
if (psf_prev)
psf_prev->sf_next = psf->sf_next;
else
- pmc->tomb = psf->sf_next;
+ *psf_list = psf->sf_next;
kfree(psf);
continue;
}
if (ndisc_mc_map(&mc->mca_addr, buf, dev, 0) == 0)
dev_mc_delete(dev, buf, dev->addr_len, 0);
}
- spin_unlock_bh(&mc->mca_lock);
if (mc->mca_flags & MAF_NOREPORT)
goto done;
+ spin_unlock_bh(&mc->mca_lock);
if (dev->flags&IFF_UP)
igmp6_leave_group(mc);
spin_lock_bh(&mc->mca_lock);
if (del_timer(&mc->mca_timer))
atomic_dec(&mc->mca_refcnt);
- spin_unlock_bh(&mc->mca_lock);
-
done:
ip6_mc_clear_src(mc);
+ spin_unlock_bh(&mc->mca_lock);
}
/*
struct net_device *dev = pmc->idev->dev;
struct mld2_report *pmr;
struct mld2_grec *pgr = 0;
- struct ip6_sf_list *psf, *psf_next, *psf_prev, *psf_list;
+ struct ip6_sf_list *psf, *psf_next, *psf_prev, **psf_list;
int scount, first, isquery, truncate;
if (pmc->mca_flags & MAF_NOREPORT)
truncate = type == MLD2_MODE_IS_EXCLUDE ||
type == MLD2_CHANGE_TO_EXCLUDE;
- psf_list = sdeleted ? pmc->mca_tomb : pmc->mca_sources;
+ psf_list = sdeleted ? &pmc->mca_tomb : &pmc->mca_sources;
- if (!psf_list) {
+ if (!*psf_list) {
if (type == MLD2_ALLOW_NEW_SOURCES ||
type == MLD2_BLOCK_OLD_SOURCES)
return skb;
first = 1;
scount = 0;
psf_prev = 0;
- for (psf=psf_list; psf; psf=psf_next) {
+ for (psf=*psf_list; psf; psf=psf_next) {
struct in6_addr *psrc;
psf_next = psf->sf_next;
if (psf_prev)
psf_prev->sf_next = psf->sf_next;
else
- pmc->mca_tomb = psf->sf_next;
+ *psf_list = psf->sf_next;
kfree(psf);
continue;
}
return -ESRCH;
}
spin_lock_bh(&pmc->mca_lock);
- read_unlock_bh(&idev->lock);
sf_markstate(pmc);
if (!delta) {
if (!pmc->mca_sfcount[sfmode]) {
spin_unlock_bh(&pmc->mca_lock);
+ read_unlock_bh(&idev->lock);
return -EINVAL;
}
pmc->mca_sfcount[sfmode]--;
} else if (sf_setstate(pmc) || changerec)
mld_ifc_event(pmc->idev);
spin_unlock_bh(&pmc->mca_lock);
+ read_unlock_bh(&idev->lock);
return err;
}
return -ESRCH;
}
spin_lock_bh(&pmc->mca_lock);
- read_unlock_bh(&idev->lock);
sf_markstate(pmc);
isexclude = pmc->mca_sfmode == MCAST_EXCLUDE;
} else if (sf_setstate(pmc))
mld_ifc_event(idev);
spin_unlock_bh(&pmc->mca_lock);
+ read_unlock_bh(&idev->lock);
return err;
}