File: [cvs.NetBSD.org] / src / sys / net / lagg / if_lagg.c (download)
Revision 1.41, Thu Mar 31 03:12:31 2022 UTC (2 years ago) by yamaguchi
Branch: MAIN
Changes since 1.40: +2 -3
lines
Make lagg interface specified "laggproto none" able to up
|
/* $NetBSD: if_lagg.c,v 1.41 2022/03/31 03:12:31 yamaguchi Exp $ */
/*
* Copyright (c) 2005, 2006 Reyk Floeter <reyk@openbsd.org>
* Copyright (c) 2007 Andrew Thompson <thompsa@FreeBSD.org>
* Copyright (c) 2014, 2016 Marcelo Araujo <araujo@FreeBSD.org>
* Copyright (c) 2021, Internet Initiative Japan Inc.
*
* Permission to use, copy, modify, and distribute this software for any
* purpose with or without fee is hereby granted, provided that the above
* copyright notice and this permission notice appear in all copies.
*
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
* ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
* ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
* OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
*/
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: if_lagg.c,v 1.41 2022/03/31 03:12:31 yamaguchi Exp $");
#ifdef _KERNEL_OPT
#include "opt_inet.h"
#include "opt_lagg.h"
#endif
#include <sys/param.h>
#include <sys/types.h>
#include <sys/cprng.h>
#include <sys/cpu.h>
#include <sys/device.h>
#include <sys/evcnt.h>
#include <sys/hash.h>
#include <sys/kmem.h>
#include <sys/module.h>
#include <sys/pserialize.h>
#include <sys/pslist.h>
#include <sys/psref.h>
#include <sys/sysctl.h>
#include <sys/syslog.h>
#include <sys/workqueue.h>
#include <net/bpf.h>
#include <net/if.h>
#include <net/if_dl.h>
#include <net/if_ether.h>
#include <net/if_media.h>
#include <net/if_types.h>
#include <net/if_vlanvar.h>
#include <netinet/ip.h>
#include <netinet/ip6.h>
#include <netinet/tcp.h>
#include <netinet/udp.h>
#if defined(INET) || defined(INET6)
#include <netinet/in.h>
#endif
#ifdef INET6
#include <netinet6/in6_ifattach.h>
#include <netinet6/in6_var.h>
#endif
#include <net/lagg/if_lagg.h>
#include <net/lagg/if_laggvar.h>
#include <net/lagg/if_laggproto.h>
#include "ioconf.h"
enum lagg_portctrl {
LAGG_PORTCTRL_ALLOC,
LAGG_PORTCTRL_FREE,
LAGG_PORTCTRL_START,
LAGG_PORTCTRL_STOP
};
enum lagg_iftypes {
LAGG_IF_TYPE_ETHERNET,
};
static const struct lagg_proto lagg_protos[] = {
[LAGG_PROTO_NONE] = {
.pr_num = LAGG_PROTO_NONE,
.pr_attach = lagg_none_attach,
},
[LAGG_PROTO_LACP] = {
.pr_num = LAGG_PROTO_LACP,
.pr_attach = lacp_attach,
.pr_detach = lacp_detach,
.pr_up = lacp_up,
.pr_down = lacp_down,
.pr_transmit = lacp_transmit,
.pr_input = lacp_input,
.pr_allocport = lacp_allocport,
.pr_freeport = lacp_freeport,
.pr_startport = lacp_startport,
.pr_stopport = lacp_stopport,
.pr_protostat = lacp_protostat,
.pr_portstat = lacp_portstat,
.pr_linkstate = lacp_linkstate_ifnet_locked,
.pr_ioctl = lacp_ioctl,
},
[LAGG_PROTO_FAILOVER] = {
.pr_num = LAGG_PROTO_FAILOVER,
.pr_attach = lagg_fail_attach,
.pr_detach = lagg_common_detach,
.pr_transmit = lagg_fail_transmit,
.pr_input = lagg_fail_input,
.pr_allocport = lagg_common_allocport,
.pr_freeport = lagg_common_freeport,
.pr_startport = lagg_common_startport,
.pr_stopport = lagg_common_stopport,
.pr_portstat = lagg_fail_portstat,
.pr_linkstate = lagg_common_linkstate,
.pr_ioctl = lagg_fail_ioctl,
},
[LAGG_PROTO_LOADBALANCE] = {
.pr_num = LAGG_PROTO_LOADBALANCE,
.pr_attach = lagg_lb_attach,
.pr_detach = lagg_common_detach,
.pr_transmit = lagg_lb_transmit,
.pr_input = lagg_lb_input,
.pr_allocport = lagg_common_allocport,
.pr_freeport = lagg_common_freeport,
.pr_startport = lagg_lb_startport,
.pr_stopport = lagg_lb_stopport,
.pr_portstat = lagg_lb_portstat,
.pr_linkstate = lagg_common_linkstate,
},
};
static int lagg_chg_sadl(struct ifnet *, uint8_t *, size_t);
static struct mbuf *
lagg_input_ethernet(struct ifnet *, struct mbuf *);
static int lagg_clone_create(struct if_clone *, int);
static int lagg_clone_destroy(struct ifnet *);
static int lagg_init(struct ifnet *);
static int lagg_init_locked(struct lagg_softc *);
static void lagg_stop(struct ifnet *, int);
static void lagg_stop_locked(struct lagg_softc *);
static int lagg_ioctl(struct ifnet *, u_long, void *);
static int lagg_transmit(struct ifnet *, struct mbuf *);
static void lagg_start(struct ifnet *);
static int lagg_media_change(struct ifnet *);
static void lagg_media_status(struct ifnet *, struct ifmediareq *);
static int lagg_vlan_cb(struct ethercom *, uint16_t, bool);
static void lagg_linkstate_changed(void *);
static void lagg_ifdetach(void *);
static struct lagg_softc *
lagg_softc_alloc(enum lagg_iftypes);
static void lagg_softc_free(struct lagg_softc *);
static int lagg_setup_sysctls(struct lagg_softc *);
static void lagg_teardown_sysctls(struct lagg_softc *);
static int lagg_proto_attach(struct lagg_softc *, lagg_proto,
struct lagg_proto_softc **);
static void lagg_proto_detach(struct lagg_variant *);
static int lagg_proto_up(struct lagg_softc *);
static void lagg_proto_down(struct lagg_softc *);
static int lagg_proto_allocport(struct lagg_softc *, struct lagg_port *);
static void lagg_proto_freeport(struct lagg_softc *, struct lagg_port *);
static void lagg_proto_startport(struct lagg_softc *,
struct lagg_port *);
static void lagg_proto_stopport(struct lagg_softc *,
struct lagg_port *);
static struct mbuf *
lagg_proto_input(struct lagg_softc *, struct lagg_port *,
struct mbuf *);
static void lagg_proto_linkstate(struct lagg_softc *, struct lagg_port *);
static int lagg_proto_ioctl(struct lagg_softc *, struct lagg_req *);
static int lagg_get_stats(struct lagg_softc *, struct lagg_req *, size_t);
static int lagg_pr_attach(struct lagg_softc *, lagg_proto);
static void lagg_pr_detach(struct lagg_softc *);
static int lagg_addport(struct lagg_softc *, struct ifnet *);
static int lagg_delport(struct lagg_softc *, struct ifnet *);
static int lagg_delport_all(struct lagg_softc *);
static int lagg_port_ioctl(struct ifnet *, u_long, void *);
static int lagg_port_output(struct ifnet *, struct mbuf *,
const struct sockaddr *, const struct rtentry *);
static void lagg_config_promisc(struct lagg_softc *, struct lagg_port *);
static void lagg_unconfig_promisc(struct lagg_softc *, struct lagg_port *);
static struct lagg_variant *
lagg_variant_getref(struct lagg_softc *, struct psref *);
static void lagg_variant_putref(struct lagg_variant *, struct psref *);
static int lagg_ether_addmulti(struct lagg_softc *, struct ifreq *);
static int lagg_ether_delmulti(struct lagg_softc *, struct ifreq *);
static void lagg_port_syncmulti(struct lagg_softc *, struct lagg_port *);
static void lagg_port_purgemulti(struct lagg_softc *, struct lagg_port *);
static int lagg_port_setup(struct lagg_softc *, struct lagg_port *,
struct ifnet *);
static void lagg_port_teardown(struct lagg_softc *, struct lagg_port *,
bool);
static void lagg_port_syncvlan(struct lagg_softc *, struct lagg_port *);
static void lagg_port_purgevlan(struct lagg_softc *, struct lagg_port *);
static void lagg_lladdr_update(struct lagg_softc *);
static void lagg_capabilities_update(struct lagg_softc *);
static void lagg_sync_ifcaps(struct lagg_softc *);
static void lagg_sync_ethcaps(struct lagg_softc *);
static struct if_clone lagg_cloner =
IF_CLONE_INITIALIZER("lagg", lagg_clone_create, lagg_clone_destroy);
static unsigned int lagg_count;
static struct psref_class
*lagg_psref_class __read_mostly;
static struct psref_class
*lagg_port_psref_class __read_mostly;
static enum lagg_iftypes
lagg_iftype = LAGG_IF_TYPE_ETHERNET;
#ifdef LAGG_DEBUG
#define LAGG_DPRINTF(_sc, _fmt, _args...) do { \
printf("%s: " _fmt, (_sc) != NULL ? \
(_sc)->sc_if.if_xname : "lagg", ##_args); \
} while (0)
#else
#define LAGG_DPRINTF(_sc, _fmt, _args...) __nothing
#endif
#ifndef LAGG_SETCAPS_RETRY
#define LAGG_SETCAPS_RETRY (LAGG_MAX_PORTS * 2)
#endif
static size_t
lagg_sizeof_softc(enum lagg_iftypes ift)
{
struct lagg_softc *_dummy = NULL;
size_t s;
s = sizeof(*_dummy) - sizeof(_dummy->sc_if);
switch (ift) {
case LAGG_IF_TYPE_ETHERNET:
s += sizeof(struct ethercom);
break;
default:
s += sizeof(struct ifnet);
break;
}
return s;
}
static bool
lagg_debug_enable(struct lagg_softc *sc)
{
if (__predict_false(ISSET(sc->sc_if.if_flags, IFF_DEBUG)))
return true;
return false;
}
static void
lagg_evcnt_attach(struct lagg_softc *sc,
struct evcnt *ev, const char *name)
{
evcnt_attach_dynamic(ev, EVCNT_TYPE_MISC, NULL,
sc->sc_evgroup, name);
}
static void
lagg_in6_ifattach(struct ifnet *ifp)
{
#ifdef INET6
KERNEL_LOCK_UNLESS_NET_MPSAFE();
if (in6_present) {
if (ISSET(ifp->if_flags, IFF_UP))
in6_ifattach(ifp, NULL);
}
KERNEL_UNLOCK_UNLESS_NET_MPSAFE();
#endif
}
static void
lagg_in6_ifdetach(struct ifnet *ifp)
{
#ifdef INET6
KERNEL_LOCK_UNLESS_NET_MPSAFE();
if (in6_present)
in6_ifdetach(ifp);
KERNEL_UNLOCK_UNLESS_NET_MPSAFE();
#endif
}
static int
lagg_lp_ioctl(struct lagg_port *lp, u_long cmd, void *data)
{
struct ifnet *ifp_port;
int error;
if (lp->lp_ioctl == NULL)
return EINVAL;
ifp_port = lp->lp_ifp;
IFNET_LOCK(ifp_port);
error = lp->lp_ioctl(ifp_port, cmd, data);
IFNET_UNLOCK(ifp_port);
return error;
}
static bool
lagg_lladdr_equal(const uint8_t *a, const uint8_t *b)
{
if (memcmp(a, b, ETHER_ADDR_LEN) == 0)
return true;
return false;
}
static void
lagg_lladdr_cpy(uint8_t *dst, const uint8_t *src)
{
memcpy(dst, src, ETHER_ADDR_LEN);
}
void
laggattach(int n)
{
/*
* Nothing to do here, initialization is handled by the
* module initialization code in lagginit() below).
*/
}
static void
lagginit(void)
{
size_t i;
lagg_psref_class = psref_class_create("laggvariant", IPL_SOFTNET);
lagg_port_psref_class = psref_class_create("laggport", IPL_SOFTNET);
for (i = 0; i < LAGG_PROTO_MAX; i++) {
if (lagg_protos[i].pr_init != NULL)
lagg_protos[i].pr_init();
}
lagg_input_ethernet_p = lagg_input_ethernet;
if_clone_attach(&lagg_cloner);
}
static int
laggdetach(void)
{
size_t i;
if (lagg_count > 0)
return EBUSY;
if_clone_detach(&lagg_cloner);
lagg_input_ethernet_p = NULL;
for (i = 0; i < LAGG_PROTO_MAX; i++) {
if (lagg_protos[i].pr_fini != NULL)
lagg_protos[i].pr_fini();
}
psref_class_destroy(lagg_port_psref_class);
psref_class_destroy(lagg_psref_class);
return 0;
}
static int
lagg_clone_create(struct if_clone *ifc, int unit)
{
struct lagg_softc *sc;
struct ifnet *ifp;
int error;
sc = lagg_softc_alloc(lagg_iftype);
ifp = &sc->sc_if;
mutex_init(&sc->sc_lock, MUTEX_DEFAULT, IPL_SOFTNET);
sc->sc_psz = pserialize_create();
SIMPLEQ_INIT(&sc->sc_ports);
LIST_INIT(&sc->sc_mclist);
TAILQ_INIT(&sc->sc_vtags);
sc->sc_hash_mac = true;
sc->sc_hash_ipaddr = true;
sc->sc_hash_ip6addr = true;
sc->sc_hash_tcp = true;
sc->sc_hash_udp = true;
if_initname(ifp, ifc->ifc_name, unit);
ifp->if_softc = sc;
ifp->if_init = lagg_init;
ifp->if_stop = lagg_stop;
ifp->if_ioctl = lagg_ioctl;
ifp->if_flags = IFF_SIMPLEX | IFF_BROADCAST | IFF_MULTICAST;
ifp->if_extflags = IFEF_MPSAFE;
ifp->if_transmit = lagg_transmit;
ifp->if_start = lagg_start;
IFQ_SET_READY(&ifp->if_snd);
error = lagg_setup_sysctls(sc);
if (error != 0)
goto destroy_psz;
/*XXX dependent on ethernet */
ifmedia_init_with_lock(&sc->sc_media, 0, lagg_media_change,
lagg_media_status, &sc->sc_lock);
ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL);
ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO);
if_initialize(ifp);
switch (lagg_iftype) {
case LAGG_IF_TYPE_ETHERNET:
cprng_fast(sc->sc_lladdr_rand, sizeof(sc->sc_lladdr_rand));
sc->sc_lladdr_rand[0] &= 0xFE; /* clear I/G bit */
sc->sc_lladdr_rand[0] |= 0x02; /* set G/L bit */
lagg_lladdr_cpy(sc->sc_lladdr, sc->sc_lladdr_rand);
ether_set_vlan_cb((struct ethercom *)ifp, lagg_vlan_cb);
ether_ifattach(ifp, sc->sc_lladdr);
break;
default:
panic("unknown if type");
}
snprintf(sc->sc_evgroup, sizeof(sc->sc_evgroup),
"%s", ifp->if_xname);
lagg_evcnt_attach(sc, &sc->sc_novar, "no lagg variant");
if_link_state_change(&sc->sc_if, LINK_STATE_DOWN);
lagg_setup_sysctls(sc);
(void)lagg_pr_attach(sc, LAGG_PROTO_NONE);
if_register(ifp);
lagg_count++;
return 0;
destroy_psz:
pserialize_destroy(sc->sc_psz);
mutex_destroy(&sc->sc_lock);
lagg_softc_free(sc);
return error;
}
static int
lagg_clone_destroy(struct ifnet *ifp)
{
struct lagg_softc *sc = (struct lagg_softc *)ifp->if_softc;
struct lagg_port *lp;
lagg_stop(ifp, 1);
LAGG_LOCK(sc);
while ((lp = LAGG_PORTS_FIRST(sc)) != NULL) {
lagg_port_teardown(sc, lp, false);
}
LAGG_UNLOCK(sc);
switch (ifp->if_type) {
case IFT_ETHER:
ether_ifdetach(ifp);
KASSERT(TAILQ_EMPTY(&sc->sc_vtags));
break;
}
if_detach(ifp);
ifmedia_fini(&sc->sc_media);
lagg_pr_detach(sc);
evcnt_detach(&sc->sc_novar);
lagg_teardown_sysctls(sc);
pserialize_destroy(sc->sc_psz);
mutex_destroy(&sc->sc_lock);
lagg_softc_free(sc);
if (lagg_count > 0)
lagg_count--;
return 0;
}
static int
lagg_init(struct ifnet *ifp)
{
struct lagg_softc *sc;
int rv;
sc = ifp->if_softc;
LAGG_LOCK(sc);
rv = lagg_init_locked(sc);
LAGG_UNLOCK(sc);
return rv;
}
static int
lagg_init_locked(struct lagg_softc *sc)
{
struct ifnet *ifp = &sc->sc_if;
int rv;
KASSERT(LAGG_LOCKED(sc));
if (ISSET(ifp->if_flags, IFF_RUNNING))
lagg_stop_locked(sc);
lagg_lladdr_update(sc);
SET(ifp->if_flags, IFF_RUNNING);
rv = lagg_proto_up(sc);
if (rv != 0)
lagg_stop_locked(sc);
return rv;
}
static void
lagg_stop(struct ifnet *ifp, int disable __unused)
{
struct lagg_softc *sc;
sc = ifp->if_softc;
LAGG_LOCK(sc);
lagg_stop_locked(sc);
LAGG_UNLOCK(sc);
}
static void
lagg_stop_locked(struct lagg_softc *sc)
{
struct ifnet *ifp = &sc->sc_if;
KASSERT(LAGG_LOCKED(sc));
if (!ISSET(ifp->if_flags, IFF_RUNNING))
return;
CLR(ifp->if_flags, IFF_RUNNING);
lagg_proto_down(sc);
}
static int
lagg_config(struct lagg_softc *sc, struct lagg_req *lrq)
{
struct ifnet *ifp_port;
struct laggreqport *rp;
struct lagg_port *lp;
struct psref psref;
size_t i;
int error, bound;
error = 0;
bound = curlwp_bind();
switch (lrq->lrq_ioctl) {
case LAGGIOC_SETPROTO:
if (lrq->lrq_proto >= LAGG_PROTO_MAX) {
error = EPROTONOSUPPORT;
break;
}
error = lagg_delport_all(sc);
if (error != 0)
break;
error = lagg_pr_attach(sc, lrq->lrq_proto);
if (error != 0)
break;
for (i = 0; i < lrq->lrq_nports; i++) {
rp = &lrq->lrq_reqports[i];
ifp_port = if_get(rp->rp_portname, &psref);
if (ifp_port == NULL) {
error = ENOENT;
break; /* break for */
}
error = lagg_addport(sc, ifp_port);
if_put(ifp_port, &psref);
if (error != 0)
break; /* break for */
}
break; /* break switch */
case LAGGIOC_ADDPORT:
rp = &lrq->lrq_reqports[0];
ifp_port = if_get(rp->rp_portname, &psref);
if (ifp_port == NULL) {
error = ENOENT;
break;
}
error = lagg_addport(sc, ifp_port);
if_put(ifp_port, &psref);
break;
case LAGGIOC_DELPORT:
rp = &lrq->lrq_reqports[0];
ifp_port = if_get(rp->rp_portname, &psref);
if (ifp_port == NULL) {
error = ENOENT;
break;
}
error = lagg_delport(sc, ifp_port);
if_put(ifp_port, &psref);
break;
case LAGGIOC_SETPORTPRI:
rp = &lrq->lrq_reqports[0];
ifp_port = if_get(rp->rp_portname, &psref);
if (ifp_port == NULL) {
error = ENOENT;
break;
}
lp = ifp_port->if_lagg;
if (lp == NULL || lp->lp_softc != sc) {
if_put(ifp_port, &psref);
error = ENOENT;
break;
}
lp->lp_prio = rp->rp_prio;
/* restart protocol */
LAGG_LOCK(sc);
lagg_proto_stopport(sc, lp);
lagg_proto_startport(sc, lp);
LAGG_UNLOCK(sc);
if_put(ifp_port, &psref);
break;
case LAGGIOC_SETPROTOOPT:
error = lagg_proto_ioctl(sc, lrq);
break;
default:
error = ENOTTY;
}
curlwp_bindx(bound);
return error;
}
static int
lagg_ioctl(struct ifnet *ifp, u_long cmd, void *data)
{
struct lagg_softc *sc;
struct ifreq *ifr = (struct ifreq *)data;
struct lagg_req laggreq, *laggresp;
struct lagg_port *lp;
size_t allocsiz, outlen, nports;
char *outbuf;
void *buf;
int error = 0, rv;
sc = ifp->if_softc;
switch (cmd) {
case SIOCGLAGG:
error = copyin(ifr->ifr_data, &laggreq, sizeof(laggreq));
if (error != 0)
break;
nports = sc->sc_nports;
nports = MIN(nports, laggreq.lrq_nports);
allocsiz = sizeof(*laggresp)
+ sizeof(laggresp->lrq_reqports[0]) * nports;
laggresp = kmem_zalloc(allocsiz, KM_SLEEP);
rv = lagg_get_stats(sc, laggresp, nports);
outbuf = (char *)laggresp;
nports = MIN(laggresp->lrq_nports, nports);
outlen = sizeof(*laggresp)
+ sizeof(laggresp->lrq_reqports[0]) * nports;
error = copyout(outbuf, ifr->ifr_data, outlen);
kmem_free(outbuf, allocsiz);
if (error == 0 && rv != 0)
error = rv;
break;
case SIOCSLAGG:
error = copyin(ifr->ifr_data, &laggreq, sizeof(laggreq));
if (error != 0)
break;
nports = laggreq.lrq_nports;
if (nports > 1) {
allocsiz = sizeof(struct lagg_req)
+ sizeof(struct laggreqport) * nports;
buf = kmem_alloc(allocsiz, KM_SLEEP);
error = copyin(ifr->ifr_data, buf, allocsiz);
if (error != 0) {
kmem_free(buf, allocsiz);
break;
}
} else {
buf = (void *)&laggreq;
allocsiz = 0;
}
error = lagg_config(sc, buf);
if (allocsiz > 0)
kmem_free(buf, allocsiz);
break;
case SIOCSIFFLAGS:
error = ifioctl_common(ifp, cmd, data);
if (error != 0)
break;
switch (ifp->if_flags & (IFF_UP | IFF_RUNNING)) {
case IFF_RUNNING:
if_stop(ifp, 1);
break;
case IFF_UP:
case IFF_UP | IFF_RUNNING:
error = if_init(ifp);
break;
}
if (error != 0)
break;
/* Set flags on ports too */
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
(void)lagg_config_promisc(sc, lp);
}
LAGG_UNLOCK(sc);
break;
case SIOCSIFMTU:
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
error = lagg_lp_ioctl(lp, cmd, (void *)ifr);
if (error != 0) {
lagg_log(sc, LOG_ERR,
"failed to change MTU to %d on port %s, "
"reverting all ports to original "
"MTU(%" PRIu64 ")\n",
ifr->ifr_mtu, lp->lp_ifp->if_xname,
ifp->if_mtu);
break;
}
}
if (error == 0) {
ifp->if_mtu = ifr->ifr_mtu;
} else {
/* set every port back to the original MTU */
ifr->ifr_mtu = ifp->if_mtu;
LAGG_PORTS_FOREACH(sc, lp) {
if (lp->lp_ioctl != NULL)
lagg_lp_ioctl(lp, cmd, (void *)ifr);
}
}
LAGG_UNLOCK(sc);
break;
case SIOCADDMULTI:
if (sc->sc_if.if_type == IFT_ETHER) {
error = lagg_ether_addmulti(sc, ifr);
} else {
error = EPROTONOSUPPORT;
}
break;
case SIOCDELMULTI:
if (sc->sc_if.if_type == IFT_ETHER) {
error = lagg_ether_delmulti(sc, ifr);
} else {
error = EPROTONOSUPPORT;
}
break;
case SIOCSIFCAP:
error = ether_ioctl(ifp, cmd, data);
if (error == 0)
lagg_sync_ifcaps(sc);
break;
case SIOCSETHERCAP:
error = ether_ioctl(ifp, cmd, data);
if (error == 0)
lagg_sync_ethcaps(sc);
break;
default:
error = ether_ioctl(ifp, cmd, data);
}
return error;
}
static int
lagg_setup_sysctls(struct lagg_softc *sc)
{
struct sysctllog **log;
const struct sysctlnode **rnode, *hashnode;
const char *ifname;
int error;
log = &sc->sc_sysctllog;
rnode = &sc->sc_sysctlnode;
ifname = sc->sc_if.if_xname;
error = sysctl_createv(log, 0, NULL, rnode,
CTLFLAG_PERMANENT, CTLTYPE_NODE, ifname,
SYSCTL_DESCR("lagg information and settings"),
NULL, 0, NULL, 0, CTL_NET, CTL_CREATE, CTL_EOL);
if (error != 0)
goto done;
error = sysctl_createv(log, 0, rnode, &hashnode,
CTLFLAG_PERMANENT, CTLTYPE_NODE, "hash",
SYSCTL_DESCR("hash calculation settings"),
NULL, 0, NULL, 0, CTL_CREATE, CTL_EOL);
if (error != 0)
goto done;
error = sysctl_createv(log, 0, &hashnode, NULL,
CTLFLAG_READWRITE, CTLTYPE_BOOL, "macaddr",
SYSCTL_DESCR("use src/dst mac addresses"),
NULL, 0, &sc->sc_hash_mac, 0, CTL_CREATE, CTL_EOL);
if (error != 0)
goto done;
error = sysctl_createv(log, 0, &hashnode, NULL,
CTLFLAG_READWRITE, CTLTYPE_BOOL, "ipaddr",
SYSCTL_DESCR("use src/dst IPv4 addresses"),
NULL, 0, &sc->sc_hash_ipaddr, 0, CTL_CREATE, CTL_EOL);
if (error != 0)
goto done;
error = sysctl_createv(log, 0, &hashnode, NULL,
CTLFLAG_READWRITE, CTLTYPE_BOOL, "ip6addr",
SYSCTL_DESCR("use src/dst IPv6 addresses"),
NULL, 0, &sc->sc_hash_ip6addr, 0, CTL_CREATE, CTL_EOL);
if (error != 0)
goto done;
error = sysctl_createv(log, 0, &hashnode, NULL,
CTLFLAG_READWRITE, CTLTYPE_BOOL, "tcp",
SYSCTL_DESCR("use TCP src/dst port"),
NULL, 0, &sc->sc_hash_tcp, 0, CTL_CREATE, CTL_EOL);
if (error != 0)
goto done;
error = sysctl_createv(log, 0, &hashnode, NULL,
CTLFLAG_READWRITE, CTLTYPE_BOOL, "udp",
SYSCTL_DESCR("use UDP src/dst port"),
NULL, 0, &sc->sc_hash_udp, 0, CTL_CREATE, CTL_EOL);
done:
if (error != 0) {
lagg_log(sc, LOG_ERR, "unable to create sysctl node\n");
sysctl_teardown(log);
}
return error;
}
static void
lagg_teardown_sysctls(struct lagg_softc *sc)
{
sc->sc_sysctlnode = NULL;
sysctl_teardown(&sc->sc_sysctllog);
}
uint32_t
lagg_hashmbuf(struct lagg_softc *sc, struct mbuf *m)
{
union {
struct ether_header _eh;
struct ether_vlan_header _evl;
struct ip _ip;
struct ip6_hdr _ip6;
struct tcphdr _th;
struct udphdr _uh;
} buf;
const struct ether_header *eh;
const struct ether_vlan_header *evl;
const struct ip *ip;
const struct ip6_hdr *ip6;
const struct tcphdr *th;
const struct udphdr *uh;
uint32_t hash, hash_src, hash_dst;
uint32_t flowlabel;
uint16_t etype, vlantag;
uint8_t proto;
size_t off;
KASSERT(ISSET(m->m_flags, M_PKTHDR));
hash = HASH32_BUF_INIT;
hash_src = HASH32_BUF_INIT;
hash_dst = HASH32_BUF_INIT;
#define LAGG_HASH_ADD(hp, v) do { \
*(hp) = hash32_buf(&(v), sizeof(v), *(hp)); \
} while(0)
eh = lagg_m_extract(m, 0, sizeof(*eh), &buf);
if (eh == NULL)
goto out;
off = ETHER_HDR_LEN;
etype = ntohs(eh->ether_type);
if (etype == ETHERTYPE_VLAN) {
evl = lagg_m_extract(m, 0, sizeof(*evl), &buf);
if (evl == NULL)
goto out;
vlantag = ntohs(evl->evl_tag);
etype = ntohs(evl->evl_proto);
off += ETHER_VLAN_ENCAP_LEN;
} else if (vlan_has_tag(m)) {
vlantag = vlan_get_tag(m);
} else {
vlantag = 0;
}
if (sc->sc_hash_mac) {
LAGG_HASH_ADD(&hash_dst, eh->ether_dhost);
LAGG_HASH_ADD(&hash_src, eh->ether_shost);
LAGG_HASH_ADD(&hash, vlantag);
}
switch (etype) {
case ETHERTYPE_IP:
ip = lagg_m_extract(m, off, sizeof(*ip), &buf);
if (ip == NULL)
goto out;
if (sc->sc_hash_ipaddr) {
LAGG_HASH_ADD(&hash_src, ip->ip_src);
LAGG_HASH_ADD(&hash_dst, ip->ip_dst);
LAGG_HASH_ADD(&hash, ip->ip_p);
}
off += ip->ip_hl << 2;
proto = ip->ip_p;
break;
case ETHERTYPE_IPV6:
ip6 = lagg_m_extract(m, off, sizeof(*ip6), &buf);
if (ip6 == NULL)
goto out;
if (sc->sc_hash_ip6addr) {
LAGG_HASH_ADD(&hash_src, ip6->ip6_src);
LAGG_HASH_ADD(&hash_dst, ip6->ip6_dst);
flowlabel = ip6->ip6_flow & IPV6_FLOWLABEL_MASK;
LAGG_HASH_ADD(&hash, flowlabel);
}
proto = ip6->ip6_nxt;
off += sizeof(*ip6);
break;
default:
return hash;
}
switch (proto) {
case IPPROTO_TCP:
th = lagg_m_extract(m, off, sizeof(*th), &buf);
if (th == NULL)
goto out;
if (sc->sc_hash_tcp) {
LAGG_HASH_ADD(&hash_src, th->th_sport);
LAGG_HASH_ADD(&hash_dst, th->th_dport);
}
break;
case IPPROTO_UDP:
uh = lagg_m_extract(m, off, sizeof(*uh), &buf);
if (uh == NULL)
goto out;
if (sc->sc_hash_udp) {
LAGG_HASH_ADD(&hash_src, uh->uh_sport);
LAGG_HASH_ADD(&hash_dst, uh->uh_dport);
}
break;
}
out:
hash_src ^= hash_dst;
LAGG_HASH_ADD(&hash, hash_src);
#undef LAGG_HASH_ADD
return hash;
}
static int
lagg_tx_common(struct ifnet *ifp, struct mbuf *m)
{
struct lagg_variant *var;
lagg_proto pr;
struct psref psref;
int error;
var = lagg_variant_getref(ifp->if_softc, &psref);
if (__predict_false(var == NULL)) {
m_freem(m);
if_statinc(ifp, if_oerrors);
return ENOENT;
}
pr = var->lv_proto;
if (__predict_true(lagg_protos[pr].pr_transmit != NULL)) {
error = lagg_protos[pr].pr_transmit(var->lv_psc, m);
/* mbuf is already freed */
} else {
m_freem(m);
if_statinc(ifp, if_oerrors);
error = ENOBUFS;
}
lagg_variant_putref(var, &psref);
return error;
}
static int
lagg_transmit(struct ifnet *ifp, struct mbuf *m)
{
return lagg_tx_common(ifp, m);
}
static void
lagg_start(struct ifnet *ifp)
{
struct mbuf *m;
for (;;) {
IFQ_DEQUEUE(&ifp->if_snd, m);
if (m == NULL)
break;
(void)lagg_tx_common(ifp, m);
}
}
void
lagg_enqueue(struct lagg_softc *sc, struct lagg_port *lp, struct mbuf *m)
{
struct ifnet *ifp;
int len, error;
short mflags;
ifp = &sc->sc_if;
len = m->m_pkthdr.len;
mflags = m->m_flags;
error = lagg_port_xmit(lp, m);
if (error) {
/* mbuf is already freed */
if_statinc(ifp, if_oerrors);
}
net_stat_ref_t nsr = IF_STAT_GETREF(ifp);
if_statinc_ref(nsr, if_opackets);
if_statadd_ref(nsr, if_obytes, len);
if (mflags & M_MCAST)
if_statinc_ref(nsr, if_omcasts);
IF_STAT_PUTREF(ifp);
}
static struct mbuf *
lagg_proto_input(struct lagg_softc *sc, struct lagg_port *lp, struct mbuf *m)
{
struct psref psref;
struct lagg_variant *var;
lagg_proto pr;
var = lagg_variant_getref(sc, &psref);
if (var == NULL) {
sc->sc_novar.ev_count++;
m_freem(m);
return NULL;
}
pr = var->lv_proto;
if (lagg_protos[pr].pr_input != NULL) {
m = lagg_protos[pr].pr_input(var->lv_psc, lp, m);
} else {
m_freem(m);
m = NULL;
}
lagg_variant_putref(var, &psref);
return m;
}
static struct mbuf *
lagg_input_ethernet(struct ifnet *ifp_port, struct mbuf *m)
{
struct ifnet *ifp;
struct psref psref;
struct lagg_port *lp;
int s;
/* sanity check */
s = pserialize_read_enter();
lp = atomic_load_consume(&ifp_port->if_lagg);
if (lp == NULL) {
/* This interface is not a member of lagg */
pserialize_read_exit(s);
return m;
}
lagg_port_getref(lp, &psref);
pserialize_read_exit(s);
ifp = &lp->lp_softc->sc_if;
/*
* Drop promiscuously received packets
* if we are not in promiscuous mode.
*/
if ((m->m_flags & (M_BCAST | M_MCAST)) == 0 &&
(ifp_port->if_flags & IFF_PROMISC) != 0 &&
(ifp->if_flags & IFF_PROMISC) == 0) {
struct ether_header *eh;
eh = mtod(m, struct ether_header *);
if (memcmp(CLLADDR(ifp->if_sadl),
eh->ether_dhost, ETHER_ADDR_LEN) != 0) {
m_freem(m);
m = NULL;
if_statinc(ifp, if_ierrors);
goto out;
}
}
if (pfil_run_hooks(ifp_port->if_pfil, &m,
ifp_port, PFIL_IN) != 0)
goto out;
m = lagg_proto_input(lp->lp_softc, lp, m);
if (m != NULL) {
m_set_rcvif(m, ifp);
m->m_flags &= ~M_PROMISC;
if_input(ifp, m);
m = NULL;
}
out:
lagg_port_putref(lp, &psref);
return m;
}
static int
lagg_media_change(struct ifnet *ifp)
{
if (ISSET(ifp->if_flags, IFF_DEBUG))
printf("%s: ignore media change\n", ifp->if_xname);
return 0;
}
static void
lagg_media_status(struct ifnet *ifp, struct ifmediareq *imr)
{
struct lagg_softc *sc;
struct lagg_port *lp;
sc = ifp->if_softc;
imr->ifm_status = IFM_AVALID;
imr->ifm_active = IFM_ETHER | IFM_AUTO;
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
if (lagg_portactive(lp))
imr->ifm_status |= IFM_ACTIVE;
}
LAGG_UNLOCK(sc);
}
static int
lagg_port_vlan_cb(struct lagg_port *lp,
struct lagg_vlantag *lvt, bool set)
{
struct ifnet *ifp_port;
int error;
if (lp->lp_iftype != IFT_ETHER)
return 0;
error = 0;
ifp_port = lp->lp_ifp;
if (set) {
error = ether_add_vlantag(ifp_port,
lvt->lvt_vtag, NULL);
} else {
error = ether_del_vlantag(ifp_port,
lvt->lvt_vtag);
}
return error;
}
static int
lagg_vlan_cb(struct ethercom *ec, uint16_t vtag, bool set)
{
struct ifnet *ifp;
struct lagg_softc *sc;
struct lagg_vlantag *lvt, *lvt0;
struct lagg_port *lp;
int error;
ifp = (struct ifnet *)ec;
sc = ifp->if_softc;
if (set) {
lvt = kmem_zalloc(sizeof(*lvt), KM_SLEEP);
lvt->lvt_vtag = vtag;
TAILQ_INSERT_TAIL(&sc->sc_vtags, lvt, lvt_entry);
} else {
TAILQ_FOREACH_SAFE(lvt, &sc->sc_vtags, lvt_entry, lvt0) {
if (lvt->lvt_vtag == vtag) {
TAILQ_REMOVE(&sc->sc_vtags, lvt, lvt_entry);
break;
}
}
if (lvt == NULL)
return ENOENT;
}
KASSERT(lvt != NULL);
LAGG_PORTS_FOREACH(sc, lp) {
error = lagg_port_vlan_cb(lp, lvt, set);
if (error != 0) {
lagg_log(sc, LOG_WARNING,
"%s failed to configure vlan on %d\n",
lp->lp_ifp->if_xname, error);
}
}
return 0;
}
static struct lagg_softc *
lagg_softc_alloc(enum lagg_iftypes ift)
{
struct lagg_softc *sc;
size_t s;
s = lagg_sizeof_softc(ift);
KASSERT(s > 0);
sc = kmem_zalloc(s, KM_SLEEP);
KASSERT(sc != NULL);
return sc;
}
static void
lagg_softc_free(struct lagg_softc *sc)
{
kmem_free(sc,
lagg_sizeof_softc(sc->sc_iftype));
}
static void
lagg_variant_update(struct lagg_softc *sc, struct lagg_variant *newvar)
{
struct lagg_variant *oldvar;
KASSERT(LAGG_LOCKED(sc));
psref_target_init(&newvar->lv_psref, lagg_psref_class);
oldvar = sc->sc_var;
atomic_store_release(&sc->sc_var, newvar);
pserialize_perform(sc->sc_psz);
if (__predict_true(oldvar != NULL))
psref_target_destroy(&oldvar->lv_psref, lagg_psref_class);
}
static struct lagg_variant *
lagg_variant_getref(struct lagg_softc *sc, struct psref *psref)
{
struct lagg_variant *var;
int s;
s = pserialize_read_enter();
var = atomic_load_consume(&sc->sc_var);
if (var == NULL) {
pserialize_read_exit(s);
return NULL;
}
psref_acquire(psref, &var->lv_psref, lagg_psref_class);
pserialize_read_exit(s);
return var;
}
static void
lagg_variant_putref(struct lagg_variant *var, struct psref *psref)
{
if (__predict_false(var == NULL))
return;
psref_release(psref, &var->lv_psref, lagg_psref_class);
}
static int
lagg_proto_attach(struct lagg_softc *sc, lagg_proto pr,
struct lagg_proto_softc **psc)
{
KASSERT(lagg_protos[pr].pr_attach != NULL);
return lagg_protos[pr].pr_attach(sc, psc);
}
static void
lagg_proto_detach(struct lagg_variant *oldvar)
{
lagg_proto pr;
pr = oldvar->lv_proto;
if (lagg_protos[pr].pr_detach == NULL)
return;
lagg_protos[pr].pr_detach(oldvar->lv_psc);
}
static int
lagg_proto_updown(struct lagg_softc *sc, bool is_up)
{
struct lagg_variant *var;
struct psref psref;
lagg_proto pr;
int error, bound;
error = 0;
bound = curlwp_bind();
var = lagg_variant_getref(sc, &psref);
if (var == NULL) {
curlwp_bindx(bound);
return ENXIO;
}
pr = var->lv_proto;
if (is_up && lagg_protos[pr].pr_up != NULL) {
error = lagg_protos[pr].pr_up(var->lv_psc);
} else if (!is_up && lagg_protos[pr].pr_down != NULL) {
lagg_protos[pr].pr_down(var->lv_psc);
}
lagg_variant_putref(var, &psref);
curlwp_bindx(bound);
return error;
}
static int
lagg_proto_up(struct lagg_softc *sc)
{
return lagg_proto_updown(sc, true);
}
static void
lagg_proto_down(struct lagg_softc *sc)
{
(void)lagg_proto_updown(sc, false);
}
static int
lagg_proto_portctrl(struct lagg_softc *sc, struct lagg_port *lp,
enum lagg_portctrl ctrl)
{
struct lagg_variant *var;
struct psref psref;
lagg_proto pr;
int error, bound;
error = 0;
bound = curlwp_bind();
var = lagg_variant_getref(sc, &psref);
if (var == NULL) {
curlwp_bindx(bound);
return ENXIO;
}
pr = var->lv_proto;
switch (ctrl) {
case LAGG_PORTCTRL_ALLOC:
if (lagg_protos[pr].pr_allocport == NULL) {
goto nosupport;
}
error = lagg_protos[pr].pr_allocport(var->lv_psc, lp);
break;
case LAGG_PORTCTRL_FREE:
if (lagg_protos[pr].pr_freeport == NULL) {
goto nosupport;
}
lagg_protos[pr].pr_freeport(var->lv_psc, lp);
break;
case LAGG_PORTCTRL_START:
if (lagg_protos[pr].pr_startport == NULL) {
goto nosupport;
}
lagg_protos[pr].pr_startport(var->lv_psc, lp);
break;
case LAGG_PORTCTRL_STOP:
if (lagg_protos[pr].pr_stopport == NULL) {
goto nosupport;
}
lagg_protos[pr].pr_stopport(var->lv_psc, lp);
break;
default:
goto nosupport;
}
lagg_variant_putref(var, &psref);
curlwp_bindx(bound);
return error;
nosupport:
lagg_variant_putref(var, &psref);
curlwp_bindx(bound);
return EPROTONOSUPPORT;
}
static int
lagg_proto_allocport(struct lagg_softc *sc, struct lagg_port *lp)
{
return lagg_proto_portctrl(sc, lp, LAGG_PORTCTRL_ALLOC);
}
static void
lagg_proto_freeport(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_proto_portctrl(sc, lp, LAGG_PORTCTRL_FREE);
}
static void
lagg_proto_startport(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_proto_portctrl(sc, lp, LAGG_PORTCTRL_START);
}
static void
lagg_proto_stopport(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_proto_portctrl(sc, lp, LAGG_PORTCTRL_STOP);
}
static void
lagg_proto_linkstate(struct lagg_softc *sc, struct lagg_port *lp)
{
struct lagg_variant *var;
struct psref psref;
lagg_proto pr;
int bound;
KASSERT(IFNET_LOCKED(lp->lp_ifp));
bound = curlwp_bind();
var = lagg_variant_getref(sc, &psref);
if (var == NULL) {
curlwp_bindx(bound);
return;
}
pr = var->lv_proto;
if (lagg_protos[pr].pr_linkstate)
lagg_protos[pr].pr_linkstate(var->lv_psc, lp);
lagg_variant_putref(var, &psref);
curlwp_bindx(bound);
}
static void
lagg_proto_stat(struct lagg_variant *var, struct laggreqproto *resp)
{
lagg_proto pr;
pr = var->lv_proto;
if (lagg_protos[pr].pr_protostat != NULL)
lagg_protos[pr].pr_protostat(var->lv_psc, resp);
}
static void
lagg_proto_portstat(struct lagg_variant *var, struct lagg_port *lp,
struct laggreqport *resp)
{
lagg_proto pr;
pr = var->lv_proto;
if (lagg_protos[pr].pr_portstat != NULL)
lagg_protos[pr].pr_portstat(var->lv_psc, lp, resp);
}
static int
lagg_proto_ioctl(struct lagg_softc *sc, struct lagg_req *lreq)
{
struct lagg_variant *var;
struct psref psref;
lagg_proto pr;
int bound, error;
error = ENOTTY;
bound = curlwp_bind();
var = lagg_variant_getref(sc, &psref);
if (var == NULL) {
error = ENXIO;
goto done;
}
pr = var->lv_proto;
if (pr != lreq->lrq_proto) {
error = EBUSY;
goto done;
}
if (lagg_protos[pr].pr_ioctl != NULL) {
error = lagg_protos[pr].pr_ioctl(var->lv_psc,
&lreq->lrq_reqproto);
}
done:
if (var != NULL)
lagg_variant_putref(var, &psref);
curlwp_bindx(bound);
return error;
}
static int
lagg_pr_attach(struct lagg_softc *sc, lagg_proto pr)
{
struct lagg_variant *newvar, *oldvar;
struct lagg_proto_softc *psc;
bool cleanup_oldvar;
int error;
error = 0;
cleanup_oldvar = false;
newvar = kmem_alloc(sizeof(*newvar), KM_SLEEP);
LAGG_LOCK(sc);
oldvar = sc->sc_var;
if (oldvar != NULL && oldvar->lv_proto == pr) {
error = 0;
goto done;
}
error = lagg_proto_attach(sc, pr, &psc);
if (error != 0)
goto done;
newvar->lv_proto = pr;
newvar->lv_psc = psc;
lagg_variant_update(sc, newvar);
newvar = NULL;
if (oldvar != NULL) {
lagg_proto_detach(oldvar);
cleanup_oldvar = true;
}
done:
LAGG_UNLOCK(sc);
if (newvar != NULL)
kmem_free(newvar, sizeof(*newvar));
if (cleanup_oldvar)
kmem_free(oldvar, sizeof(*oldvar));
return error;
}
static void
lagg_pr_detach(struct lagg_softc *sc)
{
struct lagg_variant *var;
LAGG_LOCK(sc);
var = sc->sc_var;
atomic_store_release(&sc->sc_var, NULL);
pserialize_perform(sc->sc_psz);
if (var != NULL)
lagg_proto_detach(var);
LAGG_UNLOCK(sc);
if (var != NULL)
kmem_free(var, sizeof(*var));
}
static int
lagg_ether_addmulti(struct lagg_softc *sc, struct ifreq *ifr)
{
struct lagg_port *lp;
struct lagg_mc_entry *mc;
struct ethercom *ec;
const struct sockaddr *sa;
uint8_t addrlo[ETHER_ADDR_LEN], addrhi[ETHER_ADDR_LEN];
int error;
if (sc->sc_if.if_type != IFT_ETHER)
return EPROTONOSUPPORT;
ec = (struct ethercom *)&sc->sc_if;
sa = ifreq_getaddr(SIOCADDMULTI, ifr);
error = ether_addmulti(sa, ec);
if (error != ENETRESET)
return error;
error = ether_multiaddr(sa, addrlo, addrhi);
KASSERT(error == 0);
mc = kmem_zalloc(sizeof(*mc), KM_SLEEP);
ETHER_LOCK(ec);
mc->mc_enm = ether_lookup_multi(addrlo, addrhi, ec);
ETHER_UNLOCK(ec);
KASSERT(mc->mc_enm != NULL);
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
(void)lagg_lp_ioctl(lp, SIOCADDMULTI, (void *)ifr);
}
LAGG_UNLOCK(sc);
KASSERT(sa->sa_len <= sizeof(mc->mc_addr));
memcpy(&mc->mc_addr, sa, sa->sa_len);
LIST_INSERT_HEAD(&sc->sc_mclist, mc, mc_entry);
return 0;
}
static int
lagg_ether_delmulti(struct lagg_softc *sc, struct ifreq *ifr)
{
struct lagg_port *lp;
struct lagg_mc_entry *mc;
const struct sockaddr *sa;
struct ethercom *ec;
struct ether_multi *enm;
uint8_t addrlo[ETHER_ADDR_LEN], addrhi[ETHER_ADDR_LEN];
int error;
ec = (struct ethercom *)&sc->sc_if;
sa = ifreq_getaddr(SIOCDELMULTI, ifr);
error = ether_multiaddr(sa, addrlo, addrhi);
if (error != 0)
return error;
ETHER_LOCK(ec);
enm = ether_lookup_multi(addrlo, addrhi, ec);
ETHER_UNLOCK(ec);
if (enm == NULL)
return ENOENT;
LIST_FOREACH(mc, &sc->sc_mclist, mc_entry) {
if (mc->mc_enm == enm)
break;
}
if (mc == NULL)
return ENOENT;
error = ether_delmulti(sa, ec);
if (error != ENETRESET)
return error;
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
(void)lagg_lp_ioctl(lp, SIOCDELMULTI, (void *)ifr);
}
LAGG_UNLOCK(sc);
LIST_REMOVE(mc, mc_entry);
kmem_free(mc, sizeof(*mc));
return 0;
}
static void
lagg_port_multi(struct lagg_softc *sc, struct lagg_port *lp,
u_long cmd)
{
struct lagg_mc_entry *mc;
struct ifreq ifr;
struct ifnet *ifp_port;
const struct sockaddr *sa;
ifp_port = lp->lp_ifp;
memset(&ifr, 0, sizeof(ifr));
strlcpy(ifr.ifr_name, ifp_port->if_xname, sizeof(ifr.ifr_name));
LIST_FOREACH(mc, &sc->sc_mclist, mc_entry) {
sa = (struct sockaddr *)&mc->mc_addr;
KASSERT(sizeof(ifr.ifr_space) >= sa->sa_len);
memcpy(&ifr.ifr_addr, sa, sa->sa_len);
(void)lagg_lp_ioctl(lp, cmd, (void *)&ifr);
}
}
static void
lagg_port_syncmulti(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_port_multi(sc, lp, SIOCADDMULTI);
}
static void
lagg_port_purgemulti(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_port_multi(sc, lp, SIOCDELMULTI);
}
static void
lagg_port_vlan(struct lagg_softc *sc, struct lagg_port *lp,
bool set)
{
struct lagg_vlantag *lvt;
int error;
TAILQ_FOREACH(lvt, &sc->sc_vtags, lvt_entry) {
error = lagg_port_vlan_cb(lp, lvt, set);
if (error != 0) {
lagg_log(sc, LOG_WARNING,
"%s failed to configure vlan on %d\n",
lp->lp_ifp->if_xname, error);
}
}
}
static void
lagg_port_syncvlan(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_port_vlan(sc, lp, true);
}
static void
lagg_port_purgevlan(struct lagg_softc *sc, struct lagg_port *lp)
{
lagg_port_vlan(sc, lp, false);
}
static int
lagg_setifcaps(struct lagg_port *lp, uint64_t cap)
{
struct ifcapreq ifcr;
int error;
if (lp->lp_ifp->if_capenable == cap)
return 0;
memset(&ifcr, 0, sizeof(ifcr));
ifcr.ifcr_capenable = cap;
IFNET_LOCK(lp->lp_ifp);
error = LAGG_PORT_IOCTL(lp, SIOCSIFCAP, &ifcr);
IFNET_UNLOCK(lp->lp_ifp);
return error;
}
static void
lagg_sync_ifcaps(struct lagg_softc *sc)
{
struct lagg_port *lp;
struct ifnet *ifp;
int error = 0;
ifp = (struct ifnet *)&sc->sc_if;
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
error = lagg_setifcaps(lp, ifp->if_capenable);
if (error != 0) {
lagg_log(sc, LOG_WARNING,
"failed to update capabilities "
"of %s, error=%d",
lp->lp_ifp->if_xname, error);
}
}
LAGG_UNLOCK(sc);
}
static int
lagg_setethcaps(struct lagg_port *lp, int cap)
{
struct ethercom *ec;
struct eccapreq eccr;
int error;
KASSERT(lp->lp_iftype == IFT_ETHER);
ec = (struct ethercom *)lp->lp_ifp;
if (ec->ec_capenable == cap)
return 0;
memset(&eccr, 0, sizeof(eccr));
eccr.eccr_capenable = cap;
IFNET_LOCK(lp->lp_ifp);
error = LAGG_PORT_IOCTL(lp, SIOCSETHERCAP, &eccr);
IFNET_UNLOCK(lp->lp_ifp);
return error;
}
static void
lagg_sync_ethcaps(struct lagg_softc *sc)
{
struct ethercom *ec;
struct lagg_port *lp;
int error;
ec = (struct ethercom *)&sc->sc_if;
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
if (lp->lp_iftype != IFT_ETHER)
continue;
error = lagg_setethcaps(lp, ec->ec_capenable);
if (error != 0) {
lagg_log(sc, LOG_WARNING,
"failed to update ether "
"capabilities"" of %s, error=%d",
lp->lp_ifp->if_xname, error);
}
}
LAGG_UNLOCK(sc);
}
static void
lagg_ifcap_update(struct lagg_softc *sc)
{
struct ifnet *ifp;
struct lagg_port *lp;
uint64_t cap, ena, pena;
size_t i;
KASSERT(LAGG_LOCKED(sc));
/* Get common capabilities for the lagg ports */
ena = ~(uint64_t)0;
cap = ~(uint64_t)0;
LAGG_PORTS_FOREACH(sc, lp) {
ena &= lp->lp_ifp->if_capenable;
cap &= lp->lp_ifp->if_capabilities;
}
if (ena == ~(uint64_t)0)
ena = 0;
if (cap == ~(uint64_t)0)
cap = 0;
/*
* Apply common enabled capabilities back to the lagg ports.
* May require several iterations if they are dependent.
*/
for (i = 0; i < LAGG_SETCAPS_RETRY; i++) {
pena = ena;
LAGG_PORTS_FOREACH(sc, lp) {
lagg_setifcaps(lp, ena);
ena &= lp->lp_ifp->if_capenable;
}
if (pena == ena)
break;
}
if (pena != ena) {
lagg_log(sc, LOG_DEBUG, "couldn't set "
"capabilities 0x%08"PRIx64, pena);
}
ifp = &sc->sc_if;
if (ifp->if_capabilities != cap ||
ifp->if_capenable != ena) {
ifp->if_capabilities = cap;
ifp->if_capenable = ena;
lagg_log(sc, LOG_DEBUG,"capabilities "
"0x%08"PRIx64" enabled 0x%08"PRIx64,
cap, ena);
}
}
static void
lagg_ethercap_update(struct lagg_softc *sc)
{
struct ethercom *ec;
struct lagg_port *lp;
int cap, ena, pena;
size_t i;
KASSERT(LAGG_LOCKED(sc));
if (sc->sc_if.if_type != IFT_ETHER)
return;
/* Get common enabled capabilities for the lagg ports */
ena = ~0;
cap = ~0;
LAGG_PORTS_FOREACH(sc, lp) {
switch (lp->lp_iftype) {
case IFT_ETHER:
ec = (struct ethercom *)lp->lp_ifp;
ena &= ec->ec_capenable;
cap &= ec->ec_capabilities;
break;
case IFT_L2TP:
ena &= (ETHERCAP_VLAN_MTU | ETHERCAP_JUMBO_MTU);
cap &= (ETHERCAP_VLAN_MTU | ETHERCAP_JUMBO_MTU);
break;
default:
ena = 0;
cap = 0;
}
}
if (ena == ~0)
ena = 0;
if (cap == ~0)
cap = 0;
/*
* Apply common enabled capabilities back to the lagg ports.
* May require several iterations if they are dependent.
*/
for (i = 0; i < LAGG_SETCAPS_RETRY; i++) {
pena = ena;
LAGG_PORTS_FOREACH(sc, lp) {
if (lp->lp_iftype != IFT_ETHER)
continue;
ec = (struct ethercom *)lp->lp_ifp;
lagg_setethcaps(lp, ena);
ena &= ec->ec_capenable;
}
if (pena == ena)
break;
}
if (pena != ena) {
lagg_log(sc, LOG_DEBUG, "couldn't set "
"ether capabilities 0x%08x", pena);
}
ec = (struct ethercom *)&sc->sc_if;
if (ec->ec_capabilities != cap ||
ec->ec_capenable != ena) {
ec->ec_capabilities = cap;
ec->ec_capenable = ena;
lagg_log(sc, LOG_DEBUG,
"ether capabilities 0x%08x"
" enabled 0x%08x", cap, ena);
}
}
static void
lagg_capabilities_update(struct lagg_softc *sc)
{
lagg_ifcap_update(sc);
lagg_ethercap_update(sc);
}
static int
lagg_setup_mtu(struct lagg_softc *sc, struct lagg_port *lp)
{
struct ifnet *ifp_port;
struct ifreq ifr;
int error;
ifp_port = lp->lp_ifp;
KASSERT(IFNET_LOCKED(ifp_port));
error = 0;
memset(&ifr, 0, sizeof(ifr));
if (SIMPLEQ_EMPTY(&sc->sc_ports)) {
ifr.ifr_mtu = lp->lp_mtu;
} else {
ifr.ifr_mtu = sc->sc_if.if_mtu;
}
if (sc->sc_if.if_mtu != (uint64_t)ifr.ifr_mtu)
sc->sc_if.if_mtu = ifr.ifr_mtu;
if (lp->lp_mtu != (uint64_t)ifr.ifr_mtu) {
if (lp->lp_ioctl == NULL) {
LAGG_DPRINTF(sc, "cannot change MTU for %s\n",
ifp_port->if_xname);
return EINVAL;
}
strlcpy(ifr.ifr_name, ifp_port->if_xname, sizeof(ifr.ifr_name));
error = lp->lp_ioctl(ifp_port, SIOCSIFMTU, (void *)&ifr);
if (error != 0) {
LAGG_DPRINTF(sc, "invalid MTU %d for %s\n",
ifr.ifr_mtu, ifp_port->if_xname);
return error;
}
}
return 0;
}
static void
lagg_teardown_mtu(struct lagg_softc *sc, struct lagg_port *lp)
{
struct ifnet *ifp_port;
struct ifreq ifr;
int error;
if (lp->lp_ioctl == NULL)
return;
ifp_port = lp->lp_ifp;
KASSERT(IFNET_LOCKED(ifp_port));
if (SIMPLEQ_EMPTY(&sc->sc_ports))
sc->sc_if.if_mtu = 0;
if (ifp_port->if_mtu != lp->lp_mtu) {
memset(&ifr, 0, sizeof(ifr));
strlcpy(ifr.ifr_name, ifp_port->if_xname, sizeof(ifr.ifr_name));
ifr.ifr_mtu = lp->lp_mtu;
error = lp->lp_ioctl(ifp_port, SIOCSIFMTU, (void *)&ifr);
if (error != 0) {
lagg_log(sc, LOG_WARNING,
"failed to reset MTU %d to %s\n",
ifr.ifr_mtu, ifp_port->if_xname);
}
}
}
static void
lagg_port_setsadl(struct lagg_port *lp, uint8_t *lladdr,
bool iftype_changed)
{
struct ifnet *ifp_port;
bool lladdr_changed;
int error;
ifp_port = lp->lp_ifp;
KASSERT(LAGG_LOCKED(lp->lp_softc));
KASSERT(IFNET_LOCKED(ifp_port));
switch (lp->lp_iftype) {
case IFT_ETHER:
lladdr_changed = lagg_lladdr_equal(lladdr,
CLLADDR(ifp_port->if_sadl)) ? false : true;
if (lladdr_changed == false &&
iftype_changed == false) {
break;
}
lagg_chg_sadl(ifp_port,
lladdr, ETHER_ADDR_LEN);
if (ifp_port->if_init != NULL) {
error = 0;
if (ISSET(ifp_port->if_flags, IFF_RUNNING))
error = if_init(ifp_port);
if (error != 0) {
lagg_log(lp->lp_softc, LOG_WARNING,
"%s failed to if_init() on %d\n",
ifp_port->if_xname, error);
}
} else {
if (lp->lp_promisc == false) {
ifpromisc_locked(ifp_port, 1);
lp->lp_promisc = true;
}
}
break;
default:
if_alloc_sadl(ifp_port);
break;
}
}
static void
lagg_port_unsetsadl(struct lagg_port *lp)
{
struct ifnet *ifp_port;
int error;
ifp_port = lp->lp_ifp;
KASSERT(LAGG_LOCKED(lp->lp_softc));
KASSERT(IFNET_LOCKED(ifp_port));
switch (lp->lp_iftype) {
case IFT_ETHER:
/* reset if_type before changing ifp->if_sadl */
ifp_port->if_type = lp->lp_iftype;
lagg_chg_sadl(ifp_port,
lp->lp_lladdr, ETHER_ADDR_LEN);
if (ifp_port->if_init != NULL) {
error = 0;
if (ISSET(ifp_port->if_flags, IFF_RUNNING))
error = if_init(ifp_port);
if (error != 0) {
lagg_log(lp->lp_softc, LOG_WARNING,
"%s failed to if_init() on %d\n",
ifp_port->if_xname, error);
}
} else {
if (lp->lp_promisc == true) {
ifpromisc_locked(ifp_port, 0);
lp->lp_promisc = false;
}
}
break;
default:
/* reset if_type before if_alloc_sadl */
ifp_port->if_type = lp->lp_iftype;
if_alloc_sadl(ifp_port);
break;
}
}
static void
lagg_setup_lladdr(struct lagg_softc *sc, struct lagg_port *lp)
{
KASSERT(LAGG_LOCKED(sc));
if (lagg_lladdr_equal(sc->sc_lladdr, sc->sc_lladdr_rand))
lagg_lladdr_cpy(sc->sc_lladdr, lp->lp_lladdr);
}
static void
lagg_teardown_lladdr(struct lagg_softc *sc, struct lagg_port *lp)
{
struct lagg_port *lp0;
uint8_t *lladdr_next;
KASSERT(LAGG_LOCKED(sc));
if (lagg_lladdr_equal(sc->sc_lladdr,
lp->lp_lladdr) == false) {
return;
}
lladdr_next = sc->sc_lladdr_rand;
LAGG_PORTS_FOREACH(sc, lp0) {
if (lp0->lp_iftype == IFT_ETHER) {
lladdr_next = lp0->lp_lladdr;
break;
}
}
lagg_lladdr_cpy(sc->sc_lladdr, lladdr_next);
}
static void
lagg_lladdr_update(struct lagg_softc *sc)
{
struct ifnet *ifp;
struct lagg_port *lp;
const uint8_t *lladdr;
ifp = &sc->sc_if;
KASSERT(LAGG_LOCKED(sc));
KASSERT(IFNET_LOCKED(ifp));
KASSERT(!ISSET(ifp->if_flags, IFF_RUNNING));
lladdr = CLLADDR(ifp->if_sadl);
if (lagg_lladdr_equal(sc->sc_lladdr, lladdr))
return;
lagg_lladdr_cpy(sc->sc_lladdr, lladdr);
LAGG_PORTS_FOREACH(sc, lp) {
IFNET_LOCK(lp->lp_ifp);
lagg_port_setsadl(lp, sc->sc_lladdr, false);
IFNET_UNLOCK(lp->lp_ifp);
}
}
static void
lagg_sadl_update(struct lagg_softc *sc, uint8_t *lladdr_prev)
{
struct ifnet *ifp;
struct lagg_port *lp;
const uint8_t *lladdr;
ifp = &sc->sc_if;
KASSERT(LAGG_LOCKED(sc));
KASSERT(IFNET_LOCKED(ifp));
lladdr = CLLADDR(ifp->if_sadl);
if (lagg_lladdr_equal(sc->sc_lladdr, lladdr))
return;
if (lagg_lladdr_equal(lladdr_prev, lladdr) == false)
return;
lagg_chg_sadl(ifp, sc->sc_lladdr, ETHER_ADDR_LEN);
LAGG_PORTS_FOREACH(sc, lp) {
IFNET_LOCK(lp->lp_ifp);
lagg_port_setsadl(lp, sc->sc_lladdr, false);
IFNET_UNLOCK(lp->lp_ifp);
}
LAGG_UNLOCK(sc);
lagg_in6_ifdetach(ifp);
lagg_in6_ifattach(ifp);
LAGG_LOCK(sc);
}
static int
lagg_port_setup(struct lagg_softc *sc,
struct lagg_port *lp, struct ifnet *ifp_port)
{
u_char if_type;
int error;
bool stopped, iftype_changed;
KASSERT(LAGG_LOCKED(sc));
IFNET_ASSERT_UNLOCKED(ifp_port);
if (&sc->sc_if == ifp_port) {
LAGG_DPRINTF(sc, "cannot add a lagg to itself as a port\n");
return EINVAL;
}
if (sc->sc_nports > LAGG_MAX_PORTS)
return ENOSPC;
if (ifp_port->if_lagg != NULL) {
lp = (struct lagg_port *)ifp_port->if_lagg;
if (lp->lp_softc == sc)
return EEXIST;
return EBUSY;
}
switch (ifp_port->if_type) {
case IFT_ETHER:
case IFT_L2TP:
if_type = IFT_IEEE8023ADLAG;
break;
default:
return ENOTSUP;
}
error = 0;
stopped = false;
lp->lp_softc = sc;
lp->lp_prio = LAGG_PORT_PRIO;
lp->lp_linkstate_hook = if_linkstate_change_establish(ifp_port,
lagg_linkstate_changed, ifp_port);
lp->lp_ifdetach_hook = ether_ifdetachhook_establish(ifp_port,
lagg_ifdetach, ifp_port);
psref_target_init(&lp->lp_psref, lagg_port_psref_class);
IFNET_LOCK(ifp_port);
lp->lp_iftype = ifp_port->if_type;
lp->lp_ioctl = ifp_port->if_ioctl;
lp->lp_output = ifp_port->if_output;
lp->lp_ifcapenable = ifp_port->if_capenable;
lp->lp_mtu = ifp_port->if_mtu;
if (lp->lp_iftype == IFT_ETHER) {
struct ethercom *ec;
ec = (struct ethercom *)ifp_port;
lagg_lladdr_cpy(lp->lp_lladdr, CLLADDR(ifp_port->if_sadl));
lp->lp_eccapenable = ec->ec_capenable;
}
ifp_port->if_type = if_type;
ifp_port->if_ioctl = lagg_port_ioctl;
iftype_changed = (lp->lp_iftype != ifp_port->if_type);
if (ISSET(ifp_port->if_flags, IFF_RUNNING) &&
ifp_port->if_init != NULL) {
if_stop(ifp_port, 0);
stopped = true;
}
/* to delete ipv6 link local address */
lagg_in6_ifdetach(ifp_port);
error = lagg_setup_mtu(sc, lp);
if (error != 0)
goto restore_ipv6lla;
if (lp->lp_iftype == IFT_ETHER)
lagg_setup_lladdr(sc, lp);
lagg_port_setsadl(lp, sc->sc_lladdr, iftype_changed);
IFNET_UNLOCK(ifp_port);
error = lagg_proto_allocport(sc, lp);
if (error != 0)
goto teardown_lladdr;
atomic_store_release(&ifp_port->if_lagg, (void *)lp);
SIMPLEQ_INSERT_TAIL(&sc->sc_ports, lp, lp_entry);
sc->sc_nports++;
lagg_port_syncmulti(sc, lp);
lagg_port_syncvlan(sc, lp);
IFNET_LOCK(ifp_port);
ifp_port->if_output = lagg_port_output;
if (stopped) {
if (!ISSET(ifp_port->if_flags, IFF_RUNNING)) {
error = if_init(ifp_port);
if (error != 0)
goto remove_port;
}
}
IFNET_UNLOCK(ifp_port);
lagg_config_promisc(sc, lp);
lagg_proto_startport(sc, lp);
lagg_capabilities_update(sc);
return 0;
remove_port:
SIMPLEQ_REMOVE(&sc->sc_ports, lp, lagg_port, lp_entry);
sc->sc_nports--;
IFNET_LOCK(ifp_port);
ifp_port->if_output = lp->lp_output;
IFNET_UNLOCK(ifp_port);
atomic_store_release(&ifp_port->if_lagg, NULL);
pserialize_perform(sc->sc_psz);
lagg_port_purgemulti(sc, lp);
lagg_port_purgevlan(sc, lp);
lagg_proto_freeport(sc, lp);
teardown_lladdr:
IFNET_LOCK(ifp_port);
lagg_teardown_mtu(sc, lp);
lagg_port_unsetsadl(lp);
if (lp->lp_iftype == IFT_ETHER)
lagg_teardown_lladdr(sc, lp);
restore_ipv6lla:
KASSERT(IFNET_LOCKED(ifp_port));
lagg_in6_ifdetach(ifp_port);
if (stopped) {
if (if_init(ifp_port) != 0) {
lagg_log(sc, LOG_WARNING,
"couldn't re-start port %s\n",
ifp_port->if_xname);
}
}
ifp_port->if_type = lp->lp_iftype;
if (ifp_port->if_ioctl == lagg_port_ioctl)
ifp_port->if_ioctl = lp->lp_ioctl;
IFNET_UNLOCK(ifp_port);
psref_target_destroy(&lp->lp_psref, lagg_port_psref_class);
if_linkstate_change_disestablish(ifp_port,
lp->lp_linkstate_hook, NULL);
ether_ifdetachhook_disestablish(ifp_port,
lp->lp_ifdetach_hook, &sc->sc_lock);
return error;
}
static void
lagg_port_teardown(struct lagg_softc *sc, struct lagg_port *lp,
bool is_ifdetach)
{
struct ifnet *ifp_port;
bool stopped;
KASSERT(LAGG_LOCKED(sc));
ifp_port = lp->lp_ifp;
stopped = false;
ether_ifdetachhook_disestablish(ifp_port,
lp->lp_ifdetach_hook, &sc->sc_lock);
if (ifp_port->if_lagg == NULL) {
/* already done in lagg_ifdetach() */
return;
}
lagg_proto_stopport(sc, lp);
IFNET_LOCK(ifp_port);
if (ISSET(ifp_port->if_flags, IFF_RUNNING) &&
ifp_port->if_init != NULL) {
if_stop(ifp_port, 0);
stopped = true;
}
ifp_port->if_output = lp->lp_output;
IFNET_UNLOCK(ifp_port);
SIMPLEQ_REMOVE(&sc->sc_ports, lp, lagg_port, lp_entry);
sc->sc_nports--;
atomic_store_release(&ifp_port->if_lagg, NULL);
pserialize_perform(sc->sc_psz);
if_linkstate_change_disestablish(ifp_port,
lp->lp_linkstate_hook, NULL);
psref_target_destroy(&lp->lp_psref, lagg_port_psref_class);
lagg_port_purgemulti(sc, lp);
lagg_port_purgevlan(sc, lp);
lagg_teardown_lladdr(sc, lp);
IFNET_LOCK(ifp_port);
ifp_port->if_type = lp->lp_iftype;
if (ifp_port->if_ioctl == lagg_port_ioctl)
ifp_port->if_ioctl = lp->lp_ioctl;
lagg_teardown_mtu(sc, lp);
if (stopped)
if_init(ifp_port);
IFNET_UNLOCK(ifp_port);
if (is_ifdetach == false) {
lagg_unconfig_promisc(sc, lp);
lagg_setifcaps(lp, lp->lp_ifcapenable);
if (lp->lp_iftype == IFT_ETHER)
lagg_setethcaps(lp, lp->lp_eccapenable);
IFNET_LOCK(ifp_port);
lagg_port_unsetsadl(lp);
lagg_in6_ifattach(ifp_port);
IFNET_UNLOCK(ifp_port);
}
lagg_capabilities_update(sc);
lagg_proto_freeport(sc, lp);
kmem_free(lp, sizeof(*lp));
}
static int
lagg_addport(struct lagg_softc *sc, struct ifnet *ifp_port)
{
struct lagg_port *lp;
uint8_t lladdr[ETHER_ADDR_LEN];
int error;
lp = kmem_zalloc(sizeof(*lp), KM_SLEEP);
lp->lp_ifp = ifp_port;
LAGG_LOCK(sc);
lagg_lladdr_cpy(lladdr, sc->sc_lladdr);
error = lagg_port_setup(sc, lp, ifp_port);
if (error == 0)
lagg_sadl_update(sc, lladdr);
LAGG_UNLOCK(sc);
if (error != 0)
kmem_free(lp, sizeof(*lp));
return error;
}
static int
lagg_delport(struct lagg_softc *sc, struct ifnet *ifp_port)
{
struct lagg_port *lp;
uint8_t lladdr[ETHER_ADDR_LEN];
int error;
KASSERT(IFNET_LOCKED(&sc->sc_if));
error = 0;
LAGG_LOCK(sc);
lp = ifp_port->if_lagg;
if (lp == NULL || lp->lp_softc != sc) {
error = ENOENT;
goto out;
}
if (lp->lp_ifdetaching) {
error = EBUSY;
goto out;
}
lagg_lladdr_cpy(lladdr, sc->sc_lladdr);
lagg_port_teardown(sc, lp, false);
lagg_sadl_update(sc, lladdr);
out:
LAGG_UNLOCK(sc);
return error;
}
static int
lagg_delport_all(struct lagg_softc *sc)
{
struct lagg_port *lp;
uint8_t lladdr[ETHER_ADDR_LEN];
int error;
KASSERT(IFNET_LOCKED(&sc->sc_if));
error = 0;
LAGG_LOCK(sc);
lagg_lladdr_cpy(lladdr, sc->sc_lladdr);
while ((lp = LAGG_PORTS_FIRST(sc)) != NULL) {
if (lp->lp_ifdetaching) {
error = EBUSY;
continue;
}
lagg_port_teardown(sc, lp, false);
}
lagg_sadl_update(sc, lladdr);
LAGG_UNLOCK(sc);
return error;
}
static int
lagg_get_stats(struct lagg_softc *sc, struct lagg_req *resp,
size_t nports)
{
struct lagg_variant *var;
struct lagg_port *lp;
struct laggreqport *port;
struct psref psref;
struct ifnet *ifp;
int bound;
size_t n;
bound = curlwp_bind();
var = lagg_variant_getref(sc, &psref);
if (var == NULL) {
curlwp_bindx(bound);
return ENOENT;
}
resp->lrq_proto = var->lv_proto;
lagg_proto_stat(var, &resp->lrq_reqproto);
n = 0;
LAGG_LOCK(sc);
LAGG_PORTS_FOREACH(sc, lp) {
if (n < nports) {
port = &resp->lrq_reqports[n];
ifp = lp->lp_ifp;
strlcpy(port->rp_portname, ifp->if_xname,
sizeof(port->rp_portname));
port->rp_prio = lp->lp_prio;
port->rp_flags = lp->lp_flags;
lagg_proto_portstat(var, lp, port);
}
n++;
}
LAGG_UNLOCK(sc);
resp->lrq_nports = n;
lagg_variant_putref(var, &psref);
curlwp_bindx(bound);
if (resp->lrq_nports > nports) {
return ENOBUFS;
}
return 0;
}
static void
lagg_config_promisc(struct lagg_softc *sc, struct lagg_port *lp)
{
struct ifnet *ifp, *ifp_port;
int error;
bool promisc;
KASSERT(LAGG_LOCKED(sc));
ifp = &sc->sc_if;
ifp_port = lp->lp_ifp;
if (lp->lp_iftype == IFT_ETHER) {
promisc = ISSET(ifp->if_flags, IFF_PROMISC) ?
true : false;
} else {
promisc = true;
}
if (lp->lp_promisc == promisc)
return;
error = ifpromisc(ifp_port, promisc ? 1 : 0);
if (error == ENETRESET) {
error = ifp_port->if_init(ifp_port);
}
if (error == 0) {
lp->lp_promisc = promisc;
} else {
lagg_log(sc, LOG_WARNING,
"couldn't %s promisc on %s\n",
promisc ? "set" : "unset",
ifp_port->if_xname);
}
}
static void
lagg_unconfig_promisc(struct lagg_softc *sc, struct lagg_port *lp)
{
struct ifnet *ifp_port;
int error;
KASSERT(LAGG_LOCKED(sc));
ifp_port = lp->lp_ifp;
if (lp->lp_promisc == false)
return;
error = ifpromisc(ifp_port, 0);
if (error == ENETRESET) {
error = ifp_port->if_init(ifp_port);
}
if (error != 0) {
lagg_log(sc, LOG_WARNING,
"couldn't unset promisc on %s\n",
ifp_port->if_xname);
}
}
static int
lagg_port_ioctl(struct ifnet *ifp, u_long cmd, void *data)
{
struct lagg_softc *sc;
struct lagg_port *lp;
int error = 0;
u_int ifflags;
if ((lp = ifp->if_lagg) == NULL ||
(sc = lp->lp_softc) == NULL) {
goto fallback;
}
KASSERT(IFNET_LOCKED(lp->lp_ifp));
switch (cmd) {
case SIOCSIFCAP:
case SIOCSIFMTU:
case SIOCSETHERCAP:
/* Do not allow the setting to be cahanged once joined */
error = EINVAL;
break;
case SIOCSIFFLAGS:
ifflags = ifp->if_flags;
error = LAGG_PORT_IOCTL(lp, cmd, data);
ifflags ^= ifp->if_flags;
if ((ifflags & (IFF_UP | IFF_RUNNING)) != 0)
lagg_proto_linkstate(sc, lp);
break;
default:
goto fallback;
}
return error;
fallback:
if (lp != NULL) {
error = LAGG_PORT_IOCTL(lp, cmd, data);
} else {
error = ENOTTY;
}
return error;
}
static int
lagg_port_output(struct ifnet *ifp, struct mbuf *m,
const struct sockaddr *dst, const struct rtentry *rt)
{
struct lagg_port *lp = ifp->if_lagg;
int error = 0;
switch (dst->sa_family) {
case pseudo_AF_HDRCMPLT:
case AF_UNSPEC:
if (lp != NULL)
error = lp->lp_output(ifp, m, dst, rt);
else
error = ENETDOWN;
break;
default:
m_freem(m);
error = ENETDOWN;
}
return error;
}
void
lagg_ifdetach(void *xifp_port)
{
struct ifnet *ifp_port = xifp_port;
struct lagg_port *lp;
struct lagg_softc *sc;
uint8_t lladdr[ETHER_ADDR_LEN];
int s;
IFNET_ASSERT_UNLOCKED(ifp_port);
s = pserialize_read_enter();
lp = atomic_load_consume(&ifp_port->if_lagg);
if (lp == NULL) {
pserialize_read_exit(s);
return;
}
sc = lp->lp_softc;
if (sc == NULL) {
pserialize_read_exit(s);
return;
}
pserialize_read_exit(s);
LAGG_LOCK(sc);
lp = ifp_port->if_lagg;
if (lp == NULL) {
LAGG_UNLOCK(sc);
return;
}
/*
* mark as a detaching to prevent other
* lagg_port_teardown() processings with IFNET_LOCK() held
*/
lp->lp_ifdetaching = true;
LAGG_UNLOCK(sc);
IFNET_LOCK(&sc->sc_if);
LAGG_LOCK(sc);
lp = ifp_port->if_lagg;
if (lp != NULL) {
lagg_lladdr_cpy(lladdr, sc->sc_lladdr);
lagg_port_teardown(sc, lp, true);
lagg_sadl_update(sc, lladdr);
}
LAGG_UNLOCK(sc);
IFNET_UNLOCK(&sc->sc_if);
}
void
lagg_linkstate_changed(void *xifp)
{
struct ifnet *ifp = xifp;
struct lagg_port *lp;
struct psref psref;
int s, bound;
s = pserialize_read_enter();
lp = atomic_load_consume(&ifp->if_lagg);
if (lp != NULL) {
bound = curlwp_bind();
lagg_port_getref(lp, &psref);
} else {
pserialize_read_exit(s);
return;
}
pserialize_read_exit(s);
IFNET_LOCK(lp->lp_ifp);
lagg_proto_linkstate(lp->lp_softc, lp);
IFNET_UNLOCK(lp->lp_ifp);
lagg_port_putref(lp, &psref);
curlwp_bindx(bound);
}
void
lagg_port_getref(struct lagg_port *lp, struct psref *psref)
{
psref_acquire(psref, &lp->lp_psref, lagg_port_psref_class);
}
void
lagg_port_putref(struct lagg_port *lp, struct psref *psref)
{
psref_release(psref, &lp->lp_psref, lagg_port_psref_class);
}
void
lagg_log(struct lagg_softc *sc, int lvl, const char *fmt, ...)
{
va_list ap;
if (lvl == LOG_DEBUG && !lagg_debug_enable(sc))
return;
log(lvl, "%s: ", sc->sc_if.if_xname);
va_start(ap, fmt);
vlog(lvl, fmt, ap);
va_end(ap);
}
static void
lagg_workq_work(struct work *wk, void *context)
{
struct lagg_work *lw;
lw = container_of(wk, struct lagg_work, lw_cookie);
atomic_cas_uint(&lw->lw_state, LAGG_WORK_ENQUEUED, LAGG_WORK_IDLE);
lw->lw_func(lw, lw->lw_arg);
}
struct workqueue *
lagg_workq_create(const char *name, pri_t prio, int ipl, int flags)
{
struct workqueue *wq;
int error;
error = workqueue_create(&wq, name, lagg_workq_work,
NULL, prio, ipl, flags);
if (error)
return NULL;
return wq;
}
void
lagg_workq_destroy(struct workqueue *wq)
{
workqueue_destroy(wq);
}
void
lagg_workq_add(struct workqueue *wq, struct lagg_work *lw)
{
if (atomic_cas_uint(&lw->lw_state, LAGG_WORK_IDLE,
LAGG_WORK_ENQUEUED) != LAGG_WORK_IDLE)
return;
KASSERT(lw->lw_func != NULL);
kpreempt_disable();
workqueue_enqueue(wq, &lw->lw_cookie, NULL);
kpreempt_enable();
}
void
lagg_workq_wait(struct workqueue *wq, struct lagg_work *lw)
{
atomic_swap_uint(&lw->lw_state, LAGG_WORK_STOPPING);
workqueue_wait(wq, &lw->lw_cookie);
}
static int
lagg_chg_sadl(struct ifnet *ifp, uint8_t *lla, size_t lla_len)
{
struct psref psref_cur, psref_next;
struct ifaddr *ifa_cur, *ifa_next, *ifa_lla;
const struct sockaddr_dl *sdl, *nsdl;
int s, error;
KASSERT(!cpu_intr_p() && !cpu_softintr_p());
KASSERT(IFNET_LOCKED(ifp));
KASSERT(ifp->if_addrlen == lla_len);
error = 0;
ifa_lla = NULL;
while (1) {
s = pserialize_read_enter();
IFADDR_READER_FOREACH(ifa_cur, ifp) {
sdl = satocsdl(ifa_cur->ifa_addr);
if (sdl->sdl_family != AF_LINK)
continue;
if (sdl->sdl_type != ifp->if_type) {
ifa_acquire(ifa_cur, &psref_cur);
break;
}
}
pserialize_read_exit(s);
if (ifa_cur == NULL)
break;
ifa_next = if_dl_create(ifp, &nsdl);
if (ifa_next == NULL) {
error = ENOMEM;
ifa_release(ifa_cur, &psref_cur);
goto done;
}
ifa_acquire(ifa_next, &psref_next);
(void)sockaddr_dl_setaddr(__UNCONST(nsdl), nsdl->sdl_len,
CLLADDR(sdl), ifp->if_addrlen);
ifa_insert(ifp, ifa_next);
if (ifa_lla == NULL &&
memcmp(CLLADDR(sdl), lla, lla_len) == 0) {
ifa_lla = ifa_next;
ifaref(ifa_lla);
}
if (ifa_cur == ifp->if_dl)
if_activate_sadl(ifp, ifa_next, nsdl);
if (ifa_cur == ifp->if_hwdl) {
ifp->if_hwdl = ifa_next;
ifaref(ifa_next);
ifafree(ifa_cur);
}
ifaref(ifa_cur);
ifa_release(ifa_cur, &psref_cur);
ifa_remove(ifp, ifa_cur);
KASSERTMSG(ifa_cur->ifa_refcnt == 1,
"ifa_refcnt=%d", ifa_cur->ifa_refcnt);
ifafree(ifa_cur);
ifa_release(ifa_next, &psref_next);
}
if (ifa_lla != NULL) {
ifa_next = ifa_lla;
ifa_acquire(ifa_next, &psref_next);
ifafree(ifa_lla);
nsdl = satocsdl(ifa_next->ifa_addr);
} else {
ifa_next = if_dl_create(ifp, &nsdl);
if (ifa_next == NULL) {
error = ENOMEM;
goto done;
}
ifa_acquire(ifa_next, &psref_next);
(void)sockaddr_dl_setaddr(__UNCONST(nsdl),
nsdl->sdl_len, lla, ifp->if_addrlen);
ifa_insert(ifp, ifa_next);
}
if (ifa_next != ifp->if_dl) {
ifa_cur = ifp->if_dl;
if (ifa_cur != NULL)
ifa_acquire(ifa_cur, &psref_cur);
if_activate_sadl(ifp, ifa_next, nsdl);
if (ifa_cur != NULL) {
if (ifa_cur != ifp->if_hwdl) {
ifaref(ifa_cur);
ifa_release(ifa_cur, &psref_cur);
ifa_remove(ifp, ifa_cur);
KASSERTMSG(ifa_cur->ifa_refcnt == 1,
"ifa_refcnt=%d",
ifa_cur->ifa_refcnt);
ifafree(ifa_cur);
} else {
ifa_release(ifa_cur, &psref_cur);
}
}
}
ifa_release(ifa_next, &psref_next);
done:
return error;
}
/*
* Module infrastructure
*/
#include <net/if_module.h>
IF_MODULE(MODULE_CLASS_DRIVER, lagg, NULL)