/* $NetBSD: if_bridge.c,v 1.196 2024/12/16 05:21:24 ozaki-r Exp $ */ /* * Copyright 2001 Wasabi Systems, Inc. * All rights reserved. * * Written by Jason R. Thorpe for Wasabi Systems, Inc. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed for the NetBSD Project by * Wasabi Systems, Inc. * 4. The name of Wasabi Systems, Inc. may not be used to endorse * or promote products derived from this software without specific prior * written permission. * * THIS SOFTWARE IS PROVIDED BY WASABI SYSTEMS, INC. ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL WASABI SYSTEMS, INC * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ /* * Copyright (c) 1999, 2000 Jason L. Wright (jason@thought.net) * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software * must display the following acknowledgement: * This product includes software developed by Jason L. Wright * 4. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. * * OpenBSD: if_bridge.c,v 1.60 2001/06/15 03:38:33 itojun Exp */ /* * Network interface bridge support. * * TODO: * * - Currently only supports Ethernet-like interfaces (Ethernet, * 802.11, VLANs on Ethernet, etc.) Figure out a nice way * to bridge other types of interfaces (FDDI-FDDI, and maybe * consider heterogenous bridges). */ #include __KERNEL_RCSID(0, "$NetBSD: if_bridge.c,v 1.196 2024/12/16 05:21:24 ozaki-r Exp $"); #ifdef _KERNEL_OPT #include "opt_inet.h" #include "opt_net_mpsafe.h" #endif /* _KERNEL_OPT */ #include #include #include #include #include #include /* for softnet_lock */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /* Used for bridge_ip[6]_checkbasic */ #include #include #include #include #include /* XXX */ #include #include #include #include /* XXX */ /* * Size of the route hash table. Must be a power of two. */ #ifndef BRIDGE_RTHASH_SIZE #define BRIDGE_RTHASH_SIZE 1024 #endif #define BRIDGE_RTHASH_MASK (BRIDGE_RTHASH_SIZE - 1) #include "carp.h" #if NCARP > 0 #include #include #include #endif #include "ioconf.h" __CTASSERT(sizeof(struct ifbifconf) == sizeof(struct ifbaconf)); __CTASSERT(offsetof(struct ifbifconf, ifbic_len) == offsetof(struct ifbaconf, ifbac_len)); __CTASSERT(offsetof(struct ifbifconf, ifbic_buf) == offsetof(struct ifbaconf, ifbac_buf)); /* * Maximum number of addresses to cache. */ #ifndef BRIDGE_RTABLE_MAX #define BRIDGE_RTABLE_MAX 100 #endif /* * Spanning tree defaults. */ #define BSTP_DEFAULT_MAX_AGE (20 * 256) #define BSTP_DEFAULT_HELLO_TIME (2 * 256) #define BSTP_DEFAULT_FORWARD_DELAY (15 * 256) #define BSTP_DEFAULT_HOLD_TIME (1 * 256) #define BSTP_DEFAULT_BRIDGE_PRIORITY 0x8000 #define BSTP_DEFAULT_PORT_PRIORITY 0x80 #define BSTP_DEFAULT_PATH_COST 55 /* * Timeout (in seconds) for entries learned dynamically. */ #ifndef BRIDGE_RTABLE_TIMEOUT #define BRIDGE_RTABLE_TIMEOUT (20 * 60) /* same as ARP */ #endif /* * Number of seconds between walks of the route list. */ #ifndef BRIDGE_RTABLE_PRUNE_PERIOD #define BRIDGE_RTABLE_PRUNE_PERIOD (5 * 60) #endif #define BRIDGE_RT_LOCK(_sc) mutex_enter((_sc)->sc_rtlist_lock) #define BRIDGE_RT_UNLOCK(_sc) mutex_exit((_sc)->sc_rtlist_lock) #define BRIDGE_RT_LOCKED(_sc) mutex_owned((_sc)->sc_rtlist_lock) #define BRIDGE_RT_PSZ_PERFORM(_sc) \ pserialize_perform((_sc)->sc_rtlist_psz) #define BRIDGE_RTLIST_READER_FOREACH(_brt, _sc) \ PSLIST_READER_FOREACH((_brt), &((_sc)->sc_rtlist), \ struct bridge_rtnode, brt_list) #define BRIDGE_RTLIST_WRITER_FOREACH(_brt, _sc) \ PSLIST_WRITER_FOREACH((_brt), &((_sc)->sc_rtlist), \ struct bridge_rtnode, brt_list) #define BRIDGE_RTLIST_WRITER_INSERT_HEAD(_sc, _brt) \ PSLIST_WRITER_INSERT_HEAD(&(_sc)->sc_rtlist, brt, brt_list) #define BRIDGE_RTLIST_WRITER_REMOVE(_brt) \ PSLIST_WRITER_REMOVE((_brt), brt_list) #define BRIDGE_RTHASH_READER_FOREACH(_brt, _sc, _hash) \ PSLIST_READER_FOREACH((_brt), &(_sc)->sc_rthash[(_hash)], \ struct bridge_rtnode, brt_hash) #define BRIDGE_RTHASH_WRITER_FOREACH(_brt, _sc, _hash) \ PSLIST_WRITER_FOREACH((_brt), &(_sc)->sc_rthash[(_hash)], \ struct bridge_rtnode, brt_hash) #define BRIDGE_RTHASH_WRITER_INSERT_HEAD(_sc, _hash, _brt) \ PSLIST_WRITER_INSERT_HEAD(&(_sc)->sc_rthash[(_hash)], brt, brt_hash) #define BRIDGE_RTHASH_WRITER_INSERT_AFTER(_brt, _new) \ PSLIST_WRITER_INSERT_AFTER((_brt), (_new), brt_hash) #define BRIDGE_RTHASH_WRITER_REMOVE(_brt) \ PSLIST_WRITER_REMOVE((_brt), brt_hash) #ifdef NET_MPSAFE #define DECLARE_LOCK_VARIABLE #define ACQUIRE_GLOBAL_LOCKS() do { } while (0) #define RELEASE_GLOBAL_LOCKS() do { } while (0) #else #define DECLARE_LOCK_VARIABLE int __s #define ACQUIRE_GLOBAL_LOCKS() do { \ KERNEL_LOCK(1, NULL); \ mutex_enter(softnet_lock); \ __s = splsoftnet(); \ } while (0) #define RELEASE_GLOBAL_LOCKS() do { \ splx(__s); \ mutex_exit(softnet_lock); \ KERNEL_UNLOCK_ONE(NULL); \ } while (0) #endif struct psref_class *bridge_psref_class __read_mostly; int bridge_rtable_prune_period = BRIDGE_RTABLE_PRUNE_PERIOD; static struct pool bridge_rtnode_pool; static int bridge_clone_create(struct if_clone *, int); static int bridge_clone_destroy(struct ifnet *); static int bridge_ioctl(struct ifnet *, u_long, void *); static int bridge_init(struct ifnet *); static void bridge_stop(struct ifnet *, int); static void bridge_start(struct ifnet *); static void bridge_ifdetach(void *); static void bridge_input(struct ifnet *, struct mbuf *); static void bridge_forward(struct bridge_softc *, struct mbuf *); static void bridge_timer(void *); static void bridge_broadcast(struct bridge_softc *, struct ifnet *, bool, struct mbuf *); static int bridge_rtupdate(struct bridge_softc *, const uint8_t *, struct ifnet *, int, uint8_t); static struct ifnet *bridge_rtlookup(struct bridge_softc *, const uint8_t *); static void bridge_rttrim(struct bridge_softc *); static void bridge_rtage(struct bridge_softc *); static void bridge_rtage_work(struct work *, void *); static void bridge_rtflush(struct bridge_softc *, int); static int bridge_rtdaddr(struct bridge_softc *, const uint8_t *); static void bridge_rtdelete(struct bridge_softc *, struct ifnet *ifp); static void bridge_rtable_init(struct bridge_softc *); static void bridge_rtable_fini(struct bridge_softc *); static struct bridge_rtnode *bridge_rtnode_lookup(struct bridge_softc *, const uint8_t *); static int bridge_rtnode_insert(struct bridge_softc *, struct bridge_rtnode *); static void bridge_rtnode_remove(struct bridge_softc *, struct bridge_rtnode *); static void bridge_rtnode_destroy(struct bridge_rtnode *); static struct bridge_iflist *bridge_lookup_member(struct bridge_softc *, const char *name, struct psref *); static struct bridge_iflist *bridge_lookup_member_if(struct bridge_softc *, struct ifnet *ifp, struct psref *); static void bridge_release_member(struct bridge_softc *, struct bridge_iflist *, struct psref *); static void bridge_delete_member(struct bridge_softc *, struct bridge_iflist *); static void bridge_acquire_member(struct bridge_softc *sc, struct bridge_iflist *, struct psref *); static int bridge_ioctl_add(struct bridge_softc *, void *); static int bridge_ioctl_del(struct bridge_softc *, void *); static int bridge_ioctl_gifflags(struct bridge_softc *, void *); static int bridge_ioctl_sifflags(struct bridge_softc *, void *); static int bridge_ioctl_scache(struct bridge_softc *, void *); static int bridge_ioctl_gcache(struct bridge_softc *, void *); static int bridge_ioctl_gifs(struct bridge_softc *, void *); static int bridge_ioctl_rts(struct bridge_softc *, void *); static int bridge_ioctl_saddr(struct bridge_softc *, void *); static int bridge_ioctl_sto(struct bridge_softc *, void *); static int bridge_ioctl_gto(struct bridge_softc *, void *); static int bridge_ioctl_daddr(struct bridge_softc *, void *); static int bridge_ioctl_flush(struct bridge_softc *, void *); static int bridge_ioctl_gpri(struct bridge_softc *, void *); static int bridge_ioctl_spri(struct bridge_softc *, void *); static int bridge_ioctl_ght(struct bridge_softc *, void *); static int bridge_ioctl_sht(struct bridge_softc *, void *); static int bridge_ioctl_gfd(struct bridge_softc *, void *); static int bridge_ioctl_sfd(struct bridge_softc *, void *); static int bridge_ioctl_gma(struct bridge_softc *, void *); static int bridge_ioctl_sma(struct bridge_softc *, void *); static int bridge_ioctl_sifprio(struct bridge_softc *, void *); static int bridge_ioctl_sifcost(struct bridge_softc *, void *); static int bridge_ioctl_gfilt(struct bridge_softc *, void *); static int bridge_ioctl_sfilt(struct bridge_softc *, void *); static int bridge_ipf(void *, struct mbuf **, struct ifnet *, int); static int bridge_ip_checkbasic(struct mbuf **mp); # ifdef INET6 static int bridge_ip6_checkbasic(struct mbuf **mp); # endif /* INET6 */ struct bridge_control { int (*bc_func)(struct bridge_softc *, void *); int bc_argsize; int bc_flags; }; #define BC_F_COPYIN 0x01 /* copy arguments in */ #define BC_F_COPYOUT 0x02 /* copy arguments out */ #define BC_F_SUSER 0x04 /* do super-user check */ #define BC_F_XLATEIN 0x08 /* xlate arguments in */ #define BC_F_XLATEOUT 0x10 /* xlate arguments out */ static const struct bridge_control bridge_control_table[] = { [BRDGADD] = {bridge_ioctl_add, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_SUSER}, [BRDGDEL] = {bridge_ioctl_del, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_SUSER}, [BRDGGIFFLGS] = {bridge_ioctl_gifflags, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_COPYOUT}, [BRDGSIFFLGS] = {bridge_ioctl_sifflags, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_SUSER}, [BRDGSCACHE] = {bridge_ioctl_scache, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGGCACHE] = {bridge_ioctl_gcache, sizeof(struct ifbrparam), BC_F_COPYOUT}, [OBRDGGIFS] = {bridge_ioctl_gifs, sizeof(struct ifbifconf), BC_F_COPYIN|BC_F_COPYOUT}, [OBRDGRTS] = {bridge_ioctl_rts, sizeof(struct ifbaconf), BC_F_COPYIN|BC_F_COPYOUT}, [BRDGSADDR] = {bridge_ioctl_saddr, sizeof(struct ifbareq), BC_F_COPYIN|BC_F_SUSER}, [BRDGSTO] = {bridge_ioctl_sto, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGGTO] = {bridge_ioctl_gto, sizeof(struct ifbrparam), BC_F_COPYOUT}, [BRDGDADDR] = {bridge_ioctl_daddr, sizeof(struct ifbareq), BC_F_COPYIN|BC_F_SUSER}, [BRDGFLUSH] = {bridge_ioctl_flush, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_SUSER}, [BRDGGPRI] = {bridge_ioctl_gpri, sizeof(struct ifbrparam), BC_F_COPYOUT}, [BRDGSPRI] = {bridge_ioctl_spri, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGGHT] = {bridge_ioctl_ght, sizeof(struct ifbrparam), BC_F_COPYOUT}, [BRDGSHT] = {bridge_ioctl_sht, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGGFD] = {bridge_ioctl_gfd, sizeof(struct ifbrparam), BC_F_COPYOUT}, [BRDGSFD] = {bridge_ioctl_sfd, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGGMA] = {bridge_ioctl_gma, sizeof(struct ifbrparam), BC_F_COPYOUT}, [BRDGSMA] = {bridge_ioctl_sma, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGSIFPRIO] = {bridge_ioctl_sifprio, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_SUSER}, [BRDGSIFCOST] = {bridge_ioctl_sifcost, sizeof(struct ifbreq), BC_F_COPYIN|BC_F_SUSER}, [BRDGGFILT] = {bridge_ioctl_gfilt, sizeof(struct ifbrparam), BC_F_COPYOUT}, [BRDGSFILT] = {bridge_ioctl_sfilt, sizeof(struct ifbrparam), BC_F_COPYIN|BC_F_SUSER}, [BRDGGIFS] = {bridge_ioctl_gifs, sizeof(struct ifbifconf), BC_F_XLATEIN|BC_F_XLATEOUT}, [BRDGRTS] = {bridge_ioctl_rts, sizeof(struct ifbaconf), BC_F_XLATEIN|BC_F_XLATEOUT}, }; static const int bridge_control_table_size = __arraycount(bridge_control_table); static struct if_clone bridge_cloner = IF_CLONE_INITIALIZER("bridge", bridge_clone_create, bridge_clone_destroy); /* * bridgeattach: * * Pseudo-device attach routine. */ void bridgeattach(int n) { pool_init(&bridge_rtnode_pool, sizeof(struct bridge_rtnode), 0, 0, 0, "brtpl", NULL, IPL_NET); bridge_psref_class = psref_class_create("bridge", IPL_SOFTNET); if_clone_attach(&bridge_cloner); } /* * bridge_clone_create: * * Create a new bridge instance. */ static int bridge_clone_create(struct if_clone *ifc, int unit) { struct bridge_softc *sc; struct ifnet *ifp; int error; sc = kmem_zalloc(sizeof(*sc), KM_SLEEP); ifp = &sc->sc_if; sc->sc_brtmax = BRIDGE_RTABLE_MAX; sc->sc_brttimeout = BRIDGE_RTABLE_TIMEOUT; sc->sc_bridge_max_age = BSTP_DEFAULT_MAX_AGE; sc->sc_bridge_hello_time = BSTP_DEFAULT_HELLO_TIME; sc->sc_bridge_forward_delay = BSTP_DEFAULT_FORWARD_DELAY; sc->sc_bridge_priority = BSTP_DEFAULT_BRIDGE_PRIORITY; sc->sc_hold_time = BSTP_DEFAULT_HOLD_TIME; sc->sc_filter_flags = 0; /* Initialize our routing table. */ bridge_rtable_init(sc); error = workqueue_create(&sc->sc_rtage_wq, "bridge_rtage", bridge_rtage_work, sc, PRI_SOFTNET, IPL_SOFTNET, WQ_MPSAFE); if (error) panic("%s: workqueue_create %d\n", __func__, error); callout_init(&sc->sc_brcallout, CALLOUT_MPSAFE); callout_init(&sc->sc_bstpcallout, CALLOUT_MPSAFE); mutex_init(&sc->sc_iflist_psref.bip_lock, MUTEX_DEFAULT, IPL_NONE); PSLIST_INIT(&sc->sc_iflist_psref.bip_iflist); sc->sc_iflist_psref.bip_psz = pserialize_create(); if_initname(ifp, ifc->ifc_name, unit); ifp->if_softc = sc; #ifdef NET_MPSAFE ifp->if_extflags = IFEF_MPSAFE; #endif ifp->if_mtu = ETHERMTU; ifp->if_ioctl = bridge_ioctl; ifp->if_output = bridge_output; ifp->if_start = bridge_start; ifp->if_stop = bridge_stop; ifp->if_init = bridge_init; ifp->if_type = IFT_BRIDGE; ifp->if_addrlen = 0; ifp->if_dlt = DLT_EN10MB; ifp->if_hdrlen = ETHER_HDR_LEN; if_initialize(ifp); /* * Set the link state to down. * When interfaces are added the link state will reflect * the best link state of the combined interfaces. */ ifp->if_link_state = LINK_STATE_DOWN; if_alloc_sadl(ifp); if_register(ifp); return 0; } /* * bridge_clone_destroy: * * Destroy a bridge instance. */ static int bridge_clone_destroy(struct ifnet *ifp) { struct bridge_softc *sc = ifp->if_softc; struct bridge_iflist *bif; if ((ifp->if_flags & IFF_RUNNING) != 0) bridge_stop(ifp, 1); BRIDGE_LOCK(sc); for (;;) { bif = PSLIST_WRITER_FIRST(&sc->sc_iflist_psref.bip_iflist, struct bridge_iflist, bif_next); if (bif == NULL) break; bridge_delete_member(sc, bif); } PSLIST_DESTROY(&sc->sc_iflist_psref.bip_iflist); BRIDGE_UNLOCK(sc); if_detach(ifp); /* Tear down the routing table. */ bridge_rtable_fini(sc); pserialize_destroy(sc->sc_iflist_psref.bip_psz); mutex_destroy(&sc->sc_iflist_psref.bip_lock); callout_destroy(&sc->sc_brcallout); callout_destroy(&sc->sc_bstpcallout); workqueue_destroy(sc->sc_rtage_wq); kmem_free(sc, sizeof(*sc)); return 0; } /* * bridge_ioctl: * * Handle a control request from the operator. */ static int bridge_ioctl(struct ifnet *ifp, u_long cmd, void *data) { struct bridge_softc *sc = ifp->if_softc; struct lwp *l = curlwp; /* XXX */ union { struct ifbreq ifbreq; struct ifbifconf ifbifconf; struct ifbareq ifbareq; struct ifbaconf ifbaconf; struct ifbrparam ifbrparam; } args; struct ifdrv *ifd = (struct ifdrv *) data; const struct bridge_control *bc = NULL; /* XXXGCC */ int error = 0; /* Authorize command before calling splsoftnet(). */ switch (cmd) { case SIOCGDRVSPEC: case SIOCSDRVSPEC: if (ifd->ifd_cmd >= bridge_control_table_size || (bc = &bridge_control_table[ifd->ifd_cmd]) == NULL) { error = EINVAL; return error; } /* We only care about BC_F_SUSER at this point. */ if ((bc->bc_flags & BC_F_SUSER) == 0) break; error = kauth_authorize_network(l->l_cred, KAUTH_NETWORK_INTERFACE_BRIDGE, cmd == SIOCGDRVSPEC ? KAUTH_REQ_NETWORK_INTERFACE_BRIDGE_GETPRIV : KAUTH_REQ_NETWORK_INTERFACE_BRIDGE_SETPRIV, ifd, NULL, NULL); if (error) return error; break; } const int s = splsoftnet(); switch (cmd) { case SIOCGDRVSPEC: case SIOCSDRVSPEC: KASSERT(bc != NULL); if (cmd == SIOCGDRVSPEC && (bc->bc_flags & (BC_F_COPYOUT|BC_F_XLATEOUT)) == 0) { error = EINVAL; break; } else if (cmd == SIOCSDRVSPEC && (bc->bc_flags & (BC_F_COPYOUT|BC_F_XLATEOUT)) != 0) { error = EINVAL; break; } /* BC_F_SUSER is checked above, before splsoftnet(). */ if ((bc->bc_flags & (BC_F_XLATEIN|BC_F_XLATEOUT)) == 0 && (ifd->ifd_len != bc->bc_argsize || ifd->ifd_len > sizeof(args))) { error = EINVAL; break; } memset(&args, 0, sizeof(args)); if (bc->bc_flags & BC_F_COPYIN) { error = copyin(ifd->ifd_data, &args, ifd->ifd_len); if (error) break; } else if (bc->bc_flags & BC_F_XLATEIN) { args.ifbifconf.ifbic_len = ifd->ifd_len; args.ifbifconf.ifbic_buf = ifd->ifd_data; } error = (*bc->bc_func)(sc, &args); if (error) break; if (bc->bc_flags & BC_F_COPYOUT) { error = copyout(&args, ifd->ifd_data, ifd->ifd_len); } else if (bc->bc_flags & BC_F_XLATEOUT) { ifd->ifd_len = args.ifbifconf.ifbic_len; ifd->ifd_data = args.ifbifconf.ifbic_buf; } break; case SIOCSIFFLAGS: if ((error = ifioctl_common(ifp, cmd, data)) != 0) break; switch (ifp->if_flags & (IFF_UP|IFF_RUNNING)) { case IFF_RUNNING: /* * If interface is marked down and it is running, * then stop and disable it. */ if_stop(ifp, 1); break; case IFF_UP: /* * If interface is marked up and it is stopped, then * start it. */ error = if_init(ifp); break; default: break; } break; case SIOCSIFMTU: if ((error = ifioctl_common(ifp, cmd, data)) == ENETRESET) error = 0; break; case SIOCGIFCAP: { struct ifcapreq *ifcr = (struct ifcapreq *)data; ifcr->ifcr_capabilities = sc->sc_capenable; ifcr->ifcr_capenable = sc->sc_capenable; break; } default: error = ifioctl_common(ifp, cmd, data); break; } splx(s); return error; } /* * bridge_lookup_member: * * Lookup a bridge member interface. */ static struct bridge_iflist * bridge_lookup_member(struct bridge_softc *sc, const char *name, struct psref *psref) { struct bridge_iflist *bif; struct ifnet *ifp; int s; s = pserialize_read_enter(); BRIDGE_IFLIST_READER_FOREACH(bif, sc) { ifp = bif->bif_ifp; if (strcmp(ifp->if_xname, name) == 0) break; } if (bif != NULL) bridge_acquire_member(sc, bif, psref); pserialize_read_exit(s); return bif; } /* * bridge_lookup_member_if: * * Lookup a bridge member interface by ifnet*. */ static struct bridge_iflist * bridge_lookup_member_if(struct bridge_softc *sc, struct ifnet *member_ifp, struct psref *psref) { struct bridge_iflist *bif; int s; s = pserialize_read_enter(); bif = member_ifp->if_bridgeif; if (bif != NULL) { psref_acquire(psref, &bif->bif_psref, bridge_psref_class); } pserialize_read_exit(s); return bif; } static void bridge_acquire_member(struct bridge_softc *sc, struct bridge_iflist *bif, struct psref *psref) { psref_acquire(psref, &bif->bif_psref, bridge_psref_class); } /* * bridge_release_member: * * Release the specified member interface. */ static void bridge_release_member(struct bridge_softc *sc, struct bridge_iflist *bif, struct psref *psref) { psref_release(psref, &bif->bif_psref, bridge_psref_class); } /* * bridge_delete_member: * * Delete the specified member interface. */ static void bridge_delete_member(struct bridge_softc *sc, struct bridge_iflist *bif) { struct ifnet *ifs = bif->bif_ifp; KASSERT(BRIDGE_LOCKED(sc)); ifs->_if_input = ether_input; ifs->if_bridge = NULL; ifs->if_bridgeif = NULL; PSLIST_WRITER_REMOVE(bif, bif_next); BRIDGE_PSZ_PERFORM(sc); if_linkstate_change_disestablish(ifs, bif->bif_linkstate_hook, BRIDGE_LOCK_OBJ(sc)); ether_ifdetachhook_disestablish(ifs, bif->bif_ifdetach_hook, BRIDGE_LOCK_OBJ(sc)); BRIDGE_UNLOCK(sc); switch (ifs->if_type) { case IFT_ETHER: case IFT_L2TP: /* * Take the interface out of promiscuous mode. * Don't call it with holding a spin lock. */ (void) ifpromisc(ifs, 0); IFNET_LOCK(ifs); (void) ether_disable_vlan_mtu(ifs); IFNET_UNLOCK(ifs); break; default: #ifdef DIAGNOSTIC panic("%s: impossible", __func__); #endif break; } psref_target_destroy(&bif->bif_psref, bridge_psref_class); PSLIST_ENTRY_DESTROY(bif, bif_next); kmem_free(bif, sizeof(*bif)); BRIDGE_LOCK(sc); } /* * bridge_calc_csum_flags: * * Calculate logical and b/w csum flags each member interface supports. */ void bridge_calc_csum_flags(struct bridge_softc *sc) { struct bridge_iflist *bif; struct ifnet *ifs = NULL; int flags = ~0; int capenable = ~0; BRIDGE_LOCK(sc); BRIDGE_IFLIST_READER_FOREACH(bif, sc) { ifs = bif->bif_ifp; flags &= ifs->if_csum_flags_tx; capenable &= ifs->if_capenable; } sc->sc_csum_flags_tx = flags; sc->sc_capenable = (ifs != NULL) ? capenable : 0; BRIDGE_UNLOCK(sc); } /* * bridge_calc_link_state: * * Calculate the link state based on each member interface. */ static void bridge_calc_link_state(void *xsc) { struct bridge_softc *sc = xsc; struct bridge_iflist *bif; struct ifnet *ifs; int link_state = LINK_STATE_DOWN; BRIDGE_LOCK(sc); BRIDGE_IFLIST_READER_FOREACH(bif, sc) { ifs = bif->bif_ifp; if (ifs->if_link_state == LINK_STATE_UP) { link_state = LINK_STATE_UP; break; } if (ifs->if_link_state == LINK_STATE_UNKNOWN) link_state = LINK_STATE_UNKNOWN; } if_link_state_change(&sc->sc_if, link_state); BRIDGE_UNLOCK(sc); } static int bridge_ioctl_add(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; struct bridge_iflist *bif = NULL; struct ifnet *ifs; int error = 0; struct psref psref; ifs = if_get(req->ifbr_ifsname, &psref); if (ifs == NULL) return ENOENT; if (ifs->if_bridge == sc) { error = EEXIST; goto out; } if (ifs->if_bridge != NULL) { error = EBUSY; goto out; } if (ifs->_if_input != ether_input) { error = EINVAL; goto out; } /* FIXME: doesn't work with non-IFF_SIMPLEX interfaces */ if ((ifs->if_flags & IFF_SIMPLEX) == 0) { error = EINVAL; goto out; } bif = kmem_alloc(sizeof(*bif), KM_SLEEP); switch (ifs->if_type) { case IFT_ETHER: if (sc->sc_if.if_mtu != ifs->if_mtu) { /* Change MTU of added interface to bridge MTU */ struct ifreq ifr; memset(&ifr, 0, sizeof(ifr)); ifr.ifr_mtu = sc->sc_if.if_mtu; IFNET_LOCK(ifs); error = if_ioctl(ifs, SIOCSIFMTU, &ifr); IFNET_UNLOCK(ifs); if (error != 0) goto out; } /* FALLTHROUGH */ case IFT_L2TP: IFNET_LOCK(ifs); error = ether_enable_vlan_mtu(ifs); IFNET_UNLOCK(ifs); if (error > 0) goto out; /* * Place the interface into promiscuous mode. */ error = ifpromisc(ifs, 1); if (error) goto out; break; default: error = EINVAL; goto out; } bif->bif_ifp = ifs; bif->bif_flags = IFBIF_LEARNING | IFBIF_DISCOVER; bif->bif_priority = BSTP_DEFAULT_PORT_PRIORITY; bif->bif_path_cost = BSTP_DEFAULT_PATH_COST; bif->bif_linkstate_hook = if_linkstate_change_establish(ifs, bridge_calc_link_state, sc); PSLIST_ENTRY_INIT(bif, bif_next); psref_target_init(&bif->bif_psref, bridge_psref_class); BRIDGE_LOCK(sc); ifs->if_bridge = sc; ifs->if_bridgeif = bif; PSLIST_WRITER_INSERT_HEAD(&sc->sc_iflist_psref.bip_iflist, bif, bif_next); ifs->_if_input = bridge_input; BRIDGE_UNLOCK(sc); bif->bif_ifdetach_hook = ether_ifdetachhook_establish(ifs, bridge_ifdetach, (void *)ifs); bridge_calc_csum_flags(sc); bridge_calc_link_state(sc); if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); else bstp_stop(sc); out: if_put(ifs, &psref); if (error) { if (bif != NULL) kmem_free(bif, sizeof(*bif)); } return error; } static int bridge_ioctl_del(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; const char *name = req->ifbr_ifsname; struct bridge_iflist *bif; struct ifnet *ifs; BRIDGE_LOCK(sc); /* * Don't use bridge_lookup_member. We want to get a member * with bif_refs == 0. */ BRIDGE_IFLIST_WRITER_FOREACH(bif, sc) { ifs = bif->bif_ifp; if (strcmp(ifs->if_xname, name) == 0) break; } if (bif == NULL) { BRIDGE_UNLOCK(sc); return ENOENT; } bridge_delete_member(sc, bif); BRIDGE_UNLOCK(sc); bridge_rtdelete(sc, ifs); bridge_calc_csum_flags(sc); bridge_calc_link_state(sc); if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); return 0; } static int bridge_ioctl_gifflags(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; struct bridge_iflist *bif; struct psref psref; bif = bridge_lookup_member(sc, req->ifbr_ifsname, &psref); if (bif == NULL) return ENOENT; req->ifbr_ifsflags = bif->bif_flags; req->ifbr_state = bif->bif_state; req->ifbr_priority = bif->bif_priority; req->ifbr_path_cost = bif->bif_path_cost; req->ifbr_portno = bif->bif_ifp->if_index & 0xff; bridge_release_member(sc, bif, &psref); return 0; } static int bridge_ioctl_sifflags(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; struct bridge_iflist *bif; struct psref psref; bif = bridge_lookup_member(sc, req->ifbr_ifsname, &psref); if (bif == NULL) return ENOENT; if (req->ifbr_ifsflags & IFBIF_STP) { switch (bif->bif_ifp->if_type) { case IFT_ETHER: case IFT_L2TP: /* These can do spanning tree. */ break; default: /* Nothing else can. */ bridge_release_member(sc, bif, &psref); return EINVAL; } } if (bif->bif_flags & IFBIF_PROTECTED) { if ((req->ifbr_ifsflags & IFBIF_PROTECTED) == 0) { log(LOG_INFO, "%s: disabling protection on %s\n", sc->sc_if.if_xname, bif->bif_ifp->if_xname); } } else { if (req->ifbr_ifsflags & IFBIF_PROTECTED) { log(LOG_INFO, "%s: enabling protection on %s\n", sc->sc_if.if_xname, bif->bif_ifp->if_xname); } } bif->bif_flags = req->ifbr_ifsflags; bridge_release_member(sc, bif, &psref); if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); return 0; } static int bridge_ioctl_scache(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; sc->sc_brtmax = param->ifbrp_csize; bridge_rttrim(sc); return 0; } static int bridge_ioctl_gcache(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_csize = sc->sc_brtmax; return 0; } static int bridge_ioctl_gifs(struct bridge_softc *sc, void *arg) { struct ifbifconf *bifc = arg; struct bridge_iflist *bif; struct ifbreq *breqs; int i, count, error = 0; retry: BRIDGE_LOCK(sc); count = 0; BRIDGE_IFLIST_WRITER_FOREACH(bif, sc) count++; BRIDGE_UNLOCK(sc); if (count == 0) { bifc->ifbic_len = 0; return 0; } if (bifc->ifbic_len == 0 || bifc->ifbic_len < (sizeof(*breqs) * count)) { /* Tell that a larger buffer is needed */ bifc->ifbic_len = sizeof(*breqs) * count; return 0; } breqs = kmem_alloc(sizeof(*breqs) * count, KM_SLEEP); BRIDGE_LOCK(sc); i = 0; BRIDGE_IFLIST_WRITER_FOREACH(bif, sc) i++; if (i > count) { /* * The number of members has been increased. * We need more memory! */ BRIDGE_UNLOCK(sc); kmem_free(breqs, sizeof(*breqs) * count); goto retry; } i = 0; BRIDGE_IFLIST_WRITER_FOREACH(bif, sc) { struct ifbreq *breq = &breqs[i++]; memset(breq, 0, sizeof(*breq)); strlcpy(breq->ifbr_ifsname, bif->bif_ifp->if_xname, sizeof(breq->ifbr_ifsname)); breq->ifbr_ifsflags = bif->bif_flags; breq->ifbr_state = bif->bif_state; breq->ifbr_priority = bif->bif_priority; breq->ifbr_path_cost = bif->bif_path_cost; breq->ifbr_portno = bif->bif_ifp->if_index & 0xff; } /* Don't call copyout with holding the mutex */ BRIDGE_UNLOCK(sc); for (i = 0; i < count; i++) { error = copyout(&breqs[i], bifc->ifbic_req + i, sizeof(*breqs)); if (error) break; } bifc->ifbic_len = sizeof(*breqs) * i; kmem_free(breqs, sizeof(*breqs) * count); return error; } static int bridge_ioctl_rts(struct bridge_softc *sc, void *arg) { struct ifbaconf *bac = arg; struct bridge_rtnode *brt; struct ifbareq bareq; int count = 0, error = 0, len; if (bac->ifbac_len == 0) return 0; BRIDGE_RT_LOCK(sc); /* The passed buffer is not enough, tell a required size. */ if (bac->ifbac_len < (sizeof(bareq) * sc->sc_brtcnt)) { count = sc->sc_brtcnt; goto out; } len = bac->ifbac_len; BRIDGE_RTLIST_WRITER_FOREACH(brt, sc) { if (len < sizeof(bareq)) goto out; memset(&bareq, 0, sizeof(bareq)); strlcpy(bareq.ifba_ifsname, brt->brt_ifp->if_xname, sizeof(bareq.ifba_ifsname)); memcpy(bareq.ifba_dst, brt->brt_addr, sizeof(brt->brt_addr)); if ((brt->brt_flags & IFBAF_TYPEMASK) == IFBAF_DYNAMIC) { bareq.ifba_expire = brt->brt_expire - time_uptime; } else bareq.ifba_expire = 0; bareq.ifba_flags = brt->brt_flags; error = copyout(&bareq, bac->ifbac_req + count, sizeof(bareq)); if (error) goto out; count++; len -= sizeof(bareq); } out: BRIDGE_RT_UNLOCK(sc); bac->ifbac_len = sizeof(bareq) * count; return error; } static int bridge_ioctl_saddr(struct bridge_softc *sc, void *arg) { struct ifbareq *req = arg; struct bridge_iflist *bif; int error; struct psref psref; bif = bridge_lookup_member(sc, req->ifba_ifsname, &psref); if (bif == NULL) return ENOENT; error = bridge_rtupdate(sc, req->ifba_dst, bif->bif_ifp, 1, req->ifba_flags); bridge_release_member(sc, bif, &psref); return error; } static int bridge_ioctl_sto(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; sc->sc_brttimeout = param->ifbrp_ctime; return 0; } static int bridge_ioctl_gto(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_ctime = sc->sc_brttimeout; return 0; } static int bridge_ioctl_daddr(struct bridge_softc *sc, void *arg) { struct ifbareq *req = arg; return (bridge_rtdaddr(sc, req->ifba_dst)); } static int bridge_ioctl_flush(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; bridge_rtflush(sc, req->ifbr_ifsflags); return 0; } static int bridge_ioctl_gpri(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_prio = sc->sc_bridge_priority; return 0; } static int bridge_ioctl_spri(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; sc->sc_bridge_priority = param->ifbrp_prio; if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); return 0; } static int bridge_ioctl_ght(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_hellotime = sc->sc_bridge_hello_time >> 8; return 0; } static int bridge_ioctl_sht(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; if (param->ifbrp_hellotime == 0) return EINVAL; sc->sc_bridge_hello_time = param->ifbrp_hellotime << 8; if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); return 0; } static int bridge_ioctl_gfd(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_fwddelay = sc->sc_bridge_forward_delay >> 8; return 0; } static int bridge_ioctl_sfd(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; if (param->ifbrp_fwddelay == 0) return EINVAL; sc->sc_bridge_forward_delay = param->ifbrp_fwddelay << 8; if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); return 0; } static int bridge_ioctl_gma(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_maxage = sc->sc_bridge_max_age >> 8; return 0; } static int bridge_ioctl_sma(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; if (param->ifbrp_maxage == 0) return EINVAL; sc->sc_bridge_max_age = param->ifbrp_maxage << 8; if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); return 0; } static int bridge_ioctl_sifprio(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; struct bridge_iflist *bif; struct psref psref; bif = bridge_lookup_member(sc, req->ifbr_ifsname, &psref); if (bif == NULL) return ENOENT; bif->bif_priority = req->ifbr_priority; if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); bridge_release_member(sc, bif, &psref); return 0; } static int bridge_ioctl_gfilt(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; param->ifbrp_filter = sc->sc_filter_flags; return 0; } static int bridge_ioctl_sfilt(struct bridge_softc *sc, void *arg) { struct ifbrparam *param = arg; uint32_t nflags, oflags; if (param->ifbrp_filter & ~IFBF_FILT_MASK) return EINVAL; nflags = param->ifbrp_filter; oflags = sc->sc_filter_flags; if ((nflags & IFBF_FILT_USEIPF) && !(oflags & IFBF_FILT_USEIPF)) { pfil_add_hook((void *)bridge_ipf, NULL, PFIL_IN|PFIL_OUT, sc->sc_if.if_pfil); } if (!(nflags & IFBF_FILT_USEIPF) && (oflags & IFBF_FILT_USEIPF)) { pfil_remove_hook((void *)bridge_ipf, NULL, PFIL_IN|PFIL_OUT, sc->sc_if.if_pfil); } sc->sc_filter_flags = nflags; return 0; } static int bridge_ioctl_sifcost(struct bridge_softc *sc, void *arg) { struct ifbreq *req = arg; struct bridge_iflist *bif; struct psref psref; bif = bridge_lookup_member(sc, req->ifbr_ifsname, &psref); if (bif == NULL) return ENOENT; bif->bif_path_cost = req->ifbr_path_cost; if (sc->sc_if.if_flags & IFF_RUNNING) bstp_initialization(sc); bridge_release_member(sc, bif, &psref); return 0; } /* * bridge_ifdetach: * * Detach an interface from a bridge. Called when a member * interface is detaching. */ static void bridge_ifdetach(void *xifs) { struct ifnet *ifs; struct bridge_softc *sc; struct ifbreq breq; ifs = (struct ifnet *)xifs; sc = ifs->if_bridge; /* ioctl_lock should prevent this from happening */ KASSERT(sc != NULL); memset(&breq, 0, sizeof(breq)); strlcpy(breq.ifbr_ifsname, ifs->if_xname, sizeof(breq.ifbr_ifsname)); (void) bridge_ioctl_del(sc, &breq); } /* * bridge_init: * * Initialize a bridge interface. */ static int bridge_init(struct ifnet *ifp) { struct bridge_softc *sc = ifp->if_softc; KASSERT((ifp->if_flags & IFF_RUNNING) == 0); callout_reset(&sc->sc_brcallout, bridge_rtable_prune_period * hz, bridge_timer, sc); bstp_initialization(sc); ifp->if_flags |= IFF_RUNNING; return 0; } /* * bridge_stop: * * Stop the bridge interface. */ static void bridge_stop(struct ifnet *ifp, int disable) { struct bridge_softc *sc = ifp->if_softc; KASSERT((ifp->if_flags & IFF_RUNNING) != 0); ifp->if_flags &= ~IFF_RUNNING; callout_halt(&sc->sc_brcallout, NULL); workqueue_wait(sc->sc_rtage_wq, &sc->sc_rtage_wk); bstp_stop(sc); bridge_rtflush(sc, IFBF_FLUSHDYN); } /* * bridge_enqueue: * * Enqueue a packet on a bridge member interface. */ void bridge_enqueue(struct bridge_softc *sc, struct ifnet *dst_ifp, struct mbuf *m, int runfilt) { int len, error; short mflags; if (runfilt) { if (pfil_run_hooks(sc->sc_if.if_pfil, &m, dst_ifp, PFIL_OUT) != 0) { m_freem(m); return; } if (m == NULL) return; } #ifdef ALTQ KERNEL_LOCK(1, NULL); /* * If ALTQ is enabled on the member interface, do * classification; the queueing discipline might * not require classification, but might require * the address family/header pointer in the pktattr. */ if (ALTQ_IS_ENABLED(&dst_ifp->if_snd)) { /* XXX IFT_ETHER */ altq_etherclassify(&dst_ifp->if_snd, m); } KERNEL_UNLOCK_ONE(NULL); #endif /* ALTQ */ if (vlan_has_tag(m) && !vlan_is_hwtag_enabled(dst_ifp)) { (void)ether_inject_vlantag(&m, ETHERTYPE_VLAN, vlan_get_tag(m)); if (m == NULL) { if_statinc(&sc->sc_if, if_oerrors); return; } } len = m->m_pkthdr.len; mflags = m->m_flags; error = if_transmit_lock(dst_ifp, m); if (error) { /* mbuf is already freed */ if_statinc(&sc->sc_if, if_oerrors); return; } net_stat_ref_t nsr = IF_STAT_GETREF(&sc->sc_if); if_statinc_ref(&sc->sc_if, nsr, if_opackets); if_statadd_ref(&sc->sc_if, nsr, if_obytes, len); if (mflags & M_MCAST) if_statinc_ref(&sc->sc_if, nsr, if_omcasts); IF_STAT_PUTREF(&sc->sc_if); } /* * bridge_output: * * Send output from a bridge member interface. This * performs the bridging function for locally originated * packets. * * The mbuf has the Ethernet header already attached. We must * enqueue or free the mbuf before returning. */ int bridge_output(struct ifnet *ifp, struct mbuf *m, const struct sockaddr *sa, const struct rtentry *rt) { struct ether_header *eh; struct ifnet *dst_if; struct bridge_softc *sc; struct mbuf *n; int s, bound; /* * bridge_output() is called from ether_output(), furthermore * ifp argument doesn't point to bridge(4). So, don't assert * IFEF_MPSAFE here. */ KASSERT(m->m_len >= ETHER_HDR_LEN); eh = mtod(m, struct ether_header *); sc = ifp->if_bridge; if (ETHER_IS_MULTICAST(eh->ether_dhost)) { if (memcmp(etherbroadcastaddr, eh->ether_dhost, ETHER_ADDR_LEN) == 0) m->m_flags |= M_BCAST; else m->m_flags |= M_MCAST; } /* * If bridge is down, but the original output interface is up, * go ahead and send out that interface. Otherwise, the packet * is dropped below. */ if (__predict_false(sc == NULL) || (sc->sc_if.if_flags & IFF_RUNNING) == 0) { dst_if = ifp; goto unicast_asis; } /* * If the packet is a multicast, or we don't know a better way to * get there, send to all interfaces. */ if ((m->m_flags & (M_MCAST | M_BCAST)) != 0) dst_if = NULL; else dst_if = bridge_rtlookup(sc, eh->ether_dhost); /* * In general, we need to handle TX offload in software before * enqueueing a packet. However, we can send it as is in the * cases of unicast via (1) the source interface, or (2) an * interface which supports the specified offload options. * For multicast or broadcast, send it as is only if (3) all * the member interfaces support the specified options. */ /* * Unicast via the source interface. */ if (dst_if == ifp) goto unicast_asis; /* * Unicast via other interface. */ if (dst_if != NULL) { KASSERT(m->m_flags & M_PKTHDR); if (TX_OFFLOAD_SUPPORTED(dst_if->if_csum_flags_tx, m->m_pkthdr.csum_flags)) { /* * Unicast via an interface which supports the * specified offload options. */ goto unicast_asis; } /* * Handle TX offload in software. For TSO, a packet is * split into multiple chunks. Thus, the return value of * ether_sw_offload_tx() is mbuf queue consists of them. */ m = ether_sw_offload_tx(ifp, m); if (m == NULL) return 0; do { n = m->m_nextpkt; if ((dst_if->if_flags & IFF_RUNNING) == 0) m_freem(m); else bridge_enqueue(sc, dst_if, m, 0); m = n; } while (m != NULL); return 0; } /* * Multicast or broadcast. */ if (TX_OFFLOAD_SUPPORTED(sc->sc_csum_flags_tx, m->m_pkthdr.csum_flags)) { /* * Specified TX offload options are supported by all * the member interfaces of this bridge. */ m->m_nextpkt = NULL; /* XXX */ } else { /* * Otherwise, handle TX offload in software. */ m = ether_sw_offload_tx(ifp, m); if (m == NULL) return 0; } /* * When we use pppoe over bridge, bridge_output() can be called * in a lwp context by pppoe_timeout_wk(). */ bound = curlwp_bind(); do { /* XXX Should call bridge_broadcast, but there are locking * issues which need resolving first. */ struct bridge_iflist *bif; struct mbuf *mc; bool used = false; n = m->m_nextpkt; s = pserialize_read_enter(); BRIDGE_IFLIST_READER_FOREACH(bif, sc) { struct psref psref; bridge_acquire_member(sc, bif, &psref); pserialize_read_exit(s); dst_if = bif->bif_ifp; if ((dst_if->if_flags & IFF_RUNNING) == 0) goto next; /* * If this is not the original output interface, * and the interface is participating in spanning * tree, make sure the port is in a state that * allows forwarding. */ if (dst_if != ifp && (bif->bif_flags & IFBIF_STP) != 0) { switch (bif->bif_state) { case BSTP_IFSTATE_BLOCKING: case BSTP_IFSTATE_LISTENING: case BSTP_IFSTATE_DISABLED: goto next; } } if (PSLIST_READER_NEXT(bif, struct bridge_iflist, bif_next) == NULL && ((m->m_flags & (M_MCAST | M_BCAST)) == 0 || dst_if == ifp)) { used = true; mc = m; } else { mc = m_copypacket(m, M_DONTWAIT); if (mc == NULL) { if_statinc(&sc->sc_if, if_oerrors); goto next; } } bridge_enqueue(sc, dst_if, mc, 0); if ((m->m_flags & (M_MCAST | M_BCAST)) != 0 && dst_if != ifp) { if (PSLIST_READER_NEXT(bif, struct bridge_iflist, bif_next) == NULL) { used = true; mc = m; } else { mc = m_copypacket(m, M_DONTWAIT); if (mc == NULL) { if_statinc(&sc->sc_if, if_oerrors); goto next; } } m_set_rcvif(mc, dst_if); mc->m_flags &= ~M_PROMISC; const int _s = splsoftnet(); KERNEL_LOCK_UNLESS_IFP_MPSAFE(dst_if); ether_input(dst_if, mc); KERNEL_UNLOCK_UNLESS_IFP_MPSAFE(dst_if); splx(_s); } next: s = pserialize_read_enter(); bridge_release_member(sc, bif, &psref); /* Guarantee we don't re-enter the loop as we already * decided we're at the end. */ if (used) break; } pserialize_read_exit(s); if (!used) m_freem(m); m = n; } while (m != NULL); curlwp_bindx(bound); return 0; unicast_asis: /* * XXX Spanning tree consideration here? */ if ((dst_if->if_flags & IFF_RUNNING) == 0) m_freem(m); else bridge_enqueue(sc, dst_if, m, 0); return 0; } /* * bridge_start: * * Start output on a bridge. * * NOTE: This routine should never be called in this implementation. */ static void bridge_start(struct ifnet *ifp) { printf("%s: bridge_start() called\n", ifp->if_xname); } /* * bridge_forward: * * The forwarding function of the bridge. */ static void bridge_forward(struct bridge_softc *sc, struct mbuf *m) { struct bridge_iflist *bif; struct ifnet *src_if, *dst_if; struct ether_header *eh; struct psref psref; struct psref psref_src; DECLARE_LOCK_VARIABLE; bool src_if_protected; src_if = m_get_rcvif_psref(m, &psref_src); if (src_if == NULL) { /* Interface is being destroyed? */ goto discard; } if_statadd2(&sc->sc_if, if_ipackets, 1, if_ibytes, m->m_pkthdr.len); /* * Look up the bridge_iflist. */ bif = bridge_lookup_member_if(sc, src_if, &psref); if (bif == NULL) { /* Interface is not a bridge member (anymore?) */ goto discard; } if (bif->bif_flags & IFBIF_STP) { switch (bif->bif_state) { case BSTP_IFSTATE_BLOCKING: case BSTP_IFSTATE_LISTENING: case BSTP_IFSTATE_DISABLED: bridge_release_member(sc, bif, &psref); goto discard; } } eh = mtod(m, struct ether_header *); /* * If the interface is learning, and the source * address is valid and not multicast, record * the address. */ if ((bif->bif_flags & IFBIF_LEARNING) != 0 && ETHER_IS_MULTICAST(eh->ether_shost) == 0 && (eh->ether_shost[0] == 0 && eh->ether_shost[1] == 0 && eh->ether_shost[2] == 0 && eh->ether_shost[3] == 0 && eh->ether_shost[4] == 0 && eh->ether_shost[5] == 0) == 0) { (void) bridge_rtupdate(sc, eh->ether_shost, src_if, 0, IFBAF_DYNAMIC); } if ((bif->bif_flags & IFBIF_STP) != 0 && bif->bif_state == BSTP_IFSTATE_LEARNING) { bridge_release_member(sc, bif, &psref); goto discard; } src_if_protected = ((bif->bif_flags & IFBIF_PROTECTED) != 0); bridge_release_member(sc, bif, &psref); /* * At this point, the port either doesn't participate * in spanning tree or it is in the forwarding state. */ /* * If the packet is unicast, destined for someone on * "this" side of the bridge, drop it. */ if ((m->m_flags & (M_BCAST|M_MCAST)) == 0) { dst_if = bridge_rtlookup(sc, eh->ether_dhost); if (src_if == dst_if) goto discard; } else { /* ...forward it to all interfaces. */ if_statinc(&sc->sc_if, if_imcasts); dst_if = NULL; } if (pfil_run_hooks(sc->sc_if.if_pfil, &m, src_if, PFIL_IN) != 0 || m == NULL) { goto discard; } if (dst_if == NULL) { bridge_broadcast(sc, src_if, src_if_protected, m); goto out; } m_put_rcvif_psref(src_if, &psref_src); src_if = NULL; /* * At this point, we're dealing with a unicast frame * going to a different interface. */ if ((dst_if->if_flags & IFF_RUNNING) == 0) goto discard; bif = bridge_lookup_member_if(sc, dst_if, &psref); if (bif == NULL) { /* Not a member of the bridge (anymore?) */ goto discard; } if (bif->bif_flags & IFBIF_STP) { switch (bif->bif_state) { case BSTP_IFSTATE_DISABLED: case BSTP_IFSTATE_BLOCKING: bridge_release_member(sc, bif, &psref); goto discard; } } if ((bif->bif_flags & IFBIF_PROTECTED) && src_if_protected) { bridge_release_member(sc, bif, &psref); goto discard; } bridge_release_member(sc, bif, &psref); /* * Before enqueueing this packet to the destination interface, * clear any in-bound checksum flags to prevent them from being * misused as out-bound flags. */ m->m_pkthdr.csum_flags = 0; ACQUIRE_GLOBAL_LOCKS(); bridge_enqueue(sc, dst_if, m, 1); RELEASE_GLOBAL_LOCKS(); out: if (src_if != NULL) m_put_rcvif_psref(src_if, &psref_src); return; discard: m_freem(m); goto out; } static bool bstp_state_before_learning(struct bridge_iflist *bif) { if (bif->bif_flags & IFBIF_STP) { switch (bif->bif_state) { case BSTP_IFSTATE_BLOCKING: case BSTP_IFSTATE_LISTENING: case BSTP_IFSTATE_DISABLED: return true; } } return false; } static bool bridge_ourether(struct bridge_iflist *bif, struct ether_header *eh, int src) { uint8_t *ether = src ? eh->ether_shost : eh->ether_dhost; if (memcmp(CLLADDR(bif->bif_ifp->if_sadl), ether, ETHER_ADDR_LEN) == 0 #if NCARP > 0 || (bif->bif_ifp->if_carp && carp_ourether(bif->bif_ifp->if_carp, eh, IFT_ETHER, src) != NULL) #endif /* NCARP > 0 */ ) return true; return false; } /* * bridge_input: * * Receive input from a member interface. Queue the packet for * bridging if it is not for us. */ static void bridge_input(struct ifnet *ifp, struct mbuf *m) { struct bridge_softc *sc = ifp->if_bridge; struct bridge_iflist *bif; struct ether_header *eh; struct psref psref; int bound; DECLARE_LOCK_VARIABLE; KASSERT(!cpu_intr_p()); if (__predict_false(sc == NULL) || (sc->sc_if.if_flags & IFF_RUNNING) == 0) { ACQUIRE_GLOBAL_LOCKS(); ether_input(ifp, m); RELEASE_GLOBAL_LOCKS(); return; } bound = curlwp_bind(); bif = bridge_lookup_member_if(sc, ifp, &psref); if (bif == NULL) { curlwp_bindx(bound); ACQUIRE_GLOBAL_LOCKS(); ether_input(ifp, m); RELEASE_GLOBAL_LOCKS(); return; } eh = mtod(m, struct ether_header *); if (ETHER_IS_MULTICAST(eh->ether_dhost)) { if (memcmp(etherbroadcastaddr, eh->ether_dhost, ETHER_ADDR_LEN) == 0) m->m_flags |= M_BCAST; else m->m_flags |= M_MCAST; } /* * A 'fast' path for packets addressed to interfaces that are * part of this bridge. */ if (!(m->m_flags & (M_BCAST|M_MCAST)) && !bstp_state_before_learning(bif)) { struct bridge_iflist *_bif; struct ifnet *_ifp = NULL; int s; struct psref _psref; s = pserialize_read_enter(); BRIDGE_IFLIST_READER_FOREACH(_bif, sc) { /* It is destined for us. */ if (bridge_ourether(_bif, eh, 0)) { bridge_acquire_member(sc, _bif, &_psref); pserialize_read_exit(s); if (_bif->bif_flags & IFBIF_LEARNING) (void) bridge_rtupdate(sc, eh->ether_shost, ifp, 0, IFBAF_DYNAMIC); m_set_rcvif(m, _bif->bif_ifp); _ifp = _bif->bif_ifp; bridge_release_member(sc, _bif, &_psref); goto out; } /* We just received a packet that we sent out. */ if (bridge_ourether(_bif, eh, 1)) break; } pserialize_read_exit(s); out: if (_bif != NULL) { bridge_release_member(sc, bif, &psref); curlwp_bindx(bound); if (_ifp != NULL) { m->m_flags &= ~M_PROMISC; ACQUIRE_GLOBAL_LOCKS(); ether_input(_ifp, m); RELEASE_GLOBAL_LOCKS(); } else m_freem(m); return; } } /* Tap off 802.1D packets; they do not get forwarded. */ if (bif->bif_flags & IFBIF_STP && memcmp(eh->ether_dhost, bstp_etheraddr, ETHER_ADDR_LEN) == 0) { bstp_input(sc, bif, m); bridge_release_member(sc, bif, &psref); curlwp_bindx(bound); return; } /* * A normal switch would discard the packet here, but that's not what * we've done historically. This also prevents some obnoxious behaviour. */ if (bstp_state_before_learning(bif)) { bridge_release_member(sc, bif, &psref); curlwp_bindx(bound); ACQUIRE_GLOBAL_LOCKS(); ether_input(ifp, m); RELEASE_GLOBAL_LOCKS(); return; } bridge_release_member(sc, bif, &psref); bridge_forward(sc, m); curlwp_bindx(bound); } /* * bridge_broadcast: * * Send a frame to all interfaces that are members of * the bridge, except for the one on which the packet * arrived. */ static void bridge_broadcast(struct bridge_softc *sc, struct ifnet *src_if, bool src_if_protected, struct mbuf *m) { struct bridge_iflist *bif; struct mbuf *mc; struct ifnet *dst_if; bool bmcast; int s; DECLARE_LOCK_VARIABLE; bmcast = m->m_flags & (M_BCAST|M_MCAST); s = pserialize_read_enter(); BRIDGE_IFLIST_READER_FOREACH(bif, sc) { struct psref psref; bridge_acquire_member(sc, bif, &psref); pserialize_read_exit(s); dst_if = bif->bif_ifp; if (bif->bif_flags & IFBIF_STP) { switch (bif->bif_state) { case BSTP_IFSTATE_BLOCKING: case BSTP_IFSTATE_DISABLED: goto next; } } if ((bif->bif_flags & IFBIF_DISCOVER) == 0 && !bmcast) goto next; if ((dst_if->if_flags & IFF_RUNNING) == 0) goto next; if (dst_if != src_if) { if ((bif->bif_flags & IFBIF_PROTECTED) && src_if_protected) { goto next; } mc = m_copypacket(m, M_DONTWAIT); if (mc == NULL) { if_statinc(&sc->sc_if, if_oerrors); goto next; } /* * Before enqueueing this packet to the destination * interface, clear any in-bound checksum flags to * prevent them from being misused as out-bound flags. */ mc->m_pkthdr.csum_flags = 0; ACQUIRE_GLOBAL_LOCKS(); bridge_enqueue(sc, dst_if, mc, 1); RELEASE_GLOBAL_LOCKS(); } if (bmcast) { mc = m_copypacket(m, M_DONTWAIT); if (mc == NULL) { if_statinc(&sc->sc_if, if_oerrors); goto next; } /* * Before enqueueing this packet to the destination * interface, clear any in-bound checksum flags to * prevent them from being misused as out-bound flags. */ mc->m_pkthdr.csum_flags = 0; m_set_rcvif(mc, dst_if); mc->m_flags &= ~M_PROMISC; ACQUIRE_GLOBAL_LOCKS(); ether_input(dst_if, mc); RELEASE_GLOBAL_LOCKS(); } next: s = pserialize_read_enter(); bridge_release_member(sc, bif, &psref); } pserialize_read_exit(s); m_freem(m); } static int bridge_rtalloc(struct bridge_softc *sc, const uint8_t *dst, struct bridge_rtnode **brtp) { struct bridge_rtnode *brt; int error; if (sc->sc_brtcnt >= sc->sc_brtmax) return ENOSPC; /* * Allocate a new bridge forwarding node, and * initialize the expiration time and Ethernet * address. */ brt = pool_get(&bridge_rtnode_pool, PR_NOWAIT); if (brt == NULL) return ENOMEM; memset(brt, 0, sizeof(*brt)); brt->brt_expire = time_uptime + sc->sc_brttimeout; brt->brt_flags = IFBAF_DYNAMIC; memcpy(brt->brt_addr, dst, ETHER_ADDR_LEN); PSLIST_ENTRY_INIT(brt, brt_list); PSLIST_ENTRY_INIT(brt, brt_hash); BRIDGE_RT_LOCK(sc); error = bridge_rtnode_insert(sc, brt); BRIDGE_RT_UNLOCK(sc); if (error != 0) { pool_put(&bridge_rtnode_pool, brt); return error; } *brtp = brt; return 0; } /* * bridge_rtupdate: * * Add a bridge routing entry. */ static int bridge_rtupdate(struct bridge_softc *sc, const uint8_t *dst, struct ifnet *dst_if, int setflags, uint8_t flags) { struct bridge_rtnode *brt; int s; again: /* * A route for this destination might already exist. If so, * update it, otherwise create a new one. */ s = pserialize_read_enter(); brt = bridge_rtnode_lookup(sc, dst); if (brt != NULL) { brt->brt_ifp = dst_if; if (setflags) { brt->brt_flags = flags; if (flags & IFBAF_STATIC) brt->brt_expire = 0; else brt->brt_expire = time_uptime + sc->sc_brttimeout; } else { if ((brt->brt_flags & IFBAF_TYPEMASK) == IFBAF_DYNAMIC) brt->brt_expire = time_uptime + sc->sc_brttimeout; } } pserialize_read_exit(s); if (brt == NULL) { int r; r = bridge_rtalloc(sc, dst, &brt); if (r != 0) return r; goto again; } return 0; } /* * bridge_rtlookup: * * Lookup the destination interface for an address. */ static struct ifnet * bridge_rtlookup(struct bridge_softc *sc, const uint8_t *addr) { struct bridge_rtnode *brt; struct ifnet *ifs = NULL; int s; s = pserialize_read_enter(); brt = bridge_rtnode_lookup(sc, addr); if (brt != NULL) ifs = brt->brt_ifp; pserialize_read_exit(s); return ifs; } typedef bool (*bridge_iterate_cb_t) (struct bridge_softc *, struct bridge_rtnode *, bool *, void *); /* * bridge_rtlist_iterate_remove: * * It iterates on sc->sc_rtlist and removes rtnodes of it which func * callback judges to remove. Removals of rtnodes are done in a manner * of pserialize. To this end, all kmem_* operations are placed out of * mutexes. */ static void bridge_rtlist_iterate_remove(struct bridge_softc *sc, bridge_iterate_cb_t func, void *arg) { struct bridge_rtnode *brt; struct bridge_rtnode **brt_list; int i, count; retry: count = sc->sc_brtcnt; if (count == 0) return; brt_list = kmem_alloc(sizeof(*brt_list) * count, KM_SLEEP); BRIDGE_RT_LOCK(sc); if (__predict_false(sc->sc_brtcnt > count)) { /* The rtnodes increased, we need more memory */ BRIDGE_RT_UNLOCK(sc); kmem_free(brt_list, sizeof(*brt_list) * count); goto retry; } i = 0; /* * We don't need to use a _SAFE variant here because we know * that a removed item keeps its next pointer as-is thanks to * pslist(9) and isn't freed in the loop. */ BRIDGE_RTLIST_WRITER_FOREACH(brt, sc) { bool need_break = false; if (func(sc, brt, &need_break, arg)) { bridge_rtnode_remove(sc, brt); brt_list[i++] = brt; } if (need_break) break; } if (i > 0) BRIDGE_RT_PSZ_PERFORM(sc); BRIDGE_RT_UNLOCK(sc); while (--i >= 0) bridge_rtnode_destroy(brt_list[i]); kmem_free(brt_list, sizeof(*brt_list) * count); } static bool bridge_rttrim0_cb(struct bridge_softc *sc, struct bridge_rtnode *brt, bool *need_break, void *arg) { if ((brt->brt_flags & IFBAF_TYPEMASK) == IFBAF_DYNAMIC) { /* Take into account of the subsequent removal */ if ((sc->sc_brtcnt - 1) <= sc->sc_brtmax) *need_break = true; return true; } else return false; } static void bridge_rttrim0(struct bridge_softc *sc) { bridge_rtlist_iterate_remove(sc, bridge_rttrim0_cb, NULL); } /* * bridge_rttrim: * * Trim the routine table so that we have a number * of routing entries less than or equal to the * maximum number. */ static void bridge_rttrim(struct bridge_softc *sc) { /* Make sure we actually need to do this. */ if (sc->sc_brtcnt <= sc->sc_brtmax) return; /* Force an aging cycle; this might trim enough addresses. */ bridge_rtage(sc); if (sc->sc_brtcnt <= sc->sc_brtmax) return; bridge_rttrim0(sc); return; } /* * bridge_timer: * * Aging timer for the bridge. */ static void bridge_timer(void *arg) { struct bridge_softc *sc = arg; workqueue_enqueue(sc->sc_rtage_wq, &sc->sc_rtage_wk, NULL); } static void bridge_rtage_work(struct work *wk, void *arg) { struct bridge_softc *sc = arg; KASSERT(wk == &sc->sc_rtage_wk); bridge_rtage(sc); if (sc->sc_if.if_flags & IFF_RUNNING) callout_reset(&sc->sc_brcallout, bridge_rtable_prune_period * hz, bridge_timer, sc); } static bool bridge_rtage_cb(struct bridge_softc *sc, struct bridge_rtnode *brt, bool *need_break, void *arg) { if ((brt->brt_flags & IFBAF_TYPEMASK) == IFBAF_DYNAMIC && time_uptime >= brt->brt_expire) return true; else return false; } /* * bridge_rtage: * * Perform an aging cycle. */ static void bridge_rtage(struct bridge_softc *sc) { bridge_rtlist_iterate_remove(sc, bridge_rtage_cb, NULL); } static bool bridge_rtflush_cb(struct bridge_softc *sc, struct bridge_rtnode *brt, bool *need_break, void *arg) { int full = *(int*)arg; if (full || (brt->brt_flags & IFBAF_TYPEMASK) == IFBAF_DYNAMIC) return true; else return false; } /* * bridge_rtflush: * * Remove all dynamic addresses from the bridge. */ static void bridge_rtflush(struct bridge_softc *sc, int full) { bridge_rtlist_iterate_remove(sc, bridge_rtflush_cb, &full); } /* * bridge_rtdaddr: * * Remove an address from the table. */ static int bridge_rtdaddr(struct bridge_softc *sc, const uint8_t *addr) { struct bridge_rtnode *brt; BRIDGE_RT_LOCK(sc); if ((brt = bridge_rtnode_lookup(sc, addr)) == NULL) { BRIDGE_RT_UNLOCK(sc); return ENOENT; } bridge_rtnode_remove(sc, brt); BRIDGE_RT_PSZ_PERFORM(sc); BRIDGE_RT_UNLOCK(sc); bridge_rtnode_destroy(brt); return 0; } /* * bridge_rtdelete: * * Delete routes to a speicifc member interface. */ static void bridge_rtdelete(struct bridge_softc *sc, struct ifnet *ifp) { struct bridge_rtnode *brt; /* XXX pserialize_perform for each entry is slow */ again: BRIDGE_RT_LOCK(sc); BRIDGE_RTLIST_WRITER_FOREACH(brt, sc) { if (brt->brt_ifp == ifp) break; } if (brt == NULL) { BRIDGE_RT_UNLOCK(sc); return; } bridge_rtnode_remove(sc, brt); BRIDGE_RT_PSZ_PERFORM(sc); BRIDGE_RT_UNLOCK(sc); bridge_rtnode_destroy(brt); goto again; } /* * bridge_rtable_init: * * Initialize the route table for this bridge. */ static void bridge_rtable_init(struct bridge_softc *sc) { int i; sc->sc_rthash = kmem_alloc(sizeof(*sc->sc_rthash) * BRIDGE_RTHASH_SIZE, KM_SLEEP); for (i = 0; i < BRIDGE_RTHASH_SIZE; i++) PSLIST_INIT(&sc->sc_rthash[i]); sc->sc_rthash_key = cprng_fast32(); PSLIST_INIT(&sc->sc_rtlist); sc->sc_rtlist_psz = pserialize_create(); sc->sc_rtlist_lock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_SOFTNET); } /* * bridge_rtable_fini: * * Deconstruct the route table for this bridge. */ static void bridge_rtable_fini(struct bridge_softc *sc) { kmem_free(sc->sc_rthash, sizeof(*sc->sc_rthash) * BRIDGE_RTHASH_SIZE); mutex_obj_free(sc->sc_rtlist_lock); pserialize_destroy(sc->sc_rtlist_psz); } /* * The following hash function is adapted from "Hash Functions" by Bob Jenkins * ("Algorithm Alley", Dr. Dobbs Journal, September 1997). */ #define mix(a, b, c) \ do { \ a -= b; a -= c; a ^= (c >> 13); \ b -= c; b -= a; b ^= (a << 8); \ c -= a; c -= b; c ^= (b >> 13); \ a -= b; a -= c; a ^= (c >> 12); \ b -= c; b -= a; b ^= (a << 16); \ c -= a; c -= b; c ^= (b >> 5); \ a -= b; a -= c; a ^= (c >> 3); \ b -= c; b -= a; b ^= (a << 10); \ c -= a; c -= b; c ^= (b >> 15); \ } while (/*CONSTCOND*/0) static inline uint32_t bridge_rthash(struct bridge_softc *sc, const uint8_t *addr) { uint32_t a = 0x9e3779b9, b = 0x9e3779b9, c = sc->sc_rthash_key; b += addr[5] << 8; b += addr[4]; a += (uint32_t)addr[3] << 24; a += addr[2] << 16; a += addr[1] << 8; a += addr[0]; mix(a, b, c); return (c & BRIDGE_RTHASH_MASK); } #undef mix /* * bridge_rtnode_lookup: * * Look up a bridge route node for the specified destination. */ static struct bridge_rtnode * bridge_rtnode_lookup(struct bridge_softc *sc, const uint8_t *addr) { struct bridge_rtnode *brt; uint32_t hash; int dir; hash = bridge_rthash(sc, addr); BRIDGE_RTHASH_READER_FOREACH(brt, sc, hash) { dir = memcmp(addr, brt->brt_addr, ETHER_ADDR_LEN); if (dir == 0) return brt; if (dir > 0) return NULL; } return NULL; } /* * bridge_rtnode_insert: * * Insert the specified bridge node into the route table. We * assume the entry is not already in the table. */ static int bridge_rtnode_insert(struct bridge_softc *sc, struct bridge_rtnode *brt) { struct bridge_rtnode *lbrt, *prev = NULL; uint32_t hash; KASSERT(BRIDGE_RT_LOCKED(sc)); hash = bridge_rthash(sc, brt->brt_addr); BRIDGE_RTHASH_WRITER_FOREACH(lbrt, sc, hash) { int dir = memcmp(brt->brt_addr, lbrt->brt_addr, ETHER_ADDR_LEN); if (dir == 0) return EEXIST; if (dir > 0) break; prev = lbrt; } if (prev == NULL) BRIDGE_RTHASH_WRITER_INSERT_HEAD(sc, hash, brt); else BRIDGE_RTHASH_WRITER_INSERT_AFTER(prev, brt); BRIDGE_RTLIST_WRITER_INSERT_HEAD(sc, brt); sc->sc_brtcnt++; return 0; } /* * bridge_rtnode_remove: * * Remove a bridge rtnode from the rthash and the rtlist of a bridge. */ static void bridge_rtnode_remove(struct bridge_softc *sc, struct bridge_rtnode *brt) { KASSERT(BRIDGE_RT_LOCKED(sc)); BRIDGE_RTHASH_WRITER_REMOVE(brt); BRIDGE_RTLIST_WRITER_REMOVE(brt); sc->sc_brtcnt--; } /* * bridge_rtnode_destroy: * * Destroy a bridge rtnode. */ static void bridge_rtnode_destroy(struct bridge_rtnode *brt) { PSLIST_ENTRY_DESTROY(brt, brt_list); PSLIST_ENTRY_DESTROY(brt, brt_hash); pool_put(&bridge_rtnode_pool, brt); } extern pfil_head_t *inet_pfil_hook; /* XXX */ extern pfil_head_t *inet6_pfil_hook; /* XXX */ /* * Send bridge packets through IPF if they are one of the types IPF can deal * with, or if they are ARP or REVARP. (IPF will pass ARP and REVARP without * question.) */ static int bridge_ipf(void *arg, struct mbuf **mp, struct ifnet *ifp, int dir) { int snap, error; struct ether_header *eh1, eh2; struct llc llc1; uint16_t ether_type; snap = 0; error = -1; /* Default error if not error == 0 */ eh1 = mtod(*mp, struct ether_header *); ether_type = ntohs(eh1->ether_type); /* * Check for SNAP/LLC. */ if (ether_type < ETHERMTU) { struct llc *llc2 = (struct llc *)(eh1 + 1); if ((*mp)->m_len >= ETHER_HDR_LEN + 8 && llc2->llc_dsap == LLC_SNAP_LSAP && llc2->llc_ssap == LLC_SNAP_LSAP && llc2->llc_control == LLC_UI) { ether_type = htons(llc2->llc_un.type_snap.ether_type); snap = 1; } } /* drop VLAN traffic untagged by hardware offloading */ if (vlan_has_tag(*mp)) goto bad; /* * If we're trying to filter bridge traffic, don't look at anything * other than IP and ARP traffic. If the filter doesn't understand * IPv6, don't allow IPv6 through the bridge either. This is lame * since if we really wanted, say, an AppleTalk filter, we are hosed, * but of course we don't have an AppleTalk filter to begin with. * (Note that since IPF doesn't understand ARP it will pass *ALL* * ARP traffic.) */ switch (ether_type) { case ETHERTYPE_ARP: case ETHERTYPE_REVARP: return 0; /* Automatically pass */ case ETHERTYPE_IP: # ifdef INET6 case ETHERTYPE_IPV6: # endif /* INET6 */ break; default: goto bad; } /* Strip off the Ethernet header and keep a copy. */ m_copydata(*mp, 0, ETHER_HDR_LEN, (void *) &eh2); m_adj(*mp, ETHER_HDR_LEN); /* Strip off snap header, if present */ if (snap) { m_copydata(*mp, 0, sizeof(struct llc), (void *) &llc1); m_adj(*mp, sizeof(struct llc)); } /* * Check basic packet sanity and run IPF through pfil. */ KASSERT(!cpu_intr_p()); switch (ether_type) { case ETHERTYPE_IP : error = bridge_ip_checkbasic(mp); if (error == 0) error = pfil_run_hooks(inet_pfil_hook, mp, ifp, dir); break; # ifdef INET6 case ETHERTYPE_IPV6 : error = bridge_ip6_checkbasic(mp); if (error == 0) error = pfil_run_hooks(inet6_pfil_hook, mp, ifp, dir); break; # endif default : error = 0; break; } if (*mp == NULL) return error; if (error != 0) goto bad; error = -1; /* * Finally, put everything back the way it was and return */ if (snap) { M_PREPEND(*mp, sizeof(struct llc), M_DONTWAIT); if (*mp == NULL) return error; bcopy(&llc1, mtod(*mp, void *), sizeof(struct llc)); } M_PREPEND(*mp, ETHER_HDR_LEN, M_DONTWAIT); if (*mp == NULL) return error; bcopy(&eh2, mtod(*mp, void *), ETHER_HDR_LEN); return 0; bad: m_freem(*mp); *mp = NULL; return error; } /* * Perform basic checks on header size since * IPF assumes ip_input has already processed * it for it. Cut-and-pasted from ip_input.c. * Given how simple the IPv6 version is, * does the IPv4 version really need to be * this complicated? * * XXX Should we update ipstat here, or not? * XXX Right now we update ipstat but not * XXX csum_counter. */ static int bridge_ip_checkbasic(struct mbuf **mp) { struct mbuf *m = *mp; struct ip *ip; int len, hlen; if (*mp == NULL) return -1; if (M_GET_ALIGNED_HDR(&m, struct ip, true) != 0) { /* XXXJRT new stat, please */ ip_statinc(IP_STAT_TOOSMALL); goto bad; } ip = mtod(m, struct ip *); if (ip == NULL) goto bad; if (ip->ip_v != IPVERSION) { ip_statinc(IP_STAT_BADVERS); goto bad; } hlen = ip->ip_hl << 2; if (hlen < sizeof(struct ip)) { /* minimum header length */ ip_statinc(IP_STAT_BADHLEN); goto bad; } if (hlen > m->m_len) { if ((m = m_pullup(m, hlen)) == 0) { ip_statinc(IP_STAT_BADHLEN); goto bad; } ip = mtod(m, struct ip *); if (ip == NULL) goto bad; } switch (m->m_pkthdr.csum_flags & ((m_get_rcvif_NOMPSAFE(m)->if_csum_flags_rx & M_CSUM_IPv4) | M_CSUM_IPv4_BAD)) { case M_CSUM_IPv4|M_CSUM_IPv4_BAD: /* INET_CSUM_COUNTER_INCR(&ip_hwcsum_bad); */ goto bad; case M_CSUM_IPv4: /* Checksum was okay. */ /* INET_CSUM_COUNTER_INCR(&ip_hwcsum_ok); */ break; default: /* Must compute it ourselves. */ /* INET_CSUM_COUNTER_INCR(&ip_swcsum); */ if (in_cksum(m, hlen) != 0) goto bad; break; } /* Retrieve the packet length. */ len = ntohs(ip->ip_len); /* * Check for additional length bogosity */ if (len < hlen) { ip_statinc(IP_STAT_BADLEN); goto bad; } /* * Check that the amount of data in the buffers * is as at least much as the IP header would have us expect. * Drop packet if shorter than we expect. */ if (m->m_pkthdr.len < len) { ip_statinc(IP_STAT_TOOSHORT); goto bad; } /* Checks out, proceed */ *mp = m; return 0; bad: *mp = m; return -1; } # ifdef INET6 /* * Same as above, but for IPv6. * Cut-and-pasted from ip6_input.c. * XXX Should we update ip6stat, or not? */ static int bridge_ip6_checkbasic(struct mbuf **mp) { struct mbuf *m = *mp; struct ip6_hdr *ip6; /* * If the IPv6 header is not aligned, slurp it up into a new * mbuf with space for link headers, in the event we forward * it. Otherwise, if it is aligned, make sure the entire base * IPv6 header is in the first mbuf of the chain. */ if (M_GET_ALIGNED_HDR(&m, struct ip6_hdr, true) != 0) { struct ifnet *inifp = m_get_rcvif_NOMPSAFE(m); /* XXXJRT new stat, please */ ip6_statinc(IP6_STAT_TOOSMALL); in6_ifstat_inc(inifp, ifs6_in_hdrerr); goto bad; } ip6 = mtod(m, struct ip6_hdr *); if ((ip6->ip6_vfc & IPV6_VERSION_MASK) != IPV6_VERSION) { ip6_statinc(IP6_STAT_BADVERS); in6_ifstat_inc(m_get_rcvif_NOMPSAFE(m), ifs6_in_hdrerr); goto bad; } /* Checks out, proceed */ *mp = m; return 0; bad: *mp = m; return -1; } # endif /* INET6 */