123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825 |
- // SPDX-License-Identifier: GPL-2.0
- #include <linux/kernel.h>
- #include <linux/list.h>
- #include <linux/netdevice.h>
- #include <linux/rtnetlink.h>
- #include <linux/skbuff.h>
- #include <net/ip.h>
- #include <net/switchdev.h>
- #include "br_private.h"
- static struct static_key_false br_switchdev_tx_fwd_offload;
- static bool nbp_switchdev_can_offload_tx_fwd(const struct net_bridge_port *p,
- const struct sk_buff *skb)
- {
- if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
- return false;
- return (p->flags & BR_TX_FWD_OFFLOAD) &&
- (p->hwdom != BR_INPUT_SKB_CB(skb)->src_hwdom);
- }
- bool br_switchdev_frame_uses_tx_fwd_offload(struct sk_buff *skb)
- {
- if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
- return false;
- return BR_INPUT_SKB_CB(skb)->tx_fwd_offload;
- }
- void br_switchdev_frame_set_offload_fwd_mark(struct sk_buff *skb)
- {
- skb->offload_fwd_mark = br_switchdev_frame_uses_tx_fwd_offload(skb);
- }
- /* Mark the frame for TX forwarding offload if this egress port supports it */
- void nbp_switchdev_frame_mark_tx_fwd_offload(const struct net_bridge_port *p,
- struct sk_buff *skb)
- {
- if (nbp_switchdev_can_offload_tx_fwd(p, skb))
- BR_INPUT_SKB_CB(skb)->tx_fwd_offload = true;
- }
- /* Lazily adds the hwdom of the egress bridge port to the bit mask of hwdoms
- * that the skb has been already forwarded to, to avoid further cloning to
- * other ports in the same hwdom by making nbp_switchdev_allowed_egress()
- * return false.
- */
- void nbp_switchdev_frame_mark_tx_fwd_to_hwdom(const struct net_bridge_port *p,
- struct sk_buff *skb)
- {
- if (nbp_switchdev_can_offload_tx_fwd(p, skb))
- set_bit(p->hwdom, &BR_INPUT_SKB_CB(skb)->fwd_hwdoms);
- }
- void nbp_switchdev_frame_mark(const struct net_bridge_port *p,
- struct sk_buff *skb)
- {
- if (p->hwdom)
- BR_INPUT_SKB_CB(skb)->src_hwdom = p->hwdom;
- }
- bool nbp_switchdev_allowed_egress(const struct net_bridge_port *p,
- const struct sk_buff *skb)
- {
- struct br_input_skb_cb *cb = BR_INPUT_SKB_CB(skb);
- return !test_bit(p->hwdom, &cb->fwd_hwdoms) &&
- (!skb->offload_fwd_mark || cb->src_hwdom != p->hwdom);
- }
- /* Flags that can be offloaded to hardware */
- #define BR_PORT_FLAGS_HW_OFFLOAD (BR_LEARNING | BR_FLOOD | \
- BR_MCAST_FLOOD | BR_BCAST_FLOOD | BR_PORT_LOCKED | \
- BR_HAIRPIN_MODE | BR_ISOLATED | BR_MULTICAST_TO_UNICAST)
- int br_switchdev_set_port_flag(struct net_bridge_port *p,
- unsigned long flags,
- unsigned long mask,
- struct netlink_ext_ack *extack)
- {
- struct switchdev_attr attr = {
- .orig_dev = p->dev,
- };
- struct switchdev_notifier_port_attr_info info = {
- .attr = &attr,
- };
- int err;
- mask &= BR_PORT_FLAGS_HW_OFFLOAD;
- if (!mask)
- return 0;
- attr.id = SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS;
- attr.u.brport_flags.val = flags;
- attr.u.brport_flags.mask = mask;
- /* We run from atomic context here */
- err = call_switchdev_notifiers(SWITCHDEV_PORT_ATTR_SET, p->dev,
- &info.info, extack);
- err = notifier_to_errno(err);
- if (err == -EOPNOTSUPP)
- return 0;
- if (err) {
- if (extack && !extack->_msg)
- NL_SET_ERR_MSG_MOD(extack,
- "bridge flag offload is not supported");
- return -EOPNOTSUPP;
- }
- attr.id = SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS;
- attr.flags = SWITCHDEV_F_DEFER;
- err = switchdev_port_attr_set(p->dev, &attr, extack);
- if (err) {
- if (extack && !extack->_msg)
- NL_SET_ERR_MSG_MOD(extack,
- "error setting offload flag on port");
- return err;
- }
- return 0;
- }
- static void br_switchdev_fdb_populate(struct net_bridge *br,
- struct switchdev_notifier_fdb_info *item,
- const struct net_bridge_fdb_entry *fdb,
- const void *ctx)
- {
- const struct net_bridge_port *p = READ_ONCE(fdb->dst);
- item->addr = fdb->key.addr.addr;
- item->vid = fdb->key.vlan_id;
- item->added_by_user = test_bit(BR_FDB_ADDED_BY_USER, &fdb->flags);
- item->offloaded = test_bit(BR_FDB_OFFLOADED, &fdb->flags);
- item->is_local = test_bit(BR_FDB_LOCAL, &fdb->flags);
- item->info.dev = (!p || item->is_local) ? br->dev : p->dev;
- item->info.ctx = ctx;
- }
- void
- br_switchdev_fdb_notify(struct net_bridge *br,
- const struct net_bridge_fdb_entry *fdb, int type)
- {
- struct switchdev_notifier_fdb_info item;
- /* Entries with these flags were created using ndm_state == NUD_REACHABLE,
- * ndm_flags == NTF_MASTER( | NTF_STICKY), ext_flags == 0 by something
- * equivalent to 'bridge fdb add ... master dynamic (sticky)'.
- * Drivers don't know how to deal with these, so don't notify them to
- * avoid confusing them.
- */
- if (test_bit(BR_FDB_ADDED_BY_USER, &fdb->flags) &&
- !test_bit(BR_FDB_STATIC, &fdb->flags) &&
- !test_bit(BR_FDB_ADDED_BY_EXT_LEARN, &fdb->flags))
- return;
- br_switchdev_fdb_populate(br, &item, fdb, NULL);
- switch (type) {
- case RTM_DELNEIGH:
- call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_DEVICE,
- item.info.dev, &item.info, NULL);
- break;
- case RTM_NEWNEIGH:
- call_switchdev_notifiers(SWITCHDEV_FDB_ADD_TO_DEVICE,
- item.info.dev, &item.info, NULL);
- break;
- }
- }
- int br_switchdev_port_vlan_add(struct net_device *dev, u16 vid, u16 flags,
- bool changed, struct netlink_ext_ack *extack)
- {
- struct switchdev_obj_port_vlan v = {
- .obj.orig_dev = dev,
- .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
- .flags = flags,
- .vid = vid,
- .changed = changed,
- };
- return switchdev_port_obj_add(dev, &v.obj, extack);
- }
- int br_switchdev_port_vlan_del(struct net_device *dev, u16 vid)
- {
- struct switchdev_obj_port_vlan v = {
- .obj.orig_dev = dev,
- .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
- .vid = vid,
- };
- return switchdev_port_obj_del(dev, &v.obj);
- }
- static int nbp_switchdev_hwdom_set(struct net_bridge_port *joining)
- {
- struct net_bridge *br = joining->br;
- struct net_bridge_port *p;
- int hwdom;
- /* joining is yet to be added to the port list. */
- list_for_each_entry(p, &br->port_list, list) {
- if (netdev_phys_item_id_same(&joining->ppid, &p->ppid)) {
- joining->hwdom = p->hwdom;
- return 0;
- }
- }
- hwdom = find_next_zero_bit(&br->busy_hwdoms, BR_HWDOM_MAX, 1);
- if (hwdom >= BR_HWDOM_MAX)
- return -EBUSY;
- set_bit(hwdom, &br->busy_hwdoms);
- joining->hwdom = hwdom;
- return 0;
- }
- static void nbp_switchdev_hwdom_put(struct net_bridge_port *leaving)
- {
- struct net_bridge *br = leaving->br;
- struct net_bridge_port *p;
- /* leaving is no longer in the port list. */
- list_for_each_entry(p, &br->port_list, list) {
- if (p->hwdom == leaving->hwdom)
- return;
- }
- clear_bit(leaving->hwdom, &br->busy_hwdoms);
- }
- static int nbp_switchdev_add(struct net_bridge_port *p,
- struct netdev_phys_item_id ppid,
- bool tx_fwd_offload,
- struct netlink_ext_ack *extack)
- {
- int err;
- if (p->offload_count) {
- /* Prevent unsupported configurations such as a bridge port
- * which is a bonding interface, and the member ports are from
- * different hardware switches.
- */
- if (!netdev_phys_item_id_same(&p->ppid, &ppid)) {
- NL_SET_ERR_MSG_MOD(extack,
- "Same bridge port cannot be offloaded by two physical switches");
- return -EBUSY;
- }
- /* Tolerate drivers that call switchdev_bridge_port_offload()
- * more than once for the same bridge port, such as when the
- * bridge port is an offloaded bonding/team interface.
- */
- p->offload_count++;
- return 0;
- }
- p->ppid = ppid;
- p->offload_count = 1;
- err = nbp_switchdev_hwdom_set(p);
- if (err)
- return err;
- if (tx_fwd_offload) {
- p->flags |= BR_TX_FWD_OFFLOAD;
- static_branch_inc(&br_switchdev_tx_fwd_offload);
- }
- return 0;
- }
- static void nbp_switchdev_del(struct net_bridge_port *p)
- {
- if (WARN_ON(!p->offload_count))
- return;
- p->offload_count--;
- if (p->offload_count)
- return;
- if (p->hwdom)
- nbp_switchdev_hwdom_put(p);
- if (p->flags & BR_TX_FWD_OFFLOAD) {
- p->flags &= ~BR_TX_FWD_OFFLOAD;
- static_branch_dec(&br_switchdev_tx_fwd_offload);
- }
- }
- static int
- br_switchdev_fdb_replay_one(struct net_bridge *br, struct notifier_block *nb,
- const struct net_bridge_fdb_entry *fdb,
- unsigned long action, const void *ctx)
- {
- struct switchdev_notifier_fdb_info item;
- int err;
- br_switchdev_fdb_populate(br, &item, fdb, ctx);
- err = nb->notifier_call(nb, action, &item);
- return notifier_to_errno(err);
- }
- static int
- br_switchdev_fdb_replay(const struct net_device *br_dev, const void *ctx,
- bool adding, struct notifier_block *nb)
- {
- struct net_bridge_fdb_entry *fdb;
- struct net_bridge *br;
- unsigned long action;
- int err = 0;
- if (!nb)
- return 0;
- if (!netif_is_bridge_master(br_dev))
- return -EINVAL;
- br = netdev_priv(br_dev);
- if (adding)
- action = SWITCHDEV_FDB_ADD_TO_DEVICE;
- else
- action = SWITCHDEV_FDB_DEL_TO_DEVICE;
- rcu_read_lock();
- hlist_for_each_entry_rcu(fdb, &br->fdb_list, fdb_node) {
- err = br_switchdev_fdb_replay_one(br, nb, fdb, action, ctx);
- if (err)
- break;
- }
- rcu_read_unlock();
- return err;
- }
- static int br_switchdev_vlan_attr_replay(struct net_device *br_dev,
- const void *ctx,
- struct notifier_block *nb,
- struct netlink_ext_ack *extack)
- {
- struct switchdev_notifier_port_attr_info attr_info = {
- .info = {
- .dev = br_dev,
- .extack = extack,
- .ctx = ctx,
- },
- };
- struct net_bridge *br = netdev_priv(br_dev);
- struct net_bridge_vlan_group *vg;
- struct switchdev_attr attr;
- struct net_bridge_vlan *v;
- int err;
- attr_info.attr = &attr;
- attr.orig_dev = br_dev;
- vg = br_vlan_group(br);
- if (!vg)
- return 0;
- list_for_each_entry(v, &vg->vlan_list, vlist) {
- if (v->msti) {
- attr.id = SWITCHDEV_ATTR_ID_VLAN_MSTI;
- attr.u.vlan_msti.vid = v->vid;
- attr.u.vlan_msti.msti = v->msti;
- err = nb->notifier_call(nb, SWITCHDEV_PORT_ATTR_SET,
- &attr_info);
- err = notifier_to_errno(err);
- if (err)
- return err;
- }
- }
- return 0;
- }
- static int
- br_switchdev_vlan_replay_one(struct notifier_block *nb,
- struct net_device *dev,
- struct switchdev_obj_port_vlan *vlan,
- const void *ctx, unsigned long action,
- struct netlink_ext_ack *extack)
- {
- struct switchdev_notifier_port_obj_info obj_info = {
- .info = {
- .dev = dev,
- .extack = extack,
- .ctx = ctx,
- },
- .obj = &vlan->obj,
- };
- int err;
- err = nb->notifier_call(nb, action, &obj_info);
- return notifier_to_errno(err);
- }
- static int br_switchdev_vlan_replay_group(struct notifier_block *nb,
- struct net_device *dev,
- struct net_bridge_vlan_group *vg,
- const void *ctx, unsigned long action,
- struct netlink_ext_ack *extack)
- {
- struct net_bridge_vlan *v;
- int err = 0;
- u16 pvid;
- if (!vg)
- return 0;
- pvid = br_get_pvid(vg);
- list_for_each_entry(v, &vg->vlan_list, vlist) {
- struct switchdev_obj_port_vlan vlan = {
- .obj.orig_dev = dev,
- .obj.id = SWITCHDEV_OBJ_ID_PORT_VLAN,
- .flags = br_vlan_flags(v, pvid),
- .vid = v->vid,
- };
- if (!br_vlan_should_use(v))
- continue;
- err = br_switchdev_vlan_replay_one(nb, dev, &vlan, ctx,
- action, extack);
- if (err)
- return err;
- }
- return 0;
- }
- static int br_switchdev_vlan_replay(struct net_device *br_dev,
- const void *ctx, bool adding,
- struct notifier_block *nb,
- struct netlink_ext_ack *extack)
- {
- struct net_bridge *br = netdev_priv(br_dev);
- struct net_bridge_port *p;
- unsigned long action;
- int err;
- ASSERT_RTNL();
- if (!nb)
- return 0;
- if (!netif_is_bridge_master(br_dev))
- return -EINVAL;
- if (adding)
- action = SWITCHDEV_PORT_OBJ_ADD;
- else
- action = SWITCHDEV_PORT_OBJ_DEL;
- err = br_switchdev_vlan_replay_group(nb, br_dev, br_vlan_group(br),
- ctx, action, extack);
- if (err)
- return err;
- list_for_each_entry(p, &br->port_list, list) {
- struct net_device *dev = p->dev;
- err = br_switchdev_vlan_replay_group(nb, dev,
- nbp_vlan_group(p),
- ctx, action, extack);
- if (err)
- return err;
- }
- if (adding) {
- err = br_switchdev_vlan_attr_replay(br_dev, ctx, nb, extack);
- if (err)
- return err;
- }
- return 0;
- }
- #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
- struct br_switchdev_mdb_complete_info {
- struct net_bridge_port *port;
- struct br_ip ip;
- };
- static void br_switchdev_mdb_complete(struct net_device *dev, int err, void *priv)
- {
- struct br_switchdev_mdb_complete_info *data = priv;
- struct net_bridge_port_group __rcu **pp;
- struct net_bridge_port_group *p;
- struct net_bridge_mdb_entry *mp;
- struct net_bridge_port *port = data->port;
- struct net_bridge *br = port->br;
- if (err)
- goto err;
- spin_lock_bh(&br->multicast_lock);
- mp = br_mdb_ip_get(br, &data->ip);
- if (!mp)
- goto out;
- for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
- pp = &p->next) {
- if (p->key.port != port)
- continue;
- p->flags |= MDB_PG_FLAGS_OFFLOAD;
- }
- out:
- spin_unlock_bh(&br->multicast_lock);
- err:
- kfree(priv);
- }
- static void br_switchdev_mdb_populate(struct switchdev_obj_port_mdb *mdb,
- const struct net_bridge_mdb_entry *mp)
- {
- if (mp->addr.proto == htons(ETH_P_IP))
- ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
- #if IS_ENABLED(CONFIG_IPV6)
- else if (mp->addr.proto == htons(ETH_P_IPV6))
- ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
- #endif
- else
- ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
- mdb->vid = mp->addr.vid;
- }
- static void br_switchdev_host_mdb_one(struct net_device *dev,
- struct net_device *lower_dev,
- struct net_bridge_mdb_entry *mp,
- int type)
- {
- struct switchdev_obj_port_mdb mdb = {
- .obj = {
- .id = SWITCHDEV_OBJ_ID_HOST_MDB,
- .flags = SWITCHDEV_F_DEFER,
- .orig_dev = dev,
- },
- };
- br_switchdev_mdb_populate(&mdb, mp);
- switch (type) {
- case RTM_NEWMDB:
- switchdev_port_obj_add(lower_dev, &mdb.obj, NULL);
- break;
- case RTM_DELMDB:
- switchdev_port_obj_del(lower_dev, &mdb.obj);
- break;
- }
- }
- static void br_switchdev_host_mdb(struct net_device *dev,
- struct net_bridge_mdb_entry *mp, int type)
- {
- struct net_device *lower_dev;
- struct list_head *iter;
- netdev_for_each_lower_dev(dev, lower_dev, iter)
- br_switchdev_host_mdb_one(dev, lower_dev, mp, type);
- }
- static int
- br_switchdev_mdb_replay_one(struct notifier_block *nb, struct net_device *dev,
- const struct switchdev_obj_port_mdb *mdb,
- unsigned long action, const void *ctx,
- struct netlink_ext_ack *extack)
- {
- struct switchdev_notifier_port_obj_info obj_info = {
- .info = {
- .dev = dev,
- .extack = extack,
- .ctx = ctx,
- },
- .obj = &mdb->obj,
- };
- int err;
- err = nb->notifier_call(nb, action, &obj_info);
- return notifier_to_errno(err);
- }
- static int br_switchdev_mdb_queue_one(struct list_head *mdb_list,
- enum switchdev_obj_id id,
- const struct net_bridge_mdb_entry *mp,
- struct net_device *orig_dev)
- {
- struct switchdev_obj_port_mdb *mdb;
- mdb = kzalloc(sizeof(*mdb), GFP_ATOMIC);
- if (!mdb)
- return -ENOMEM;
- mdb->obj.id = id;
- mdb->obj.orig_dev = orig_dev;
- br_switchdev_mdb_populate(mdb, mp);
- list_add_tail(&mdb->obj.list, mdb_list);
- return 0;
- }
- void br_switchdev_mdb_notify(struct net_device *dev,
- struct net_bridge_mdb_entry *mp,
- struct net_bridge_port_group *pg,
- int type)
- {
- struct br_switchdev_mdb_complete_info *complete_info;
- struct switchdev_obj_port_mdb mdb = {
- .obj = {
- .id = SWITCHDEV_OBJ_ID_PORT_MDB,
- .flags = SWITCHDEV_F_DEFER,
- },
- };
- if (!pg)
- return br_switchdev_host_mdb(dev, mp, type);
- br_switchdev_mdb_populate(&mdb, mp);
- mdb.obj.orig_dev = pg->key.port->dev;
- switch (type) {
- case RTM_NEWMDB:
- complete_info = kmalloc(sizeof(*complete_info), GFP_ATOMIC);
- if (!complete_info)
- break;
- complete_info->port = pg->key.port;
- complete_info->ip = mp->addr;
- mdb.obj.complete_priv = complete_info;
- mdb.obj.complete = br_switchdev_mdb_complete;
- if (switchdev_port_obj_add(pg->key.port->dev, &mdb.obj, NULL))
- kfree(complete_info);
- break;
- case RTM_DELMDB:
- switchdev_port_obj_del(pg->key.port->dev, &mdb.obj);
- break;
- }
- }
- #endif
- static int
- br_switchdev_mdb_replay(struct net_device *br_dev, struct net_device *dev,
- const void *ctx, bool adding, struct notifier_block *nb,
- struct netlink_ext_ack *extack)
- {
- #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
- const struct net_bridge_mdb_entry *mp;
- struct switchdev_obj *obj, *tmp;
- struct net_bridge *br;
- unsigned long action;
- LIST_HEAD(mdb_list);
- int err = 0;
- ASSERT_RTNL();
- if (!nb)
- return 0;
- if (!netif_is_bridge_master(br_dev) || !netif_is_bridge_port(dev))
- return -EINVAL;
- br = netdev_priv(br_dev);
- if (!br_opt_get(br, BROPT_MULTICAST_ENABLED))
- return 0;
- /* We cannot walk over br->mdb_list protected just by the rtnl_mutex,
- * because the write-side protection is br->multicast_lock. But we
- * need to emulate the [ blocking ] calling context of a regular
- * switchdev event, so since both br->multicast_lock and RCU read side
- * critical sections are atomic, we have no choice but to pick the RCU
- * read side lock, queue up all our events, leave the critical section
- * and notify switchdev from blocking context.
- */
- rcu_read_lock();
- hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
- struct net_bridge_port_group __rcu * const *pp;
- const struct net_bridge_port_group *p;
- if (mp->host_joined) {
- err = br_switchdev_mdb_queue_one(&mdb_list,
- SWITCHDEV_OBJ_ID_HOST_MDB,
- mp, br_dev);
- if (err) {
- rcu_read_unlock();
- goto out_free_mdb;
- }
- }
- for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
- pp = &p->next) {
- if (p->key.port->dev != dev)
- continue;
- err = br_switchdev_mdb_queue_one(&mdb_list,
- SWITCHDEV_OBJ_ID_PORT_MDB,
- mp, dev);
- if (err) {
- rcu_read_unlock();
- goto out_free_mdb;
- }
- }
- }
- rcu_read_unlock();
- if (adding)
- action = SWITCHDEV_PORT_OBJ_ADD;
- else
- action = SWITCHDEV_PORT_OBJ_DEL;
- list_for_each_entry(obj, &mdb_list, list) {
- err = br_switchdev_mdb_replay_one(nb, dev,
- SWITCHDEV_OBJ_PORT_MDB(obj),
- action, ctx, extack);
- if (err)
- goto out_free_mdb;
- }
- out_free_mdb:
- list_for_each_entry_safe(obj, tmp, &mdb_list, list) {
- list_del(&obj->list);
- kfree(SWITCHDEV_OBJ_PORT_MDB(obj));
- }
- if (err)
- return err;
- #endif
- return 0;
- }
- static int nbp_switchdev_sync_objs(struct net_bridge_port *p, const void *ctx,
- struct notifier_block *atomic_nb,
- struct notifier_block *blocking_nb,
- struct netlink_ext_ack *extack)
- {
- struct net_device *br_dev = p->br->dev;
- struct net_device *dev = p->dev;
- int err;
- err = br_switchdev_vlan_replay(br_dev, ctx, true, blocking_nb, extack);
- if (err && err != -EOPNOTSUPP)
- return err;
- err = br_switchdev_mdb_replay(br_dev, dev, ctx, true, blocking_nb,
- extack);
- if (err && err != -EOPNOTSUPP)
- return err;
- err = br_switchdev_fdb_replay(br_dev, ctx, true, atomic_nb);
- if (err && err != -EOPNOTSUPP)
- return err;
- return 0;
- }
- static void nbp_switchdev_unsync_objs(struct net_bridge_port *p,
- const void *ctx,
- struct notifier_block *atomic_nb,
- struct notifier_block *blocking_nb)
- {
- struct net_device *br_dev = p->br->dev;
- struct net_device *dev = p->dev;
- br_switchdev_fdb_replay(br_dev, ctx, false, atomic_nb);
- br_switchdev_mdb_replay(br_dev, dev, ctx, false, blocking_nb, NULL);
- br_switchdev_vlan_replay(br_dev, ctx, false, blocking_nb, NULL);
- }
- /* Let the bridge know that this port is offloaded, so that it can assign a
- * switchdev hardware domain to it.
- */
- int br_switchdev_port_offload(struct net_bridge_port *p,
- struct net_device *dev, const void *ctx,
- struct notifier_block *atomic_nb,
- struct notifier_block *blocking_nb,
- bool tx_fwd_offload,
- struct netlink_ext_ack *extack)
- {
- struct netdev_phys_item_id ppid;
- int err;
- err = dev_get_port_parent_id(dev, &ppid, false);
- if (err)
- return err;
- err = nbp_switchdev_add(p, ppid, tx_fwd_offload, extack);
- if (err)
- return err;
- err = nbp_switchdev_sync_objs(p, ctx, atomic_nb, blocking_nb, extack);
- if (err)
- goto out_switchdev_del;
- return 0;
- out_switchdev_del:
- nbp_switchdev_del(p);
- return err;
- }
- void br_switchdev_port_unoffload(struct net_bridge_port *p, const void *ctx,
- struct notifier_block *atomic_nb,
- struct notifier_block *blocking_nb)
- {
- nbp_switchdev_unsync_objs(p, ctx, atomic_nb, blocking_nb);
- nbp_switchdev_del(p);
- }
|