2 * Bridge netlink control interface
5 * Stephen Hemminger <shemminger@osdl.org>
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License
9 * as published by the Free Software Foundation; either version
10 * 2 of the License, or (at your option) any later version.
13 #include <linux/kernel.h>
14 #include <linux/slab.h>
15 #include <linux/etherdevice.h>
16 #include <net/rtnetlink.h>
17 #include <net/net_namespace.h>
19 #include <uapi/linux/if_bridge.h>
21 #include "br_private.h"
22 #include "br_private_stp.h"
23 #include "br_private_tunnel.h"
25 static int __get_num_vlan_infos(struct net_bridge_vlan_group *vg,
28 struct net_bridge_vlan *v;
29 u16 vid_range_start = 0, vid_range_end = 0, vid_range_flags = 0;
33 if (!(filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED))
36 pvid = br_get_pvid(vg);
37 /* Count number of vlan infos */
38 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) {
40 /* only a context, bridge vlan not activated */
41 if (!br_vlan_should_use(v))
44 flags |= BRIDGE_VLAN_INFO_PVID;
46 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
47 flags |= BRIDGE_VLAN_INFO_UNTAGGED;
49 if (vid_range_start == 0) {
51 } else if ((v->vid - vid_range_end) == 1 &&
52 flags == vid_range_flags) {
53 vid_range_end = v->vid;
56 if ((vid_range_end - vid_range_start) > 0)
62 vid_range_start = v->vid;
63 vid_range_end = v->vid;
64 vid_range_flags = flags;
67 if (vid_range_start != 0) {
68 if ((vid_range_end - vid_range_start) > 0)
77 static int br_get_num_vlan_infos(struct net_bridge_vlan_group *vg,
85 if (filter_mask & RTEXT_FILTER_BRVLAN)
89 num_vlans = __get_num_vlan_infos(vg, filter_mask);
95 static size_t br_get_link_af_size_filtered(const struct net_device *dev,
98 struct net_bridge_vlan_group *vg = NULL;
99 struct net_bridge_port *p = NULL;
100 struct net_bridge *br;
105 if (br_port_exists(dev)) {
106 p = br_port_get_rcu(dev);
107 vg = nbp_vlan_group_rcu(p);
108 } else if (dev->priv_flags & IFF_EBRIDGE) {
109 br = netdev_priv(dev);
110 vg = br_vlan_group_rcu(br);
112 num_vlan_infos = br_get_num_vlan_infos(vg, filter_mask);
115 if (p && (p->flags & BR_VLAN_TUNNEL))
116 vinfo_sz += br_get_vlan_tunnel_info_size(vg);
118 /* Each VLAN is returned in bridge_vlan_info along with flags */
119 vinfo_sz += num_vlan_infos * nla_total_size(sizeof(struct bridge_vlan_info));
124 static inline size_t br_port_info_size(void)
126 return nla_total_size(1) /* IFLA_BRPORT_STATE */
127 + nla_total_size(2) /* IFLA_BRPORT_PRIORITY */
128 + nla_total_size(4) /* IFLA_BRPORT_COST */
129 + nla_total_size(1) /* IFLA_BRPORT_MODE */
130 + nla_total_size(1) /* IFLA_BRPORT_GUARD */
131 + nla_total_size(1) /* IFLA_BRPORT_PROTECT */
132 + nla_total_size(1) /* IFLA_BRPORT_FAST_LEAVE */
133 + nla_total_size(1) /* IFLA_BRPORT_MCAST_TO_UCAST */
134 + nla_total_size(1) /* IFLA_BRPORT_LEARNING */
135 + nla_total_size(1) /* IFLA_BRPORT_UNICAST_FLOOD */
136 + nla_total_size(1) /* IFLA_BRPORT_MCAST_FLOOD */
137 + nla_total_size(1) /* IFLA_BRPORT_BCAST_FLOOD */
138 + nla_total_size(1) /* IFLA_BRPORT_PROXYARP */
139 + nla_total_size(1) /* IFLA_BRPORT_PROXYARP_WIFI */
140 + nla_total_size(1) /* IFLA_BRPORT_VLAN_TUNNEL */
141 + nla_total_size(1) /* IFLA_BRPORT_NEIGH_SUPPRESS */
142 + nla_total_size(sizeof(struct ifla_bridge_id)) /* IFLA_BRPORT_ROOT_ID */
143 + nla_total_size(sizeof(struct ifla_bridge_id)) /* IFLA_BRPORT_BRIDGE_ID */
144 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_DESIGNATED_PORT */
145 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_DESIGNATED_COST */
146 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_ID */
147 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_NO */
148 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_TOPOLOGY_CHANGE_ACK */
149 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_CONFIG_PENDING */
150 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_MESSAGE_AGE_TIMER */
151 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_FORWARD_DELAY_TIMER */
152 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_HOLD_TIMER */
153 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
154 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_MULTICAST_ROUTER */
156 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_GROUP_FWD_MASK */
160 static inline size_t br_nlmsg_size(struct net_device *dev, u32 filter_mask)
162 return NLMSG_ALIGN(sizeof(struct ifinfomsg))
163 + nla_total_size(IFNAMSIZ) /* IFLA_IFNAME */
164 + nla_total_size(MAX_ADDR_LEN) /* IFLA_ADDRESS */
165 + nla_total_size(4) /* IFLA_MASTER */
166 + nla_total_size(4) /* IFLA_MTU */
167 + nla_total_size(4) /* IFLA_LINK */
168 + nla_total_size(1) /* IFLA_OPERSTATE */
169 + nla_total_size(br_port_info_size()) /* IFLA_PROTINFO */
170 + nla_total_size(br_get_link_af_size_filtered(dev,
171 filter_mask)); /* IFLA_AF_SPEC */
174 static int br_port_fill_attrs(struct sk_buff *skb,
175 const struct net_bridge_port *p)
177 u8 mode = !!(p->flags & BR_HAIRPIN_MODE);
180 if (nla_put_u8(skb, IFLA_BRPORT_STATE, p->state) ||
181 nla_put_u16(skb, IFLA_BRPORT_PRIORITY, p->priority) ||
182 nla_put_u32(skb, IFLA_BRPORT_COST, p->path_cost) ||
183 nla_put_u8(skb, IFLA_BRPORT_MODE, mode) ||
184 nla_put_u8(skb, IFLA_BRPORT_GUARD, !!(p->flags & BR_BPDU_GUARD)) ||
185 nla_put_u8(skb, IFLA_BRPORT_PROTECT,
186 !!(p->flags & BR_ROOT_BLOCK)) ||
187 nla_put_u8(skb, IFLA_BRPORT_FAST_LEAVE,
188 !!(p->flags & BR_MULTICAST_FAST_LEAVE)) ||
189 nla_put_u8(skb, IFLA_BRPORT_MCAST_TO_UCAST,
190 !!(p->flags & BR_MULTICAST_TO_UNICAST)) ||
191 nla_put_u8(skb, IFLA_BRPORT_LEARNING, !!(p->flags & BR_LEARNING)) ||
192 nla_put_u8(skb, IFLA_BRPORT_UNICAST_FLOOD,
193 !!(p->flags & BR_FLOOD)) ||
194 nla_put_u8(skb, IFLA_BRPORT_MCAST_FLOOD,
195 !!(p->flags & BR_MCAST_FLOOD)) ||
196 nla_put_u8(skb, IFLA_BRPORT_BCAST_FLOOD,
197 !!(p->flags & BR_BCAST_FLOOD)) ||
198 nla_put_u8(skb, IFLA_BRPORT_PROXYARP, !!(p->flags & BR_PROXYARP)) ||
199 nla_put_u8(skb, IFLA_BRPORT_PROXYARP_WIFI,
200 !!(p->flags & BR_PROXYARP_WIFI)) ||
201 nla_put(skb, IFLA_BRPORT_ROOT_ID, sizeof(struct ifla_bridge_id),
202 &p->designated_root) ||
203 nla_put(skb, IFLA_BRPORT_BRIDGE_ID, sizeof(struct ifla_bridge_id),
204 &p->designated_bridge) ||
205 nla_put_u16(skb, IFLA_BRPORT_DESIGNATED_PORT, p->designated_port) ||
206 nla_put_u16(skb, IFLA_BRPORT_DESIGNATED_COST, p->designated_cost) ||
207 nla_put_u16(skb, IFLA_BRPORT_ID, p->port_id) ||
208 nla_put_u16(skb, IFLA_BRPORT_NO, p->port_no) ||
209 nla_put_u8(skb, IFLA_BRPORT_TOPOLOGY_CHANGE_ACK,
210 p->topology_change_ack) ||
211 nla_put_u8(skb, IFLA_BRPORT_CONFIG_PENDING, p->config_pending) ||
212 nla_put_u8(skb, IFLA_BRPORT_VLAN_TUNNEL, !!(p->flags &
214 nla_put_u16(skb, IFLA_BRPORT_GROUP_FWD_MASK, p->group_fwd_mask) ||
215 nla_put_u8(skb, IFLA_BRPORT_NEIGH_SUPPRESS,
216 !!(p->flags & BR_NEIGH_SUPPRESS)))
219 timerval = br_timer_value(&p->message_age_timer);
220 if (nla_put_u64_64bit(skb, IFLA_BRPORT_MESSAGE_AGE_TIMER, timerval,
223 timerval = br_timer_value(&p->forward_delay_timer);
224 if (nla_put_u64_64bit(skb, IFLA_BRPORT_FORWARD_DELAY_TIMER, timerval,
227 timerval = br_timer_value(&p->hold_timer);
228 if (nla_put_u64_64bit(skb, IFLA_BRPORT_HOLD_TIMER, timerval,
232 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
233 if (nla_put_u8(skb, IFLA_BRPORT_MULTICAST_ROUTER,
234 p->multicast_router))
241 static int br_fill_ifvlaninfo_range(struct sk_buff *skb, u16 vid_start,
242 u16 vid_end, u16 flags)
244 struct bridge_vlan_info vinfo;
246 if ((vid_end - vid_start) > 0) {
247 /* add range to skb */
248 vinfo.vid = vid_start;
249 vinfo.flags = flags | BRIDGE_VLAN_INFO_RANGE_BEGIN;
250 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
251 sizeof(vinfo), &vinfo))
252 goto nla_put_failure;
255 vinfo.flags = flags | BRIDGE_VLAN_INFO_RANGE_END;
256 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
257 sizeof(vinfo), &vinfo))
258 goto nla_put_failure;
260 vinfo.vid = vid_start;
262 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
263 sizeof(vinfo), &vinfo))
264 goto nla_put_failure;
273 static int br_fill_ifvlaninfo_compressed(struct sk_buff *skb,
274 struct net_bridge_vlan_group *vg)
276 struct net_bridge_vlan *v;
277 u16 vid_range_start = 0, vid_range_end = 0, vid_range_flags = 0;
281 /* Pack IFLA_BRIDGE_VLAN_INFO's for every vlan
282 * and mark vlan info with begin and end flags
283 * if vlaninfo represents a range
285 pvid = br_get_pvid(vg);
286 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) {
288 if (!br_vlan_should_use(v))
291 flags |= BRIDGE_VLAN_INFO_PVID;
293 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
294 flags |= BRIDGE_VLAN_INFO_UNTAGGED;
296 if (vid_range_start == 0) {
298 } else if ((v->vid - vid_range_end) == 1 &&
299 flags == vid_range_flags) {
300 vid_range_end = v->vid;
303 err = br_fill_ifvlaninfo_range(skb, vid_range_start,
311 vid_range_start = v->vid;
312 vid_range_end = v->vid;
313 vid_range_flags = flags;
316 if (vid_range_start != 0) {
317 /* Call it once more to send any left over vlans */
318 err = br_fill_ifvlaninfo_range(skb, vid_range_start,
328 static int br_fill_ifvlaninfo(struct sk_buff *skb,
329 struct net_bridge_vlan_group *vg)
331 struct bridge_vlan_info vinfo;
332 struct net_bridge_vlan *v;
335 pvid = br_get_pvid(vg);
336 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) {
337 if (!br_vlan_should_use(v))
343 vinfo.flags |= BRIDGE_VLAN_INFO_PVID;
345 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED)
346 vinfo.flags |= BRIDGE_VLAN_INFO_UNTAGGED;
348 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO,
349 sizeof(vinfo), &vinfo))
350 goto nla_put_failure;
360 * Create one netlink message for one interface
361 * Contains port and master info as well as carrier and bridge state.
363 static int br_fill_ifinfo(struct sk_buff *skb,
364 struct net_bridge_port *port,
365 u32 pid, u32 seq, int event, unsigned int flags,
366 u32 filter_mask, const struct net_device *dev)
368 struct net_bridge *br;
369 struct ifinfomsg *hdr;
370 struct nlmsghdr *nlh;
371 u8 operstate = netif_running(dev) ? dev->operstate : IF_OPER_DOWN;
376 br = netdev_priv(dev);
378 br_debug(br, "br_fill_info event %d port %s master %s\n",
379 event, dev->name, br->dev->name);
381 nlh = nlmsg_put(skb, pid, seq, event, sizeof(*hdr), flags);
385 hdr = nlmsg_data(nlh);
386 hdr->ifi_family = AF_BRIDGE;
388 hdr->ifi_type = dev->type;
389 hdr->ifi_index = dev->ifindex;
390 hdr->ifi_flags = dev_get_flags(dev);
393 if (nla_put_string(skb, IFLA_IFNAME, dev->name) ||
394 nla_put_u32(skb, IFLA_MASTER, br->dev->ifindex) ||
395 nla_put_u32(skb, IFLA_MTU, dev->mtu) ||
396 nla_put_u8(skb, IFLA_OPERSTATE, operstate) ||
398 nla_put(skb, IFLA_ADDRESS, dev->addr_len, dev->dev_addr)) ||
399 (dev->ifindex != dev_get_iflink(dev) &&
400 nla_put_u32(skb, IFLA_LINK, dev_get_iflink(dev))))
401 goto nla_put_failure;
403 if (event == RTM_NEWLINK && port) {
405 = nla_nest_start(skb, IFLA_PROTINFO | NLA_F_NESTED);
407 if (nest == NULL || br_port_fill_attrs(skb, port) < 0)
408 goto nla_put_failure;
409 nla_nest_end(skb, nest);
412 /* Check if the VID information is requested */
413 if ((filter_mask & RTEXT_FILTER_BRVLAN) ||
414 (filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)) {
415 struct net_bridge_vlan_group *vg;
419 /* RCU needed because of the VLAN locking rules (rcu || rtnl) */
422 vg = nbp_vlan_group_rcu(port);
424 vg = br_vlan_group_rcu(br);
426 if (!vg || !vg->num_vlans) {
430 af = nla_nest_start(skb, IFLA_AF_SPEC);
433 goto nla_put_failure;
435 if (filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)
436 err = br_fill_ifvlaninfo_compressed(skb, vg);
438 err = br_fill_ifvlaninfo(skb, vg);
440 if (port && (port->flags & BR_VLAN_TUNNEL))
441 err = br_fill_vlan_tunnel_info(skb, vg);
444 goto nla_put_failure;
445 nla_nest_end(skb, af);
453 nlmsg_cancel(skb, nlh);
458 * Notify listeners of a change in port information
460 void br_ifinfo_notify(int event, struct net_bridge_port *port)
465 u32 filter = RTEXT_FILTER_BRVLAN_COMPRESSED;
470 net = dev_net(port->dev);
471 br_debug(port->br, "port %u(%s) event %d\n",
472 (unsigned int)port->port_no, port->dev->name, event);
474 skb = nlmsg_new(br_nlmsg_size(port->dev, filter), GFP_ATOMIC);
478 err = br_fill_ifinfo(skb, port, 0, 0, event, 0, filter, port->dev);
480 /* -EMSGSIZE implies BUG in br_nlmsg_size() */
481 WARN_ON(err == -EMSGSIZE);
485 rtnl_notify(skb, net, 0, RTNLGRP_LINK, NULL, GFP_ATOMIC);
488 rtnl_set_sk_err(net, RTNLGRP_LINK, err);
493 * Dump information about all ports, in response to GETLINK
495 int br_getlink(struct sk_buff *skb, u32 pid, u32 seq,
496 struct net_device *dev, u32 filter_mask, int nlflags)
498 struct net_bridge_port *port = br_port_get_rtnl(dev);
500 if (!port && !(filter_mask & RTEXT_FILTER_BRVLAN) &&
501 !(filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED))
504 return br_fill_ifinfo(skb, port, pid, seq, RTM_NEWLINK, nlflags,
508 static int br_vlan_info(struct net_bridge *br, struct net_bridge_port *p,
509 int cmd, struct bridge_vlan_info *vinfo)
516 /* if the MASTER flag is set this will act on the global
517 * per-VLAN entry as well
519 err = nbp_vlan_add(p, vinfo->vid, vinfo->flags);
523 vinfo->flags |= BRIDGE_VLAN_INFO_BRENTRY;
524 err = br_vlan_add(br, vinfo->vid, vinfo->flags);
530 nbp_vlan_delete(p, vinfo->vid);
531 if (vinfo->flags & BRIDGE_VLAN_INFO_MASTER)
532 br_vlan_delete(p->br, vinfo->vid);
534 br_vlan_delete(br, vinfo->vid);
542 static int br_process_vlan_info(struct net_bridge *br,
543 struct net_bridge_port *p, int cmd,
544 struct bridge_vlan_info *vinfo_curr,
545 struct bridge_vlan_info **vinfo_last)
547 if (!vinfo_curr->vid || vinfo_curr->vid >= VLAN_VID_MASK)
550 if (vinfo_curr->flags & BRIDGE_VLAN_INFO_RANGE_BEGIN) {
551 /* check if we are already processing a range */
554 *vinfo_last = vinfo_curr;
555 /* don't allow range of pvids */
556 if ((*vinfo_last)->flags & BRIDGE_VLAN_INFO_PVID)
562 struct bridge_vlan_info tmp_vinfo;
565 if (!(vinfo_curr->flags & BRIDGE_VLAN_INFO_RANGE_END))
568 if (vinfo_curr->vid <= (*vinfo_last)->vid)
571 memcpy(&tmp_vinfo, *vinfo_last,
572 sizeof(struct bridge_vlan_info));
573 for (v = (*vinfo_last)->vid; v <= vinfo_curr->vid; v++) {
575 err = br_vlan_info(br, p, cmd, &tmp_vinfo);
584 return br_vlan_info(br, p, cmd, vinfo_curr);
587 static int br_afspec(struct net_bridge *br,
588 struct net_bridge_port *p,
589 struct nlattr *af_spec,
592 struct bridge_vlan_info *vinfo_curr = NULL;
593 struct bridge_vlan_info *vinfo_last = NULL;
595 struct vtunnel_info tinfo_last = {};
596 struct vtunnel_info tinfo_curr = {};
599 nla_for_each_nested(attr, af_spec, rem) {
601 switch (nla_type(attr)) {
602 case IFLA_BRIDGE_VLAN_TUNNEL_INFO:
603 if (!p || !(p->flags & BR_VLAN_TUNNEL))
605 err = br_parse_vlan_tunnel_info(attr, &tinfo_curr);
608 err = br_process_vlan_tunnel_info(br, p, cmd,
614 case IFLA_BRIDGE_VLAN_INFO:
615 if (nla_len(attr) != sizeof(struct bridge_vlan_info))
617 vinfo_curr = nla_data(attr);
618 err = br_process_vlan_info(br, p, cmd, vinfo_curr,
629 static const struct nla_policy br_port_policy[IFLA_BRPORT_MAX + 1] = {
630 [IFLA_BRPORT_STATE] = { .type = NLA_U8 },
631 [IFLA_BRPORT_COST] = { .type = NLA_U32 },
632 [IFLA_BRPORT_PRIORITY] = { .type = NLA_U16 },
633 [IFLA_BRPORT_MODE] = { .type = NLA_U8 },
634 [IFLA_BRPORT_GUARD] = { .type = NLA_U8 },
635 [IFLA_BRPORT_PROTECT] = { .type = NLA_U8 },
636 [IFLA_BRPORT_FAST_LEAVE]= { .type = NLA_U8 },
637 [IFLA_BRPORT_LEARNING] = { .type = NLA_U8 },
638 [IFLA_BRPORT_UNICAST_FLOOD] = { .type = NLA_U8 },
639 [IFLA_BRPORT_PROXYARP] = { .type = NLA_U8 },
640 [IFLA_BRPORT_PROXYARP_WIFI] = { .type = NLA_U8 },
641 [IFLA_BRPORT_MULTICAST_ROUTER] = { .type = NLA_U8 },
642 [IFLA_BRPORT_MCAST_TO_UCAST] = { .type = NLA_U8 },
643 [IFLA_BRPORT_MCAST_FLOOD] = { .type = NLA_U8 },
644 [IFLA_BRPORT_BCAST_FLOOD] = { .type = NLA_U8 },
645 [IFLA_BRPORT_GROUP_FWD_MASK] = { .type = NLA_U16 },
648 /* Change the state of the port and notify spanning tree */
649 static int br_set_port_state(struct net_bridge_port *p, u8 state)
651 if (state > BR_STATE_BLOCKING)
654 /* if kernel STP is running, don't allow changes */
655 if (p->br->stp_enabled == BR_KERNEL_STP)
658 /* if device is not up, change is not allowed
659 * if link is not present, only allowable state is disabled
661 if (!netif_running(p->dev) ||
662 (!netif_oper_up(p->dev) && state != BR_STATE_DISABLED))
665 br_set_state(p, state);
666 br_port_state_selection(p->br);
670 /* Set/clear or port flags based on attribute */
671 static int br_set_port_flag(struct net_bridge_port *p, struct nlattr *tb[],
672 int attrtype, unsigned long mask)
680 if (nla_get_u8(tb[attrtype]))
681 flags = p->flags | mask;
683 flags = p->flags & ~mask;
685 err = br_switchdev_set_port_flag(p, flags, mask);
693 /* Process bridge protocol info on port */
694 static int br_setport(struct net_bridge_port *p, struct nlattr *tb[])
696 unsigned long old_flags = p->flags;
697 bool br_vlan_tunnel_old = false;
700 err = br_set_port_flag(p, tb, IFLA_BRPORT_MODE, BR_HAIRPIN_MODE);
704 err = br_set_port_flag(p, tb, IFLA_BRPORT_GUARD, BR_BPDU_GUARD);
708 err = br_set_port_flag(p, tb, IFLA_BRPORT_FAST_LEAVE, BR_MULTICAST_FAST_LEAVE);
712 err = br_set_port_flag(p, tb, IFLA_BRPORT_PROTECT, BR_ROOT_BLOCK);
716 err = br_set_port_flag(p, tb, IFLA_BRPORT_LEARNING, BR_LEARNING);
720 err = br_set_port_flag(p, tb, IFLA_BRPORT_UNICAST_FLOOD, BR_FLOOD);
724 err = br_set_port_flag(p, tb, IFLA_BRPORT_MCAST_FLOOD, BR_MCAST_FLOOD);
728 err = br_set_port_flag(p, tb, IFLA_BRPORT_MCAST_TO_UCAST, BR_MULTICAST_TO_UNICAST);
732 err = br_set_port_flag(p, tb, IFLA_BRPORT_BCAST_FLOOD, BR_BCAST_FLOOD);
736 err = br_set_port_flag(p, tb, IFLA_BRPORT_PROXYARP, BR_PROXYARP);
740 err = br_set_port_flag(p, tb, IFLA_BRPORT_PROXYARP_WIFI, BR_PROXYARP_WIFI);
744 br_vlan_tunnel_old = (p->flags & BR_VLAN_TUNNEL) ? true : false;
745 err = br_set_port_flag(p, tb, IFLA_BRPORT_VLAN_TUNNEL, BR_VLAN_TUNNEL);
749 if (br_vlan_tunnel_old && !(p->flags & BR_VLAN_TUNNEL))
750 nbp_vlan_tunnel_info_flush(p);
752 if (tb[IFLA_BRPORT_COST]) {
753 err = br_stp_set_path_cost(p, nla_get_u32(tb[IFLA_BRPORT_COST]));
758 if (tb[IFLA_BRPORT_PRIORITY]) {
759 err = br_stp_set_port_priority(p, nla_get_u16(tb[IFLA_BRPORT_PRIORITY]));
764 if (tb[IFLA_BRPORT_STATE]) {
765 err = br_set_port_state(p, nla_get_u8(tb[IFLA_BRPORT_STATE]));
770 if (tb[IFLA_BRPORT_FLUSH])
771 br_fdb_delete_by_port(p->br, p, 0, 0);
773 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
774 if (tb[IFLA_BRPORT_MULTICAST_ROUTER]) {
775 u8 mcast_router = nla_get_u8(tb[IFLA_BRPORT_MULTICAST_ROUTER]);
777 err = br_multicast_set_port_router(p, mcast_router);
783 if (tb[IFLA_BRPORT_GROUP_FWD_MASK]) {
784 u16 fwd_mask = nla_get_u16(tb[IFLA_BRPORT_GROUP_FWD_MASK]);
786 if (fwd_mask & BR_GROUPFWD_MACPAUSE)
788 p->group_fwd_mask = fwd_mask;
791 err = br_set_port_flag(p, tb, IFLA_BRPORT_NEIGH_SUPPRESS,
796 br_port_flags_change(p, old_flags ^ p->flags);
800 /* Change state and parameters on port. */
801 int br_setlink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags)
803 struct nlattr *protinfo;
804 struct nlattr *afspec;
805 struct net_bridge_port *p;
806 struct nlattr *tb[IFLA_BRPORT_MAX + 1];
809 protinfo = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_PROTINFO);
810 afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
811 if (!protinfo && !afspec)
814 p = br_port_get_rtnl(dev);
815 /* We want to accept dev as bridge itself if the AF_SPEC
816 * is set to see if someone is setting vlan info on the bridge
822 if (protinfo->nla_type & NLA_F_NESTED) {
823 err = nla_parse_nested(tb, IFLA_BRPORT_MAX, protinfo,
824 br_port_policy, NULL);
828 spin_lock_bh(&p->br->lock);
829 err = br_setport(p, tb);
830 spin_unlock_bh(&p->br->lock);
832 /* Binary compatibility with old RSTP */
833 if (nla_len(protinfo) < sizeof(u8))
836 spin_lock_bh(&p->br->lock);
837 err = br_set_port_state(p, nla_get_u8(protinfo));
838 spin_unlock_bh(&p->br->lock);
845 err = br_afspec((struct net_bridge *)netdev_priv(dev), p,
846 afspec, RTM_SETLINK);
850 br_ifinfo_notify(RTM_NEWLINK, p);
855 /* Delete port information */
856 int br_dellink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags)
858 struct nlattr *afspec;
859 struct net_bridge_port *p;
862 afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
866 p = br_port_get_rtnl(dev);
867 /* We want to accept dev as bridge itself as well */
868 if (!p && !(dev->priv_flags & IFF_EBRIDGE))
871 err = br_afspec((struct net_bridge *)netdev_priv(dev), p,
872 afspec, RTM_DELLINK);
874 /* Send RTM_NEWLINK because userspace
875 * expects RTM_NEWLINK for vlan dels
877 br_ifinfo_notify(RTM_NEWLINK, p);
882 static int br_validate(struct nlattr *tb[], struct nlattr *data[],
883 struct netlink_ext_ack *extack)
885 if (tb[IFLA_ADDRESS]) {
886 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
888 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
889 return -EADDRNOTAVAIL;
895 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
896 if (data[IFLA_BR_VLAN_PROTOCOL]) {
897 switch (nla_get_be16(data[IFLA_BR_VLAN_PROTOCOL])) {
898 case htons(ETH_P_8021Q):
899 case htons(ETH_P_8021AD):
902 return -EPROTONOSUPPORT;
906 if (data[IFLA_BR_VLAN_DEFAULT_PVID]) {
907 __u16 defpvid = nla_get_u16(data[IFLA_BR_VLAN_DEFAULT_PVID]);
909 if (defpvid >= VLAN_VID_MASK)
917 static int br_port_slave_changelink(struct net_device *brdev,
918 struct net_device *dev,
920 struct nlattr *data[],
921 struct netlink_ext_ack *extack)
923 struct net_bridge *br = netdev_priv(brdev);
929 spin_lock_bh(&br->lock);
930 ret = br_setport(br_port_get_rtnl(dev), data);
931 spin_unlock_bh(&br->lock);
936 static int br_port_fill_slave_info(struct sk_buff *skb,
937 const struct net_device *brdev,
938 const struct net_device *dev)
940 return br_port_fill_attrs(skb, br_port_get_rtnl(dev));
943 static size_t br_port_get_slave_size(const struct net_device *brdev,
944 const struct net_device *dev)
946 return br_port_info_size();
949 static const struct nla_policy br_policy[IFLA_BR_MAX + 1] = {
950 [IFLA_BR_FORWARD_DELAY] = { .type = NLA_U32 },
951 [IFLA_BR_HELLO_TIME] = { .type = NLA_U32 },
952 [IFLA_BR_MAX_AGE] = { .type = NLA_U32 },
953 [IFLA_BR_AGEING_TIME] = { .type = NLA_U32 },
954 [IFLA_BR_STP_STATE] = { .type = NLA_U32 },
955 [IFLA_BR_PRIORITY] = { .type = NLA_U16 },
956 [IFLA_BR_VLAN_FILTERING] = { .type = NLA_U8 },
957 [IFLA_BR_VLAN_PROTOCOL] = { .type = NLA_U16 },
958 [IFLA_BR_GROUP_FWD_MASK] = { .type = NLA_U16 },
959 [IFLA_BR_GROUP_ADDR] = { .type = NLA_BINARY,
961 [IFLA_BR_MCAST_ROUTER] = { .type = NLA_U8 },
962 [IFLA_BR_MCAST_SNOOPING] = { .type = NLA_U8 },
963 [IFLA_BR_MCAST_QUERY_USE_IFADDR] = { .type = NLA_U8 },
964 [IFLA_BR_MCAST_QUERIER] = { .type = NLA_U8 },
965 [IFLA_BR_MCAST_HASH_ELASTICITY] = { .type = NLA_U32 },
966 [IFLA_BR_MCAST_HASH_MAX] = { .type = NLA_U32 },
967 [IFLA_BR_MCAST_LAST_MEMBER_CNT] = { .type = NLA_U32 },
968 [IFLA_BR_MCAST_STARTUP_QUERY_CNT] = { .type = NLA_U32 },
969 [IFLA_BR_MCAST_LAST_MEMBER_INTVL] = { .type = NLA_U64 },
970 [IFLA_BR_MCAST_MEMBERSHIP_INTVL] = { .type = NLA_U64 },
971 [IFLA_BR_MCAST_QUERIER_INTVL] = { .type = NLA_U64 },
972 [IFLA_BR_MCAST_QUERY_INTVL] = { .type = NLA_U64 },
973 [IFLA_BR_MCAST_QUERY_RESPONSE_INTVL] = { .type = NLA_U64 },
974 [IFLA_BR_MCAST_STARTUP_QUERY_INTVL] = { .type = NLA_U64 },
975 [IFLA_BR_NF_CALL_IPTABLES] = { .type = NLA_U8 },
976 [IFLA_BR_NF_CALL_IP6TABLES] = { .type = NLA_U8 },
977 [IFLA_BR_NF_CALL_ARPTABLES] = { .type = NLA_U8 },
978 [IFLA_BR_VLAN_DEFAULT_PVID] = { .type = NLA_U16 },
979 [IFLA_BR_VLAN_STATS_ENABLED] = { .type = NLA_U8 },
980 [IFLA_BR_MCAST_STATS_ENABLED] = { .type = NLA_U8 },
981 [IFLA_BR_MCAST_IGMP_VERSION] = { .type = NLA_U8 },
982 [IFLA_BR_MCAST_MLD_VERSION] = { .type = NLA_U8 },
985 static int br_changelink(struct net_device *brdev, struct nlattr *tb[],
986 struct nlattr *data[],
987 struct netlink_ext_ack *extack)
989 struct net_bridge *br = netdev_priv(brdev);
995 if (data[IFLA_BR_FORWARD_DELAY]) {
996 err = br_set_forward_delay(br, nla_get_u32(data[IFLA_BR_FORWARD_DELAY]));
1001 if (data[IFLA_BR_HELLO_TIME]) {
1002 err = br_set_hello_time(br, nla_get_u32(data[IFLA_BR_HELLO_TIME]));
1007 if (data[IFLA_BR_MAX_AGE]) {
1008 err = br_set_max_age(br, nla_get_u32(data[IFLA_BR_MAX_AGE]));
1013 if (data[IFLA_BR_AGEING_TIME]) {
1014 err = br_set_ageing_time(br, nla_get_u32(data[IFLA_BR_AGEING_TIME]));
1019 if (data[IFLA_BR_STP_STATE]) {
1020 u32 stp_enabled = nla_get_u32(data[IFLA_BR_STP_STATE]);
1022 br_stp_set_enabled(br, stp_enabled);
1025 if (data[IFLA_BR_PRIORITY]) {
1026 u32 priority = nla_get_u16(data[IFLA_BR_PRIORITY]);
1028 br_stp_set_bridge_priority(br, priority);
1031 if (data[IFLA_BR_VLAN_FILTERING]) {
1032 u8 vlan_filter = nla_get_u8(data[IFLA_BR_VLAN_FILTERING]);
1034 err = __br_vlan_filter_toggle(br, vlan_filter);
1039 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
1040 if (data[IFLA_BR_VLAN_PROTOCOL]) {
1041 __be16 vlan_proto = nla_get_be16(data[IFLA_BR_VLAN_PROTOCOL]);
1043 err = __br_vlan_set_proto(br, vlan_proto);
1048 if (data[IFLA_BR_VLAN_DEFAULT_PVID]) {
1049 __u16 defpvid = nla_get_u16(data[IFLA_BR_VLAN_DEFAULT_PVID]);
1051 err = __br_vlan_set_default_pvid(br, defpvid);
1056 if (data[IFLA_BR_VLAN_STATS_ENABLED]) {
1057 __u8 vlan_stats = nla_get_u8(data[IFLA_BR_VLAN_STATS_ENABLED]);
1059 err = br_vlan_set_stats(br, vlan_stats);
1065 if (data[IFLA_BR_GROUP_FWD_MASK]) {
1066 u16 fwd_mask = nla_get_u16(data[IFLA_BR_GROUP_FWD_MASK]);
1068 if (fwd_mask & BR_GROUPFWD_RESTRICTED)
1070 br->group_fwd_mask = fwd_mask;
1073 if (data[IFLA_BR_GROUP_ADDR]) {
1074 u8 new_addr[ETH_ALEN];
1076 if (nla_len(data[IFLA_BR_GROUP_ADDR]) != ETH_ALEN)
1078 memcpy(new_addr, nla_data(data[IFLA_BR_GROUP_ADDR]), ETH_ALEN);
1079 if (!is_link_local_ether_addr(new_addr))
1081 if (new_addr[5] == 1 || /* 802.3x Pause address */
1082 new_addr[5] == 2 || /* 802.3ad Slow protocols */
1083 new_addr[5] == 3) /* 802.1X PAE address */
1085 spin_lock_bh(&br->lock);
1086 memcpy(br->group_addr, new_addr, sizeof(br->group_addr));
1087 spin_unlock_bh(&br->lock);
1088 br->group_addr_set = true;
1089 br_recalculate_fwd_mask(br);
1092 if (data[IFLA_BR_FDB_FLUSH])
1095 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1096 if (data[IFLA_BR_MCAST_ROUTER]) {
1097 u8 multicast_router = nla_get_u8(data[IFLA_BR_MCAST_ROUTER]);
1099 err = br_multicast_set_router(br, multicast_router);
1104 if (data[IFLA_BR_MCAST_SNOOPING]) {
1105 u8 mcast_snooping = nla_get_u8(data[IFLA_BR_MCAST_SNOOPING]);
1107 err = br_multicast_toggle(br, mcast_snooping);
1112 if (data[IFLA_BR_MCAST_QUERY_USE_IFADDR]) {
1115 val = nla_get_u8(data[IFLA_BR_MCAST_QUERY_USE_IFADDR]);
1116 br->multicast_query_use_ifaddr = !!val;
1119 if (data[IFLA_BR_MCAST_QUERIER]) {
1120 u8 mcast_querier = nla_get_u8(data[IFLA_BR_MCAST_QUERIER]);
1122 err = br_multicast_set_querier(br, mcast_querier);
1127 if (data[IFLA_BR_MCAST_HASH_ELASTICITY]) {
1128 u32 val = nla_get_u32(data[IFLA_BR_MCAST_HASH_ELASTICITY]);
1130 br->hash_elasticity = val;
1133 if (data[IFLA_BR_MCAST_HASH_MAX]) {
1134 u32 hash_max = nla_get_u32(data[IFLA_BR_MCAST_HASH_MAX]);
1136 err = br_multicast_set_hash_max(br, hash_max);
1141 if (data[IFLA_BR_MCAST_LAST_MEMBER_CNT]) {
1142 u32 val = nla_get_u32(data[IFLA_BR_MCAST_LAST_MEMBER_CNT]);
1144 br->multicast_last_member_count = val;
1147 if (data[IFLA_BR_MCAST_STARTUP_QUERY_CNT]) {
1148 u32 val = nla_get_u32(data[IFLA_BR_MCAST_STARTUP_QUERY_CNT]);
1150 br->multicast_startup_query_count = val;
1153 if (data[IFLA_BR_MCAST_LAST_MEMBER_INTVL]) {
1154 u64 val = nla_get_u64(data[IFLA_BR_MCAST_LAST_MEMBER_INTVL]);
1156 br->multicast_last_member_interval = clock_t_to_jiffies(val);
1159 if (data[IFLA_BR_MCAST_MEMBERSHIP_INTVL]) {
1160 u64 val = nla_get_u64(data[IFLA_BR_MCAST_MEMBERSHIP_INTVL]);
1162 br->multicast_membership_interval = clock_t_to_jiffies(val);
1165 if (data[IFLA_BR_MCAST_QUERIER_INTVL]) {
1166 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERIER_INTVL]);
1168 br->multicast_querier_interval = clock_t_to_jiffies(val);
1171 if (data[IFLA_BR_MCAST_QUERY_INTVL]) {
1172 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERY_INTVL]);
1174 br->multicast_query_interval = clock_t_to_jiffies(val);
1177 if (data[IFLA_BR_MCAST_QUERY_RESPONSE_INTVL]) {
1178 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERY_RESPONSE_INTVL]);
1180 br->multicast_query_response_interval = clock_t_to_jiffies(val);
1183 if (data[IFLA_BR_MCAST_STARTUP_QUERY_INTVL]) {
1184 u64 val = nla_get_u64(data[IFLA_BR_MCAST_STARTUP_QUERY_INTVL]);
1186 br->multicast_startup_query_interval = clock_t_to_jiffies(val);
1189 if (data[IFLA_BR_MCAST_STATS_ENABLED]) {
1192 mcast_stats = nla_get_u8(data[IFLA_BR_MCAST_STATS_ENABLED]);
1193 br->multicast_stats_enabled = !!mcast_stats;
1196 if (data[IFLA_BR_MCAST_IGMP_VERSION]) {
1199 igmp_version = nla_get_u8(data[IFLA_BR_MCAST_IGMP_VERSION]);
1200 err = br_multicast_set_igmp_version(br, igmp_version);
1205 #if IS_ENABLED(CONFIG_IPV6)
1206 if (data[IFLA_BR_MCAST_MLD_VERSION]) {
1209 mld_version = nla_get_u8(data[IFLA_BR_MCAST_MLD_VERSION]);
1210 err = br_multicast_set_mld_version(br, mld_version);
1216 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
1217 if (data[IFLA_BR_NF_CALL_IPTABLES]) {
1218 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_IPTABLES]);
1220 br->nf_call_iptables = val ? true : false;
1223 if (data[IFLA_BR_NF_CALL_IP6TABLES]) {
1224 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_IP6TABLES]);
1226 br->nf_call_ip6tables = val ? true : false;
1229 if (data[IFLA_BR_NF_CALL_ARPTABLES]) {
1230 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_ARPTABLES]);
1232 br->nf_call_arptables = val ? true : false;
1239 static int br_dev_newlink(struct net *src_net, struct net_device *dev,
1240 struct nlattr *tb[], struct nlattr *data[],
1241 struct netlink_ext_ack *extack)
1243 struct net_bridge *br = netdev_priv(dev);
1246 if (tb[IFLA_ADDRESS]) {
1247 spin_lock_bh(&br->lock);
1248 br_stp_change_bridge_id(br, nla_data(tb[IFLA_ADDRESS]));
1249 spin_unlock_bh(&br->lock);
1252 err = register_netdevice(dev);
1256 err = br_changelink(dev, tb, data, extack);
1258 unregister_netdevice(dev);
1262 static size_t br_get_size(const struct net_device *brdev)
1264 return nla_total_size(sizeof(u32)) + /* IFLA_BR_FORWARD_DELAY */
1265 nla_total_size(sizeof(u32)) + /* IFLA_BR_HELLO_TIME */
1266 nla_total_size(sizeof(u32)) + /* IFLA_BR_MAX_AGE */
1267 nla_total_size(sizeof(u32)) + /* IFLA_BR_AGEING_TIME */
1268 nla_total_size(sizeof(u32)) + /* IFLA_BR_STP_STATE */
1269 nla_total_size(sizeof(u16)) + /* IFLA_BR_PRIORITY */
1270 nla_total_size(sizeof(u8)) + /* IFLA_BR_VLAN_FILTERING */
1271 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
1272 nla_total_size(sizeof(__be16)) + /* IFLA_BR_VLAN_PROTOCOL */
1273 nla_total_size(sizeof(u16)) + /* IFLA_BR_VLAN_DEFAULT_PVID */
1274 nla_total_size(sizeof(u8)) + /* IFLA_BR_VLAN_STATS_ENABLED */
1276 nla_total_size(sizeof(u16)) + /* IFLA_BR_GROUP_FWD_MASK */
1277 nla_total_size(sizeof(struct ifla_bridge_id)) + /* IFLA_BR_ROOT_ID */
1278 nla_total_size(sizeof(struct ifla_bridge_id)) + /* IFLA_BR_BRIDGE_ID */
1279 nla_total_size(sizeof(u16)) + /* IFLA_BR_ROOT_PORT */
1280 nla_total_size(sizeof(u32)) + /* IFLA_BR_ROOT_PATH_COST */
1281 nla_total_size(sizeof(u8)) + /* IFLA_BR_TOPOLOGY_CHANGE */
1282 nla_total_size(sizeof(u8)) + /* IFLA_BR_TOPOLOGY_CHANGE_DETECTED */
1283 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_HELLO_TIMER */
1284 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_TCN_TIMER */
1285 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_TOPOLOGY_CHANGE_TIMER */
1286 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_GC_TIMER */
1287 nla_total_size(ETH_ALEN) + /* IFLA_BR_GROUP_ADDR */
1288 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1289 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_ROUTER */
1290 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_SNOOPING */
1291 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_QUERY_USE_IFADDR */
1292 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_QUERIER */
1293 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_STATS_ENABLED */
1294 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_HASH_ELASTICITY */
1295 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_HASH_MAX */
1296 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_LAST_MEMBER_CNT */
1297 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_STARTUP_QUERY_CNT */
1298 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_LAST_MEMBER_INTVL */
1299 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_MEMBERSHIP_INTVL */
1300 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERIER_INTVL */
1301 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERY_INTVL */
1302 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERY_RESPONSE_INTVL */
1303 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_STARTUP_QUERY_INTVL */
1304 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_IGMP_VERSION */
1305 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_MLD_VERSION */
1307 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
1308 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_IPTABLES */
1309 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_IP6TABLES */
1310 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_ARPTABLES */
1315 static int br_fill_info(struct sk_buff *skb, const struct net_device *brdev)
1317 struct net_bridge *br = netdev_priv(brdev);
1318 u32 forward_delay = jiffies_to_clock_t(br->forward_delay);
1319 u32 hello_time = jiffies_to_clock_t(br->hello_time);
1320 u32 age_time = jiffies_to_clock_t(br->max_age);
1321 u32 ageing_time = jiffies_to_clock_t(br->ageing_time);
1322 u32 stp_enabled = br->stp_enabled;
1323 u16 priority = (br->bridge_id.prio[0] << 8) | br->bridge_id.prio[1];
1324 u8 vlan_enabled = br_vlan_enabled(br->dev);
1327 clockval = br_timer_value(&br->hello_timer);
1328 if (nla_put_u64_64bit(skb, IFLA_BR_HELLO_TIMER, clockval, IFLA_BR_PAD))
1330 clockval = br_timer_value(&br->tcn_timer);
1331 if (nla_put_u64_64bit(skb, IFLA_BR_TCN_TIMER, clockval, IFLA_BR_PAD))
1333 clockval = br_timer_value(&br->topology_change_timer);
1334 if (nla_put_u64_64bit(skb, IFLA_BR_TOPOLOGY_CHANGE_TIMER, clockval,
1337 clockval = br_timer_value(&br->gc_work.timer);
1338 if (nla_put_u64_64bit(skb, IFLA_BR_GC_TIMER, clockval, IFLA_BR_PAD))
1341 if (nla_put_u32(skb, IFLA_BR_FORWARD_DELAY, forward_delay) ||
1342 nla_put_u32(skb, IFLA_BR_HELLO_TIME, hello_time) ||
1343 nla_put_u32(skb, IFLA_BR_MAX_AGE, age_time) ||
1344 nla_put_u32(skb, IFLA_BR_AGEING_TIME, ageing_time) ||
1345 nla_put_u32(skb, IFLA_BR_STP_STATE, stp_enabled) ||
1346 nla_put_u16(skb, IFLA_BR_PRIORITY, priority) ||
1347 nla_put_u8(skb, IFLA_BR_VLAN_FILTERING, vlan_enabled) ||
1348 nla_put_u16(skb, IFLA_BR_GROUP_FWD_MASK, br->group_fwd_mask) ||
1349 nla_put(skb, IFLA_BR_BRIDGE_ID, sizeof(struct ifla_bridge_id),
1351 nla_put(skb, IFLA_BR_ROOT_ID, sizeof(struct ifla_bridge_id),
1352 &br->designated_root) ||
1353 nla_put_u16(skb, IFLA_BR_ROOT_PORT, br->root_port) ||
1354 nla_put_u32(skb, IFLA_BR_ROOT_PATH_COST, br->root_path_cost) ||
1355 nla_put_u8(skb, IFLA_BR_TOPOLOGY_CHANGE, br->topology_change) ||
1356 nla_put_u8(skb, IFLA_BR_TOPOLOGY_CHANGE_DETECTED,
1357 br->topology_change_detected) ||
1358 nla_put(skb, IFLA_BR_GROUP_ADDR, ETH_ALEN, br->group_addr))
1361 #ifdef CONFIG_BRIDGE_VLAN_FILTERING
1362 if (nla_put_be16(skb, IFLA_BR_VLAN_PROTOCOL, br->vlan_proto) ||
1363 nla_put_u16(skb, IFLA_BR_VLAN_DEFAULT_PVID, br->default_pvid) ||
1364 nla_put_u8(skb, IFLA_BR_VLAN_STATS_ENABLED, br->vlan_stats_enabled))
1367 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1368 if (nla_put_u8(skb, IFLA_BR_MCAST_ROUTER, br->multicast_router) ||
1369 nla_put_u8(skb, IFLA_BR_MCAST_SNOOPING, !br->multicast_disabled) ||
1370 nla_put_u8(skb, IFLA_BR_MCAST_QUERY_USE_IFADDR,
1371 br->multicast_query_use_ifaddr) ||
1372 nla_put_u8(skb, IFLA_BR_MCAST_QUERIER, br->multicast_querier) ||
1373 nla_put_u8(skb, IFLA_BR_MCAST_STATS_ENABLED,
1374 br->multicast_stats_enabled) ||
1375 nla_put_u32(skb, IFLA_BR_MCAST_HASH_ELASTICITY,
1376 br->hash_elasticity) ||
1377 nla_put_u32(skb, IFLA_BR_MCAST_HASH_MAX, br->hash_max) ||
1378 nla_put_u32(skb, IFLA_BR_MCAST_LAST_MEMBER_CNT,
1379 br->multicast_last_member_count) ||
1380 nla_put_u32(skb, IFLA_BR_MCAST_STARTUP_QUERY_CNT,
1381 br->multicast_startup_query_count) ||
1382 nla_put_u8(skb, IFLA_BR_MCAST_IGMP_VERSION,
1383 br->multicast_igmp_version))
1385 #if IS_ENABLED(CONFIG_IPV6)
1386 if (nla_put_u8(skb, IFLA_BR_MCAST_MLD_VERSION,
1387 br->multicast_mld_version))
1390 clockval = jiffies_to_clock_t(br->multicast_last_member_interval);
1391 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_LAST_MEMBER_INTVL, clockval,
1394 clockval = jiffies_to_clock_t(br->multicast_membership_interval);
1395 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_MEMBERSHIP_INTVL, clockval,
1398 clockval = jiffies_to_clock_t(br->multicast_querier_interval);
1399 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERIER_INTVL, clockval,
1402 clockval = jiffies_to_clock_t(br->multicast_query_interval);
1403 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERY_INTVL, clockval,
1406 clockval = jiffies_to_clock_t(br->multicast_query_response_interval);
1407 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERY_RESPONSE_INTVL, clockval,
1410 clockval = jiffies_to_clock_t(br->multicast_startup_query_interval);
1411 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_STARTUP_QUERY_INTVL, clockval,
1415 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
1416 if (nla_put_u8(skb, IFLA_BR_NF_CALL_IPTABLES,
1417 br->nf_call_iptables ? 1 : 0) ||
1418 nla_put_u8(skb, IFLA_BR_NF_CALL_IP6TABLES,
1419 br->nf_call_ip6tables ? 1 : 0) ||
1420 nla_put_u8(skb, IFLA_BR_NF_CALL_ARPTABLES,
1421 br->nf_call_arptables ? 1 : 0))
1428 static size_t br_get_linkxstats_size(const struct net_device *dev, int attr)
1430 struct net_bridge_port *p = NULL;
1431 struct net_bridge_vlan_group *vg;
1432 struct net_bridge_vlan *v;
1433 struct net_bridge *br;
1437 case IFLA_STATS_LINK_XSTATS:
1438 br = netdev_priv(dev);
1439 vg = br_vlan_group(br);
1441 case IFLA_STATS_LINK_XSTATS_SLAVE:
1442 p = br_port_get_rtnl(dev);
1446 vg = nbp_vlan_group(p);
1453 /* we need to count all, even placeholder entries */
1454 list_for_each_entry(v, &vg->vlan_list, vlist)
1458 return numvls * nla_total_size(sizeof(struct bridge_vlan_xstats)) +
1459 nla_total_size(sizeof(struct br_mcast_stats)) +
1463 static int br_fill_linkxstats(struct sk_buff *skb,
1464 const struct net_device *dev,
1465 int *prividx, int attr)
1467 struct nlattr *nla __maybe_unused;
1468 struct net_bridge_port *p = NULL;
1469 struct net_bridge_vlan_group *vg;
1470 struct net_bridge_vlan *v;
1471 struct net_bridge *br;
1472 struct nlattr *nest;
1476 case IFLA_STATS_LINK_XSTATS:
1477 br = netdev_priv(dev);
1478 vg = br_vlan_group(br);
1480 case IFLA_STATS_LINK_XSTATS_SLAVE:
1481 p = br_port_get_rtnl(dev);
1485 vg = nbp_vlan_group(p);
1491 nest = nla_nest_start(skb, LINK_XSTATS_TYPE_BRIDGE);
1498 pvid = br_get_pvid(vg);
1499 list_for_each_entry(v, &vg->vlan_list, vlist) {
1500 struct bridge_vlan_xstats vxi;
1501 struct br_vlan_stats stats;
1503 if (++vl_idx < *prividx)
1505 memset(&vxi, 0, sizeof(vxi));
1507 vxi.flags = v->flags;
1509 vxi.flags |= BRIDGE_VLAN_INFO_PVID;
1510 br_vlan_get_stats(v, &stats);
1511 vxi.rx_bytes = stats.rx_bytes;
1512 vxi.rx_packets = stats.rx_packets;
1513 vxi.tx_bytes = stats.tx_bytes;
1514 vxi.tx_packets = stats.tx_packets;
1516 if (nla_put(skb, BRIDGE_XSTATS_VLAN, sizeof(vxi), &vxi))
1517 goto nla_put_failure;
1521 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING
1522 if (++vl_idx >= *prividx) {
1523 nla = nla_reserve_64bit(skb, BRIDGE_XSTATS_MCAST,
1524 sizeof(struct br_mcast_stats),
1527 goto nla_put_failure;
1528 br_multicast_get_stats(br, p, nla_data(nla));
1531 nla_nest_end(skb, nest);
1537 nla_nest_end(skb, nest);
1543 static struct rtnl_af_ops br_af_ops __read_mostly = {
1544 .family = AF_BRIDGE,
1545 .get_link_af_size = br_get_link_af_size_filtered,
1548 struct rtnl_link_ops br_link_ops __read_mostly = {
1550 .priv_size = sizeof(struct net_bridge),
1551 .setup = br_dev_setup,
1552 .maxtype = IFLA_BR_MAX,
1553 .policy = br_policy,
1554 .validate = br_validate,
1555 .newlink = br_dev_newlink,
1556 .changelink = br_changelink,
1557 .dellink = br_dev_delete,
1558 .get_size = br_get_size,
1559 .fill_info = br_fill_info,
1560 .fill_linkxstats = br_fill_linkxstats,
1561 .get_linkxstats_size = br_get_linkxstats_size,
1563 .slave_maxtype = IFLA_BRPORT_MAX,
1564 .slave_policy = br_port_policy,
1565 .slave_changelink = br_port_slave_changelink,
1566 .get_slave_size = br_port_get_slave_size,
1567 .fill_slave_info = br_port_fill_slave_info,
1570 int __init br_netlink_init(void)
1575 rtnl_af_register(&br_af_ops);
1577 err = rtnl_link_register(&br_link_ops);
1584 rtnl_af_unregister(&br_af_ops);
1589 void br_netlink_fini(void)
1592 rtnl_af_unregister(&br_af_ops);
1593 rtnl_link_unregister(&br_link_ops);