xref: /openbmc/linux/net/dsa/slave.c (revision 76ce0265)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * net/dsa/slave.c - Slave device handling
4  * Copyright (c) 2008-2009 Marvell Semiconductor
5  */
6 
7 #include <linux/list.h>
8 #include <linux/etherdevice.h>
9 #include <linux/netdevice.h>
10 #include <linux/phy.h>
11 #include <linux/phy_fixed.h>
12 #include <linux/phylink.h>
13 #include <linux/of_net.h>
14 #include <linux/of_mdio.h>
15 #include <linux/mdio.h>
16 #include <net/rtnetlink.h>
17 #include <net/pkt_cls.h>
18 #include <net/tc_act/tc_mirred.h>
19 #include <linux/if_bridge.h>
20 #include <linux/netpoll.h>
21 #include <linux/ptp_classify.h>
22 
23 #include "dsa_priv.h"
24 
25 /* slave mii_bus handling ***************************************************/
26 static int dsa_slave_phy_read(struct mii_bus *bus, int addr, int reg)
27 {
28 	struct dsa_switch *ds = bus->priv;
29 
30 	if (ds->phys_mii_mask & (1 << addr))
31 		return ds->ops->phy_read(ds, addr, reg);
32 
33 	return 0xffff;
34 }
35 
36 static int dsa_slave_phy_write(struct mii_bus *bus, int addr, int reg, u16 val)
37 {
38 	struct dsa_switch *ds = bus->priv;
39 
40 	if (ds->phys_mii_mask & (1 << addr))
41 		return ds->ops->phy_write(ds, addr, reg, val);
42 
43 	return 0;
44 }
45 
46 void dsa_slave_mii_bus_init(struct dsa_switch *ds)
47 {
48 	ds->slave_mii_bus->priv = (void *)ds;
49 	ds->slave_mii_bus->name = "dsa slave smi";
50 	ds->slave_mii_bus->read = dsa_slave_phy_read;
51 	ds->slave_mii_bus->write = dsa_slave_phy_write;
52 	snprintf(ds->slave_mii_bus->id, MII_BUS_ID_SIZE, "dsa-%d.%d",
53 		 ds->dst->index, ds->index);
54 	ds->slave_mii_bus->parent = ds->dev;
55 	ds->slave_mii_bus->phy_mask = ~ds->phys_mii_mask;
56 }
57 
58 
59 /* slave device handling ****************************************************/
60 static int dsa_slave_get_iflink(const struct net_device *dev)
61 {
62 	return dsa_slave_to_master(dev)->ifindex;
63 }
64 
65 static int dsa_slave_open(struct net_device *dev)
66 {
67 	struct net_device *master = dsa_slave_to_master(dev);
68 	struct dsa_port *dp = dsa_slave_to_port(dev);
69 	int err;
70 
71 	if (!(master->flags & IFF_UP))
72 		return -ENETDOWN;
73 
74 	if (!ether_addr_equal(dev->dev_addr, master->dev_addr)) {
75 		err = dev_uc_add(master, dev->dev_addr);
76 		if (err < 0)
77 			goto out;
78 	}
79 
80 	if (dev->flags & IFF_ALLMULTI) {
81 		err = dev_set_allmulti(master, 1);
82 		if (err < 0)
83 			goto del_unicast;
84 	}
85 	if (dev->flags & IFF_PROMISC) {
86 		err = dev_set_promiscuity(master, 1);
87 		if (err < 0)
88 			goto clear_allmulti;
89 	}
90 
91 	err = dsa_port_enable(dp, dev->phydev);
92 	if (err)
93 		goto clear_promisc;
94 
95 	phylink_start(dp->pl);
96 
97 	return 0;
98 
99 clear_promisc:
100 	if (dev->flags & IFF_PROMISC)
101 		dev_set_promiscuity(master, -1);
102 clear_allmulti:
103 	if (dev->flags & IFF_ALLMULTI)
104 		dev_set_allmulti(master, -1);
105 del_unicast:
106 	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
107 		dev_uc_del(master, dev->dev_addr);
108 out:
109 	return err;
110 }
111 
112 static int dsa_slave_close(struct net_device *dev)
113 {
114 	struct net_device *master = dsa_slave_to_master(dev);
115 	struct dsa_port *dp = dsa_slave_to_port(dev);
116 
117 	phylink_stop(dp->pl);
118 
119 	dsa_port_disable(dp);
120 
121 	dev_mc_unsync(master, dev);
122 	dev_uc_unsync(master, dev);
123 	if (dev->flags & IFF_ALLMULTI)
124 		dev_set_allmulti(master, -1);
125 	if (dev->flags & IFF_PROMISC)
126 		dev_set_promiscuity(master, -1);
127 
128 	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
129 		dev_uc_del(master, dev->dev_addr);
130 
131 	return 0;
132 }
133 
134 static void dsa_slave_change_rx_flags(struct net_device *dev, int change)
135 {
136 	struct net_device *master = dsa_slave_to_master(dev);
137 	if (dev->flags & IFF_UP) {
138 		if (change & IFF_ALLMULTI)
139 			dev_set_allmulti(master,
140 					 dev->flags & IFF_ALLMULTI ? 1 : -1);
141 		if (change & IFF_PROMISC)
142 			dev_set_promiscuity(master,
143 					    dev->flags & IFF_PROMISC ? 1 : -1);
144 	}
145 }
146 
147 static void dsa_slave_set_rx_mode(struct net_device *dev)
148 {
149 	struct net_device *master = dsa_slave_to_master(dev);
150 
151 	dev_mc_sync(master, dev);
152 	dev_uc_sync(master, dev);
153 }
154 
155 static int dsa_slave_set_mac_address(struct net_device *dev, void *a)
156 {
157 	struct net_device *master = dsa_slave_to_master(dev);
158 	struct sockaddr *addr = a;
159 	int err;
160 
161 	if (!is_valid_ether_addr(addr->sa_data))
162 		return -EADDRNOTAVAIL;
163 
164 	if (!(dev->flags & IFF_UP))
165 		goto out;
166 
167 	if (!ether_addr_equal(addr->sa_data, master->dev_addr)) {
168 		err = dev_uc_add(master, addr->sa_data);
169 		if (err < 0)
170 			return err;
171 	}
172 
173 	if (!ether_addr_equal(dev->dev_addr, master->dev_addr))
174 		dev_uc_del(master, dev->dev_addr);
175 
176 out:
177 	ether_addr_copy(dev->dev_addr, addr->sa_data);
178 
179 	return 0;
180 }
181 
182 struct dsa_slave_dump_ctx {
183 	struct net_device *dev;
184 	struct sk_buff *skb;
185 	struct netlink_callback *cb;
186 	int idx;
187 };
188 
189 static int
190 dsa_slave_port_fdb_do_dump(const unsigned char *addr, u16 vid,
191 			   bool is_static, void *data)
192 {
193 	struct dsa_slave_dump_ctx *dump = data;
194 	u32 portid = NETLINK_CB(dump->cb->skb).portid;
195 	u32 seq = dump->cb->nlh->nlmsg_seq;
196 	struct nlmsghdr *nlh;
197 	struct ndmsg *ndm;
198 
199 	if (dump->idx < dump->cb->args[2])
200 		goto skip;
201 
202 	nlh = nlmsg_put(dump->skb, portid, seq, RTM_NEWNEIGH,
203 			sizeof(*ndm), NLM_F_MULTI);
204 	if (!nlh)
205 		return -EMSGSIZE;
206 
207 	ndm = nlmsg_data(nlh);
208 	ndm->ndm_family  = AF_BRIDGE;
209 	ndm->ndm_pad1    = 0;
210 	ndm->ndm_pad2    = 0;
211 	ndm->ndm_flags   = NTF_SELF;
212 	ndm->ndm_type    = 0;
213 	ndm->ndm_ifindex = dump->dev->ifindex;
214 	ndm->ndm_state   = is_static ? NUD_NOARP : NUD_REACHABLE;
215 
216 	if (nla_put(dump->skb, NDA_LLADDR, ETH_ALEN, addr))
217 		goto nla_put_failure;
218 
219 	if (vid && nla_put_u16(dump->skb, NDA_VLAN, vid))
220 		goto nla_put_failure;
221 
222 	nlmsg_end(dump->skb, nlh);
223 
224 skip:
225 	dump->idx++;
226 	return 0;
227 
228 nla_put_failure:
229 	nlmsg_cancel(dump->skb, nlh);
230 	return -EMSGSIZE;
231 }
232 
233 static int
234 dsa_slave_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb,
235 		   struct net_device *dev, struct net_device *filter_dev,
236 		   int *idx)
237 {
238 	struct dsa_port *dp = dsa_slave_to_port(dev);
239 	struct dsa_slave_dump_ctx dump = {
240 		.dev = dev,
241 		.skb = skb,
242 		.cb = cb,
243 		.idx = *idx,
244 	};
245 	int err;
246 
247 	err = dsa_port_fdb_dump(dp, dsa_slave_port_fdb_do_dump, &dump);
248 	*idx = dump.idx;
249 
250 	return err;
251 }
252 
253 static int dsa_slave_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
254 {
255 	struct dsa_slave_priv *p = netdev_priv(dev);
256 	struct dsa_switch *ds = p->dp->ds;
257 	int port = p->dp->index;
258 
259 	/* Pass through to switch driver if it supports timestamping */
260 	switch (cmd) {
261 	case SIOCGHWTSTAMP:
262 		if (ds->ops->port_hwtstamp_get)
263 			return ds->ops->port_hwtstamp_get(ds, port, ifr);
264 		break;
265 	case SIOCSHWTSTAMP:
266 		if (ds->ops->port_hwtstamp_set)
267 			return ds->ops->port_hwtstamp_set(ds, port, ifr);
268 		break;
269 	}
270 
271 	return phylink_mii_ioctl(p->dp->pl, ifr, cmd);
272 }
273 
274 static int dsa_slave_port_attr_set(struct net_device *dev,
275 				   const struct switchdev_attr *attr,
276 				   struct switchdev_trans *trans)
277 {
278 	struct dsa_port *dp = dsa_slave_to_port(dev);
279 	int ret;
280 
281 	switch (attr->id) {
282 	case SWITCHDEV_ATTR_ID_PORT_STP_STATE:
283 		ret = dsa_port_set_state(dp, attr->u.stp_state, trans);
284 		break;
285 	case SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING:
286 		ret = dsa_port_vlan_filtering(dp, attr->u.vlan_filtering,
287 					      trans);
288 		break;
289 	case SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME:
290 		ret = dsa_port_ageing_time(dp, attr->u.ageing_time, trans);
291 		break;
292 	case SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS:
293 		ret = dsa_port_pre_bridge_flags(dp, attr->u.brport_flags,
294 						trans);
295 		break;
296 	case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS:
297 		ret = dsa_port_bridge_flags(dp, attr->u.brport_flags, trans);
298 		break;
299 	case SWITCHDEV_ATTR_ID_BRIDGE_MROUTER:
300 		ret = dsa_port_mrouter(dp->cpu_dp, attr->u.mrouter, trans);
301 		break;
302 	default:
303 		ret = -EOPNOTSUPP;
304 		break;
305 	}
306 
307 	return ret;
308 }
309 
310 static int dsa_slave_vlan_add(struct net_device *dev,
311 			      const struct switchdev_obj *obj,
312 			      struct switchdev_trans *trans)
313 {
314 	struct dsa_port *dp = dsa_slave_to_port(dev);
315 	struct switchdev_obj_port_vlan vlan;
316 	int err;
317 
318 	if (obj->orig_dev != dev)
319 		return -EOPNOTSUPP;
320 
321 	if (dp->bridge_dev && !br_vlan_enabled(dp->bridge_dev))
322 		return 0;
323 
324 	vlan = *SWITCHDEV_OBJ_PORT_VLAN(obj);
325 
326 	err = dsa_port_vlan_add(dp, &vlan, trans);
327 	if (err)
328 		return err;
329 
330 	/* We need the dedicated CPU port to be a member of the VLAN as well.
331 	 * Even though drivers often handle CPU membership in special ways,
332 	 * it doesn't make sense to program a PVID, so clear this flag.
333 	 */
334 	vlan.flags &= ~BRIDGE_VLAN_INFO_PVID;
335 
336 	err = dsa_port_vlan_add(dp->cpu_dp, &vlan, trans);
337 	if (err)
338 		return err;
339 
340 	return 0;
341 }
342 
343 static int dsa_slave_port_obj_add(struct net_device *dev,
344 				  const struct switchdev_obj *obj,
345 				  struct switchdev_trans *trans,
346 				  struct netlink_ext_ack *extack)
347 {
348 	struct dsa_port *dp = dsa_slave_to_port(dev);
349 	int err;
350 
351 	/* For the prepare phase, ensure the full set of changes is feasable in
352 	 * one go in order to signal a failure properly. If an operation is not
353 	 * supported, return -EOPNOTSUPP.
354 	 */
355 
356 	switch (obj->id) {
357 	case SWITCHDEV_OBJ_ID_PORT_MDB:
358 		if (obj->orig_dev != dev)
359 			return -EOPNOTSUPP;
360 		err = dsa_port_mdb_add(dp, SWITCHDEV_OBJ_PORT_MDB(obj), trans);
361 		break;
362 	case SWITCHDEV_OBJ_ID_HOST_MDB:
363 		/* DSA can directly translate this to a normal MDB add,
364 		 * but on the CPU port.
365 		 */
366 		err = dsa_port_mdb_add(dp->cpu_dp, SWITCHDEV_OBJ_PORT_MDB(obj),
367 				       trans);
368 		break;
369 	case SWITCHDEV_OBJ_ID_PORT_VLAN:
370 		err = dsa_slave_vlan_add(dev, obj, trans);
371 		break;
372 	default:
373 		err = -EOPNOTSUPP;
374 		break;
375 	}
376 
377 	return err;
378 }
379 
380 static int dsa_slave_vlan_del(struct net_device *dev,
381 			      const struct switchdev_obj *obj)
382 {
383 	struct dsa_port *dp = dsa_slave_to_port(dev);
384 
385 	if (obj->orig_dev != dev)
386 		return -EOPNOTSUPP;
387 
388 	if (dp->bridge_dev && !br_vlan_enabled(dp->bridge_dev))
389 		return 0;
390 
391 	/* Do not deprogram the CPU port as it may be shared with other user
392 	 * ports which can be members of this VLAN as well.
393 	 */
394 	return dsa_port_vlan_del(dp, SWITCHDEV_OBJ_PORT_VLAN(obj));
395 }
396 
397 static int dsa_slave_port_obj_del(struct net_device *dev,
398 				  const struct switchdev_obj *obj)
399 {
400 	struct dsa_port *dp = dsa_slave_to_port(dev);
401 	int err;
402 
403 	switch (obj->id) {
404 	case SWITCHDEV_OBJ_ID_PORT_MDB:
405 		if (obj->orig_dev != dev)
406 			return -EOPNOTSUPP;
407 		err = dsa_port_mdb_del(dp, SWITCHDEV_OBJ_PORT_MDB(obj));
408 		break;
409 	case SWITCHDEV_OBJ_ID_HOST_MDB:
410 		/* DSA can directly translate this to a normal MDB add,
411 		 * but on the CPU port.
412 		 */
413 		err = dsa_port_mdb_del(dp->cpu_dp, SWITCHDEV_OBJ_PORT_MDB(obj));
414 		break;
415 	case SWITCHDEV_OBJ_ID_PORT_VLAN:
416 		err = dsa_slave_vlan_del(dev, obj);
417 		break;
418 	default:
419 		err = -EOPNOTSUPP;
420 		break;
421 	}
422 
423 	return err;
424 }
425 
426 static int dsa_slave_get_port_parent_id(struct net_device *dev,
427 					struct netdev_phys_item_id *ppid)
428 {
429 	struct dsa_port *dp = dsa_slave_to_port(dev);
430 	struct dsa_switch *ds = dp->ds;
431 	struct dsa_switch_tree *dst = ds->dst;
432 
433 	/* For non-legacy ports, devlink is used and it takes
434 	 * care of the name generation. This ndo implementation
435 	 * should be removed with legacy support.
436 	 */
437 	if (dp->ds->devlink)
438 		return -EOPNOTSUPP;
439 
440 	ppid->id_len = sizeof(dst->index);
441 	memcpy(&ppid->id, &dst->index, ppid->id_len);
442 
443 	return 0;
444 }
445 
446 static inline netdev_tx_t dsa_slave_netpoll_send_skb(struct net_device *dev,
447 						     struct sk_buff *skb)
448 {
449 #ifdef CONFIG_NET_POLL_CONTROLLER
450 	struct dsa_slave_priv *p = netdev_priv(dev);
451 
452 	if (p->netpoll)
453 		netpoll_send_skb(p->netpoll, skb);
454 #else
455 	BUG();
456 #endif
457 	return NETDEV_TX_OK;
458 }
459 
460 static void dsa_skb_tx_timestamp(struct dsa_slave_priv *p,
461 				 struct sk_buff *skb)
462 {
463 	struct dsa_switch *ds = p->dp->ds;
464 	struct sk_buff *clone;
465 	unsigned int type;
466 
467 	type = ptp_classify_raw(skb);
468 	if (type == PTP_CLASS_NONE)
469 		return;
470 
471 	if (!ds->ops->port_txtstamp)
472 		return;
473 
474 	clone = skb_clone_sk(skb);
475 	if (!clone)
476 		return;
477 
478 	DSA_SKB_CB(skb)->clone = clone;
479 
480 	if (ds->ops->port_txtstamp(ds, p->dp->index, clone, type))
481 		return;
482 
483 	kfree_skb(clone);
484 }
485 
486 netdev_tx_t dsa_enqueue_skb(struct sk_buff *skb, struct net_device *dev)
487 {
488 	/* SKB for netpoll still need to be mangled with the protocol-specific
489 	 * tag to be successfully transmitted
490 	 */
491 	if (unlikely(netpoll_tx_running(dev)))
492 		return dsa_slave_netpoll_send_skb(dev, skb);
493 
494 	/* Queue the SKB for transmission on the parent interface, but
495 	 * do not modify its EtherType
496 	 */
497 	skb->dev = dsa_slave_to_master(dev);
498 	dev_queue_xmit(skb);
499 
500 	return NETDEV_TX_OK;
501 }
502 EXPORT_SYMBOL_GPL(dsa_enqueue_skb);
503 
504 static netdev_tx_t dsa_slave_xmit(struct sk_buff *skb, struct net_device *dev)
505 {
506 	struct dsa_slave_priv *p = netdev_priv(dev);
507 	struct pcpu_sw_netstats *s;
508 	struct sk_buff *nskb;
509 
510 	s = this_cpu_ptr(p->stats64);
511 	u64_stats_update_begin(&s->syncp);
512 	s->tx_packets++;
513 	s->tx_bytes += skb->len;
514 	u64_stats_update_end(&s->syncp);
515 
516 	DSA_SKB_CB(skb)->clone = NULL;
517 
518 	/* Identify PTP protocol packets, clone them, and pass them to the
519 	 * switch driver
520 	 */
521 	dsa_skb_tx_timestamp(p, skb);
522 
523 	/* Transmit function may have to reallocate the original SKB,
524 	 * in which case it must have freed it. Only free it here on error.
525 	 */
526 	nskb = p->xmit(skb, dev);
527 	if (!nskb) {
528 		kfree_skb(skb);
529 		return NETDEV_TX_OK;
530 	}
531 
532 	return dsa_enqueue_skb(nskb, dev);
533 }
534 
535 /* ethtool operations *******************************************************/
536 
537 static void dsa_slave_get_drvinfo(struct net_device *dev,
538 				  struct ethtool_drvinfo *drvinfo)
539 {
540 	strlcpy(drvinfo->driver, "dsa", sizeof(drvinfo->driver));
541 	strlcpy(drvinfo->fw_version, "N/A", sizeof(drvinfo->fw_version));
542 	strlcpy(drvinfo->bus_info, "platform", sizeof(drvinfo->bus_info));
543 }
544 
545 static int dsa_slave_get_regs_len(struct net_device *dev)
546 {
547 	struct dsa_port *dp = dsa_slave_to_port(dev);
548 	struct dsa_switch *ds = dp->ds;
549 
550 	if (ds->ops->get_regs_len)
551 		return ds->ops->get_regs_len(ds, dp->index);
552 
553 	return -EOPNOTSUPP;
554 }
555 
556 static void
557 dsa_slave_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *_p)
558 {
559 	struct dsa_port *dp = dsa_slave_to_port(dev);
560 	struct dsa_switch *ds = dp->ds;
561 
562 	if (ds->ops->get_regs)
563 		ds->ops->get_regs(ds, dp->index, regs, _p);
564 }
565 
566 static int dsa_slave_nway_reset(struct net_device *dev)
567 {
568 	struct dsa_port *dp = dsa_slave_to_port(dev);
569 
570 	return phylink_ethtool_nway_reset(dp->pl);
571 }
572 
573 static int dsa_slave_get_eeprom_len(struct net_device *dev)
574 {
575 	struct dsa_port *dp = dsa_slave_to_port(dev);
576 	struct dsa_switch *ds = dp->ds;
577 
578 	if (ds->cd && ds->cd->eeprom_len)
579 		return ds->cd->eeprom_len;
580 
581 	if (ds->ops->get_eeprom_len)
582 		return ds->ops->get_eeprom_len(ds);
583 
584 	return 0;
585 }
586 
587 static int dsa_slave_get_eeprom(struct net_device *dev,
588 				struct ethtool_eeprom *eeprom, u8 *data)
589 {
590 	struct dsa_port *dp = dsa_slave_to_port(dev);
591 	struct dsa_switch *ds = dp->ds;
592 
593 	if (ds->ops->get_eeprom)
594 		return ds->ops->get_eeprom(ds, eeprom, data);
595 
596 	return -EOPNOTSUPP;
597 }
598 
599 static int dsa_slave_set_eeprom(struct net_device *dev,
600 				struct ethtool_eeprom *eeprom, u8 *data)
601 {
602 	struct dsa_port *dp = dsa_slave_to_port(dev);
603 	struct dsa_switch *ds = dp->ds;
604 
605 	if (ds->ops->set_eeprom)
606 		return ds->ops->set_eeprom(ds, eeprom, data);
607 
608 	return -EOPNOTSUPP;
609 }
610 
611 static void dsa_slave_get_strings(struct net_device *dev,
612 				  uint32_t stringset, uint8_t *data)
613 {
614 	struct dsa_port *dp = dsa_slave_to_port(dev);
615 	struct dsa_switch *ds = dp->ds;
616 
617 	if (stringset == ETH_SS_STATS) {
618 		int len = ETH_GSTRING_LEN;
619 
620 		strncpy(data, "tx_packets", len);
621 		strncpy(data + len, "tx_bytes", len);
622 		strncpy(data + 2 * len, "rx_packets", len);
623 		strncpy(data + 3 * len, "rx_bytes", len);
624 		if (ds->ops->get_strings)
625 			ds->ops->get_strings(ds, dp->index, stringset,
626 					     data + 4 * len);
627 	}
628 }
629 
630 static void dsa_slave_get_ethtool_stats(struct net_device *dev,
631 					struct ethtool_stats *stats,
632 					uint64_t *data)
633 {
634 	struct dsa_port *dp = dsa_slave_to_port(dev);
635 	struct dsa_slave_priv *p = netdev_priv(dev);
636 	struct dsa_switch *ds = dp->ds;
637 	struct pcpu_sw_netstats *s;
638 	unsigned int start;
639 	int i;
640 
641 	for_each_possible_cpu(i) {
642 		u64 tx_packets, tx_bytes, rx_packets, rx_bytes;
643 
644 		s = per_cpu_ptr(p->stats64, i);
645 		do {
646 			start = u64_stats_fetch_begin_irq(&s->syncp);
647 			tx_packets = s->tx_packets;
648 			tx_bytes = s->tx_bytes;
649 			rx_packets = s->rx_packets;
650 			rx_bytes = s->rx_bytes;
651 		} while (u64_stats_fetch_retry_irq(&s->syncp, start));
652 		data[0] += tx_packets;
653 		data[1] += tx_bytes;
654 		data[2] += rx_packets;
655 		data[3] += rx_bytes;
656 	}
657 	if (ds->ops->get_ethtool_stats)
658 		ds->ops->get_ethtool_stats(ds, dp->index, data + 4);
659 }
660 
661 static int dsa_slave_get_sset_count(struct net_device *dev, int sset)
662 {
663 	struct dsa_port *dp = dsa_slave_to_port(dev);
664 	struct dsa_switch *ds = dp->ds;
665 
666 	if (sset == ETH_SS_STATS) {
667 		int count;
668 
669 		count = 4;
670 		if (ds->ops->get_sset_count)
671 			count += ds->ops->get_sset_count(ds, dp->index, sset);
672 
673 		return count;
674 	}
675 
676 	return -EOPNOTSUPP;
677 }
678 
679 static void dsa_slave_get_wol(struct net_device *dev, struct ethtool_wolinfo *w)
680 {
681 	struct dsa_port *dp = dsa_slave_to_port(dev);
682 	struct dsa_switch *ds = dp->ds;
683 
684 	phylink_ethtool_get_wol(dp->pl, w);
685 
686 	if (ds->ops->get_wol)
687 		ds->ops->get_wol(ds, dp->index, w);
688 }
689 
690 static int dsa_slave_set_wol(struct net_device *dev, struct ethtool_wolinfo *w)
691 {
692 	struct dsa_port *dp = dsa_slave_to_port(dev);
693 	struct dsa_switch *ds = dp->ds;
694 	int ret = -EOPNOTSUPP;
695 
696 	phylink_ethtool_set_wol(dp->pl, w);
697 
698 	if (ds->ops->set_wol)
699 		ret = ds->ops->set_wol(ds, dp->index, w);
700 
701 	return ret;
702 }
703 
704 static int dsa_slave_set_eee(struct net_device *dev, struct ethtool_eee *e)
705 {
706 	struct dsa_port *dp = dsa_slave_to_port(dev);
707 	struct dsa_switch *ds = dp->ds;
708 	int ret;
709 
710 	/* Port's PHY and MAC both need to be EEE capable */
711 	if (!dev->phydev || !dp->pl)
712 		return -ENODEV;
713 
714 	if (!ds->ops->set_mac_eee)
715 		return -EOPNOTSUPP;
716 
717 	ret = ds->ops->set_mac_eee(ds, dp->index, e);
718 	if (ret)
719 		return ret;
720 
721 	return phylink_ethtool_set_eee(dp->pl, e);
722 }
723 
724 static int dsa_slave_get_eee(struct net_device *dev, struct ethtool_eee *e)
725 {
726 	struct dsa_port *dp = dsa_slave_to_port(dev);
727 	struct dsa_switch *ds = dp->ds;
728 	int ret;
729 
730 	/* Port's PHY and MAC both need to be EEE capable */
731 	if (!dev->phydev || !dp->pl)
732 		return -ENODEV;
733 
734 	if (!ds->ops->get_mac_eee)
735 		return -EOPNOTSUPP;
736 
737 	ret = ds->ops->get_mac_eee(ds, dp->index, e);
738 	if (ret)
739 		return ret;
740 
741 	return phylink_ethtool_get_eee(dp->pl, e);
742 }
743 
744 static int dsa_slave_get_link_ksettings(struct net_device *dev,
745 					struct ethtool_link_ksettings *cmd)
746 {
747 	struct dsa_port *dp = dsa_slave_to_port(dev);
748 
749 	return phylink_ethtool_ksettings_get(dp->pl, cmd);
750 }
751 
752 static int dsa_slave_set_link_ksettings(struct net_device *dev,
753 					const struct ethtool_link_ksettings *cmd)
754 {
755 	struct dsa_port *dp = dsa_slave_to_port(dev);
756 
757 	return phylink_ethtool_ksettings_set(dp->pl, cmd);
758 }
759 
760 static void dsa_slave_get_pauseparam(struct net_device *dev,
761 				     struct ethtool_pauseparam *pause)
762 {
763 	struct dsa_port *dp = dsa_slave_to_port(dev);
764 
765 	phylink_ethtool_get_pauseparam(dp->pl, pause);
766 }
767 
768 static int dsa_slave_set_pauseparam(struct net_device *dev,
769 				    struct ethtool_pauseparam *pause)
770 {
771 	struct dsa_port *dp = dsa_slave_to_port(dev);
772 
773 	return phylink_ethtool_set_pauseparam(dp->pl, pause);
774 }
775 
776 #ifdef CONFIG_NET_POLL_CONTROLLER
777 static int dsa_slave_netpoll_setup(struct net_device *dev,
778 				   struct netpoll_info *ni)
779 {
780 	struct net_device *master = dsa_slave_to_master(dev);
781 	struct dsa_slave_priv *p = netdev_priv(dev);
782 	struct netpoll *netpoll;
783 	int err = 0;
784 
785 	netpoll = kzalloc(sizeof(*netpoll), GFP_KERNEL);
786 	if (!netpoll)
787 		return -ENOMEM;
788 
789 	err = __netpoll_setup(netpoll, master);
790 	if (err) {
791 		kfree(netpoll);
792 		goto out;
793 	}
794 
795 	p->netpoll = netpoll;
796 out:
797 	return err;
798 }
799 
800 static void dsa_slave_netpoll_cleanup(struct net_device *dev)
801 {
802 	struct dsa_slave_priv *p = netdev_priv(dev);
803 	struct netpoll *netpoll = p->netpoll;
804 
805 	if (!netpoll)
806 		return;
807 
808 	p->netpoll = NULL;
809 
810 	__netpoll_free(netpoll);
811 }
812 
813 static void dsa_slave_poll_controller(struct net_device *dev)
814 {
815 }
816 #endif
817 
818 static int dsa_slave_get_phys_port_name(struct net_device *dev,
819 					char *name, size_t len)
820 {
821 	struct dsa_port *dp = dsa_slave_to_port(dev);
822 
823 	/* For non-legacy ports, devlink is used and it takes
824 	 * care of the name generation. This ndo implementation
825 	 * should be removed with legacy support.
826 	 */
827 	if (dp->ds->devlink)
828 		return -EOPNOTSUPP;
829 
830 	if (snprintf(name, len, "p%d", dp->index) >= len)
831 		return -EINVAL;
832 
833 	return 0;
834 }
835 
836 static struct dsa_mall_tc_entry *
837 dsa_slave_mall_tc_entry_find(struct net_device *dev, unsigned long cookie)
838 {
839 	struct dsa_slave_priv *p = netdev_priv(dev);
840 	struct dsa_mall_tc_entry *mall_tc_entry;
841 
842 	list_for_each_entry(mall_tc_entry, &p->mall_tc_list, list)
843 		if (mall_tc_entry->cookie == cookie)
844 			return mall_tc_entry;
845 
846 	return NULL;
847 }
848 
849 static int dsa_slave_add_cls_matchall(struct net_device *dev,
850 				      struct tc_cls_matchall_offload *cls,
851 				      bool ingress)
852 {
853 	struct dsa_port *dp = dsa_slave_to_port(dev);
854 	struct dsa_slave_priv *p = netdev_priv(dev);
855 	struct dsa_mall_tc_entry *mall_tc_entry;
856 	__be16 protocol = cls->common.protocol;
857 	struct dsa_switch *ds = dp->ds;
858 	struct flow_action_entry *act;
859 	struct dsa_port *to_dp;
860 	int err = -EOPNOTSUPP;
861 
862 	if (!ds->ops->port_mirror_add)
863 		return err;
864 
865 	if (!flow_offload_has_one_action(&cls->rule->action))
866 		return err;
867 
868 	act = &cls->rule->action.entries[0];
869 
870 	if (act->id == FLOW_ACTION_MIRRED && protocol == htons(ETH_P_ALL)) {
871 		struct dsa_mall_mirror_tc_entry *mirror;
872 
873 		if (!act->dev)
874 			return -EINVAL;
875 
876 		if (!dsa_slave_dev_check(act->dev))
877 			return -EOPNOTSUPP;
878 
879 		mall_tc_entry = kzalloc(sizeof(*mall_tc_entry), GFP_KERNEL);
880 		if (!mall_tc_entry)
881 			return -ENOMEM;
882 
883 		mall_tc_entry->cookie = cls->cookie;
884 		mall_tc_entry->type = DSA_PORT_MALL_MIRROR;
885 		mirror = &mall_tc_entry->mirror;
886 
887 		to_dp = dsa_slave_to_port(act->dev);
888 
889 		mirror->to_local_port = to_dp->index;
890 		mirror->ingress = ingress;
891 
892 		err = ds->ops->port_mirror_add(ds, dp->index, mirror, ingress);
893 		if (err) {
894 			kfree(mall_tc_entry);
895 			return err;
896 		}
897 
898 		list_add_tail(&mall_tc_entry->list, &p->mall_tc_list);
899 	}
900 
901 	return 0;
902 }
903 
904 static void dsa_slave_del_cls_matchall(struct net_device *dev,
905 				       struct tc_cls_matchall_offload *cls)
906 {
907 	struct dsa_port *dp = dsa_slave_to_port(dev);
908 	struct dsa_mall_tc_entry *mall_tc_entry;
909 	struct dsa_switch *ds = dp->ds;
910 
911 	if (!ds->ops->port_mirror_del)
912 		return;
913 
914 	mall_tc_entry = dsa_slave_mall_tc_entry_find(dev, cls->cookie);
915 	if (!mall_tc_entry)
916 		return;
917 
918 	list_del(&mall_tc_entry->list);
919 
920 	switch (mall_tc_entry->type) {
921 	case DSA_PORT_MALL_MIRROR:
922 		ds->ops->port_mirror_del(ds, dp->index, &mall_tc_entry->mirror);
923 		break;
924 	default:
925 		WARN_ON(1);
926 	}
927 
928 	kfree(mall_tc_entry);
929 }
930 
931 static int dsa_slave_setup_tc_cls_matchall(struct net_device *dev,
932 					   struct tc_cls_matchall_offload *cls,
933 					   bool ingress)
934 {
935 	if (cls->common.chain_index)
936 		return -EOPNOTSUPP;
937 
938 	switch (cls->command) {
939 	case TC_CLSMATCHALL_REPLACE:
940 		return dsa_slave_add_cls_matchall(dev, cls, ingress);
941 	case TC_CLSMATCHALL_DESTROY:
942 		dsa_slave_del_cls_matchall(dev, cls);
943 		return 0;
944 	default:
945 		return -EOPNOTSUPP;
946 	}
947 }
948 
949 static int dsa_slave_setup_tc_block_cb(enum tc_setup_type type, void *type_data,
950 				       void *cb_priv, bool ingress)
951 {
952 	struct net_device *dev = cb_priv;
953 
954 	if (!tc_can_offload(dev))
955 		return -EOPNOTSUPP;
956 
957 	switch (type) {
958 	case TC_SETUP_CLSMATCHALL:
959 		return dsa_slave_setup_tc_cls_matchall(dev, type_data, ingress);
960 	default:
961 		return -EOPNOTSUPP;
962 	}
963 }
964 
965 static int dsa_slave_setup_tc_block_cb_ig(enum tc_setup_type type,
966 					  void *type_data, void *cb_priv)
967 {
968 	return dsa_slave_setup_tc_block_cb(type, type_data, cb_priv, true);
969 }
970 
971 static int dsa_slave_setup_tc_block_cb_eg(enum tc_setup_type type,
972 					  void *type_data, void *cb_priv)
973 {
974 	return dsa_slave_setup_tc_block_cb(type, type_data, cb_priv, false);
975 }
976 
977 static LIST_HEAD(dsa_slave_block_cb_list);
978 
979 static int dsa_slave_setup_tc_block(struct net_device *dev,
980 				    struct flow_block_offload *f)
981 {
982 	struct flow_block_cb *block_cb;
983 	flow_setup_cb_t *cb;
984 
985 	if (f->binder_type == FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS)
986 		cb = dsa_slave_setup_tc_block_cb_ig;
987 	else if (f->binder_type == FLOW_BLOCK_BINDER_TYPE_CLSACT_EGRESS)
988 		cb = dsa_slave_setup_tc_block_cb_eg;
989 	else
990 		return -EOPNOTSUPP;
991 
992 	f->driver_block_list = &dsa_slave_block_cb_list;
993 
994 	switch (f->command) {
995 	case FLOW_BLOCK_BIND:
996 		if (flow_block_cb_is_busy(cb, dev, &dsa_slave_block_cb_list))
997 			return -EBUSY;
998 
999 		block_cb = flow_block_cb_alloc(cb, dev, dev, NULL);
1000 		if (IS_ERR(block_cb))
1001 			return PTR_ERR(block_cb);
1002 
1003 		flow_block_cb_add(block_cb, f);
1004 		list_add_tail(&block_cb->driver_list, &dsa_slave_block_cb_list);
1005 		return 0;
1006 	case FLOW_BLOCK_UNBIND:
1007 		block_cb = flow_block_cb_lookup(f->block, cb, dev);
1008 		if (!block_cb)
1009 			return -ENOENT;
1010 
1011 		flow_block_cb_remove(block_cb, f);
1012 		list_del(&block_cb->driver_list);
1013 		return 0;
1014 	default:
1015 		return -EOPNOTSUPP;
1016 	}
1017 }
1018 
1019 static int dsa_slave_setup_tc(struct net_device *dev, enum tc_setup_type type,
1020 			      void *type_data)
1021 {
1022 	struct dsa_port *dp = dsa_slave_to_port(dev);
1023 	struct dsa_switch *ds = dp->ds;
1024 
1025 	if (type == TC_SETUP_BLOCK)
1026 		return dsa_slave_setup_tc_block(dev, type_data);
1027 
1028 	if (!ds->ops->port_setup_tc)
1029 		return -EOPNOTSUPP;
1030 
1031 	return ds->ops->port_setup_tc(ds, dp->index, type, type_data);
1032 }
1033 
1034 static void dsa_slave_get_stats64(struct net_device *dev,
1035 				  struct rtnl_link_stats64 *stats)
1036 {
1037 	struct dsa_slave_priv *p = netdev_priv(dev);
1038 	struct pcpu_sw_netstats *s;
1039 	unsigned int start;
1040 	int i;
1041 
1042 	netdev_stats_to_stats64(stats, &dev->stats);
1043 	for_each_possible_cpu(i) {
1044 		u64 tx_packets, tx_bytes, rx_packets, rx_bytes;
1045 
1046 		s = per_cpu_ptr(p->stats64, i);
1047 		do {
1048 			start = u64_stats_fetch_begin_irq(&s->syncp);
1049 			tx_packets = s->tx_packets;
1050 			tx_bytes = s->tx_bytes;
1051 			rx_packets = s->rx_packets;
1052 			rx_bytes = s->rx_bytes;
1053 		} while (u64_stats_fetch_retry_irq(&s->syncp, start));
1054 
1055 		stats->tx_packets += tx_packets;
1056 		stats->tx_bytes += tx_bytes;
1057 		stats->rx_packets += rx_packets;
1058 		stats->rx_bytes += rx_bytes;
1059 	}
1060 }
1061 
1062 static int dsa_slave_get_rxnfc(struct net_device *dev,
1063 			       struct ethtool_rxnfc *nfc, u32 *rule_locs)
1064 {
1065 	struct dsa_port *dp = dsa_slave_to_port(dev);
1066 	struct dsa_switch *ds = dp->ds;
1067 
1068 	if (!ds->ops->get_rxnfc)
1069 		return -EOPNOTSUPP;
1070 
1071 	return ds->ops->get_rxnfc(ds, dp->index, nfc, rule_locs);
1072 }
1073 
1074 static int dsa_slave_set_rxnfc(struct net_device *dev,
1075 			       struct ethtool_rxnfc *nfc)
1076 {
1077 	struct dsa_port *dp = dsa_slave_to_port(dev);
1078 	struct dsa_switch *ds = dp->ds;
1079 
1080 	if (!ds->ops->set_rxnfc)
1081 		return -EOPNOTSUPP;
1082 
1083 	return ds->ops->set_rxnfc(ds, dp->index, nfc);
1084 }
1085 
1086 static int dsa_slave_get_ts_info(struct net_device *dev,
1087 				 struct ethtool_ts_info *ts)
1088 {
1089 	struct dsa_slave_priv *p = netdev_priv(dev);
1090 	struct dsa_switch *ds = p->dp->ds;
1091 
1092 	if (!ds->ops->get_ts_info)
1093 		return -EOPNOTSUPP;
1094 
1095 	return ds->ops->get_ts_info(ds, p->dp->index, ts);
1096 }
1097 
1098 static int dsa_slave_vlan_rx_add_vid(struct net_device *dev, __be16 proto,
1099 				     u16 vid)
1100 {
1101 	struct dsa_port *dp = dsa_slave_to_port(dev);
1102 	struct bridge_vlan_info info;
1103 	int ret;
1104 
1105 	/* Check for a possible bridge VLAN entry now since there is no
1106 	 * need to emulate the switchdev prepare + commit phase.
1107 	 */
1108 	if (dp->bridge_dev) {
1109 		if (!br_vlan_enabled(dp->bridge_dev))
1110 			return 0;
1111 
1112 		/* br_vlan_get_info() returns -EINVAL or -ENOENT if the
1113 		 * device, respectively the VID is not found, returning
1114 		 * 0 means success, which is a failure for us here.
1115 		 */
1116 		ret = br_vlan_get_info(dp->bridge_dev, vid, &info);
1117 		if (ret == 0)
1118 			return -EBUSY;
1119 	}
1120 
1121 	ret = dsa_port_vid_add(dp, vid, 0);
1122 	if (ret)
1123 		return ret;
1124 
1125 	ret = dsa_port_vid_add(dp->cpu_dp, vid, 0);
1126 	if (ret)
1127 		return ret;
1128 
1129 	return 0;
1130 }
1131 
1132 static int dsa_slave_vlan_rx_kill_vid(struct net_device *dev, __be16 proto,
1133 				      u16 vid)
1134 {
1135 	struct dsa_port *dp = dsa_slave_to_port(dev);
1136 	struct bridge_vlan_info info;
1137 	int ret;
1138 
1139 	/* Check for a possible bridge VLAN entry now since there is no
1140 	 * need to emulate the switchdev prepare + commit phase.
1141 	 */
1142 	if (dp->bridge_dev) {
1143 		if (!br_vlan_enabled(dp->bridge_dev))
1144 			return 0;
1145 
1146 		/* br_vlan_get_info() returns -EINVAL or -ENOENT if the
1147 		 * device, respectively the VID is not found, returning
1148 		 * 0 means success, which is a failure for us here.
1149 		 */
1150 		ret = br_vlan_get_info(dp->bridge_dev, vid, &info);
1151 		if (ret == 0)
1152 			return -EBUSY;
1153 	}
1154 
1155 	/* Do not deprogram the CPU port as it may be shared with other user
1156 	 * ports which can be members of this VLAN as well.
1157 	 */
1158 	return dsa_port_vid_del(dp, vid);
1159 }
1160 
1161 static const struct ethtool_ops dsa_slave_ethtool_ops = {
1162 	.get_drvinfo		= dsa_slave_get_drvinfo,
1163 	.get_regs_len		= dsa_slave_get_regs_len,
1164 	.get_regs		= dsa_slave_get_regs,
1165 	.nway_reset		= dsa_slave_nway_reset,
1166 	.get_link		= ethtool_op_get_link,
1167 	.get_eeprom_len		= dsa_slave_get_eeprom_len,
1168 	.get_eeprom		= dsa_slave_get_eeprom,
1169 	.set_eeprom		= dsa_slave_set_eeprom,
1170 	.get_strings		= dsa_slave_get_strings,
1171 	.get_ethtool_stats	= dsa_slave_get_ethtool_stats,
1172 	.get_sset_count		= dsa_slave_get_sset_count,
1173 	.set_wol		= dsa_slave_set_wol,
1174 	.get_wol		= dsa_slave_get_wol,
1175 	.set_eee		= dsa_slave_set_eee,
1176 	.get_eee		= dsa_slave_get_eee,
1177 	.get_link_ksettings	= dsa_slave_get_link_ksettings,
1178 	.set_link_ksettings	= dsa_slave_set_link_ksettings,
1179 	.get_pauseparam		= dsa_slave_get_pauseparam,
1180 	.set_pauseparam		= dsa_slave_set_pauseparam,
1181 	.get_rxnfc		= dsa_slave_get_rxnfc,
1182 	.set_rxnfc		= dsa_slave_set_rxnfc,
1183 	.get_ts_info		= dsa_slave_get_ts_info,
1184 };
1185 
1186 /* legacy way, bypassing the bridge *****************************************/
1187 int dsa_legacy_fdb_add(struct ndmsg *ndm, struct nlattr *tb[],
1188 		       struct net_device *dev,
1189 		       const unsigned char *addr, u16 vid,
1190 		       u16 flags,
1191 		       struct netlink_ext_ack *extack)
1192 {
1193 	struct dsa_port *dp = dsa_slave_to_port(dev);
1194 
1195 	return dsa_port_fdb_add(dp, addr, vid);
1196 }
1197 
1198 int dsa_legacy_fdb_del(struct ndmsg *ndm, struct nlattr *tb[],
1199 		       struct net_device *dev,
1200 		       const unsigned char *addr, u16 vid)
1201 {
1202 	struct dsa_port *dp = dsa_slave_to_port(dev);
1203 
1204 	return dsa_port_fdb_del(dp, addr, vid);
1205 }
1206 
1207 static struct devlink_port *dsa_slave_get_devlink_port(struct net_device *dev)
1208 {
1209 	struct dsa_port *dp = dsa_slave_to_port(dev);
1210 
1211 	return dp->ds->devlink ? &dp->devlink_port : NULL;
1212 }
1213 
1214 static const struct net_device_ops dsa_slave_netdev_ops = {
1215 	.ndo_open	 	= dsa_slave_open,
1216 	.ndo_stop		= dsa_slave_close,
1217 	.ndo_start_xmit		= dsa_slave_xmit,
1218 	.ndo_change_rx_flags	= dsa_slave_change_rx_flags,
1219 	.ndo_set_rx_mode	= dsa_slave_set_rx_mode,
1220 	.ndo_set_mac_address	= dsa_slave_set_mac_address,
1221 	.ndo_fdb_add		= dsa_legacy_fdb_add,
1222 	.ndo_fdb_del		= dsa_legacy_fdb_del,
1223 	.ndo_fdb_dump		= dsa_slave_fdb_dump,
1224 	.ndo_do_ioctl		= dsa_slave_ioctl,
1225 	.ndo_get_iflink		= dsa_slave_get_iflink,
1226 #ifdef CONFIG_NET_POLL_CONTROLLER
1227 	.ndo_netpoll_setup	= dsa_slave_netpoll_setup,
1228 	.ndo_netpoll_cleanup	= dsa_slave_netpoll_cleanup,
1229 	.ndo_poll_controller	= dsa_slave_poll_controller,
1230 #endif
1231 	.ndo_get_phys_port_name	= dsa_slave_get_phys_port_name,
1232 	.ndo_setup_tc		= dsa_slave_setup_tc,
1233 	.ndo_get_stats64	= dsa_slave_get_stats64,
1234 	.ndo_get_port_parent_id	= dsa_slave_get_port_parent_id,
1235 	.ndo_vlan_rx_add_vid	= dsa_slave_vlan_rx_add_vid,
1236 	.ndo_vlan_rx_kill_vid	= dsa_slave_vlan_rx_kill_vid,
1237 	.ndo_get_devlink_port	= dsa_slave_get_devlink_port,
1238 };
1239 
1240 static struct device_type dsa_type = {
1241 	.name	= "dsa",
1242 };
1243 
1244 void dsa_port_phylink_mac_change(struct dsa_switch *ds, int port, bool up)
1245 {
1246 	const struct dsa_port *dp = dsa_to_port(ds, port);
1247 
1248 	phylink_mac_change(dp->pl, up);
1249 }
1250 EXPORT_SYMBOL_GPL(dsa_port_phylink_mac_change);
1251 
1252 static void dsa_slave_phylink_fixed_state(struct net_device *dev,
1253 					  struct phylink_link_state *state)
1254 {
1255 	struct dsa_port *dp = dsa_slave_to_port(dev);
1256 	struct dsa_switch *ds = dp->ds;
1257 
1258 	/* No need to check that this operation is valid, the callback would
1259 	 * not be called if it was not.
1260 	 */
1261 	ds->ops->phylink_fixed_state(ds, dp->index, state);
1262 }
1263 
1264 /* slave device setup *******************************************************/
1265 static int dsa_slave_phy_connect(struct net_device *slave_dev, int addr)
1266 {
1267 	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1268 	struct dsa_switch *ds = dp->ds;
1269 
1270 	slave_dev->phydev = mdiobus_get_phy(ds->slave_mii_bus, addr);
1271 	if (!slave_dev->phydev) {
1272 		netdev_err(slave_dev, "no phy at %d\n", addr);
1273 		return -ENODEV;
1274 	}
1275 
1276 	return phylink_connect_phy(dp->pl, slave_dev->phydev);
1277 }
1278 
1279 static int dsa_slave_phy_setup(struct net_device *slave_dev)
1280 {
1281 	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1282 	struct device_node *port_dn = dp->dn;
1283 	struct dsa_switch *ds = dp->ds;
1284 	phy_interface_t mode;
1285 	u32 phy_flags = 0;
1286 	int ret;
1287 
1288 	ret = of_get_phy_mode(port_dn, &mode);
1289 	if (ret)
1290 		mode = PHY_INTERFACE_MODE_NA;
1291 
1292 	dp->pl_config.dev = &slave_dev->dev;
1293 	dp->pl_config.type = PHYLINK_NETDEV;
1294 
1295 	dp->pl = phylink_create(&dp->pl_config, of_fwnode_handle(port_dn), mode,
1296 				&dsa_port_phylink_mac_ops);
1297 	if (IS_ERR(dp->pl)) {
1298 		netdev_err(slave_dev,
1299 			   "error creating PHYLINK: %ld\n", PTR_ERR(dp->pl));
1300 		return PTR_ERR(dp->pl);
1301 	}
1302 
1303 	/* Register only if the switch provides such a callback, since this
1304 	 * callback takes precedence over polling the link GPIO in PHYLINK
1305 	 * (see phylink_get_fixed_state).
1306 	 */
1307 	if (ds->ops->phylink_fixed_state)
1308 		phylink_fixed_state_cb(dp->pl, dsa_slave_phylink_fixed_state);
1309 
1310 	if (ds->ops->get_phy_flags)
1311 		phy_flags = ds->ops->get_phy_flags(ds, dp->index);
1312 
1313 	ret = phylink_of_phy_connect(dp->pl, port_dn, phy_flags);
1314 	if (ret == -ENODEV && ds->slave_mii_bus) {
1315 		/* We could not connect to a designated PHY or SFP, so try to
1316 		 * use the switch internal MDIO bus instead
1317 		 */
1318 		ret = dsa_slave_phy_connect(slave_dev, dp->index);
1319 		if (ret) {
1320 			netdev_err(slave_dev,
1321 				   "failed to connect to port %d: %d\n",
1322 				   dp->index, ret);
1323 			phylink_destroy(dp->pl);
1324 			return ret;
1325 		}
1326 	}
1327 
1328 	return ret;
1329 }
1330 
1331 int dsa_slave_suspend(struct net_device *slave_dev)
1332 {
1333 	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1334 
1335 	if (!netif_running(slave_dev))
1336 		return 0;
1337 
1338 	netif_device_detach(slave_dev);
1339 
1340 	rtnl_lock();
1341 	phylink_stop(dp->pl);
1342 	rtnl_unlock();
1343 
1344 	return 0;
1345 }
1346 
1347 int dsa_slave_resume(struct net_device *slave_dev)
1348 {
1349 	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1350 
1351 	if (!netif_running(slave_dev))
1352 		return 0;
1353 
1354 	netif_device_attach(slave_dev);
1355 
1356 	rtnl_lock();
1357 	phylink_start(dp->pl);
1358 	rtnl_unlock();
1359 
1360 	return 0;
1361 }
1362 
1363 static void dsa_slave_notify(struct net_device *dev, unsigned long val)
1364 {
1365 	struct net_device *master = dsa_slave_to_master(dev);
1366 	struct dsa_port *dp = dsa_slave_to_port(dev);
1367 	struct dsa_notifier_register_info rinfo = {
1368 		.switch_number = dp->ds->index,
1369 		.port_number = dp->index,
1370 		.master = master,
1371 		.info.dev = dev,
1372 	};
1373 
1374 	call_dsa_notifiers(val, dev, &rinfo.info);
1375 }
1376 
1377 int dsa_slave_create(struct dsa_port *port)
1378 {
1379 	const struct dsa_port *cpu_dp = port->cpu_dp;
1380 	struct net_device *master = cpu_dp->master;
1381 	struct dsa_switch *ds = port->ds;
1382 	const char *name = port->name;
1383 	struct net_device *slave_dev;
1384 	struct dsa_slave_priv *p;
1385 	int ret;
1386 
1387 	if (!ds->num_tx_queues)
1388 		ds->num_tx_queues = 1;
1389 
1390 	slave_dev = alloc_netdev_mqs(sizeof(struct dsa_slave_priv), name,
1391 				     NET_NAME_UNKNOWN, ether_setup,
1392 				     ds->num_tx_queues, 1);
1393 	if (slave_dev == NULL)
1394 		return -ENOMEM;
1395 
1396 	slave_dev->features = master->vlan_features | NETIF_F_HW_TC;
1397 	if (ds->ops->port_vlan_add && ds->ops->port_vlan_del)
1398 		slave_dev->features |= NETIF_F_HW_VLAN_CTAG_FILTER;
1399 	slave_dev->hw_features |= NETIF_F_HW_TC;
1400 	slave_dev->ethtool_ops = &dsa_slave_ethtool_ops;
1401 	if (!IS_ERR_OR_NULL(port->mac))
1402 		ether_addr_copy(slave_dev->dev_addr, port->mac);
1403 	else
1404 		eth_hw_addr_inherit(slave_dev, master);
1405 	slave_dev->priv_flags |= IFF_NO_QUEUE;
1406 	slave_dev->netdev_ops = &dsa_slave_netdev_ops;
1407 	slave_dev->min_mtu = 0;
1408 	slave_dev->max_mtu = ETH_MAX_MTU;
1409 	SET_NETDEV_DEVTYPE(slave_dev, &dsa_type);
1410 
1411 	SET_NETDEV_DEV(slave_dev, port->ds->dev);
1412 	slave_dev->dev.of_node = port->dn;
1413 	slave_dev->vlan_features = master->vlan_features;
1414 
1415 	p = netdev_priv(slave_dev);
1416 	p->stats64 = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
1417 	if (!p->stats64) {
1418 		free_netdev(slave_dev);
1419 		return -ENOMEM;
1420 	}
1421 	p->dp = port;
1422 	INIT_LIST_HEAD(&p->mall_tc_list);
1423 	p->xmit = cpu_dp->tag_ops->xmit;
1424 	port->slave = slave_dev;
1425 
1426 	netif_carrier_off(slave_dev);
1427 
1428 	ret = dsa_slave_phy_setup(slave_dev);
1429 	if (ret) {
1430 		netdev_err(master, "error %d setting up slave phy\n", ret);
1431 		goto out_free;
1432 	}
1433 
1434 	dsa_slave_notify(slave_dev, DSA_PORT_REGISTER);
1435 
1436 	ret = register_netdev(slave_dev);
1437 	if (ret) {
1438 		netdev_err(master, "error %d registering interface %s\n",
1439 			   ret, slave_dev->name);
1440 		goto out_phy;
1441 	}
1442 
1443 	return 0;
1444 
1445 out_phy:
1446 	rtnl_lock();
1447 	phylink_disconnect_phy(p->dp->pl);
1448 	rtnl_unlock();
1449 	phylink_destroy(p->dp->pl);
1450 out_free:
1451 	free_percpu(p->stats64);
1452 	free_netdev(slave_dev);
1453 	port->slave = NULL;
1454 	return ret;
1455 }
1456 
1457 void dsa_slave_destroy(struct net_device *slave_dev)
1458 {
1459 	struct dsa_port *dp = dsa_slave_to_port(slave_dev);
1460 	struct dsa_slave_priv *p = netdev_priv(slave_dev);
1461 
1462 	netif_carrier_off(slave_dev);
1463 	rtnl_lock();
1464 	phylink_disconnect_phy(dp->pl);
1465 	rtnl_unlock();
1466 
1467 	dsa_slave_notify(slave_dev, DSA_PORT_UNREGISTER);
1468 	unregister_netdev(slave_dev);
1469 	phylink_destroy(dp->pl);
1470 	free_percpu(p->stats64);
1471 	free_netdev(slave_dev);
1472 }
1473 
1474 bool dsa_slave_dev_check(const struct net_device *dev)
1475 {
1476 	return dev->netdev_ops == &dsa_slave_netdev_ops;
1477 }
1478 
1479 static int dsa_slave_changeupper(struct net_device *dev,
1480 				 struct netdev_notifier_changeupper_info *info)
1481 {
1482 	struct dsa_port *dp = dsa_slave_to_port(dev);
1483 	int err = NOTIFY_DONE;
1484 
1485 	if (netif_is_bridge_master(info->upper_dev)) {
1486 		if (info->linking) {
1487 			err = dsa_port_bridge_join(dp, info->upper_dev);
1488 			err = notifier_from_errno(err);
1489 		} else {
1490 			dsa_port_bridge_leave(dp, info->upper_dev);
1491 			err = NOTIFY_OK;
1492 		}
1493 	}
1494 
1495 	return err;
1496 }
1497 
1498 static int dsa_slave_upper_vlan_check(struct net_device *dev,
1499 				      struct netdev_notifier_changeupper_info *
1500 				      info)
1501 {
1502 	struct netlink_ext_ack *ext_ack;
1503 	struct net_device *slave;
1504 	struct dsa_port *dp;
1505 
1506 	ext_ack = netdev_notifier_info_to_extack(&info->info);
1507 
1508 	if (!is_vlan_dev(dev))
1509 		return NOTIFY_DONE;
1510 
1511 	slave = vlan_dev_real_dev(dev);
1512 	if (!dsa_slave_dev_check(slave))
1513 		return NOTIFY_DONE;
1514 
1515 	dp = dsa_slave_to_port(slave);
1516 	if (!dp->bridge_dev)
1517 		return NOTIFY_DONE;
1518 
1519 	/* Deny enslaving a VLAN device into a VLAN-aware bridge */
1520 	if (br_vlan_enabled(dp->bridge_dev) &&
1521 	    netif_is_bridge_master(info->upper_dev) && info->linking) {
1522 		NL_SET_ERR_MSG_MOD(ext_ack,
1523 				   "Cannot enslave VLAN device into VLAN aware bridge");
1524 		return notifier_from_errno(-EINVAL);
1525 	}
1526 
1527 	return NOTIFY_DONE;
1528 }
1529 
1530 static int dsa_slave_netdevice_event(struct notifier_block *nb,
1531 				     unsigned long event, void *ptr)
1532 {
1533 	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
1534 
1535 	if (event == NETDEV_CHANGEUPPER) {
1536 		if (!dsa_slave_dev_check(dev))
1537 			return dsa_slave_upper_vlan_check(dev, ptr);
1538 
1539 		return dsa_slave_changeupper(dev, ptr);
1540 	}
1541 
1542 	return NOTIFY_DONE;
1543 }
1544 
1545 struct dsa_switchdev_event_work {
1546 	struct work_struct work;
1547 	struct switchdev_notifier_fdb_info fdb_info;
1548 	struct net_device *dev;
1549 	unsigned long event;
1550 };
1551 
1552 static void dsa_slave_switchdev_event_work(struct work_struct *work)
1553 {
1554 	struct dsa_switchdev_event_work *switchdev_work =
1555 		container_of(work, struct dsa_switchdev_event_work, work);
1556 	struct net_device *dev = switchdev_work->dev;
1557 	struct switchdev_notifier_fdb_info *fdb_info;
1558 	struct dsa_port *dp = dsa_slave_to_port(dev);
1559 	int err;
1560 
1561 	rtnl_lock();
1562 	switch (switchdev_work->event) {
1563 	case SWITCHDEV_FDB_ADD_TO_DEVICE:
1564 		fdb_info = &switchdev_work->fdb_info;
1565 		if (!fdb_info->added_by_user)
1566 			break;
1567 
1568 		err = dsa_port_fdb_add(dp, fdb_info->addr, fdb_info->vid);
1569 		if (err) {
1570 			netdev_dbg(dev, "fdb add failed err=%d\n", err);
1571 			break;
1572 		}
1573 		fdb_info->offloaded = true;
1574 		call_switchdev_notifiers(SWITCHDEV_FDB_OFFLOADED, dev,
1575 					 &fdb_info->info, NULL);
1576 		break;
1577 
1578 	case SWITCHDEV_FDB_DEL_TO_DEVICE:
1579 		fdb_info = &switchdev_work->fdb_info;
1580 		if (!fdb_info->added_by_user)
1581 			break;
1582 
1583 		err = dsa_port_fdb_del(dp, fdb_info->addr, fdb_info->vid);
1584 		if (err) {
1585 			netdev_dbg(dev, "fdb del failed err=%d\n", err);
1586 			dev_close(dev);
1587 		}
1588 		break;
1589 	}
1590 	rtnl_unlock();
1591 
1592 	kfree(switchdev_work->fdb_info.addr);
1593 	kfree(switchdev_work);
1594 	dev_put(dev);
1595 }
1596 
1597 static int
1598 dsa_slave_switchdev_fdb_work_init(struct dsa_switchdev_event_work *
1599 				  switchdev_work,
1600 				  const struct switchdev_notifier_fdb_info *
1601 				  fdb_info)
1602 {
1603 	memcpy(&switchdev_work->fdb_info, fdb_info,
1604 	       sizeof(switchdev_work->fdb_info));
1605 	switchdev_work->fdb_info.addr = kzalloc(ETH_ALEN, GFP_ATOMIC);
1606 	if (!switchdev_work->fdb_info.addr)
1607 		return -ENOMEM;
1608 	ether_addr_copy((u8 *)switchdev_work->fdb_info.addr,
1609 			fdb_info->addr);
1610 	return 0;
1611 }
1612 
1613 /* Called under rcu_read_lock() */
1614 static int dsa_slave_switchdev_event(struct notifier_block *unused,
1615 				     unsigned long event, void *ptr)
1616 {
1617 	struct net_device *dev = switchdev_notifier_info_to_dev(ptr);
1618 	struct dsa_switchdev_event_work *switchdev_work;
1619 	int err;
1620 
1621 	if (event == SWITCHDEV_PORT_ATTR_SET) {
1622 		err = switchdev_handle_port_attr_set(dev, ptr,
1623 						     dsa_slave_dev_check,
1624 						     dsa_slave_port_attr_set);
1625 		return notifier_from_errno(err);
1626 	}
1627 
1628 	if (!dsa_slave_dev_check(dev))
1629 		return NOTIFY_DONE;
1630 
1631 	switchdev_work = kzalloc(sizeof(*switchdev_work), GFP_ATOMIC);
1632 	if (!switchdev_work)
1633 		return NOTIFY_BAD;
1634 
1635 	INIT_WORK(&switchdev_work->work,
1636 		  dsa_slave_switchdev_event_work);
1637 	switchdev_work->dev = dev;
1638 	switchdev_work->event = event;
1639 
1640 	switch (event) {
1641 	case SWITCHDEV_FDB_ADD_TO_DEVICE: /* fall through */
1642 	case SWITCHDEV_FDB_DEL_TO_DEVICE:
1643 		if (dsa_slave_switchdev_fdb_work_init(switchdev_work, ptr))
1644 			goto err_fdb_work_init;
1645 		dev_hold(dev);
1646 		break;
1647 	default:
1648 		kfree(switchdev_work);
1649 		return NOTIFY_DONE;
1650 	}
1651 
1652 	dsa_schedule_work(&switchdev_work->work);
1653 	return NOTIFY_OK;
1654 
1655 err_fdb_work_init:
1656 	kfree(switchdev_work);
1657 	return NOTIFY_BAD;
1658 }
1659 
1660 static int dsa_slave_switchdev_blocking_event(struct notifier_block *unused,
1661 					      unsigned long event, void *ptr)
1662 {
1663 	struct net_device *dev = switchdev_notifier_info_to_dev(ptr);
1664 	int err;
1665 
1666 	switch (event) {
1667 	case SWITCHDEV_PORT_OBJ_ADD:
1668 		err = switchdev_handle_port_obj_add(dev, ptr,
1669 						    dsa_slave_dev_check,
1670 						    dsa_slave_port_obj_add);
1671 		return notifier_from_errno(err);
1672 	case SWITCHDEV_PORT_OBJ_DEL:
1673 		err = switchdev_handle_port_obj_del(dev, ptr,
1674 						    dsa_slave_dev_check,
1675 						    dsa_slave_port_obj_del);
1676 		return notifier_from_errno(err);
1677 	case SWITCHDEV_PORT_ATTR_SET:
1678 		err = switchdev_handle_port_attr_set(dev, ptr,
1679 						     dsa_slave_dev_check,
1680 						     dsa_slave_port_attr_set);
1681 		return notifier_from_errno(err);
1682 	}
1683 
1684 	return NOTIFY_DONE;
1685 }
1686 
1687 static struct notifier_block dsa_slave_nb __read_mostly = {
1688 	.notifier_call  = dsa_slave_netdevice_event,
1689 };
1690 
1691 static struct notifier_block dsa_slave_switchdev_notifier = {
1692 	.notifier_call = dsa_slave_switchdev_event,
1693 };
1694 
1695 static struct notifier_block dsa_slave_switchdev_blocking_notifier = {
1696 	.notifier_call = dsa_slave_switchdev_blocking_event,
1697 };
1698 
1699 int dsa_slave_register_notifier(void)
1700 {
1701 	struct notifier_block *nb;
1702 	int err;
1703 
1704 	err = register_netdevice_notifier(&dsa_slave_nb);
1705 	if (err)
1706 		return err;
1707 
1708 	err = register_switchdev_notifier(&dsa_slave_switchdev_notifier);
1709 	if (err)
1710 		goto err_switchdev_nb;
1711 
1712 	nb = &dsa_slave_switchdev_blocking_notifier;
1713 	err = register_switchdev_blocking_notifier(nb);
1714 	if (err)
1715 		goto err_switchdev_blocking_nb;
1716 
1717 	return 0;
1718 
1719 err_switchdev_blocking_nb:
1720 	unregister_switchdev_notifier(&dsa_slave_switchdev_notifier);
1721 err_switchdev_nb:
1722 	unregister_netdevice_notifier(&dsa_slave_nb);
1723 	return err;
1724 }
1725 
1726 void dsa_slave_unregister_notifier(void)
1727 {
1728 	struct notifier_block *nb;
1729 	int err;
1730 
1731 	nb = &dsa_slave_switchdev_blocking_notifier;
1732 	err = unregister_switchdev_blocking_notifier(nb);
1733 	if (err)
1734 		pr_err("DSA: failed to unregister switchdev blocking notifier (%d)\n", err);
1735 
1736 	err = unregister_switchdev_notifier(&dsa_slave_switchdev_notifier);
1737 	if (err)
1738 		pr_err("DSA: failed to unregister switchdev notifier (%d)\n", err);
1739 
1740 	err = unregister_netdevice_notifier(&dsa_slave_nb);
1741 	if (err)
1742 		pr_err("DSA: failed to unregister slave notifier (%d)\n", err);
1743 }
1744