1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Handling of a master device, switching frames via its switch fabric CPU port
4 *
5 * Copyright (c) 2017 Savoir-faire Linux Inc.
6 * Vivien Didelot <vivien.didelot@savoirfairelinux.com>
7 */
8
9 #include <linux/ethtool.h>
10 #include <linux/netdevice.h>
11 #include <linux/netlink.h>
12 #include <net/dsa.h>
13
14 #include "dsa.h"
15 #include "master.h"
16 #include "port.h"
17 #include "tag.h"
18
dsa_master_get_regs_len(struct net_device * dev)19 static int dsa_master_get_regs_len(struct net_device *dev)
20 {
21 struct dsa_port *cpu_dp = dev->dsa_ptr;
22 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops;
23 struct dsa_switch *ds = cpu_dp->ds;
24 int port = cpu_dp->index;
25 int ret = 0;
26 int len;
27
28 if (ops->get_regs_len) {
29 len = ops->get_regs_len(dev);
30 if (len < 0)
31 return len;
32 ret += len;
33 }
34
35 ret += sizeof(struct ethtool_drvinfo);
36 ret += sizeof(struct ethtool_regs);
37
38 if (ds->ops->get_regs_len) {
39 len = ds->ops->get_regs_len(ds, port);
40 if (len < 0)
41 return len;
42 ret += len;
43 }
44
45 return ret;
46 }
47
dsa_master_get_regs(struct net_device * dev,struct ethtool_regs * regs,void * data)48 static void dsa_master_get_regs(struct net_device *dev,
49 struct ethtool_regs *regs, void *data)
50 {
51 struct dsa_port *cpu_dp = dev->dsa_ptr;
52 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops;
53 struct dsa_switch *ds = cpu_dp->ds;
54 struct ethtool_drvinfo *cpu_info;
55 struct ethtool_regs *cpu_regs;
56 int port = cpu_dp->index;
57 int len;
58
59 if (ops->get_regs_len && ops->get_regs) {
60 len = ops->get_regs_len(dev);
61 if (len < 0)
62 return;
63 regs->len = len;
64 ops->get_regs(dev, regs, data);
65 data += regs->len;
66 }
67
68 cpu_info = (struct ethtool_drvinfo *)data;
69 strscpy(cpu_info->driver, "dsa", sizeof(cpu_info->driver));
70 data += sizeof(*cpu_info);
71 cpu_regs = (struct ethtool_regs *)data;
72 data += sizeof(*cpu_regs);
73
74 if (ds->ops->get_regs_len && ds->ops->get_regs) {
75 len = ds->ops->get_regs_len(ds, port);
76 if (len < 0)
77 return;
78 cpu_regs->len = len;
79 ds->ops->get_regs(ds, port, cpu_regs, data);
80 }
81 }
82
dsa_master_get_ethtool_stats(struct net_device * dev,struct ethtool_stats * stats,uint64_t * data)83 static void dsa_master_get_ethtool_stats(struct net_device *dev,
84 struct ethtool_stats *stats,
85 uint64_t *data)
86 {
87 struct dsa_port *cpu_dp = dev->dsa_ptr;
88 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops;
89 struct dsa_switch *ds = cpu_dp->ds;
90 int port = cpu_dp->index;
91 int count = 0;
92
93 if (ops->get_sset_count && ops->get_ethtool_stats) {
94 count = ops->get_sset_count(dev, ETH_SS_STATS);
95 ops->get_ethtool_stats(dev, stats, data);
96 }
97
98 if (ds->ops->get_ethtool_stats)
99 ds->ops->get_ethtool_stats(ds, port, data + count);
100 }
101
dsa_master_get_ethtool_phy_stats(struct net_device * dev,struct ethtool_stats * stats,uint64_t * data)102 static void dsa_master_get_ethtool_phy_stats(struct net_device *dev,
103 struct ethtool_stats *stats,
104 uint64_t *data)
105 {
106 struct dsa_port *cpu_dp = dev->dsa_ptr;
107 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops;
108 struct dsa_switch *ds = cpu_dp->ds;
109 int port = cpu_dp->index;
110 int count = 0;
111
112 if (dev->phydev && !ops->get_ethtool_phy_stats) {
113 count = phy_ethtool_get_sset_count(dev->phydev);
114 if (count >= 0)
115 phy_ethtool_get_stats(dev->phydev, stats, data);
116 } else if (ops->get_sset_count && ops->get_ethtool_phy_stats) {
117 count = ops->get_sset_count(dev, ETH_SS_PHY_STATS);
118 ops->get_ethtool_phy_stats(dev, stats, data);
119 }
120
121 if (count < 0)
122 count = 0;
123
124 if (ds->ops->get_ethtool_phy_stats)
125 ds->ops->get_ethtool_phy_stats(ds, port, data + count);
126 }
127
dsa_master_get_sset_count(struct net_device * dev,int sset)128 static int dsa_master_get_sset_count(struct net_device *dev, int sset)
129 {
130 struct dsa_port *cpu_dp = dev->dsa_ptr;
131 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops;
132 struct dsa_switch *ds = cpu_dp->ds;
133 int count = 0;
134
135 if (sset == ETH_SS_PHY_STATS && dev->phydev &&
136 !ops->get_ethtool_phy_stats)
137 count = phy_ethtool_get_sset_count(dev->phydev);
138 else if (ops->get_sset_count)
139 count = ops->get_sset_count(dev, sset);
140
141 if (count < 0)
142 count = 0;
143
144 if (ds->ops->get_sset_count)
145 count += ds->ops->get_sset_count(ds, cpu_dp->index, sset);
146
147 return count;
148 }
149
dsa_master_get_strings(struct net_device * dev,uint32_t stringset,uint8_t * data)150 static void dsa_master_get_strings(struct net_device *dev, uint32_t stringset,
151 uint8_t *data)
152 {
153 struct dsa_port *cpu_dp = dev->dsa_ptr;
154 const struct ethtool_ops *ops = cpu_dp->orig_ethtool_ops;
155 struct dsa_switch *ds = cpu_dp->ds;
156 int port = cpu_dp->index;
157 int len = ETH_GSTRING_LEN;
158 int mcount = 0, count, i;
159 uint8_t pfx[4];
160 uint8_t *ndata;
161
162 snprintf(pfx, sizeof(pfx), "p%.2d", port);
163 /* We do not want to be NULL-terminated, since this is a prefix */
164 pfx[sizeof(pfx) - 1] = '_';
165
166 if (stringset == ETH_SS_PHY_STATS && dev->phydev &&
167 !ops->get_ethtool_phy_stats) {
168 mcount = phy_ethtool_get_sset_count(dev->phydev);
169 if (mcount < 0)
170 mcount = 0;
171 else
172 phy_ethtool_get_strings(dev->phydev, data);
173 } else if (ops->get_sset_count && ops->get_strings) {
174 mcount = ops->get_sset_count(dev, stringset);
175 if (mcount < 0)
176 mcount = 0;
177 ops->get_strings(dev, stringset, data);
178 }
179
180 if (ds->ops->get_strings) {
181 ndata = data + mcount * len;
182 /* This function copies ETH_GSTRINGS_LEN bytes, we will mangle
183 * the output after to prepend our CPU port prefix we
184 * constructed earlier
185 */
186 ds->ops->get_strings(ds, port, stringset, ndata);
187 count = ds->ops->get_sset_count(ds, port, stringset);
188 if (count < 0)
189 return;
190 for (i = 0; i < count; i++) {
191 memmove(ndata + (i * len + sizeof(pfx)),
192 ndata + i * len, len - sizeof(pfx));
193 memcpy(ndata + i * len, pfx, sizeof(pfx));
194 }
195 }
196 }
197
198 /* Deny PTP operations on master if there is at least one switch in the tree
199 * that is PTP capable.
200 */
__dsa_master_hwtstamp_validate(struct net_device * dev,const struct kernel_hwtstamp_config * config,struct netlink_ext_ack * extack)201 int __dsa_master_hwtstamp_validate(struct net_device *dev,
202 const struct kernel_hwtstamp_config *config,
203 struct netlink_ext_ack *extack)
204 {
205 struct dsa_port *cpu_dp = dev->dsa_ptr;
206 struct dsa_switch *ds = cpu_dp->ds;
207 struct dsa_switch_tree *dst;
208 struct dsa_port *dp;
209
210 dst = ds->dst;
211
212 list_for_each_entry(dp, &dst->ports, list) {
213 if (dsa_port_supports_hwtstamp(dp)) {
214 NL_SET_ERR_MSG(extack,
215 "HW timestamping not allowed on DSA master when switch supports the operation");
216 return -EBUSY;
217 }
218 }
219
220 return 0;
221 }
222
dsa_master_ethtool_setup(struct net_device * dev)223 static int dsa_master_ethtool_setup(struct net_device *dev)
224 {
225 struct dsa_port *cpu_dp = dev->dsa_ptr;
226 struct dsa_switch *ds = cpu_dp->ds;
227 struct ethtool_ops *ops;
228
229 if (netif_is_lag_master(dev))
230 return 0;
231
232 ops = devm_kzalloc(ds->dev, sizeof(*ops), GFP_KERNEL);
233 if (!ops)
234 return -ENOMEM;
235
236 cpu_dp->orig_ethtool_ops = dev->ethtool_ops;
237 if (cpu_dp->orig_ethtool_ops)
238 memcpy(ops, cpu_dp->orig_ethtool_ops, sizeof(*ops));
239
240 ops->get_regs_len = dsa_master_get_regs_len;
241 ops->get_regs = dsa_master_get_regs;
242 ops->get_sset_count = dsa_master_get_sset_count;
243 ops->get_ethtool_stats = dsa_master_get_ethtool_stats;
244 ops->get_strings = dsa_master_get_strings;
245 ops->get_ethtool_phy_stats = dsa_master_get_ethtool_phy_stats;
246
247 dev->ethtool_ops = ops;
248
249 return 0;
250 }
251
dsa_master_ethtool_teardown(struct net_device * dev)252 static void dsa_master_ethtool_teardown(struct net_device *dev)
253 {
254 struct dsa_port *cpu_dp = dev->dsa_ptr;
255
256 if (netif_is_lag_master(dev))
257 return;
258
259 dev->ethtool_ops = cpu_dp->orig_ethtool_ops;
260 cpu_dp->orig_ethtool_ops = NULL;
261 }
262
263 /* Keep the master always promiscuous if the tagging protocol requires that
264 * (garbles MAC DA) or if it doesn't support unicast filtering, case in which
265 * it would revert to promiscuous mode as soon as we call dev_uc_add() on it
266 * anyway.
267 */
dsa_master_set_promiscuity(struct net_device * dev,int inc)268 static void dsa_master_set_promiscuity(struct net_device *dev, int inc)
269 {
270 const struct dsa_device_ops *ops = dev->dsa_ptr->tag_ops;
271
272 if ((dev->priv_flags & IFF_UNICAST_FLT) && !ops->promisc_on_master)
273 return;
274
275 ASSERT_RTNL();
276
277 dev_set_promiscuity(dev, inc);
278 }
279
tagging_show(struct device * d,struct device_attribute * attr,char * buf)280 static ssize_t tagging_show(struct device *d, struct device_attribute *attr,
281 char *buf)
282 {
283 struct net_device *dev = to_net_dev(d);
284 struct dsa_port *cpu_dp = dev->dsa_ptr;
285
286 return sysfs_emit(buf, "%s\n",
287 dsa_tag_protocol_to_str(cpu_dp->tag_ops));
288 }
289
tagging_store(struct device * d,struct device_attribute * attr,const char * buf,size_t count)290 static ssize_t tagging_store(struct device *d, struct device_attribute *attr,
291 const char *buf, size_t count)
292 {
293 const struct dsa_device_ops *new_tag_ops, *old_tag_ops;
294 const char *end = strchrnul(buf, '\n'), *name;
295 struct net_device *dev = to_net_dev(d);
296 struct dsa_port *cpu_dp = dev->dsa_ptr;
297 size_t len = end - buf;
298 int err;
299
300 /* Empty string passed */
301 if (!len)
302 return -ENOPROTOOPT;
303
304 name = kstrndup(buf, len, GFP_KERNEL);
305 if (!name)
306 return -ENOMEM;
307
308 old_tag_ops = cpu_dp->tag_ops;
309 new_tag_ops = dsa_tag_driver_get_by_name(name);
310 kfree(name);
311 /* Bad tagger name? */
312 if (IS_ERR(new_tag_ops))
313 return PTR_ERR(new_tag_ops);
314
315 if (new_tag_ops == old_tag_ops)
316 /* Drop the temporarily held duplicate reference, since
317 * the DSA switch tree uses this tagger.
318 */
319 goto out;
320
321 err = dsa_tree_change_tag_proto(cpu_dp->ds->dst, new_tag_ops,
322 old_tag_ops);
323 if (err) {
324 /* On failure the old tagger is restored, so we don't need the
325 * driver for the new one.
326 */
327 dsa_tag_driver_put(new_tag_ops);
328 return err;
329 }
330
331 /* On success we no longer need the module for the old tagging protocol
332 */
333 out:
334 dsa_tag_driver_put(old_tag_ops);
335 return count;
336 }
337 static DEVICE_ATTR_RW(tagging);
338
339 static struct attribute *dsa_slave_attrs[] = {
340 &dev_attr_tagging.attr,
341 NULL
342 };
343
344 static const struct attribute_group dsa_group = {
345 .name = "dsa",
346 .attrs = dsa_slave_attrs,
347 };
348
dsa_master_reset_mtu(struct net_device * dev)349 static void dsa_master_reset_mtu(struct net_device *dev)
350 {
351 int err;
352
353 err = dev_set_mtu(dev, ETH_DATA_LEN);
354 if (err)
355 netdev_dbg(dev,
356 "Unable to reset MTU to exclude DSA overheads\n");
357 }
358
dsa_master_setup(struct net_device * dev,struct dsa_port * cpu_dp)359 int dsa_master_setup(struct net_device *dev, struct dsa_port *cpu_dp)
360 {
361 const struct dsa_device_ops *tag_ops = cpu_dp->tag_ops;
362 struct dsa_switch *ds = cpu_dp->ds;
363 struct device_link *consumer_link;
364 int mtu, ret;
365
366 mtu = ETH_DATA_LEN + dsa_tag_protocol_overhead(tag_ops);
367
368 /* The DSA master must use SET_NETDEV_DEV for this to work. */
369 if (!netif_is_lag_master(dev)) {
370 consumer_link = device_link_add(ds->dev, dev->dev.parent,
371 DL_FLAG_AUTOREMOVE_CONSUMER);
372 if (!consumer_link)
373 netdev_err(dev,
374 "Failed to create a device link to DSA switch %s\n",
375 dev_name(ds->dev));
376 }
377
378 /* The switch driver may not implement ->port_change_mtu(), case in
379 * which dsa_slave_change_mtu() will not update the master MTU either,
380 * so we need to do that here.
381 */
382 ret = dev_set_mtu(dev, mtu);
383 if (ret)
384 netdev_warn(dev, "error %d setting MTU to %d to include DSA overhead\n",
385 ret, mtu);
386
387 /* If we use a tagging format that doesn't have an ethertype
388 * field, make sure that all packets from this point on get
389 * sent to the tag format's receive function.
390 */
391 wmb();
392
393 dev->dsa_ptr = cpu_dp;
394
395 dsa_master_set_promiscuity(dev, 1);
396
397 ret = dsa_master_ethtool_setup(dev);
398 if (ret)
399 goto out_err_reset_promisc;
400
401 ret = sysfs_create_group(&dev->dev.kobj, &dsa_group);
402 if (ret)
403 goto out_err_ethtool_teardown;
404
405 return ret;
406
407 out_err_ethtool_teardown:
408 dsa_master_ethtool_teardown(dev);
409 out_err_reset_promisc:
410 dsa_master_set_promiscuity(dev, -1);
411 return ret;
412 }
413
dsa_master_teardown(struct net_device * dev)414 void dsa_master_teardown(struct net_device *dev)
415 {
416 sysfs_remove_group(&dev->dev.kobj, &dsa_group);
417 dsa_master_ethtool_teardown(dev);
418 dsa_master_reset_mtu(dev);
419 dsa_master_set_promiscuity(dev, -1);
420
421 dev->dsa_ptr = NULL;
422
423 /* If we used a tagging format that doesn't have an ethertype
424 * field, make sure that all packets from this point get sent
425 * without the tag and go through the regular receive path.
426 */
427 wmb();
428 }
429
dsa_master_lag_setup(struct net_device * lag_dev,struct dsa_port * cpu_dp,struct netdev_lag_upper_info * uinfo,struct netlink_ext_ack * extack)430 int dsa_master_lag_setup(struct net_device *lag_dev, struct dsa_port *cpu_dp,
431 struct netdev_lag_upper_info *uinfo,
432 struct netlink_ext_ack *extack)
433 {
434 bool master_setup = false;
435 int err;
436
437 if (!netdev_uses_dsa(lag_dev)) {
438 err = dsa_master_setup(lag_dev, cpu_dp);
439 if (err)
440 return err;
441
442 master_setup = true;
443 }
444
445 err = dsa_port_lag_join(cpu_dp, lag_dev, uinfo, extack);
446 if (err) {
447 NL_SET_ERR_MSG_WEAK_MOD(extack, "CPU port failed to join LAG");
448 goto out_master_teardown;
449 }
450
451 return 0;
452
453 out_master_teardown:
454 if (master_setup)
455 dsa_master_teardown(lag_dev);
456 return err;
457 }
458
459 /* Tear down a master if there isn't any other user port on it,
460 * optionally also destroying LAG information.
461 */
dsa_master_lag_teardown(struct net_device * lag_dev,struct dsa_port * cpu_dp)462 void dsa_master_lag_teardown(struct net_device *lag_dev,
463 struct dsa_port *cpu_dp)
464 {
465 struct net_device *upper;
466 struct list_head *iter;
467
468 dsa_port_lag_leave(cpu_dp, lag_dev);
469
470 netdev_for_each_upper_dev_rcu(lag_dev, upper, iter)
471 if (dsa_slave_dev_check(upper))
472 return;
473
474 dsa_master_teardown(lag_dev);
475 }
476