xref: /openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c (revision 26d0dfbb16fcb17d128a79dc70f3020ea6992af0)
1  /*
2   * Copyright (c) 2016, Mellanox Technologies. All rights reserved.
3   *
4   * This software is available to you under a choice of one of two
5   * licenses.  You may choose to be licensed under the terms of the GNU
6   * General Public License (GPL) Version 2, available from the file
7   * COPYING in the main directory of this source tree, or the
8   * OpenIB.org BSD license below:
9   *
10   *     Redistribution and use in source and binary forms, with or
11   *     without modification, are permitted provided that the following
12   *     conditions are met:
13   *
14   *      - Redistributions of source code must retain the above
15   *        copyright notice, this list of conditions and the following
16   *        disclaimer.
17   *
18   *      - Redistributions in binary form must reproduce the above
19   *        copyright notice, this list of conditions and the following
20   *        disclaimer in the documentation and/or other materials
21   *        provided with the distribution.
22   *
23   * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24   * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25   * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26   * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27   * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28   * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29   * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30   * SOFTWARE.
31   */
32  
33  #include <linux/mlx5/fs.h>
34  #include "en.h"
35  #include "en/params.h"
36  #include "en/xsk/pool.h"
37  #include "en/fs_ethtool.h"
38  
39  struct mlx5e_ethtool_table {
40  	struct mlx5_flow_table *ft;
41  	int                    num_rules;
42  };
43  
44  #define ETHTOOL_NUM_L3_L4_FTS 7
45  #define ETHTOOL_NUM_L2_FTS 4
46  
47  struct mlx5e_ethtool_steering {
48  	struct mlx5e_ethtool_table      l3_l4_ft[ETHTOOL_NUM_L3_L4_FTS];
49  	struct mlx5e_ethtool_table      l2_ft[ETHTOOL_NUM_L2_FTS];
50  	struct list_head                rules;
51  	int                             tot_num_rules;
52  };
53  
54  static int flow_type_to_traffic_type(u32 flow_type);
55  
flow_type_mask(u32 flow_type)56  static u32 flow_type_mask(u32 flow_type)
57  {
58  	return flow_type & ~(FLOW_EXT | FLOW_MAC_EXT | FLOW_RSS);
59  }
60  
61  struct mlx5e_ethtool_rule {
62  	struct list_head             list;
63  	struct ethtool_rx_flow_spec  flow_spec;
64  	struct mlx5_flow_handle	     *rule;
65  	struct mlx5e_ethtool_table   *eth_ft;
66  	struct mlx5e_rss             *rss;
67  };
68  
put_flow_table(struct mlx5e_ethtool_table * eth_ft)69  static void put_flow_table(struct mlx5e_ethtool_table *eth_ft)
70  {
71  	if (!--eth_ft->num_rules) {
72  		mlx5_destroy_flow_table(eth_ft->ft);
73  		eth_ft->ft = NULL;
74  	}
75  }
76  
77  #define MLX5E_ETHTOOL_L3_L4_PRIO 0
78  #define MLX5E_ETHTOOL_L2_PRIO (MLX5E_ETHTOOL_L3_L4_PRIO + ETHTOOL_NUM_L3_L4_FTS)
79  #define MLX5E_ETHTOOL_NUM_ENTRIES 64000
80  #define MLX5E_ETHTOOL_NUM_GROUPS  10
get_flow_table(struct mlx5e_priv * priv,struct ethtool_rx_flow_spec * fs,int num_tuples)81  static struct mlx5e_ethtool_table *get_flow_table(struct mlx5e_priv *priv,
82  						  struct ethtool_rx_flow_spec *fs,
83  						  int num_tuples)
84  {
85  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(priv->fs);
86  	struct mlx5_flow_table_attr ft_attr = {};
87  	struct mlx5e_ethtool_table *eth_ft;
88  	struct mlx5_flow_namespace *ns;
89  	struct mlx5_flow_table *ft;
90  	int max_tuples;
91  	int table_size;
92  	int prio;
93  
94  	switch (flow_type_mask(fs->flow_type)) {
95  	case TCP_V4_FLOW:
96  	case UDP_V4_FLOW:
97  	case TCP_V6_FLOW:
98  	case UDP_V6_FLOW:
99  	case IP_USER_FLOW:
100  	case IPV6_USER_FLOW:
101  		max_tuples = ETHTOOL_NUM_L3_L4_FTS;
102  		prio = MLX5E_ETHTOOL_L3_L4_PRIO + (max_tuples - num_tuples);
103  		eth_ft = &ethtool->l3_l4_ft[prio];
104  		break;
105  	case ETHER_FLOW:
106  		max_tuples = ETHTOOL_NUM_L2_FTS;
107  		prio = max_tuples - num_tuples;
108  		eth_ft = &ethtool->l2_ft[prio];
109  		prio += MLX5E_ETHTOOL_L2_PRIO;
110  		break;
111  	default:
112  		return ERR_PTR(-EINVAL);
113  	}
114  
115  	eth_ft->num_rules++;
116  	if (eth_ft->ft)
117  		return eth_ft;
118  
119  	ns = mlx5_get_flow_namespace(priv->mdev,
120  				     MLX5_FLOW_NAMESPACE_ETHTOOL);
121  	if (!ns)
122  		return ERR_PTR(-EOPNOTSUPP);
123  
124  	table_size = min_t(u32, BIT(MLX5_CAP_FLOWTABLE(priv->mdev,
125  						       flow_table_properties_nic_receive.log_max_ft_size)),
126  			   MLX5E_ETHTOOL_NUM_ENTRIES);
127  
128  	ft_attr.prio = prio;
129  	ft_attr.max_fte = table_size;
130  	ft_attr.autogroup.max_num_groups = MLX5E_ETHTOOL_NUM_GROUPS;
131  	ft = mlx5_create_auto_grouped_flow_table(ns, &ft_attr);
132  	if (IS_ERR(ft))
133  		return (void *)ft;
134  
135  	eth_ft->ft = ft;
136  	return eth_ft;
137  }
138  
mask_spec(u8 * mask,u8 * val,size_t size)139  static void mask_spec(u8 *mask, u8 *val, size_t size)
140  {
141  	unsigned int i;
142  
143  	for (i = 0; i < size; i++, mask++, val++)
144  		*((u8 *)val) = *((u8 *)mask) & *((u8 *)val);
145  }
146  
147  #define MLX5E_FTE_SET(header_p, fld, v)  \
148  	MLX5_SET(fte_match_set_lyr_2_4, header_p, fld, v)
149  
150  #define MLX5E_FTE_ADDR_OF(header_p, fld) \
151  	MLX5_ADDR_OF(fte_match_set_lyr_2_4, header_p, fld)
152  
153  static void
set_ip4(void * headers_c,void * headers_v,__be32 ip4src_m,__be32 ip4src_v,__be32 ip4dst_m,__be32 ip4dst_v)154  set_ip4(void *headers_c, void *headers_v, __be32 ip4src_m,
155  	__be32 ip4src_v, __be32 ip4dst_m, __be32 ip4dst_v)
156  {
157  	if (ip4src_m) {
158  		memcpy(MLX5E_FTE_ADDR_OF(headers_v, src_ipv4_src_ipv6.ipv4_layout.ipv4),
159  		       &ip4src_v, sizeof(ip4src_v));
160  		memcpy(MLX5E_FTE_ADDR_OF(headers_c, src_ipv4_src_ipv6.ipv4_layout.ipv4),
161  		       &ip4src_m, sizeof(ip4src_m));
162  	}
163  	if (ip4dst_m) {
164  		memcpy(MLX5E_FTE_ADDR_OF(headers_v, dst_ipv4_dst_ipv6.ipv4_layout.ipv4),
165  		       &ip4dst_v, sizeof(ip4dst_v));
166  		memcpy(MLX5E_FTE_ADDR_OF(headers_c, dst_ipv4_dst_ipv6.ipv4_layout.ipv4),
167  		       &ip4dst_m, sizeof(ip4dst_m));
168  	}
169  
170  	MLX5E_FTE_SET(headers_c, ethertype, 0xffff);
171  	MLX5E_FTE_SET(headers_v, ethertype, ETH_P_IP);
172  }
173  
174  static void
set_ip6(void * headers_c,void * headers_v,__be32 ip6src_m[4],__be32 ip6src_v[4],__be32 ip6dst_m[4],__be32 ip6dst_v[4])175  set_ip6(void *headers_c, void *headers_v, __be32 ip6src_m[4],
176  	__be32 ip6src_v[4], __be32 ip6dst_m[4], __be32 ip6dst_v[4])
177  {
178  	u8 ip6_sz = MLX5_FLD_SZ_BYTES(ipv6_layout, ipv6);
179  
180  	if (!ipv6_addr_any((struct in6_addr *)ip6src_m)) {
181  		memcpy(MLX5E_FTE_ADDR_OF(headers_v, src_ipv4_src_ipv6.ipv6_layout.ipv6),
182  		       ip6src_v, ip6_sz);
183  		memcpy(MLX5E_FTE_ADDR_OF(headers_c, src_ipv4_src_ipv6.ipv6_layout.ipv6),
184  		       ip6src_m, ip6_sz);
185  	}
186  	if (!ipv6_addr_any((struct in6_addr *)ip6dst_m)) {
187  		memcpy(MLX5E_FTE_ADDR_OF(headers_v, dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
188  		       ip6dst_v, ip6_sz);
189  		memcpy(MLX5E_FTE_ADDR_OF(headers_c, dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
190  		       ip6dst_m, ip6_sz);
191  	}
192  
193  	MLX5E_FTE_SET(headers_c, ethertype, 0xffff);
194  	MLX5E_FTE_SET(headers_v, ethertype, ETH_P_IPV6);
195  }
196  
197  static void
set_tcp(void * headers_c,void * headers_v,__be16 psrc_m,__be16 psrc_v,__be16 pdst_m,__be16 pdst_v)198  set_tcp(void *headers_c, void *headers_v, __be16 psrc_m, __be16 psrc_v,
199  	__be16 pdst_m, __be16 pdst_v)
200  {
201  	if (psrc_m) {
202  		MLX5E_FTE_SET(headers_c, tcp_sport, ntohs(psrc_m));
203  		MLX5E_FTE_SET(headers_v, tcp_sport, ntohs(psrc_v));
204  	}
205  	if (pdst_m) {
206  		MLX5E_FTE_SET(headers_c, tcp_dport, ntohs(pdst_m));
207  		MLX5E_FTE_SET(headers_v, tcp_dport, ntohs(pdst_v));
208  	}
209  
210  	MLX5E_FTE_SET(headers_c, ip_protocol, 0xffff);
211  	MLX5E_FTE_SET(headers_v, ip_protocol, IPPROTO_TCP);
212  }
213  
214  static void
set_udp(void * headers_c,void * headers_v,__be16 psrc_m,__be16 psrc_v,__be16 pdst_m,__be16 pdst_v)215  set_udp(void *headers_c, void *headers_v, __be16 psrc_m, __be16 psrc_v,
216  	__be16 pdst_m, __be16 pdst_v)
217  {
218  	if (psrc_m) {
219  		MLX5E_FTE_SET(headers_c, udp_sport, ntohs(psrc_m));
220  		MLX5E_FTE_SET(headers_v, udp_sport, ntohs(psrc_v));
221  	}
222  
223  	if (pdst_m) {
224  		MLX5E_FTE_SET(headers_c, udp_dport, ntohs(pdst_m));
225  		MLX5E_FTE_SET(headers_v, udp_dport, ntohs(pdst_v));
226  	}
227  
228  	MLX5E_FTE_SET(headers_c, ip_protocol, 0xffff);
229  	MLX5E_FTE_SET(headers_v, ip_protocol, IPPROTO_UDP);
230  }
231  
232  static void
parse_tcp4(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)233  parse_tcp4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
234  {
235  	struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.tcp_ip4_spec;
236  	struct ethtool_tcpip4_spec *l4_val  = &fs->h_u.tcp_ip4_spec;
237  
238  	set_ip4(headers_c, headers_v, l4_mask->ip4src, l4_val->ip4src,
239  		l4_mask->ip4dst, l4_val->ip4dst);
240  
241  	set_tcp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc,
242  		l4_mask->pdst, l4_val->pdst);
243  }
244  
245  static void
parse_udp4(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)246  parse_udp4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
247  {
248  	struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.udp_ip4_spec;
249  	struct ethtool_tcpip4_spec *l4_val  = &fs->h_u.udp_ip4_spec;
250  
251  	set_ip4(headers_c, headers_v, l4_mask->ip4src, l4_val->ip4src,
252  		l4_mask->ip4dst, l4_val->ip4dst);
253  
254  	set_udp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc,
255  		l4_mask->pdst, l4_val->pdst);
256  }
257  
258  static void
parse_ip4(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)259  parse_ip4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
260  {
261  	struct ethtool_usrip4_spec *l3_mask = &fs->m_u.usr_ip4_spec;
262  	struct ethtool_usrip4_spec *l3_val  = &fs->h_u.usr_ip4_spec;
263  
264  	set_ip4(headers_c, headers_v, l3_mask->ip4src, l3_val->ip4src,
265  		l3_mask->ip4dst, l3_val->ip4dst);
266  
267  	if (l3_mask->proto) {
268  		MLX5E_FTE_SET(headers_c, ip_protocol, l3_mask->proto);
269  		MLX5E_FTE_SET(headers_v, ip_protocol, l3_val->proto);
270  	}
271  }
272  
273  static void
parse_ip6(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)274  parse_ip6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
275  {
276  	struct ethtool_usrip6_spec *l3_mask = &fs->m_u.usr_ip6_spec;
277  	struct ethtool_usrip6_spec *l3_val  = &fs->h_u.usr_ip6_spec;
278  
279  	set_ip6(headers_c, headers_v, l3_mask->ip6src,
280  		l3_val->ip6src, l3_mask->ip6dst, l3_val->ip6dst);
281  
282  	if (l3_mask->l4_proto) {
283  		MLX5E_FTE_SET(headers_c, ip_protocol, l3_mask->l4_proto);
284  		MLX5E_FTE_SET(headers_v, ip_protocol, l3_val->l4_proto);
285  	}
286  }
287  
288  static void
parse_tcp6(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)289  parse_tcp6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
290  {
291  	struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.tcp_ip6_spec;
292  	struct ethtool_tcpip6_spec *l4_val  = &fs->h_u.tcp_ip6_spec;
293  
294  	set_ip6(headers_c, headers_v, l4_mask->ip6src,
295  		l4_val->ip6src, l4_mask->ip6dst, l4_val->ip6dst);
296  
297  	set_tcp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc,
298  		l4_mask->pdst, l4_val->pdst);
299  }
300  
301  static void
parse_udp6(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)302  parse_udp6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
303  {
304  	struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.udp_ip6_spec;
305  	struct ethtool_tcpip6_spec *l4_val  = &fs->h_u.udp_ip6_spec;
306  
307  	set_ip6(headers_c, headers_v, l4_mask->ip6src,
308  		l4_val->ip6src, l4_mask->ip6dst, l4_val->ip6dst);
309  
310  	set_udp(headers_c, headers_v, l4_mask->psrc, l4_val->psrc,
311  		l4_mask->pdst, l4_val->pdst);
312  }
313  
314  static void
parse_ether(void * headers_c,void * headers_v,struct ethtool_rx_flow_spec * fs)315  parse_ether(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
316  {
317  	struct ethhdr *eth_mask = &fs->m_u.ether_spec;
318  	struct ethhdr *eth_val = &fs->h_u.ether_spec;
319  
320  	mask_spec((u8 *)eth_mask, (u8 *)eth_val, sizeof(*eth_mask));
321  	ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, smac_47_16), eth_mask->h_source);
322  	ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, smac_47_16), eth_val->h_source);
323  	ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, dmac_47_16), eth_mask->h_dest);
324  	ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, dmac_47_16), eth_val->h_dest);
325  	MLX5E_FTE_SET(headers_c, ethertype, ntohs(eth_mask->h_proto));
326  	MLX5E_FTE_SET(headers_v, ethertype, ntohs(eth_val->h_proto));
327  }
328  
329  static void
set_cvlan(void * headers_c,void * headers_v,__be16 vlan_tci)330  set_cvlan(void *headers_c, void *headers_v, __be16 vlan_tci)
331  {
332  	MLX5E_FTE_SET(headers_c, cvlan_tag, 1);
333  	MLX5E_FTE_SET(headers_v, cvlan_tag, 1);
334  	MLX5E_FTE_SET(headers_c, first_vid, 0xfff);
335  	MLX5E_FTE_SET(headers_v, first_vid, ntohs(vlan_tci));
336  }
337  
338  static void
set_dmac(void * headers_c,void * headers_v,unsigned char m_dest[ETH_ALEN],unsigned char v_dest[ETH_ALEN])339  set_dmac(void *headers_c, void *headers_v,
340  	 unsigned char m_dest[ETH_ALEN], unsigned char v_dest[ETH_ALEN])
341  {
342  	ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, dmac_47_16), m_dest);
343  	ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, dmac_47_16), v_dest);
344  }
345  
set_flow_attrs(u32 * match_c,u32 * match_v,struct ethtool_rx_flow_spec * fs)346  static int set_flow_attrs(u32 *match_c, u32 *match_v,
347  			  struct ethtool_rx_flow_spec *fs)
348  {
349  	void *outer_headers_c = MLX5_ADDR_OF(fte_match_param, match_c,
350  					     outer_headers);
351  	void *outer_headers_v = MLX5_ADDR_OF(fte_match_param, match_v,
352  					     outer_headers);
353  	u32 flow_type = flow_type_mask(fs->flow_type);
354  
355  	switch (flow_type) {
356  	case TCP_V4_FLOW:
357  		parse_tcp4(outer_headers_c, outer_headers_v, fs);
358  		break;
359  	case UDP_V4_FLOW:
360  		parse_udp4(outer_headers_c, outer_headers_v, fs);
361  		break;
362  	case IP_USER_FLOW:
363  		parse_ip4(outer_headers_c, outer_headers_v, fs);
364  		break;
365  	case TCP_V6_FLOW:
366  		parse_tcp6(outer_headers_c, outer_headers_v, fs);
367  		break;
368  	case UDP_V6_FLOW:
369  		parse_udp6(outer_headers_c, outer_headers_v, fs);
370  		break;
371  	case IPV6_USER_FLOW:
372  		parse_ip6(outer_headers_c, outer_headers_v, fs);
373  		break;
374  	case ETHER_FLOW:
375  		parse_ether(outer_headers_c, outer_headers_v, fs);
376  		break;
377  	default:
378  		return -EINVAL;
379  	}
380  
381  	if ((fs->flow_type & FLOW_EXT) &&
382  	    (fs->m_ext.vlan_tci & cpu_to_be16(VLAN_VID_MASK)))
383  		set_cvlan(outer_headers_c, outer_headers_v, fs->h_ext.vlan_tci);
384  
385  	if (fs->flow_type & FLOW_MAC_EXT &&
386  	    !is_zero_ether_addr(fs->m_ext.h_dest)) {
387  		mask_spec(fs->m_ext.h_dest, fs->h_ext.h_dest, ETH_ALEN);
388  		set_dmac(outer_headers_c, outer_headers_v, fs->m_ext.h_dest,
389  			 fs->h_ext.h_dest);
390  	}
391  
392  	return 0;
393  }
394  
add_rule_to_list(struct mlx5e_priv * priv,struct mlx5e_ethtool_rule * rule)395  static void add_rule_to_list(struct mlx5e_priv *priv,
396  			     struct mlx5e_ethtool_rule *rule)
397  {
398  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(priv->fs);
399  	struct list_head *head = &ethtool->rules;
400  	struct mlx5e_ethtool_rule *iter;
401  
402  	list_for_each_entry(iter, &ethtool->rules, list) {
403  		if (iter->flow_spec.location > rule->flow_spec.location)
404  			break;
405  		head = &iter->list;
406  	}
407  	ethtool->tot_num_rules++;
408  	list_add(&rule->list, head);
409  }
410  
outer_header_zero(u32 * match_criteria)411  static bool outer_header_zero(u32 *match_criteria)
412  {
413  	int size = MLX5_FLD_SZ_BYTES(fte_match_param, outer_headers);
414  	char *outer_headers_c = MLX5_ADDR_OF(fte_match_param, match_criteria,
415  					     outer_headers);
416  
417  	return outer_headers_c[0] == 0 && !memcmp(outer_headers_c,
418  						  outer_headers_c + 1,
419  						  size - 1);
420  }
421  
flow_get_tirn(struct mlx5e_priv * priv,struct mlx5e_ethtool_rule * eth_rule,struct ethtool_rx_flow_spec * fs,u32 rss_context,u32 * tirn)422  static int flow_get_tirn(struct mlx5e_priv *priv,
423  			 struct mlx5e_ethtool_rule *eth_rule,
424  			 struct ethtool_rx_flow_spec *fs,
425  			 u32 rss_context, u32 *tirn)
426  {
427  	if (fs->flow_type & FLOW_RSS) {
428  		struct mlx5e_packet_merge_param pkt_merge_param;
429  		struct mlx5e_rss *rss;
430  		u32 flow_type;
431  		int err;
432  		int tt;
433  
434  		rss = mlx5e_rx_res_rss_get(priv->rx_res, rss_context);
435  		if (!rss)
436  			return -ENOENT;
437  
438  		flow_type = flow_type_mask(fs->flow_type);
439  		tt = flow_type_to_traffic_type(flow_type);
440  		if (tt < 0)
441  			return -EINVAL;
442  
443  		pkt_merge_param = priv->channels.params.packet_merge;
444  		err = mlx5e_rss_obtain_tirn(rss, tt, &pkt_merge_param, false, tirn);
445  		if (err)
446  			return err;
447  		eth_rule->rss = rss;
448  		mlx5e_rss_refcnt_inc(eth_rule->rss);
449  	} else {
450  		*tirn = mlx5e_rx_res_get_tirn_direct(priv->rx_res, fs->ring_cookie);
451  	}
452  
453  	return 0;
454  }
455  
456  static struct mlx5_flow_handle *
add_ethtool_flow_rule(struct mlx5e_priv * priv,struct mlx5e_ethtool_rule * eth_rule,struct mlx5_flow_table * ft,struct ethtool_rx_flow_spec * fs,u32 rss_context)457  add_ethtool_flow_rule(struct mlx5e_priv *priv,
458  		      struct mlx5e_ethtool_rule *eth_rule,
459  		      struct mlx5_flow_table *ft,
460  		      struct ethtool_rx_flow_spec *fs, u32 rss_context)
461  {
462  	struct mlx5_flow_act flow_act = { .flags = FLOW_ACT_NO_APPEND };
463  	struct mlx5_flow_destination *dst = NULL;
464  	struct mlx5_flow_handle *rule;
465  	struct mlx5_flow_spec *spec;
466  	int err = 0;
467  
468  	spec = kvzalloc(sizeof(*spec), GFP_KERNEL);
469  	if (!spec)
470  		return ERR_PTR(-ENOMEM);
471  	err = set_flow_attrs(spec->match_criteria, spec->match_value,
472  			     fs);
473  	if (err)
474  		goto free;
475  
476  	if (fs->ring_cookie == RX_CLS_FLOW_DISC) {
477  		flow_act.action = MLX5_FLOW_CONTEXT_ACTION_DROP;
478  	} else {
479  		dst = kzalloc(sizeof(*dst), GFP_KERNEL);
480  		if (!dst) {
481  			err = -ENOMEM;
482  			goto free;
483  		}
484  
485  		err = flow_get_tirn(priv, eth_rule, fs, rss_context, &dst->tir_num);
486  		if (err)
487  			goto free;
488  
489  		dst->type = MLX5_FLOW_DESTINATION_TYPE_TIR;
490  		flow_act.action = MLX5_FLOW_CONTEXT_ACTION_FWD_DEST;
491  	}
492  
493  	spec->match_criteria_enable = (!outer_header_zero(spec->match_criteria));
494  	spec->flow_context.flow_tag = MLX5_FS_DEFAULT_FLOW_TAG;
495  	rule = mlx5_add_flow_rules(ft, spec, &flow_act, dst, dst ? 1 : 0);
496  	if (IS_ERR(rule)) {
497  		err = PTR_ERR(rule);
498  		netdev_err(priv->netdev, "%s: failed to add ethtool steering rule: %d\n",
499  			   __func__, err);
500  		goto free;
501  	}
502  free:
503  	kvfree(spec);
504  	kfree(dst);
505  	return err ? ERR_PTR(err) : rule;
506  }
507  
del_ethtool_rule(struct mlx5e_flow_steering * fs,struct mlx5e_ethtool_rule * eth_rule)508  static void del_ethtool_rule(struct mlx5e_flow_steering *fs,
509  			     struct mlx5e_ethtool_rule *eth_rule)
510  {
511  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs);
512  	if (eth_rule->rule)
513  		mlx5_del_flow_rules(eth_rule->rule);
514  	if (eth_rule->rss)
515  		mlx5e_rss_refcnt_dec(eth_rule->rss);
516  	list_del(&eth_rule->list);
517  	ethtool->tot_num_rules--;
518  	put_flow_table(eth_rule->eth_ft);
519  	kfree(eth_rule);
520  }
521  
find_ethtool_rule(struct mlx5e_priv * priv,int location)522  static struct mlx5e_ethtool_rule *find_ethtool_rule(struct mlx5e_priv *priv,
523  						    int location)
524  {
525  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(priv->fs);
526  	struct mlx5e_ethtool_rule *iter;
527  
528  	list_for_each_entry(iter, &ethtool->rules, list) {
529  		if (iter->flow_spec.location == location)
530  			return iter;
531  	}
532  	return NULL;
533  }
534  
get_ethtool_rule(struct mlx5e_priv * priv,int location)535  static struct mlx5e_ethtool_rule *get_ethtool_rule(struct mlx5e_priv *priv,
536  						   int location)
537  {
538  	struct mlx5e_ethtool_rule *eth_rule;
539  
540  	eth_rule = find_ethtool_rule(priv, location);
541  	if (eth_rule)
542  		del_ethtool_rule(priv->fs, eth_rule);
543  
544  	eth_rule = kzalloc(sizeof(*eth_rule), GFP_KERNEL);
545  	if (!eth_rule)
546  		return ERR_PTR(-ENOMEM);
547  
548  	add_rule_to_list(priv, eth_rule);
549  	return eth_rule;
550  }
551  
552  #define MAX_NUM_OF_ETHTOOL_RULES BIT(10)
553  
554  #define all_ones(field) (field == (__force typeof(field))-1)
555  #define all_zeros_or_all_ones(field)		\
556  	((field) == 0 || (field) == (__force typeof(field))-1)
557  
validate_ethter(struct ethtool_rx_flow_spec * fs)558  static int validate_ethter(struct ethtool_rx_flow_spec *fs)
559  {
560  	struct ethhdr *eth_mask = &fs->m_u.ether_spec;
561  	int ntuples = 0;
562  
563  	if (!is_zero_ether_addr(eth_mask->h_dest))
564  		ntuples++;
565  	if (!is_zero_ether_addr(eth_mask->h_source))
566  		ntuples++;
567  	if (eth_mask->h_proto)
568  		ntuples++;
569  	return ntuples;
570  }
571  
validate_tcpudp4(struct ethtool_rx_flow_spec * fs)572  static int validate_tcpudp4(struct ethtool_rx_flow_spec *fs)
573  {
574  	struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.tcp_ip4_spec;
575  	int ntuples = 0;
576  
577  	if (l4_mask->tos)
578  		return -EINVAL;
579  
580  	if (l4_mask->ip4src)
581  		ntuples++;
582  	if (l4_mask->ip4dst)
583  		ntuples++;
584  	if (l4_mask->psrc)
585  		ntuples++;
586  	if (l4_mask->pdst)
587  		ntuples++;
588  	/* Flow is TCP/UDP */
589  	return ++ntuples;
590  }
591  
validate_ip4(struct ethtool_rx_flow_spec * fs)592  static int validate_ip4(struct ethtool_rx_flow_spec *fs)
593  {
594  	struct ethtool_usrip4_spec *l3_mask = &fs->m_u.usr_ip4_spec;
595  	int ntuples = 0;
596  
597  	if (l3_mask->l4_4_bytes || l3_mask->tos ||
598  	    fs->h_u.usr_ip4_spec.ip_ver != ETH_RX_NFC_IP4)
599  		return -EINVAL;
600  	if (l3_mask->ip4src)
601  		ntuples++;
602  	if (l3_mask->ip4dst)
603  		ntuples++;
604  	if (l3_mask->proto)
605  		ntuples++;
606  	/* Flow is IPv4 */
607  	return ++ntuples;
608  }
609  
validate_ip6(struct ethtool_rx_flow_spec * fs)610  static int validate_ip6(struct ethtool_rx_flow_spec *fs)
611  {
612  	struct ethtool_usrip6_spec *l3_mask = &fs->m_u.usr_ip6_spec;
613  	int ntuples = 0;
614  
615  	if (l3_mask->l4_4_bytes || l3_mask->tclass)
616  		return -EINVAL;
617  	if (!ipv6_addr_any((struct in6_addr *)l3_mask->ip6src))
618  		ntuples++;
619  
620  	if (!ipv6_addr_any((struct in6_addr *)l3_mask->ip6dst))
621  		ntuples++;
622  	if (l3_mask->l4_proto)
623  		ntuples++;
624  	/* Flow is IPv6 */
625  	return ++ntuples;
626  }
627  
validate_tcpudp6(struct ethtool_rx_flow_spec * fs)628  static int validate_tcpudp6(struct ethtool_rx_flow_spec *fs)
629  {
630  	struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.tcp_ip6_spec;
631  	int ntuples = 0;
632  
633  	if (l4_mask->tclass)
634  		return -EINVAL;
635  
636  	if (!ipv6_addr_any((struct in6_addr *)l4_mask->ip6src))
637  		ntuples++;
638  
639  	if (!ipv6_addr_any((struct in6_addr *)l4_mask->ip6dst))
640  		ntuples++;
641  
642  	if (l4_mask->psrc)
643  		ntuples++;
644  	if (l4_mask->pdst)
645  		ntuples++;
646  	/* Flow is TCP/UDP */
647  	return ++ntuples;
648  }
649  
validate_vlan(struct ethtool_rx_flow_spec * fs)650  static int validate_vlan(struct ethtool_rx_flow_spec *fs)
651  {
652  	if (fs->m_ext.vlan_etype ||
653  	    fs->m_ext.vlan_tci != cpu_to_be16(VLAN_VID_MASK))
654  		return -EINVAL;
655  
656  	if (fs->m_ext.vlan_tci &&
657  	    (be16_to_cpu(fs->h_ext.vlan_tci) >= VLAN_N_VID))
658  		return -EINVAL;
659  
660  	return 1;
661  }
662  
validate_flow(struct mlx5e_priv * priv,struct ethtool_rx_flow_spec * fs)663  static int validate_flow(struct mlx5e_priv *priv,
664  			 struct ethtool_rx_flow_spec *fs)
665  {
666  	int num_tuples = 0;
667  	int ret = 0;
668  
669  	if (fs->location >= MAX_NUM_OF_ETHTOOL_RULES)
670  		return -ENOSPC;
671  
672  	if (fs->ring_cookie != RX_CLS_FLOW_DISC)
673  		if (fs->ring_cookie >= priv->channels.params.num_channels)
674  			return -EINVAL;
675  
676  	switch (flow_type_mask(fs->flow_type)) {
677  	case ETHER_FLOW:
678  		num_tuples += validate_ethter(fs);
679  		break;
680  	case TCP_V4_FLOW:
681  	case UDP_V4_FLOW:
682  		ret = validate_tcpudp4(fs);
683  		if (ret < 0)
684  			return ret;
685  		num_tuples += ret;
686  		break;
687  	case IP_USER_FLOW:
688  		ret = validate_ip4(fs);
689  		if (ret < 0)
690  			return ret;
691  		num_tuples += ret;
692  		break;
693  	case TCP_V6_FLOW:
694  	case UDP_V6_FLOW:
695  		ret = validate_tcpudp6(fs);
696  		if (ret < 0)
697  			return ret;
698  		num_tuples += ret;
699  		break;
700  	case IPV6_USER_FLOW:
701  		ret = validate_ip6(fs);
702  		if (ret < 0)
703  			return ret;
704  		num_tuples += ret;
705  		break;
706  	default:
707  		return -ENOTSUPP;
708  	}
709  	if ((fs->flow_type & FLOW_EXT)) {
710  		ret = validate_vlan(fs);
711  		if (ret < 0)
712  			return ret;
713  		num_tuples += ret;
714  	}
715  
716  	if (fs->flow_type & FLOW_MAC_EXT &&
717  	    !is_zero_ether_addr(fs->m_ext.h_dest))
718  		num_tuples++;
719  
720  	return num_tuples;
721  }
722  
723  static int
mlx5e_ethtool_flow_replace(struct mlx5e_priv * priv,struct ethtool_rx_flow_spec * fs,u32 rss_context)724  mlx5e_ethtool_flow_replace(struct mlx5e_priv *priv,
725  			   struct ethtool_rx_flow_spec *fs, u32 rss_context)
726  {
727  	struct mlx5e_ethtool_table *eth_ft;
728  	struct mlx5e_ethtool_rule *eth_rule;
729  	struct mlx5_flow_handle *rule;
730  	int num_tuples;
731  	int err;
732  
733  	num_tuples = validate_flow(priv, fs);
734  	if (num_tuples <= 0) {
735  		netdev_warn(priv->netdev, "%s: flow is not valid %d\n",
736  			    __func__, num_tuples);
737  		return num_tuples < 0 ? num_tuples : -EINVAL;
738  	}
739  
740  	eth_ft = get_flow_table(priv, fs, num_tuples);
741  	if (IS_ERR(eth_ft))
742  		return PTR_ERR(eth_ft);
743  
744  	eth_rule = get_ethtool_rule(priv, fs->location);
745  	if (IS_ERR(eth_rule)) {
746  		put_flow_table(eth_ft);
747  		return PTR_ERR(eth_rule);
748  	}
749  
750  	eth_rule->flow_spec = *fs;
751  	eth_rule->eth_ft = eth_ft;
752  
753  	rule = add_ethtool_flow_rule(priv, eth_rule, eth_ft->ft, fs, rss_context);
754  	if (IS_ERR(rule)) {
755  		err = PTR_ERR(rule);
756  		goto del_ethtool_rule;
757  	}
758  
759  	eth_rule->rule = rule;
760  
761  	return 0;
762  
763  del_ethtool_rule:
764  	del_ethtool_rule(priv->fs, eth_rule);
765  
766  	return err;
767  }
768  
769  static int
mlx5e_ethtool_flow_remove(struct mlx5e_priv * priv,int location)770  mlx5e_ethtool_flow_remove(struct mlx5e_priv *priv, int location)
771  {
772  	struct mlx5e_ethtool_rule *eth_rule;
773  	int err = 0;
774  
775  	if (location >= MAX_NUM_OF_ETHTOOL_RULES)
776  		return -ENOSPC;
777  
778  	eth_rule = find_ethtool_rule(priv, location);
779  	if (!eth_rule) {
780  		err =  -ENOENT;
781  		goto out;
782  	}
783  
784  	del_ethtool_rule(priv->fs, eth_rule);
785  out:
786  	return err;
787  }
788  
789  static int
mlx5e_ethtool_get_flow(struct mlx5e_priv * priv,struct ethtool_rxnfc * info,int location)790  mlx5e_ethtool_get_flow(struct mlx5e_priv *priv,
791  		       struct ethtool_rxnfc *info, int location)
792  {
793  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(priv->fs);
794  	struct mlx5e_ethtool_rule *eth_rule;
795  
796  	if (location < 0 || location >= MAX_NUM_OF_ETHTOOL_RULES)
797  		return -EINVAL;
798  
799  	list_for_each_entry(eth_rule, &ethtool->rules, list) {
800  		int index;
801  
802  		if (eth_rule->flow_spec.location != location)
803  			continue;
804  		if (!info)
805  			return 0;
806  		info->fs = eth_rule->flow_spec;
807  		if (!eth_rule->rss)
808  			return 0;
809  		index = mlx5e_rx_res_rss_index(priv->rx_res, eth_rule->rss);
810  		if (index < 0)
811  			return index;
812  		info->rss_context = index;
813  		return 0;
814  	}
815  
816  	return -ENOENT;
817  }
818  
819  static int
mlx5e_ethtool_get_all_flows(struct mlx5e_priv * priv,struct ethtool_rxnfc * info,u32 * rule_locs)820  mlx5e_ethtool_get_all_flows(struct mlx5e_priv *priv,
821  			    struct ethtool_rxnfc *info, u32 *rule_locs)
822  {
823  	int location = 0;
824  	int idx = 0;
825  	int err = 0;
826  
827  	info->data = MAX_NUM_OF_ETHTOOL_RULES;
828  	while ((!err || err == -ENOENT) && idx < info->rule_cnt) {
829  		err = mlx5e_ethtool_get_flow(priv, NULL, location);
830  		if (!err)
831  			rule_locs[idx++] = location;
832  		location++;
833  	}
834  	return err;
835  }
836  
mlx5e_ethtool_alloc(struct mlx5e_ethtool_steering ** ethtool)837  int mlx5e_ethtool_alloc(struct mlx5e_ethtool_steering **ethtool)
838  {
839  	*ethtool =  kvzalloc(sizeof(**ethtool), GFP_KERNEL);
840  	if (!*ethtool)
841  		return -ENOMEM;
842  	return 0;
843  }
844  
mlx5e_ethtool_free(struct mlx5e_ethtool_steering * ethtool)845  void mlx5e_ethtool_free(struct mlx5e_ethtool_steering *ethtool)
846  {
847  	kvfree(ethtool);
848  }
849  
mlx5e_ethtool_cleanup_steering(struct mlx5e_flow_steering * fs)850  void mlx5e_ethtool_cleanup_steering(struct mlx5e_flow_steering *fs)
851  {
852  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs);
853  	struct mlx5e_ethtool_rule *iter;
854  	struct mlx5e_ethtool_rule *temp;
855  
856  	list_for_each_entry_safe(iter, temp, &ethtool->rules, list)
857  		del_ethtool_rule(fs, iter);
858  }
859  
mlx5e_ethtool_init_steering(struct mlx5e_flow_steering * fs)860  void mlx5e_ethtool_init_steering(struct mlx5e_flow_steering *fs)
861  {
862  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs);
863  
864  	INIT_LIST_HEAD(&ethtool->rules);
865  }
866  
flow_type_to_traffic_type(u32 flow_type)867  static int flow_type_to_traffic_type(u32 flow_type)
868  {
869  	switch (flow_type) {
870  	case TCP_V4_FLOW:
871  		return MLX5_TT_IPV4_TCP;
872  	case TCP_V6_FLOW:
873  		return MLX5_TT_IPV6_TCP;
874  	case UDP_V4_FLOW:
875  		return MLX5_TT_IPV4_UDP;
876  	case UDP_V6_FLOW:
877  		return MLX5_TT_IPV6_UDP;
878  	case AH_V4_FLOW:
879  		return MLX5_TT_IPV4_IPSEC_AH;
880  	case AH_V6_FLOW:
881  		return MLX5_TT_IPV6_IPSEC_AH;
882  	case ESP_V4_FLOW:
883  		return MLX5_TT_IPV4_IPSEC_ESP;
884  	case ESP_V6_FLOW:
885  		return MLX5_TT_IPV6_IPSEC_ESP;
886  	case IPV4_FLOW:
887  		return MLX5_TT_IPV4;
888  	case IPV6_FLOW:
889  		return MLX5_TT_IPV6;
890  	default:
891  		return -EINVAL;
892  	}
893  }
894  
mlx5e_set_rss_hash_opt(struct mlx5e_priv * priv,struct ethtool_rxnfc * nfc)895  static int mlx5e_set_rss_hash_opt(struct mlx5e_priv *priv,
896  				  struct ethtool_rxnfc *nfc)
897  {
898  	u8 rx_hash_field = 0;
899  	u32 flow_type = 0;
900  	u32 rss_idx = 0;
901  	int err;
902  	int tt;
903  
904  	if (nfc->flow_type & FLOW_RSS)
905  		rss_idx = nfc->rss_context;
906  
907  	flow_type = flow_type_mask(nfc->flow_type);
908  	tt = flow_type_to_traffic_type(flow_type);
909  	if (tt < 0)
910  		return tt;
911  
912  	/*  RSS does not support anything other than hashing to queues
913  	 *  on src IP, dest IP, TCP/UDP src port and TCP/UDP dest
914  	 *  port.
915  	 */
916  	if (flow_type != TCP_V4_FLOW &&
917  	    flow_type != TCP_V6_FLOW &&
918  	    flow_type != UDP_V4_FLOW &&
919  	    flow_type != UDP_V6_FLOW)
920  		return -EOPNOTSUPP;
921  
922  	if (nfc->data & ~(RXH_IP_SRC | RXH_IP_DST |
923  			  RXH_L4_B_0_1 | RXH_L4_B_2_3))
924  		return -EOPNOTSUPP;
925  
926  	if (nfc->data & RXH_IP_SRC)
927  		rx_hash_field |= MLX5_HASH_FIELD_SEL_SRC_IP;
928  	if (nfc->data & RXH_IP_DST)
929  		rx_hash_field |= MLX5_HASH_FIELD_SEL_DST_IP;
930  	if (nfc->data & RXH_L4_B_0_1)
931  		rx_hash_field |= MLX5_HASH_FIELD_SEL_L4_SPORT;
932  	if (nfc->data & RXH_L4_B_2_3)
933  		rx_hash_field |= MLX5_HASH_FIELD_SEL_L4_DPORT;
934  
935  	mutex_lock(&priv->state_lock);
936  	err = mlx5e_rx_res_rss_set_hash_fields(priv->rx_res, rss_idx, tt, rx_hash_field);
937  	mutex_unlock(&priv->state_lock);
938  
939  	return err;
940  }
941  
mlx5e_get_rss_hash_opt(struct mlx5e_priv * priv,struct ethtool_rxnfc * nfc)942  static int mlx5e_get_rss_hash_opt(struct mlx5e_priv *priv,
943  				  struct ethtool_rxnfc *nfc)
944  {
945  	int hash_field = 0;
946  	u32 flow_type = 0;
947  	u32 rss_idx = 0;
948  	int tt;
949  
950  	if (nfc->flow_type & FLOW_RSS)
951  		rss_idx = nfc->rss_context;
952  
953  	flow_type = flow_type_mask(nfc->flow_type);
954  	tt = flow_type_to_traffic_type(flow_type);
955  	if (tt < 0)
956  		return tt;
957  
958  	hash_field = mlx5e_rx_res_rss_get_hash_fields(priv->rx_res, rss_idx, tt);
959  	if (hash_field < 0)
960  		return hash_field;
961  
962  	nfc->data = 0;
963  
964  	if (hash_field & MLX5_HASH_FIELD_SEL_SRC_IP)
965  		nfc->data |= RXH_IP_SRC;
966  	if (hash_field & MLX5_HASH_FIELD_SEL_DST_IP)
967  		nfc->data |= RXH_IP_DST;
968  	if (hash_field & MLX5_HASH_FIELD_SEL_L4_SPORT)
969  		nfc->data |= RXH_L4_B_0_1;
970  	if (hash_field & MLX5_HASH_FIELD_SEL_L4_DPORT)
971  		nfc->data |= RXH_L4_B_2_3;
972  
973  	return 0;
974  }
975  
mlx5e_ethtool_set_rxnfc(struct mlx5e_priv * priv,struct ethtool_rxnfc * cmd)976  int mlx5e_ethtool_set_rxnfc(struct mlx5e_priv *priv, struct ethtool_rxnfc *cmd)
977  {
978  	int err = 0;
979  
980  	switch (cmd->cmd) {
981  	case ETHTOOL_SRXCLSRLINS:
982  		err = mlx5e_ethtool_flow_replace(priv, &cmd->fs, cmd->rss_context);
983  		break;
984  	case ETHTOOL_SRXCLSRLDEL:
985  		err = mlx5e_ethtool_flow_remove(priv, cmd->fs.location);
986  		break;
987  	case ETHTOOL_SRXFH:
988  		err = mlx5e_set_rss_hash_opt(priv, cmd);
989  		break;
990  	default:
991  		err = -EOPNOTSUPP;
992  		break;
993  	}
994  
995  	return err;
996  }
997  
mlx5e_ethtool_get_rxnfc(struct mlx5e_priv * priv,struct ethtool_rxnfc * info,u32 * rule_locs)998  int mlx5e_ethtool_get_rxnfc(struct mlx5e_priv *priv,
999  			    struct ethtool_rxnfc *info, u32 *rule_locs)
1000  {
1001  	struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(priv->fs);
1002  	int err = 0;
1003  
1004  	switch (info->cmd) {
1005  	case ETHTOOL_GRXCLSRLCNT:
1006  		info->rule_cnt = ethtool->tot_num_rules;
1007  		break;
1008  	case ETHTOOL_GRXCLSRULE:
1009  		err = mlx5e_ethtool_get_flow(priv, info, info->fs.location);
1010  		break;
1011  	case ETHTOOL_GRXCLSRLALL:
1012  		err = mlx5e_ethtool_get_all_flows(priv, info, rule_locs);
1013  		break;
1014  	case ETHTOOL_GRXFH:
1015  		err =  mlx5e_get_rss_hash_opt(priv, info);
1016  		break;
1017  	default:
1018  		err = -EOPNOTSUPP;
1019  		break;
1020  	}
1021  
1022  	return err;
1023  }
1024  
1025