1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (C) 2018-2020, Intel Corporation. */
3 
4 /* flow director ethtool support for ice */
5 
6 #include "ice.h"
7 #include "ice_lib.h"
8 #include "ice_flow.h"
9 
10 static struct in6_addr full_ipv6_addr_mask = {
11 	.in6_u = {
12 		.u6_addr8 = {
13 			0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF,
14 			0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF,
15 		}
16 	}
17 };
18 
19 static struct in6_addr zero_ipv6_addr_mask = {
20 	.in6_u = {
21 		.u6_addr8 = {
22 			0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
23 			0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
24 		}
25 	}
26 };
27 
28 /* calls to ice_flow_add_prof require the number of segments in the array
29  * for segs_cnt. In this code that is one more than the index.
30  */
31 #define TNL_SEG_CNT(_TNL_) ((_TNL_) + 1)
32 
33 /**
34  * ice_fltr_to_ethtool_flow - convert filter type values to ethtool
35  * flow type values
36  * @flow: filter type to be converted
37  *
38  * Returns the corresponding ethtool flow type.
39  */
40 static int ice_fltr_to_ethtool_flow(enum ice_fltr_ptype flow)
41 {
42 	switch (flow) {
43 	case ICE_FLTR_PTYPE_NONF_IPV4_TCP:
44 		return TCP_V4_FLOW;
45 	case ICE_FLTR_PTYPE_NONF_IPV4_UDP:
46 		return UDP_V4_FLOW;
47 	case ICE_FLTR_PTYPE_NONF_IPV4_SCTP:
48 		return SCTP_V4_FLOW;
49 	case ICE_FLTR_PTYPE_NONF_IPV4_OTHER:
50 		return IPV4_USER_FLOW;
51 	case ICE_FLTR_PTYPE_NONF_IPV6_TCP:
52 		return TCP_V6_FLOW;
53 	case ICE_FLTR_PTYPE_NONF_IPV6_UDP:
54 		return UDP_V6_FLOW;
55 	case ICE_FLTR_PTYPE_NONF_IPV6_SCTP:
56 		return SCTP_V6_FLOW;
57 	case ICE_FLTR_PTYPE_NONF_IPV6_OTHER:
58 		return IPV6_USER_FLOW;
59 	default:
60 		/* 0 is undefined ethtool flow */
61 		return 0;
62 	}
63 }
64 
65 /**
66  * ice_ethtool_flow_to_fltr - convert ethtool flow type to filter enum
67  * @eth: Ethtool flow type to be converted
68  *
69  * Returns flow enum
70  */
71 static enum ice_fltr_ptype ice_ethtool_flow_to_fltr(int eth)
72 {
73 	switch (eth) {
74 	case TCP_V4_FLOW:
75 		return ICE_FLTR_PTYPE_NONF_IPV4_TCP;
76 	case UDP_V4_FLOW:
77 		return ICE_FLTR_PTYPE_NONF_IPV4_UDP;
78 	case SCTP_V4_FLOW:
79 		return ICE_FLTR_PTYPE_NONF_IPV4_SCTP;
80 	case IPV4_USER_FLOW:
81 		return ICE_FLTR_PTYPE_NONF_IPV4_OTHER;
82 	case TCP_V6_FLOW:
83 		return ICE_FLTR_PTYPE_NONF_IPV6_TCP;
84 	case UDP_V6_FLOW:
85 		return ICE_FLTR_PTYPE_NONF_IPV6_UDP;
86 	case SCTP_V6_FLOW:
87 		return ICE_FLTR_PTYPE_NONF_IPV6_SCTP;
88 	case IPV6_USER_FLOW:
89 		return ICE_FLTR_PTYPE_NONF_IPV6_OTHER;
90 	default:
91 		return ICE_FLTR_PTYPE_NONF_NONE;
92 	}
93 }
94 
95 /**
96  * ice_is_mask_valid - check mask field set
97  * @mask: full mask to check
98  * @field: field for which mask should be valid
99  *
100  * If the mask is fully set return true. If it is not valid for field return
101  * false.
102  */
103 static bool ice_is_mask_valid(u64 mask, u64 field)
104 {
105 	return (mask & field) == field;
106 }
107 
108 /**
109  * ice_get_ethtool_fdir_entry - fill ethtool structure with fdir filter data
110  * @hw: hardware structure that contains filter list
111  * @cmd: ethtool command data structure to receive the filter data
112  *
113  * Returns 0 on success and -EINVAL on failure
114  */
115 int ice_get_ethtool_fdir_entry(struct ice_hw *hw, struct ethtool_rxnfc *cmd)
116 {
117 	struct ethtool_rx_flow_spec *fsp;
118 	struct ice_fdir_fltr *rule;
119 	int ret = 0;
120 	u16 idx;
121 
122 	fsp = (struct ethtool_rx_flow_spec *)&cmd->fs;
123 
124 	mutex_lock(&hw->fdir_fltr_lock);
125 
126 	rule = ice_fdir_find_fltr_by_idx(hw, fsp->location);
127 
128 	if (!rule || fsp->location != rule->fltr_id) {
129 		ret = -EINVAL;
130 		goto release_lock;
131 	}
132 
133 	fsp->flow_type = ice_fltr_to_ethtool_flow(rule->flow_type);
134 
135 	memset(&fsp->m_u, 0, sizeof(fsp->m_u));
136 	memset(&fsp->m_ext, 0, sizeof(fsp->m_ext));
137 
138 	switch (fsp->flow_type) {
139 	case IPV4_USER_FLOW:
140 		fsp->h_u.usr_ip4_spec.ip_ver = ETH_RX_NFC_IP4;
141 		fsp->h_u.usr_ip4_spec.proto = 0;
142 		fsp->h_u.usr_ip4_spec.l4_4_bytes = rule->ip.v4.l4_header;
143 		fsp->h_u.usr_ip4_spec.tos = rule->ip.v4.tos;
144 		fsp->h_u.usr_ip4_spec.ip4src = rule->ip.v4.src_ip;
145 		fsp->h_u.usr_ip4_spec.ip4dst = rule->ip.v4.dst_ip;
146 		fsp->m_u.usr_ip4_spec.ip4src = rule->mask.v4.src_ip;
147 		fsp->m_u.usr_ip4_spec.ip4dst = rule->mask.v4.dst_ip;
148 		fsp->m_u.usr_ip4_spec.ip_ver = 0xFF;
149 		fsp->m_u.usr_ip4_spec.proto = 0;
150 		fsp->m_u.usr_ip4_spec.l4_4_bytes = rule->mask.v4.l4_header;
151 		fsp->m_u.usr_ip4_spec.tos = rule->mask.v4.tos;
152 		break;
153 	case TCP_V4_FLOW:
154 	case UDP_V4_FLOW:
155 	case SCTP_V4_FLOW:
156 		fsp->h_u.tcp_ip4_spec.psrc = rule->ip.v4.src_port;
157 		fsp->h_u.tcp_ip4_spec.pdst = rule->ip.v4.dst_port;
158 		fsp->h_u.tcp_ip4_spec.ip4src = rule->ip.v4.src_ip;
159 		fsp->h_u.tcp_ip4_spec.ip4dst = rule->ip.v4.dst_ip;
160 		fsp->m_u.tcp_ip4_spec.psrc = rule->mask.v4.src_port;
161 		fsp->m_u.tcp_ip4_spec.pdst = rule->mask.v4.dst_port;
162 		fsp->m_u.tcp_ip4_spec.ip4src = rule->mask.v4.src_ip;
163 		fsp->m_u.tcp_ip4_spec.ip4dst = rule->mask.v4.dst_ip;
164 		break;
165 	case IPV6_USER_FLOW:
166 		fsp->h_u.usr_ip6_spec.l4_4_bytes = rule->ip.v6.l4_header;
167 		fsp->h_u.usr_ip6_spec.tclass = rule->ip.v6.tc;
168 		fsp->h_u.usr_ip6_spec.l4_proto = rule->ip.v6.proto;
169 		memcpy(fsp->h_u.tcp_ip6_spec.ip6src, rule->ip.v6.src_ip,
170 		       sizeof(struct in6_addr));
171 		memcpy(fsp->h_u.tcp_ip6_spec.ip6dst, rule->ip.v6.dst_ip,
172 		       sizeof(struct in6_addr));
173 		memcpy(fsp->m_u.tcp_ip6_spec.ip6src, rule->mask.v6.src_ip,
174 		       sizeof(struct in6_addr));
175 		memcpy(fsp->m_u.tcp_ip6_spec.ip6dst, rule->mask.v6.dst_ip,
176 		       sizeof(struct in6_addr));
177 		fsp->m_u.usr_ip6_spec.l4_4_bytes = rule->mask.v6.l4_header;
178 		fsp->m_u.usr_ip6_spec.tclass = rule->mask.v6.tc;
179 		fsp->m_u.usr_ip6_spec.l4_proto = rule->mask.v6.proto;
180 		break;
181 	case TCP_V6_FLOW:
182 	case UDP_V6_FLOW:
183 	case SCTP_V6_FLOW:
184 		memcpy(fsp->h_u.tcp_ip6_spec.ip6src, rule->ip.v6.src_ip,
185 		       sizeof(struct in6_addr));
186 		memcpy(fsp->h_u.tcp_ip6_spec.ip6dst, rule->ip.v6.dst_ip,
187 		       sizeof(struct in6_addr));
188 		fsp->h_u.tcp_ip6_spec.psrc = rule->ip.v6.src_port;
189 		fsp->h_u.tcp_ip6_spec.pdst = rule->ip.v6.dst_port;
190 		memcpy(fsp->m_u.tcp_ip6_spec.ip6src,
191 		       rule->mask.v6.src_ip,
192 		       sizeof(struct in6_addr));
193 		memcpy(fsp->m_u.tcp_ip6_spec.ip6dst,
194 		       rule->mask.v6.dst_ip,
195 		       sizeof(struct in6_addr));
196 		fsp->m_u.tcp_ip6_spec.psrc = rule->mask.v6.src_port;
197 		fsp->m_u.tcp_ip6_spec.pdst = rule->mask.v6.dst_port;
198 		fsp->h_u.tcp_ip6_spec.tclass = rule->ip.v6.tc;
199 		fsp->m_u.tcp_ip6_spec.tclass = rule->mask.v6.tc;
200 		break;
201 	default:
202 		break;
203 	}
204 
205 	if (rule->dest_ctl == ICE_FLTR_PRGM_DESC_DEST_DROP_PKT)
206 		fsp->ring_cookie = RX_CLS_FLOW_DISC;
207 	else
208 		fsp->ring_cookie = rule->q_index;
209 
210 	idx = ice_ethtool_flow_to_fltr(fsp->flow_type);
211 	if (idx == ICE_FLTR_PTYPE_NONF_NONE) {
212 		dev_err(ice_hw_to_dev(hw), "Missing input index for flow_type %d\n",
213 			rule->flow_type);
214 		ret = -EINVAL;
215 	}
216 
217 release_lock:
218 	mutex_unlock(&hw->fdir_fltr_lock);
219 	return ret;
220 }
221 
222 /**
223  * ice_get_fdir_fltr_ids - fill buffer with filter IDs of active filters
224  * @hw: hardware structure containing the filter list
225  * @cmd: ethtool command data structure
226  * @rule_locs: ethtool array passed in from OS to receive filter IDs
227  *
228  * Returns 0 as expected for success by ethtool
229  */
230 int
231 ice_get_fdir_fltr_ids(struct ice_hw *hw, struct ethtool_rxnfc *cmd,
232 		      u32 *rule_locs)
233 {
234 	struct ice_fdir_fltr *f_rule;
235 	unsigned int cnt = 0;
236 	int val = 0;
237 
238 	/* report total rule count */
239 	cmd->data = ice_get_fdir_cnt_all(hw);
240 
241 	mutex_lock(&hw->fdir_fltr_lock);
242 
243 	list_for_each_entry(f_rule, &hw->fdir_list_head, fltr_node) {
244 		if (cnt == cmd->rule_cnt) {
245 			val = -EMSGSIZE;
246 			goto release_lock;
247 		}
248 		rule_locs[cnt] = f_rule->fltr_id;
249 		cnt++;
250 	}
251 
252 release_lock:
253 	mutex_unlock(&hw->fdir_fltr_lock);
254 	if (!val)
255 		cmd->rule_cnt = cnt;
256 	return val;
257 }
258 
259 /**
260  * ice_fdir_get_hw_prof - return the ice_fd_hw_proc associated with a flow
261  * @hw: hardware structure containing the filter list
262  * @blk: hardware block
263  * @flow: FDir flow type to release
264  */
265 static struct ice_fd_hw_prof *
266 ice_fdir_get_hw_prof(struct ice_hw *hw, enum ice_block blk, int flow)
267 {
268 	if (blk == ICE_BLK_FD && hw->fdir_prof)
269 		return hw->fdir_prof[flow];
270 
271 	return NULL;
272 }
273 
274 /**
275  * ice_fdir_erase_flow_from_hw - remove a flow from the HW profile tables
276  * @hw: hardware structure containing the filter list
277  * @blk: hardware block
278  * @flow: FDir flow type to release
279  */
280 static void
281 ice_fdir_erase_flow_from_hw(struct ice_hw *hw, enum ice_block blk, int flow)
282 {
283 	struct ice_fd_hw_prof *prof = ice_fdir_get_hw_prof(hw, blk, flow);
284 	int tun;
285 
286 	if (!prof)
287 		return;
288 
289 	for (tun = 0; tun < ICE_FD_HW_SEG_MAX; tun++) {
290 		u64 prof_id;
291 		int j;
292 
293 		prof_id = flow + tun * ICE_FLTR_PTYPE_MAX;
294 		for (j = 0; j < prof->cnt; j++) {
295 			u16 vsi_num;
296 
297 			if (!prof->entry_h[j][tun] || !prof->vsi_h[j])
298 				continue;
299 			vsi_num = ice_get_hw_vsi_num(hw, prof->vsi_h[j]);
300 			ice_rem_prof_id_flow(hw, blk, vsi_num, prof_id);
301 			ice_flow_rem_entry(hw, blk, prof->entry_h[j][tun]);
302 			prof->entry_h[j][tun] = 0;
303 		}
304 		ice_flow_rem_prof(hw, blk, prof_id);
305 	}
306 }
307 
308 /**
309  * ice_fdir_rem_flow - release the ice_flow structures for a filter type
310  * @hw: hardware structure containing the filter list
311  * @blk: hardware block
312  * @flow_type: FDir flow type to release
313  */
314 static void
315 ice_fdir_rem_flow(struct ice_hw *hw, enum ice_block blk,
316 		  enum ice_fltr_ptype flow_type)
317 {
318 	int flow = (int)flow_type & ~FLOW_EXT;
319 	struct ice_fd_hw_prof *prof;
320 	int tun, i;
321 
322 	prof = ice_fdir_get_hw_prof(hw, blk, flow);
323 	if (!prof)
324 		return;
325 
326 	ice_fdir_erase_flow_from_hw(hw, blk, flow);
327 	for (i = 0; i < prof->cnt; i++)
328 		prof->vsi_h[i] = 0;
329 	for (tun = 0; tun < ICE_FD_HW_SEG_MAX; tun++) {
330 		if (!prof->fdir_seg[tun])
331 			continue;
332 		devm_kfree(ice_hw_to_dev(hw), prof->fdir_seg[tun]);
333 		prof->fdir_seg[tun] = NULL;
334 	}
335 	prof->cnt = 0;
336 }
337 
338 /**
339  * ice_fdir_release_flows - release all flows in use for later replay
340  * @hw: pointer to HW instance
341  */
342 void ice_fdir_release_flows(struct ice_hw *hw)
343 {
344 	int flow;
345 
346 	/* release Flow Director HW table entries */
347 	for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++)
348 		ice_fdir_erase_flow_from_hw(hw, ICE_BLK_FD, flow);
349 }
350 
351 /**
352  * ice_fdir_replay_flows - replay HW Flow Director filter info
353  * @hw: pointer to HW instance
354  */
355 void ice_fdir_replay_flows(struct ice_hw *hw)
356 {
357 	int flow;
358 
359 	for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) {
360 		int tun;
361 
362 		if (!hw->fdir_prof[flow] || !hw->fdir_prof[flow]->cnt)
363 			continue;
364 		for (tun = 0; tun < ICE_FD_HW_SEG_MAX; tun++) {
365 			struct ice_flow_prof *hw_prof;
366 			struct ice_fd_hw_prof *prof;
367 			u64 prof_id;
368 			int j;
369 
370 			prof = hw->fdir_prof[flow];
371 			prof_id = flow + tun * ICE_FLTR_PTYPE_MAX;
372 			ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, prof_id,
373 					  prof->fdir_seg[tun], TNL_SEG_CNT(tun),
374 					  &hw_prof);
375 			for (j = 0; j < prof->cnt; j++) {
376 				enum ice_flow_priority prio;
377 				u64 entry_h = 0;
378 				int err;
379 
380 				prio = ICE_FLOW_PRIO_NORMAL;
381 				err = ice_flow_add_entry(hw, ICE_BLK_FD,
382 							 prof_id,
383 							 prof->vsi_h[0],
384 							 prof->vsi_h[j],
385 							 prio, prof->fdir_seg,
386 							 &entry_h);
387 				if (err) {
388 					dev_err(ice_hw_to_dev(hw), "Could not replay Flow Director, flow type %d\n",
389 						flow);
390 					continue;
391 				}
392 				prof->entry_h[j][tun] = entry_h;
393 			}
394 		}
395 	}
396 }
397 
398 /**
399  * ice_parse_rx_flow_user_data - deconstruct user-defined data
400  * @fsp: pointer to ethtool Rx flow specification
401  * @data: pointer to userdef data structure for storage
402  *
403  * Returns 0 on success, negative error value on failure
404  */
405 static int
406 ice_parse_rx_flow_user_data(struct ethtool_rx_flow_spec *fsp,
407 			    struct ice_rx_flow_userdef *data)
408 {
409 	u64 value, mask;
410 
411 	memset(data, 0, sizeof(*data));
412 	if (!(fsp->flow_type & FLOW_EXT))
413 		return 0;
414 
415 	value = be64_to_cpu(*((__force __be64 *)fsp->h_ext.data));
416 	mask = be64_to_cpu(*((__force __be64 *)fsp->m_ext.data));
417 	if (!mask)
418 		return 0;
419 
420 #define ICE_USERDEF_FLEX_WORD_M	GENMASK_ULL(15, 0)
421 #define ICE_USERDEF_FLEX_OFFS_S	16
422 #define ICE_USERDEF_FLEX_OFFS_M	GENMASK_ULL(31, ICE_USERDEF_FLEX_OFFS_S)
423 #define ICE_USERDEF_FLEX_FLTR_M	GENMASK_ULL(31, 0)
424 
425 	/* 0x1fe is the maximum value for offsets stored in the internal
426 	 * filtering tables.
427 	 */
428 #define ICE_USERDEF_FLEX_MAX_OFFS_VAL 0x1fe
429 
430 	if (!ice_is_mask_valid(mask, ICE_USERDEF_FLEX_FLTR_M) ||
431 	    value > ICE_USERDEF_FLEX_FLTR_M)
432 		return -EINVAL;
433 
434 	data->flex_word = value & ICE_USERDEF_FLEX_WORD_M;
435 	data->flex_offset = (value & ICE_USERDEF_FLEX_OFFS_M) >>
436 			     ICE_USERDEF_FLEX_OFFS_S;
437 	if (data->flex_offset > ICE_USERDEF_FLEX_MAX_OFFS_VAL)
438 		return -EINVAL;
439 
440 	data->flex_fltr = true;
441 
442 	return 0;
443 }
444 
445 /**
446  * ice_fdir_num_avail_fltr - return the number of unused flow director filters
447  * @hw: pointer to hardware structure
448  * @vsi: software VSI structure
449  *
450  * There are 2 filter pools: guaranteed and best effort(shared). Each VSI can
451  * use filters from either pool. The guaranteed pool is divided between VSIs.
452  * The best effort filter pool is common to all VSIs and is a device shared
453  * resource pool. The number of filters available to this VSI is the sum of
454  * the VSIs guaranteed filter pool and the global available best effort
455  * filter pool.
456  *
457  * Returns the number of available flow director filters to this VSI
458  */
459 static int ice_fdir_num_avail_fltr(struct ice_hw *hw, struct ice_vsi *vsi)
460 {
461 	u16 vsi_num = ice_get_hw_vsi_num(hw, vsi->idx);
462 	u16 num_guar;
463 	u16 num_be;
464 
465 	/* total guaranteed filters assigned to this VSI */
466 	num_guar = vsi->num_gfltr;
467 
468 	/* minus the guaranteed filters programed by this VSI */
469 	num_guar -= (rd32(hw, VSIQF_FD_CNT(vsi_num)) &
470 		     VSIQF_FD_CNT_FD_GCNT_M) >> VSIQF_FD_CNT_FD_GCNT_S;
471 
472 	/* total global best effort filters */
473 	num_be = hw->func_caps.fd_fltr_best_effort;
474 
475 	/* minus the global best effort filters programmed */
476 	num_be -= (rd32(hw, GLQF_FD_CNT) & GLQF_FD_CNT_FD_BCNT_M) >>
477 		   GLQF_FD_CNT_FD_BCNT_S;
478 
479 	return num_guar + num_be;
480 }
481 
482 /**
483  * ice_fdir_alloc_flow_prof - allocate FDir flow profile structure(s)
484  * @hw: HW structure containing the FDir flow profile structure(s)
485  * @flow: flow type to allocate the flow profile for
486  *
487  * Allocate the fdir_prof and fdir_prof[flow] if not already created. Return 0
488  * on success and negative on error.
489  */
490 static int
491 ice_fdir_alloc_flow_prof(struct ice_hw *hw, enum ice_fltr_ptype flow)
492 {
493 	if (!hw)
494 		return -EINVAL;
495 
496 	if (!hw->fdir_prof) {
497 		hw->fdir_prof = devm_kcalloc(ice_hw_to_dev(hw),
498 					     ICE_FLTR_PTYPE_MAX,
499 					     sizeof(*hw->fdir_prof),
500 					     GFP_KERNEL);
501 		if (!hw->fdir_prof)
502 			return -ENOMEM;
503 	}
504 
505 	if (!hw->fdir_prof[flow]) {
506 		hw->fdir_prof[flow] = devm_kzalloc(ice_hw_to_dev(hw),
507 						   sizeof(**hw->fdir_prof),
508 						   GFP_KERNEL);
509 		if (!hw->fdir_prof[flow])
510 			return -ENOMEM;
511 	}
512 
513 	return 0;
514 }
515 
516 /**
517  * ice_fdir_set_hw_fltr_rule - Configure HW tables to generate a FDir rule
518  * @pf: pointer to the PF structure
519  * @seg: protocol header description pointer
520  * @flow: filter enum
521  * @tun: FDir segment to program
522  */
523 static int
524 ice_fdir_set_hw_fltr_rule(struct ice_pf *pf, struct ice_flow_seg_info *seg,
525 			  enum ice_fltr_ptype flow, enum ice_fd_hw_seg tun)
526 {
527 	struct device *dev = ice_pf_to_dev(pf);
528 	struct ice_vsi *main_vsi, *ctrl_vsi;
529 	struct ice_flow_seg_info *old_seg;
530 	struct ice_flow_prof *prof = NULL;
531 	struct ice_fd_hw_prof *hw_prof;
532 	struct ice_hw *hw = &pf->hw;
533 	enum ice_status status;
534 	u64 entry1_h = 0;
535 	u64 entry2_h = 0;
536 	u64 prof_id;
537 	int err;
538 
539 	main_vsi = ice_get_main_vsi(pf);
540 	if (!main_vsi)
541 		return -EINVAL;
542 
543 	ctrl_vsi = ice_get_ctrl_vsi(pf);
544 	if (!ctrl_vsi)
545 		return -EINVAL;
546 
547 	err = ice_fdir_alloc_flow_prof(hw, flow);
548 	if (err)
549 		return err;
550 
551 	hw_prof = hw->fdir_prof[flow];
552 	old_seg = hw_prof->fdir_seg[tun];
553 	if (old_seg) {
554 		/* This flow_type already has a changed input set.
555 		 * If it matches the requested input set then we are
556 		 * done. Or, if it's different then it's an error.
557 		 */
558 		if (!memcmp(old_seg, seg, sizeof(*seg)))
559 			return -EEXIST;
560 
561 		/* if there are FDir filters using this flow,
562 		 * then return error.
563 		 */
564 		if (hw->fdir_fltr_cnt[flow]) {
565 			dev_err(dev, "Failed to add filter.  Flow director filters on each port must have the same input set.\n");
566 			return -EINVAL;
567 		}
568 
569 		if (ice_is_arfs_using_perfect_flow(hw, flow)) {
570 			dev_err(dev, "aRFS using perfect flow type %d, cannot change input set\n",
571 				flow);
572 			return -EINVAL;
573 		}
574 
575 		/* remove HW filter definition */
576 		ice_fdir_rem_flow(hw, ICE_BLK_FD, flow);
577 	}
578 
579 	/* Adding a profile, but there is only one header supported.
580 	 * That is the final parameters are 1 header (segment), no
581 	 * actions (NULL) and zero actions 0.
582 	 */
583 	prof_id = flow + tun * ICE_FLTR_PTYPE_MAX;
584 	status = ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, prof_id, seg,
585 				   TNL_SEG_CNT(tun), &prof);
586 	if (status)
587 		return ice_status_to_errno(status);
588 	status = ice_flow_add_entry(hw, ICE_BLK_FD, prof_id, main_vsi->idx,
589 				    main_vsi->idx, ICE_FLOW_PRIO_NORMAL,
590 				    seg, &entry1_h);
591 	if (status) {
592 		err = ice_status_to_errno(status);
593 		goto err_prof;
594 	}
595 	status = ice_flow_add_entry(hw, ICE_BLK_FD, prof_id, main_vsi->idx,
596 				    ctrl_vsi->idx, ICE_FLOW_PRIO_NORMAL,
597 				    seg, &entry2_h);
598 	if (status) {
599 		err = ice_status_to_errno(status);
600 		goto err_entry;
601 	}
602 
603 	hw_prof->fdir_seg[tun] = seg;
604 	hw_prof->entry_h[0][tun] = entry1_h;
605 	hw_prof->entry_h[1][tun] = entry2_h;
606 	hw_prof->vsi_h[0] = main_vsi->idx;
607 	hw_prof->vsi_h[1] = ctrl_vsi->idx;
608 	if (!hw_prof->cnt)
609 		hw_prof->cnt = 2;
610 
611 	return 0;
612 
613 err_entry:
614 	ice_rem_prof_id_flow(hw, ICE_BLK_FD,
615 			     ice_get_hw_vsi_num(hw, main_vsi->idx), prof_id);
616 	ice_flow_rem_entry(hw, ICE_BLK_FD, entry1_h);
617 err_prof:
618 	ice_flow_rem_prof(hw, ICE_BLK_FD, prof_id);
619 	dev_err(dev, "Failed to add filter.  Flow director filters on each port must have the same input set.\n");
620 
621 	return err;
622 }
623 
624 /**
625  * ice_set_init_fdir_seg
626  * @seg: flow segment for programming
627  * @l3_proto: ICE_FLOW_SEG_HDR_IPV4 or ICE_FLOW_SEG_HDR_IPV6
628  * @l4_proto: ICE_FLOW_SEG_HDR_TCP or ICE_FLOW_SEG_HDR_UDP
629  *
630  * Set the configuration for perfect filters to the provided flow segment for
631  * programming the HW filter. This is to be called only when initializing
632  * filters as this function it assumes no filters exist.
633  */
634 static int
635 ice_set_init_fdir_seg(struct ice_flow_seg_info *seg,
636 		      enum ice_flow_seg_hdr l3_proto,
637 		      enum ice_flow_seg_hdr l4_proto)
638 {
639 	enum ice_flow_field src_addr, dst_addr, src_port, dst_port;
640 
641 	if (!seg)
642 		return -EINVAL;
643 
644 	if (l3_proto == ICE_FLOW_SEG_HDR_IPV4) {
645 		src_addr = ICE_FLOW_FIELD_IDX_IPV4_SA;
646 		dst_addr = ICE_FLOW_FIELD_IDX_IPV4_DA;
647 	} else if (l3_proto == ICE_FLOW_SEG_HDR_IPV6) {
648 		src_addr = ICE_FLOW_FIELD_IDX_IPV6_SA;
649 		dst_addr = ICE_FLOW_FIELD_IDX_IPV6_DA;
650 	} else {
651 		return -EINVAL;
652 	}
653 
654 	if (l4_proto == ICE_FLOW_SEG_HDR_TCP) {
655 		src_port = ICE_FLOW_FIELD_IDX_TCP_SRC_PORT;
656 		dst_port = ICE_FLOW_FIELD_IDX_TCP_DST_PORT;
657 	} else if (l4_proto == ICE_FLOW_SEG_HDR_UDP) {
658 		src_port = ICE_FLOW_FIELD_IDX_UDP_SRC_PORT;
659 		dst_port = ICE_FLOW_FIELD_IDX_UDP_DST_PORT;
660 	} else {
661 		return -EINVAL;
662 	}
663 
664 	ICE_FLOW_SET_HDRS(seg, l3_proto | l4_proto);
665 
666 	/* IP source address */
667 	ice_flow_set_fld(seg, src_addr, ICE_FLOW_FLD_OFF_INVAL,
668 			 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL, false);
669 
670 	/* IP destination address */
671 	ice_flow_set_fld(seg, dst_addr, ICE_FLOW_FLD_OFF_INVAL,
672 			 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL, false);
673 
674 	/* Layer 4 source port */
675 	ice_flow_set_fld(seg, src_port, ICE_FLOW_FLD_OFF_INVAL,
676 			 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL, false);
677 
678 	/* Layer 4 destination port */
679 	ice_flow_set_fld(seg, dst_port, ICE_FLOW_FLD_OFF_INVAL,
680 			 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL, false);
681 
682 	return 0;
683 }
684 
685 /**
686  * ice_create_init_fdir_rule
687  * @pf: PF structure
688  * @flow: filter enum
689  *
690  * Return error value or 0 on success.
691  */
692 static int
693 ice_create_init_fdir_rule(struct ice_pf *pf, enum ice_fltr_ptype flow)
694 {
695 	struct ice_flow_seg_info *seg, *tun_seg;
696 	struct device *dev = ice_pf_to_dev(pf);
697 	struct ice_hw *hw = &pf->hw;
698 	int ret;
699 
700 	/* if there is already a filter rule for kind return -EINVAL */
701 	if (hw->fdir_prof && hw->fdir_prof[flow] &&
702 	    hw->fdir_prof[flow]->fdir_seg[0])
703 		return -EINVAL;
704 
705 	seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
706 	if (!seg)
707 		return -ENOMEM;
708 
709 	tun_seg = devm_kzalloc(dev, sizeof(*seg) * ICE_FD_HW_SEG_MAX,
710 			       GFP_KERNEL);
711 	if (!tun_seg) {
712 		devm_kfree(dev, seg);
713 		return -ENOMEM;
714 	}
715 
716 	if (flow == ICE_FLTR_PTYPE_NONF_IPV4_TCP)
717 		ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV4,
718 					    ICE_FLOW_SEG_HDR_TCP);
719 	else if (flow == ICE_FLTR_PTYPE_NONF_IPV4_UDP)
720 		ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV4,
721 					    ICE_FLOW_SEG_HDR_UDP);
722 	else if (flow == ICE_FLTR_PTYPE_NONF_IPV6_TCP)
723 		ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV6,
724 					    ICE_FLOW_SEG_HDR_TCP);
725 	else if (flow == ICE_FLTR_PTYPE_NONF_IPV6_UDP)
726 		ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV6,
727 					    ICE_FLOW_SEG_HDR_UDP);
728 	else
729 		ret = -EINVAL;
730 	if (ret)
731 		goto err_exit;
732 
733 	/* add filter for outer headers */
734 	ret = ice_fdir_set_hw_fltr_rule(pf, seg, flow, ICE_FD_HW_SEG_NON_TUN);
735 	if (ret)
736 		/* could not write filter, free memory */
737 		goto err_exit;
738 
739 	/* make tunneled filter HW entries if possible */
740 	memcpy(&tun_seg[1], seg, sizeof(*seg));
741 	ret = ice_fdir_set_hw_fltr_rule(pf, tun_seg, flow, ICE_FD_HW_SEG_TUN);
742 	if (ret)
743 		/* could not write tunnel filter, but outer header filter
744 		 * exists
745 		 */
746 		devm_kfree(dev, tun_seg);
747 
748 	set_bit(flow, hw->fdir_perfect_fltr);
749 	return ret;
750 err_exit:
751 	devm_kfree(dev, tun_seg);
752 	devm_kfree(dev, seg);
753 
754 	return -EOPNOTSUPP;
755 }
756 
757 /**
758  * ice_set_fdir_ip4_seg
759  * @seg: flow segment for programming
760  * @tcp_ip4_spec: mask data from ethtool
761  * @l4_proto: Layer 4 protocol to program
762  * @perfect_fltr: only valid on success; returns true if perfect filter,
763  *		  false if not
764  *
765  * Set the mask data into the flow segment to be used to program HW
766  * table based on provided L4 protocol for IPv4
767  */
768 static int
769 ice_set_fdir_ip4_seg(struct ice_flow_seg_info *seg,
770 		     struct ethtool_tcpip4_spec *tcp_ip4_spec,
771 		     enum ice_flow_seg_hdr l4_proto, bool *perfect_fltr)
772 {
773 	enum ice_flow_field src_port, dst_port;
774 
775 	/* make sure we don't have any empty rule */
776 	if (!tcp_ip4_spec->psrc && !tcp_ip4_spec->ip4src &&
777 	    !tcp_ip4_spec->pdst && !tcp_ip4_spec->ip4dst)
778 		return -EINVAL;
779 
780 	/* filtering on TOS not supported */
781 	if (tcp_ip4_spec->tos)
782 		return -EOPNOTSUPP;
783 
784 	if (l4_proto == ICE_FLOW_SEG_HDR_TCP) {
785 		src_port = ICE_FLOW_FIELD_IDX_TCP_SRC_PORT;
786 		dst_port = ICE_FLOW_FIELD_IDX_TCP_DST_PORT;
787 	} else if (l4_proto == ICE_FLOW_SEG_HDR_UDP) {
788 		src_port = ICE_FLOW_FIELD_IDX_UDP_SRC_PORT;
789 		dst_port = ICE_FLOW_FIELD_IDX_UDP_DST_PORT;
790 	} else if (l4_proto == ICE_FLOW_SEG_HDR_SCTP) {
791 		src_port = ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT;
792 		dst_port = ICE_FLOW_FIELD_IDX_SCTP_DST_PORT;
793 	} else {
794 		return -EOPNOTSUPP;
795 	}
796 
797 	*perfect_fltr = true;
798 	ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4 | l4_proto);
799 
800 	/* IP source address */
801 	if (tcp_ip4_spec->ip4src == htonl(0xFFFFFFFF))
802 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_SA,
803 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
804 				 ICE_FLOW_FLD_OFF_INVAL, false);
805 	else if (!tcp_ip4_spec->ip4src)
806 		*perfect_fltr = false;
807 	else
808 		return -EOPNOTSUPP;
809 
810 	/* IP destination address */
811 	if (tcp_ip4_spec->ip4dst == htonl(0xFFFFFFFF))
812 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_DA,
813 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
814 				 ICE_FLOW_FLD_OFF_INVAL, false);
815 	else if (!tcp_ip4_spec->ip4dst)
816 		*perfect_fltr = false;
817 	else
818 		return -EOPNOTSUPP;
819 
820 	/* Layer 4 source port */
821 	if (tcp_ip4_spec->psrc == htons(0xFFFF))
822 		ice_flow_set_fld(seg, src_port, ICE_FLOW_FLD_OFF_INVAL,
823 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
824 				 false);
825 	else if (!tcp_ip4_spec->psrc)
826 		*perfect_fltr = false;
827 	else
828 		return -EOPNOTSUPP;
829 
830 	/* Layer 4 destination port */
831 	if (tcp_ip4_spec->pdst == htons(0xFFFF))
832 		ice_flow_set_fld(seg, dst_port, ICE_FLOW_FLD_OFF_INVAL,
833 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
834 				 false);
835 	else if (!tcp_ip4_spec->pdst)
836 		*perfect_fltr = false;
837 	else
838 		return -EOPNOTSUPP;
839 
840 	return 0;
841 }
842 
843 /**
844  * ice_set_fdir_ip4_usr_seg
845  * @seg: flow segment for programming
846  * @usr_ip4_spec: ethtool userdef packet offset
847  * @perfect_fltr: only valid on success; returns true if perfect filter,
848  *		  false if not
849  *
850  * Set the offset data into the flow segment to be used to program HW
851  * table for IPv4
852  */
853 static int
854 ice_set_fdir_ip4_usr_seg(struct ice_flow_seg_info *seg,
855 			 struct ethtool_usrip4_spec *usr_ip4_spec,
856 			 bool *perfect_fltr)
857 {
858 	/* first 4 bytes of Layer 4 header */
859 	if (usr_ip4_spec->l4_4_bytes)
860 		return -EINVAL;
861 	if (usr_ip4_spec->tos)
862 		return -EINVAL;
863 	if (usr_ip4_spec->ip_ver)
864 		return -EINVAL;
865 	/* Filtering on Layer 4 protocol not supported */
866 	if (usr_ip4_spec->proto)
867 		return -EOPNOTSUPP;
868 	/* empty rules are not valid */
869 	if (!usr_ip4_spec->ip4src && !usr_ip4_spec->ip4dst)
870 		return -EINVAL;
871 
872 	*perfect_fltr = true;
873 	ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4);
874 
875 	/* IP source address */
876 	if (usr_ip4_spec->ip4src == htonl(0xFFFFFFFF))
877 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_SA,
878 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
879 				 ICE_FLOW_FLD_OFF_INVAL, false);
880 	else if (!usr_ip4_spec->ip4src)
881 		*perfect_fltr = false;
882 	else
883 		return -EOPNOTSUPP;
884 
885 	/* IP destination address */
886 	if (usr_ip4_spec->ip4dst == htonl(0xFFFFFFFF))
887 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_DA,
888 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
889 				 ICE_FLOW_FLD_OFF_INVAL, false);
890 	else if (!usr_ip4_spec->ip4dst)
891 		*perfect_fltr = false;
892 	else
893 		return -EOPNOTSUPP;
894 
895 	return 0;
896 }
897 
898 /**
899  * ice_set_fdir_ip6_seg
900  * @seg: flow segment for programming
901  * @tcp_ip6_spec: mask data from ethtool
902  * @l4_proto: Layer 4 protocol to program
903  * @perfect_fltr: only valid on success; returns true if perfect filter,
904  *		  false if not
905  *
906  * Set the mask data into the flow segment to be used to program HW
907  * table based on provided L4 protocol for IPv6
908  */
909 static int
910 ice_set_fdir_ip6_seg(struct ice_flow_seg_info *seg,
911 		     struct ethtool_tcpip6_spec *tcp_ip6_spec,
912 		     enum ice_flow_seg_hdr l4_proto, bool *perfect_fltr)
913 {
914 	enum ice_flow_field src_port, dst_port;
915 
916 	/* make sure we don't have any empty rule */
917 	if (!memcmp(tcp_ip6_spec->ip6src, &zero_ipv6_addr_mask,
918 		    sizeof(struct in6_addr)) &&
919 	    !memcmp(tcp_ip6_spec->ip6dst, &zero_ipv6_addr_mask,
920 		    sizeof(struct in6_addr)) &&
921 	    !tcp_ip6_spec->psrc && !tcp_ip6_spec->pdst)
922 		return -EINVAL;
923 
924 	/* filtering on TC not supported */
925 	if (tcp_ip6_spec->tclass)
926 		return -EOPNOTSUPP;
927 
928 	if (l4_proto == ICE_FLOW_SEG_HDR_TCP) {
929 		src_port = ICE_FLOW_FIELD_IDX_TCP_SRC_PORT;
930 		dst_port = ICE_FLOW_FIELD_IDX_TCP_DST_PORT;
931 	} else if (l4_proto == ICE_FLOW_SEG_HDR_UDP) {
932 		src_port = ICE_FLOW_FIELD_IDX_UDP_SRC_PORT;
933 		dst_port = ICE_FLOW_FIELD_IDX_UDP_DST_PORT;
934 	} else if (l4_proto == ICE_FLOW_SEG_HDR_SCTP) {
935 		src_port = ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT;
936 		dst_port = ICE_FLOW_FIELD_IDX_SCTP_DST_PORT;
937 	} else {
938 		return -EINVAL;
939 	}
940 
941 	*perfect_fltr = true;
942 	ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6 | l4_proto);
943 
944 	if (!memcmp(tcp_ip6_spec->ip6src, &full_ipv6_addr_mask,
945 		    sizeof(struct in6_addr)))
946 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_SA,
947 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
948 				 ICE_FLOW_FLD_OFF_INVAL, false);
949 	else if (!memcmp(tcp_ip6_spec->ip6src, &zero_ipv6_addr_mask,
950 			 sizeof(struct in6_addr)))
951 		*perfect_fltr = false;
952 	else
953 		return -EOPNOTSUPP;
954 
955 	if (!memcmp(tcp_ip6_spec->ip6dst, &full_ipv6_addr_mask,
956 		    sizeof(struct in6_addr)))
957 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_DA,
958 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
959 				 ICE_FLOW_FLD_OFF_INVAL, false);
960 	else if (!memcmp(tcp_ip6_spec->ip6dst, &zero_ipv6_addr_mask,
961 			 sizeof(struct in6_addr)))
962 		*perfect_fltr = false;
963 	else
964 		return -EOPNOTSUPP;
965 
966 	/* Layer 4 source port */
967 	if (tcp_ip6_spec->psrc == htons(0xFFFF))
968 		ice_flow_set_fld(seg, src_port, ICE_FLOW_FLD_OFF_INVAL,
969 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
970 				 false);
971 	else if (!tcp_ip6_spec->psrc)
972 		*perfect_fltr = false;
973 	else
974 		return -EOPNOTSUPP;
975 
976 	/* Layer 4 destination port */
977 	if (tcp_ip6_spec->pdst == htons(0xFFFF))
978 		ice_flow_set_fld(seg, dst_port, ICE_FLOW_FLD_OFF_INVAL,
979 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
980 				 false);
981 	else if (!tcp_ip6_spec->pdst)
982 		*perfect_fltr = false;
983 	else
984 		return -EOPNOTSUPP;
985 
986 	return 0;
987 }
988 
989 /**
990  * ice_set_fdir_ip6_usr_seg
991  * @seg: flow segment for programming
992  * @usr_ip6_spec: ethtool userdef packet offset
993  * @perfect_fltr: only valid on success; returns true if perfect filter,
994  *		  false if not
995  *
996  * Set the offset data into the flow segment to be used to program HW
997  * table for IPv6
998  */
999 static int
1000 ice_set_fdir_ip6_usr_seg(struct ice_flow_seg_info *seg,
1001 			 struct ethtool_usrip6_spec *usr_ip6_spec,
1002 			 bool *perfect_fltr)
1003 {
1004 	/* filtering on Layer 4 bytes not supported */
1005 	if (usr_ip6_spec->l4_4_bytes)
1006 		return -EOPNOTSUPP;
1007 	/* filtering on TC not supported */
1008 	if (usr_ip6_spec->tclass)
1009 		return -EOPNOTSUPP;
1010 	/* filtering on Layer 4 protocol not supported */
1011 	if (usr_ip6_spec->l4_proto)
1012 		return -EOPNOTSUPP;
1013 	/* empty rules are not valid */
1014 	if (!memcmp(usr_ip6_spec->ip6src, &zero_ipv6_addr_mask,
1015 		    sizeof(struct in6_addr)) &&
1016 	    !memcmp(usr_ip6_spec->ip6dst, &zero_ipv6_addr_mask,
1017 		    sizeof(struct in6_addr)))
1018 		return -EINVAL;
1019 
1020 	*perfect_fltr = true;
1021 	ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6);
1022 
1023 	if (!memcmp(usr_ip6_spec->ip6src, &full_ipv6_addr_mask,
1024 		    sizeof(struct in6_addr)))
1025 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_SA,
1026 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
1027 				 ICE_FLOW_FLD_OFF_INVAL, false);
1028 	else if (!memcmp(usr_ip6_spec->ip6src, &zero_ipv6_addr_mask,
1029 			 sizeof(struct in6_addr)))
1030 		*perfect_fltr = false;
1031 	else
1032 		return -EOPNOTSUPP;
1033 
1034 	if (!memcmp(usr_ip6_spec->ip6dst, &full_ipv6_addr_mask,
1035 		    sizeof(struct in6_addr)))
1036 		ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_DA,
1037 				 ICE_FLOW_FLD_OFF_INVAL, ICE_FLOW_FLD_OFF_INVAL,
1038 				 ICE_FLOW_FLD_OFF_INVAL, false);
1039 	else if (!memcmp(usr_ip6_spec->ip6dst, &zero_ipv6_addr_mask,
1040 			 sizeof(struct in6_addr)))
1041 		*perfect_fltr = false;
1042 	else
1043 		return -EOPNOTSUPP;
1044 
1045 	return 0;
1046 }
1047 
1048 /**
1049  * ice_cfg_fdir_xtrct_seq - Configure extraction sequence for the given filter
1050  * @pf: PF structure
1051  * @fsp: pointer to ethtool Rx flow specification
1052  * @user: user defined data from flow specification
1053  *
1054  * Returns 0 on success.
1055  */
1056 static int
1057 ice_cfg_fdir_xtrct_seq(struct ice_pf *pf, struct ethtool_rx_flow_spec *fsp,
1058 		       struct ice_rx_flow_userdef *user)
1059 {
1060 	struct ice_flow_seg_info *seg, *tun_seg;
1061 	struct device *dev = ice_pf_to_dev(pf);
1062 	enum ice_fltr_ptype fltr_idx;
1063 	struct ice_hw *hw = &pf->hw;
1064 	bool perfect_filter;
1065 	int ret;
1066 
1067 	seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
1068 	if (!seg)
1069 		return -ENOMEM;
1070 
1071 	tun_seg = devm_kzalloc(dev, sizeof(*seg) * ICE_FD_HW_SEG_MAX,
1072 			       GFP_KERNEL);
1073 	if (!tun_seg) {
1074 		devm_kfree(dev, seg);
1075 		return -ENOMEM;
1076 	}
1077 
1078 	switch (fsp->flow_type & ~FLOW_EXT) {
1079 	case TCP_V4_FLOW:
1080 		ret = ice_set_fdir_ip4_seg(seg, &fsp->m_u.tcp_ip4_spec,
1081 					   ICE_FLOW_SEG_HDR_TCP,
1082 					   &perfect_filter);
1083 		break;
1084 	case UDP_V4_FLOW:
1085 		ret = ice_set_fdir_ip4_seg(seg, &fsp->m_u.tcp_ip4_spec,
1086 					   ICE_FLOW_SEG_HDR_UDP,
1087 					   &perfect_filter);
1088 		break;
1089 	case SCTP_V4_FLOW:
1090 		ret = ice_set_fdir_ip4_seg(seg, &fsp->m_u.tcp_ip4_spec,
1091 					   ICE_FLOW_SEG_HDR_SCTP,
1092 					   &perfect_filter);
1093 		break;
1094 	case IPV4_USER_FLOW:
1095 		ret = ice_set_fdir_ip4_usr_seg(seg, &fsp->m_u.usr_ip4_spec,
1096 					       &perfect_filter);
1097 		break;
1098 	case TCP_V6_FLOW:
1099 		ret = ice_set_fdir_ip6_seg(seg, &fsp->m_u.tcp_ip6_spec,
1100 					   ICE_FLOW_SEG_HDR_TCP,
1101 					   &perfect_filter);
1102 		break;
1103 	case UDP_V6_FLOW:
1104 		ret = ice_set_fdir_ip6_seg(seg, &fsp->m_u.tcp_ip6_spec,
1105 					   ICE_FLOW_SEG_HDR_UDP,
1106 					   &perfect_filter);
1107 		break;
1108 	case SCTP_V6_FLOW:
1109 		ret = ice_set_fdir_ip6_seg(seg, &fsp->m_u.tcp_ip6_spec,
1110 					   ICE_FLOW_SEG_HDR_SCTP,
1111 					   &perfect_filter);
1112 		break;
1113 	case IPV6_USER_FLOW:
1114 		ret = ice_set_fdir_ip6_usr_seg(seg, &fsp->m_u.usr_ip6_spec,
1115 					       &perfect_filter);
1116 		break;
1117 	default:
1118 		ret = -EINVAL;
1119 	}
1120 	if (ret)
1121 		goto err_exit;
1122 
1123 	/* tunnel segments are shifted up one. */
1124 	memcpy(&tun_seg[1], seg, sizeof(*seg));
1125 
1126 	if (user && user->flex_fltr) {
1127 		perfect_filter = false;
1128 		ice_flow_add_fld_raw(seg, user->flex_offset,
1129 				     ICE_FLTR_PRGM_FLEX_WORD_SIZE,
1130 				     ICE_FLOW_FLD_OFF_INVAL,
1131 				     ICE_FLOW_FLD_OFF_INVAL);
1132 		ice_flow_add_fld_raw(&tun_seg[1], user->flex_offset,
1133 				     ICE_FLTR_PRGM_FLEX_WORD_SIZE,
1134 				     ICE_FLOW_FLD_OFF_INVAL,
1135 				     ICE_FLOW_FLD_OFF_INVAL);
1136 	}
1137 
1138 	/* add filter for outer headers */
1139 	fltr_idx = ice_ethtool_flow_to_fltr(fsp->flow_type & ~FLOW_EXT);
1140 	ret = ice_fdir_set_hw_fltr_rule(pf, seg, fltr_idx,
1141 					ICE_FD_HW_SEG_NON_TUN);
1142 	if (ret == -EEXIST)
1143 		/* Rule already exists, free memory and continue */
1144 		devm_kfree(dev, seg);
1145 	else if (ret)
1146 		/* could not write filter, free memory */
1147 		goto err_exit;
1148 
1149 	/* make tunneled filter HW entries if possible */
1150 	memcpy(&tun_seg[1], seg, sizeof(*seg));
1151 	ret = ice_fdir_set_hw_fltr_rule(pf, tun_seg, fltr_idx,
1152 					ICE_FD_HW_SEG_TUN);
1153 	if (ret == -EEXIST) {
1154 		/* Rule already exists, free memory and count as success */
1155 		devm_kfree(dev, tun_seg);
1156 		ret = 0;
1157 	} else if (ret) {
1158 		/* could not write tunnel filter, but outer filter exists */
1159 		devm_kfree(dev, tun_seg);
1160 	}
1161 
1162 	if (perfect_filter)
1163 		set_bit(fltr_idx, hw->fdir_perfect_fltr);
1164 	else
1165 		clear_bit(fltr_idx, hw->fdir_perfect_fltr);
1166 
1167 	return ret;
1168 
1169 err_exit:
1170 	devm_kfree(dev, tun_seg);
1171 	devm_kfree(dev, seg);
1172 
1173 	return -EOPNOTSUPP;
1174 }
1175 
1176 /**
1177  * ice_fdir_write_fltr - send a flow director filter to the hardware
1178  * @pf: PF data structure
1179  * @input: filter structure
1180  * @add: true adds filter and false removed filter
1181  * @is_tun: true adds inner filter on tunnel and false outer headers
1182  *
1183  * returns 0 on success and negative value on error
1184  */
1185 int
1186 ice_fdir_write_fltr(struct ice_pf *pf, struct ice_fdir_fltr *input, bool add,
1187 		    bool is_tun)
1188 {
1189 	struct device *dev = ice_pf_to_dev(pf);
1190 	struct ice_hw *hw = &pf->hw;
1191 	struct ice_fltr_desc desc;
1192 	struct ice_vsi *ctrl_vsi;
1193 	enum ice_status status;
1194 	u8 *pkt, *frag_pkt;
1195 	bool has_frag;
1196 	int err;
1197 
1198 	ctrl_vsi = ice_get_ctrl_vsi(pf);
1199 	if (!ctrl_vsi)
1200 		return -EINVAL;
1201 
1202 	pkt = devm_kzalloc(dev, ICE_FDIR_MAX_RAW_PKT_SIZE, GFP_KERNEL);
1203 	if (!pkt)
1204 		return -ENOMEM;
1205 	frag_pkt = devm_kzalloc(dev, ICE_FDIR_MAX_RAW_PKT_SIZE, GFP_KERNEL);
1206 	if (!frag_pkt) {
1207 		err = -ENOMEM;
1208 		goto err_free;
1209 	}
1210 
1211 	ice_fdir_get_prgm_desc(hw, input, &desc, add);
1212 	status = ice_fdir_get_gen_prgm_pkt(hw, input, pkt, false, is_tun);
1213 	if (status) {
1214 		err = ice_status_to_errno(status);
1215 		goto err_free_all;
1216 	}
1217 	err = ice_prgm_fdir_fltr(ctrl_vsi, &desc, pkt);
1218 	if (err)
1219 		goto err_free_all;
1220 
1221 	/* repeat for fragment packet */
1222 	has_frag = ice_fdir_has_frag(input->flow_type);
1223 	if (has_frag) {
1224 		/* does not return error */
1225 		ice_fdir_get_prgm_desc(hw, input, &desc, add);
1226 		status = ice_fdir_get_gen_prgm_pkt(hw, input, frag_pkt, true,
1227 						   is_tun);
1228 		if (status) {
1229 			err = ice_status_to_errno(status);
1230 			goto err_frag;
1231 		}
1232 		err = ice_prgm_fdir_fltr(ctrl_vsi, &desc, frag_pkt);
1233 		if (err)
1234 			goto err_frag;
1235 	} else {
1236 		devm_kfree(dev, frag_pkt);
1237 	}
1238 
1239 	return 0;
1240 
1241 err_free_all:
1242 	devm_kfree(dev, frag_pkt);
1243 err_free:
1244 	devm_kfree(dev, pkt);
1245 	return err;
1246 
1247 err_frag:
1248 	devm_kfree(dev, frag_pkt);
1249 	return err;
1250 }
1251 
1252 /**
1253  * ice_fdir_write_all_fltr - send a flow director filter to the hardware
1254  * @pf: PF data structure
1255  * @input: filter structure
1256  * @add: true adds filter and false removed filter
1257  *
1258  * returns 0 on success and negative value on error
1259  */
1260 static int
1261 ice_fdir_write_all_fltr(struct ice_pf *pf, struct ice_fdir_fltr *input,
1262 			bool add)
1263 {
1264 	u16 port_num;
1265 	int tun;
1266 
1267 	for (tun = 0; tun < ICE_FD_HW_SEG_MAX; tun++) {
1268 		bool is_tun = tun == ICE_FD_HW_SEG_TUN;
1269 		int err;
1270 
1271 		if (is_tun && !ice_get_open_tunnel_port(&pf->hw, TNL_ALL,
1272 							&port_num))
1273 			continue;
1274 		err = ice_fdir_write_fltr(pf, input, add, is_tun);
1275 		if (err)
1276 			return err;
1277 	}
1278 	return 0;
1279 }
1280 
1281 /**
1282  * ice_fdir_replay_fltrs - replay filters from the HW filter list
1283  * @pf: board private structure
1284  */
1285 void ice_fdir_replay_fltrs(struct ice_pf *pf)
1286 {
1287 	struct ice_fdir_fltr *f_rule;
1288 	struct ice_hw *hw = &pf->hw;
1289 
1290 	list_for_each_entry(f_rule, &hw->fdir_list_head, fltr_node) {
1291 		int err = ice_fdir_write_all_fltr(pf, f_rule, true);
1292 
1293 		if (err)
1294 			dev_dbg(ice_pf_to_dev(pf), "Flow Director error %d, could not reprogram filter %d\n",
1295 				err, f_rule->fltr_id);
1296 	}
1297 }
1298 
1299 /**
1300  * ice_fdir_create_dflt_rules - create default perfect filters
1301  * @pf: PF data structure
1302  *
1303  * Returns 0 for success or error.
1304  */
1305 int ice_fdir_create_dflt_rules(struct ice_pf *pf)
1306 {
1307 	int err;
1308 
1309 	/* Create perfect TCP and UDP rules in hardware. */
1310 	err = ice_create_init_fdir_rule(pf, ICE_FLTR_PTYPE_NONF_IPV4_TCP);
1311 	if (err)
1312 		return err;
1313 
1314 	err = ice_create_init_fdir_rule(pf, ICE_FLTR_PTYPE_NONF_IPV4_UDP);
1315 	if (err)
1316 		return err;
1317 
1318 	err = ice_create_init_fdir_rule(pf, ICE_FLTR_PTYPE_NONF_IPV6_TCP);
1319 	if (err)
1320 		return err;
1321 
1322 	err = ice_create_init_fdir_rule(pf, ICE_FLTR_PTYPE_NONF_IPV6_UDP);
1323 
1324 	return err;
1325 }
1326 
1327 /**
1328  * ice_vsi_manage_fdir - turn on/off flow director
1329  * @vsi: the VSI being changed
1330  * @ena: boolean value indicating if this is an enable or disable request
1331  */
1332 void ice_vsi_manage_fdir(struct ice_vsi *vsi, bool ena)
1333 {
1334 	struct ice_fdir_fltr *f_rule, *tmp;
1335 	struct ice_pf *pf = vsi->back;
1336 	struct ice_hw *hw = &pf->hw;
1337 	enum ice_fltr_ptype flow;
1338 
1339 	if (ena) {
1340 		set_bit(ICE_FLAG_FD_ENA, pf->flags);
1341 		ice_fdir_create_dflt_rules(pf);
1342 		return;
1343 	}
1344 
1345 	mutex_lock(&hw->fdir_fltr_lock);
1346 	if (!test_and_clear_bit(ICE_FLAG_FD_ENA, pf->flags))
1347 		goto release_lock;
1348 	list_for_each_entry_safe(f_rule, tmp, &hw->fdir_list_head, fltr_node) {
1349 		/* ignore return value */
1350 		ice_fdir_write_all_fltr(pf, f_rule, false);
1351 		ice_fdir_update_cntrs(hw, f_rule->flow_type, false);
1352 		list_del(&f_rule->fltr_node);
1353 		devm_kfree(ice_hw_to_dev(hw), f_rule);
1354 	}
1355 
1356 	if (hw->fdir_prof)
1357 		for (flow = ICE_FLTR_PTYPE_NONF_NONE; flow < ICE_FLTR_PTYPE_MAX;
1358 		     flow++)
1359 			if (hw->fdir_prof[flow])
1360 				ice_fdir_rem_flow(hw, ICE_BLK_FD, flow);
1361 
1362 release_lock:
1363 	mutex_unlock(&hw->fdir_fltr_lock);
1364 }
1365 
1366 /**
1367  * ice_fdir_do_rem_flow - delete flow and possibly add perfect flow
1368  * @pf: PF structure
1369  * @flow_type: FDir flow type to release
1370  */
1371 static void
1372 ice_fdir_do_rem_flow(struct ice_pf *pf, enum ice_fltr_ptype flow_type)
1373 {
1374 	struct ice_hw *hw = &pf->hw;
1375 	bool need_perfect = false;
1376 
1377 	if (flow_type == ICE_FLTR_PTYPE_NONF_IPV4_TCP ||
1378 	    flow_type == ICE_FLTR_PTYPE_NONF_IPV4_UDP ||
1379 	    flow_type == ICE_FLTR_PTYPE_NONF_IPV6_TCP ||
1380 	    flow_type == ICE_FLTR_PTYPE_NONF_IPV6_UDP)
1381 		need_perfect = true;
1382 
1383 	if (need_perfect && test_bit(flow_type, hw->fdir_perfect_fltr))
1384 		return;
1385 
1386 	ice_fdir_rem_flow(hw, ICE_BLK_FD, flow_type);
1387 	if (need_perfect)
1388 		ice_create_init_fdir_rule(pf, flow_type);
1389 }
1390 
1391 /**
1392  * ice_fdir_update_list_entry - add or delete a filter from the filter list
1393  * @pf: PF structure
1394  * @input: filter structure
1395  * @fltr_idx: ethtool index of filter to modify
1396  *
1397  * returns 0 on success and negative on errors
1398  */
1399 static int
1400 ice_fdir_update_list_entry(struct ice_pf *pf, struct ice_fdir_fltr *input,
1401 			   int fltr_idx)
1402 {
1403 	struct ice_fdir_fltr *old_fltr;
1404 	struct ice_hw *hw = &pf->hw;
1405 	int err = -ENOENT;
1406 
1407 	/* Do not update filters during reset */
1408 	if (ice_is_reset_in_progress(pf->state))
1409 		return -EBUSY;
1410 
1411 	old_fltr = ice_fdir_find_fltr_by_idx(hw, fltr_idx);
1412 	if (old_fltr) {
1413 		err = ice_fdir_write_all_fltr(pf, old_fltr, false);
1414 		if (err)
1415 			return err;
1416 		ice_fdir_update_cntrs(hw, old_fltr->flow_type, false);
1417 		if (!input && !hw->fdir_fltr_cnt[old_fltr->flow_type])
1418 			/* we just deleted the last filter of flow_type so we
1419 			 * should also delete the HW filter info.
1420 			 */
1421 			ice_fdir_do_rem_flow(pf, old_fltr->flow_type);
1422 		list_del(&old_fltr->fltr_node);
1423 		devm_kfree(ice_hw_to_dev(hw), old_fltr);
1424 	}
1425 	if (!input)
1426 		return err;
1427 	ice_fdir_list_add_fltr(hw, input);
1428 	ice_fdir_update_cntrs(hw, input->flow_type, true);
1429 	return 0;
1430 }
1431 
1432 /**
1433  * ice_del_fdir_ethtool - delete Flow Director filter
1434  * @vsi: pointer to target VSI
1435  * @cmd: command to add or delete Flow Director filter
1436  *
1437  * Returns 0 on success and negative values for failure
1438  */
1439 int ice_del_fdir_ethtool(struct ice_vsi *vsi, struct ethtool_rxnfc *cmd)
1440 {
1441 	struct ethtool_rx_flow_spec *fsp =
1442 		(struct ethtool_rx_flow_spec *)&cmd->fs;
1443 	struct ice_pf *pf = vsi->back;
1444 	struct ice_hw *hw = &pf->hw;
1445 	int val;
1446 
1447 	if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
1448 		return -EOPNOTSUPP;
1449 
1450 	/* Do not delete filters during reset */
1451 	if (ice_is_reset_in_progress(pf->state)) {
1452 		dev_err(ice_pf_to_dev(pf), "Device is resetting - deleting Flow Director filters not supported during reset\n");
1453 		return -EBUSY;
1454 	}
1455 
1456 	if (test_bit(__ICE_FD_FLUSH_REQ, pf->state))
1457 		return -EBUSY;
1458 
1459 	mutex_lock(&hw->fdir_fltr_lock);
1460 	val = ice_fdir_update_list_entry(pf, NULL, fsp->location);
1461 	mutex_unlock(&hw->fdir_fltr_lock);
1462 
1463 	return val;
1464 }
1465 
1466 /**
1467  * ice_set_fdir_input_set - Set the input set for Flow Director
1468  * @vsi: pointer to target VSI
1469  * @fsp: pointer to ethtool Rx flow specification
1470  * @input: filter structure
1471  */
1472 static int
1473 ice_set_fdir_input_set(struct ice_vsi *vsi, struct ethtool_rx_flow_spec *fsp,
1474 		       struct ice_fdir_fltr *input)
1475 {
1476 	u16 dest_vsi, q_index = 0;
1477 	struct ice_pf *pf;
1478 	struct ice_hw *hw;
1479 	int flow_type;
1480 	u8 dest_ctl;
1481 
1482 	if (!vsi || !fsp || !input)
1483 		return -EINVAL;
1484 
1485 	pf = vsi->back;
1486 	hw = &pf->hw;
1487 
1488 	dest_vsi = vsi->idx;
1489 	if (fsp->ring_cookie == RX_CLS_FLOW_DISC) {
1490 		dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DROP_PKT;
1491 	} else {
1492 		u32 ring = ethtool_get_flow_spec_ring(fsp->ring_cookie);
1493 		u8 vf = ethtool_get_flow_spec_ring_vf(fsp->ring_cookie);
1494 
1495 		if (vf) {
1496 			dev_err(ice_pf_to_dev(pf), "Failed to add filter. Flow director filters are not supported on VF queues.\n");
1497 			return -EINVAL;
1498 		}
1499 
1500 		if (ring >= vsi->num_rxq)
1501 			return -EINVAL;
1502 
1503 		dest_ctl = ICE_FLTR_PRGM_DESC_DEST_DIRECT_PKT_QINDEX;
1504 		q_index = ring;
1505 	}
1506 
1507 	input->fltr_id = fsp->location;
1508 	input->q_index = q_index;
1509 	flow_type = fsp->flow_type & ~FLOW_EXT;
1510 
1511 	input->dest_vsi = dest_vsi;
1512 	input->dest_ctl = dest_ctl;
1513 	input->fltr_status = ICE_FLTR_PRGM_DESC_FD_STATUS_FD_ID;
1514 	input->cnt_index = ICE_FD_SB_STAT_IDX(hw->fd_ctr_base);
1515 	input->flow_type = ice_ethtool_flow_to_fltr(flow_type);
1516 
1517 	if (fsp->flow_type & FLOW_EXT) {
1518 		memcpy(input->ext_data.usr_def, fsp->h_ext.data,
1519 		       sizeof(input->ext_data.usr_def));
1520 		input->ext_data.vlan_type = fsp->h_ext.vlan_etype;
1521 		input->ext_data.vlan_tag = fsp->h_ext.vlan_tci;
1522 		memcpy(input->ext_mask.usr_def, fsp->m_ext.data,
1523 		       sizeof(input->ext_mask.usr_def));
1524 		input->ext_mask.vlan_type = fsp->m_ext.vlan_etype;
1525 		input->ext_mask.vlan_tag = fsp->m_ext.vlan_tci;
1526 	}
1527 
1528 	switch (flow_type) {
1529 	case TCP_V4_FLOW:
1530 	case UDP_V4_FLOW:
1531 	case SCTP_V4_FLOW:
1532 		input->ip.v4.dst_port = fsp->h_u.tcp_ip4_spec.pdst;
1533 		input->ip.v4.src_port = fsp->h_u.tcp_ip4_spec.psrc;
1534 		input->ip.v4.dst_ip = fsp->h_u.tcp_ip4_spec.ip4dst;
1535 		input->ip.v4.src_ip = fsp->h_u.tcp_ip4_spec.ip4src;
1536 		input->mask.v4.dst_port = fsp->m_u.tcp_ip4_spec.pdst;
1537 		input->mask.v4.src_port = fsp->m_u.tcp_ip4_spec.psrc;
1538 		input->mask.v4.dst_ip = fsp->m_u.tcp_ip4_spec.ip4dst;
1539 		input->mask.v4.src_ip = fsp->m_u.tcp_ip4_spec.ip4src;
1540 		break;
1541 	case IPV4_USER_FLOW:
1542 		input->ip.v4.dst_ip = fsp->h_u.usr_ip4_spec.ip4dst;
1543 		input->ip.v4.src_ip = fsp->h_u.usr_ip4_spec.ip4src;
1544 		input->ip.v4.l4_header = fsp->h_u.usr_ip4_spec.l4_4_bytes;
1545 		input->ip.v4.proto = fsp->h_u.usr_ip4_spec.proto;
1546 		input->ip.v4.ip_ver = fsp->h_u.usr_ip4_spec.ip_ver;
1547 		input->ip.v4.tos = fsp->h_u.usr_ip4_spec.tos;
1548 		input->mask.v4.dst_ip = fsp->m_u.usr_ip4_spec.ip4dst;
1549 		input->mask.v4.src_ip = fsp->m_u.usr_ip4_spec.ip4src;
1550 		input->mask.v4.l4_header = fsp->m_u.usr_ip4_spec.l4_4_bytes;
1551 		input->mask.v4.proto = fsp->m_u.usr_ip4_spec.proto;
1552 		input->mask.v4.ip_ver = fsp->m_u.usr_ip4_spec.ip_ver;
1553 		input->mask.v4.tos = fsp->m_u.usr_ip4_spec.tos;
1554 		break;
1555 	case TCP_V6_FLOW:
1556 	case UDP_V6_FLOW:
1557 	case SCTP_V6_FLOW:
1558 		memcpy(input->ip.v6.dst_ip, fsp->h_u.usr_ip6_spec.ip6dst,
1559 		       sizeof(struct in6_addr));
1560 		memcpy(input->ip.v6.src_ip, fsp->h_u.usr_ip6_spec.ip6src,
1561 		       sizeof(struct in6_addr));
1562 		input->ip.v6.dst_port = fsp->h_u.tcp_ip6_spec.pdst;
1563 		input->ip.v6.src_port = fsp->h_u.tcp_ip6_spec.psrc;
1564 		input->ip.v6.tc = fsp->h_u.tcp_ip6_spec.tclass;
1565 		memcpy(input->mask.v6.dst_ip, fsp->m_u.tcp_ip6_spec.ip6dst,
1566 		       sizeof(struct in6_addr));
1567 		memcpy(input->mask.v6.src_ip, fsp->m_u.tcp_ip6_spec.ip6src,
1568 		       sizeof(struct in6_addr));
1569 		input->mask.v6.dst_port = fsp->m_u.tcp_ip6_spec.pdst;
1570 		input->mask.v6.src_port = fsp->m_u.tcp_ip6_spec.psrc;
1571 		input->mask.v6.tc = fsp->m_u.tcp_ip6_spec.tclass;
1572 		break;
1573 	case IPV6_USER_FLOW:
1574 		memcpy(input->ip.v6.dst_ip, fsp->h_u.usr_ip6_spec.ip6dst,
1575 		       sizeof(struct in6_addr));
1576 		memcpy(input->ip.v6.src_ip, fsp->h_u.usr_ip6_spec.ip6src,
1577 		       sizeof(struct in6_addr));
1578 		input->ip.v6.l4_header = fsp->h_u.usr_ip6_spec.l4_4_bytes;
1579 		input->ip.v6.tc = fsp->h_u.usr_ip6_spec.tclass;
1580 		input->ip.v6.proto = fsp->h_u.usr_ip6_spec.l4_proto;
1581 		memcpy(input->mask.v6.dst_ip, fsp->m_u.usr_ip6_spec.ip6dst,
1582 		       sizeof(struct in6_addr));
1583 		memcpy(input->mask.v6.src_ip, fsp->m_u.usr_ip6_spec.ip6src,
1584 		       sizeof(struct in6_addr));
1585 		input->mask.v6.l4_header = fsp->m_u.usr_ip6_spec.l4_4_bytes;
1586 		input->mask.v6.tc = fsp->m_u.usr_ip6_spec.tclass;
1587 		input->mask.v6.proto = fsp->m_u.usr_ip6_spec.l4_proto;
1588 		break;
1589 	default:
1590 		/* not doing un-parsed flow types */
1591 		return -EINVAL;
1592 	}
1593 
1594 	return 0;
1595 }
1596 
1597 /**
1598  * ice_add_fdir_ethtool - Add/Remove Flow Director filter
1599  * @vsi: pointer to target VSI
1600  * @cmd: command to add or delete Flow Director filter
1601  *
1602  * Returns 0 on success and negative values for failure
1603  */
1604 int ice_add_fdir_ethtool(struct ice_vsi *vsi, struct ethtool_rxnfc *cmd)
1605 {
1606 	struct ice_rx_flow_userdef userdata;
1607 	struct ethtool_rx_flow_spec *fsp;
1608 	struct ice_fdir_fltr *input;
1609 	struct device *dev;
1610 	struct ice_pf *pf;
1611 	struct ice_hw *hw;
1612 	int fltrs_needed;
1613 	u16 tunnel_port;
1614 	int ret;
1615 
1616 	if (!vsi)
1617 		return -EINVAL;
1618 
1619 	pf = vsi->back;
1620 	hw = &pf->hw;
1621 	dev = ice_pf_to_dev(pf);
1622 
1623 	if (!test_bit(ICE_FLAG_FD_ENA, pf->flags))
1624 		return -EOPNOTSUPP;
1625 
1626 	/* Do not program filters during reset */
1627 	if (ice_is_reset_in_progress(pf->state)) {
1628 		dev_err(dev, "Device is resetting - adding Flow Director filters not supported during reset\n");
1629 		return -EBUSY;
1630 	}
1631 
1632 	fsp = (struct ethtool_rx_flow_spec *)&cmd->fs;
1633 
1634 	if (ice_parse_rx_flow_user_data(fsp, &userdata))
1635 		return -EINVAL;
1636 
1637 	if (fsp->flow_type & FLOW_MAC_EXT)
1638 		return -EINVAL;
1639 
1640 	ret = ice_cfg_fdir_xtrct_seq(pf, fsp, &userdata);
1641 	if (ret)
1642 		return ret;
1643 
1644 	if (fsp->location >= ice_get_fdir_cnt_all(hw)) {
1645 		dev_err(dev, "Failed to add filter.  The maximum number of flow director filters has been reached.\n");
1646 		return -ENOSPC;
1647 	}
1648 
1649 	/* return error if not an update and no available filters */
1650 	fltrs_needed = ice_get_open_tunnel_port(hw, TNL_ALL, &tunnel_port) ?
1651 		2 : 1;
1652 	if (!ice_fdir_find_fltr_by_idx(hw, fsp->location) &&
1653 	    ice_fdir_num_avail_fltr(hw, pf->vsi[vsi->idx]) < fltrs_needed) {
1654 		dev_err(dev, "Failed to add filter.  The maximum number of flow director filters has been reached.\n");
1655 		return -ENOSPC;
1656 	}
1657 
1658 	input = devm_kzalloc(dev, sizeof(*input), GFP_KERNEL);
1659 	if (!input)
1660 		return -ENOMEM;
1661 
1662 	ret = ice_set_fdir_input_set(vsi, fsp, input);
1663 	if (ret)
1664 		goto free_input;
1665 
1666 	mutex_lock(&hw->fdir_fltr_lock);
1667 	if (ice_fdir_is_dup_fltr(hw, input)) {
1668 		ret = -EINVAL;
1669 		goto release_lock;
1670 	}
1671 
1672 	if (userdata.flex_fltr) {
1673 		input->flex_fltr = true;
1674 		input->flex_word = cpu_to_be16(userdata.flex_word);
1675 		input->flex_offset = userdata.flex_offset;
1676 	}
1677 
1678 	/* input struct is added to the HW filter list */
1679 	ice_fdir_update_list_entry(pf, input, fsp->location);
1680 
1681 	ret = ice_fdir_write_all_fltr(pf, input, true);
1682 	if (ret)
1683 		goto remove_sw_rule;
1684 
1685 	goto release_lock;
1686 
1687 remove_sw_rule:
1688 	ice_fdir_update_cntrs(hw, input->flow_type, false);
1689 	list_del(&input->fltr_node);
1690 release_lock:
1691 	mutex_unlock(&hw->fdir_fltr_lock);
1692 free_input:
1693 	if (ret)
1694 		devm_kfree(dev, input);
1695 
1696 	return ret;
1697 }
1698