xref: /openbmc/linux/net/sched/act_ctinfo.c (revision bfa87ac8)
1 // SPDX-License-Identifier: GPL-2.0+
2 /* net/sched/act_ctinfo.c  netfilter ctinfo connmark actions
3  *
4  * Copyright (c) 2019 Kevin Darbyshire-Bryant <ldir@darbyshire-bryant.me.uk>
5  */
6 
7 #include <linux/module.h>
8 #include <linux/init.h>
9 #include <linux/kernel.h>
10 #include <linux/skbuff.h>
11 #include <linux/rtnetlink.h>
12 #include <linux/pkt_cls.h>
13 #include <linux/ip.h>
14 #include <linux/ipv6.h>
15 #include <net/netlink.h>
16 #include <net/pkt_sched.h>
17 #include <net/act_api.h>
18 #include <net/pkt_cls.h>
19 #include <uapi/linux/tc_act/tc_ctinfo.h>
20 #include <net/tc_act/tc_ctinfo.h>
21 
22 #include <net/netfilter/nf_conntrack.h>
23 #include <net/netfilter/nf_conntrack_core.h>
24 #include <net/netfilter/nf_conntrack_ecache.h>
25 #include <net/netfilter/nf_conntrack_zones.h>
26 
27 static struct tc_action_ops act_ctinfo_ops;
28 
29 static void tcf_ctinfo_dscp_set(struct nf_conn *ct, struct tcf_ctinfo *ca,
30 				struct tcf_ctinfo_params *cp,
31 				struct sk_buff *skb, int wlen, int proto)
32 {
33 	u8 dscp, newdscp;
34 
35 	newdscp = (((ct->mark & cp->dscpmask) >> cp->dscpmaskshift) << 2) &
36 		     ~INET_ECN_MASK;
37 
38 	switch (proto) {
39 	case NFPROTO_IPV4:
40 		dscp = ipv4_get_dsfield(ip_hdr(skb)) & ~INET_ECN_MASK;
41 		if (dscp != newdscp) {
42 			if (likely(!skb_try_make_writable(skb, wlen))) {
43 				ipv4_change_dsfield(ip_hdr(skb),
44 						    INET_ECN_MASK,
45 						    newdscp);
46 				ca->stats_dscp_set++;
47 			} else {
48 				ca->stats_dscp_error++;
49 			}
50 		}
51 		break;
52 	case NFPROTO_IPV6:
53 		dscp = ipv6_get_dsfield(ipv6_hdr(skb)) & ~INET_ECN_MASK;
54 		if (dscp != newdscp) {
55 			if (likely(!skb_try_make_writable(skb, wlen))) {
56 				ipv6_change_dsfield(ipv6_hdr(skb),
57 						    INET_ECN_MASK,
58 						    newdscp);
59 				ca->stats_dscp_set++;
60 			} else {
61 				ca->stats_dscp_error++;
62 			}
63 		}
64 		break;
65 	default:
66 		break;
67 	}
68 }
69 
70 static void tcf_ctinfo_cpmark_set(struct nf_conn *ct, struct tcf_ctinfo *ca,
71 				  struct tcf_ctinfo_params *cp,
72 				  struct sk_buff *skb)
73 {
74 	ca->stats_cpmark_set++;
75 	skb->mark = ct->mark & cp->cpmarkmask;
76 }
77 
78 static int tcf_ctinfo_act(struct sk_buff *skb, const struct tc_action *a,
79 			  struct tcf_result *res)
80 {
81 	const struct nf_conntrack_tuple_hash *thash = NULL;
82 	struct tcf_ctinfo *ca = to_ctinfo(a);
83 	struct nf_conntrack_tuple tuple;
84 	struct nf_conntrack_zone zone;
85 	enum ip_conntrack_info ctinfo;
86 	struct tcf_ctinfo_params *cp;
87 	struct nf_conn *ct;
88 	int proto, wlen;
89 	int action;
90 
91 	cp = rcu_dereference_bh(ca->params);
92 
93 	tcf_lastuse_update(&ca->tcf_tm);
94 	bstats_update(&ca->tcf_bstats, skb);
95 	action = READ_ONCE(ca->tcf_action);
96 
97 	wlen = skb_network_offset(skb);
98 	switch (skb_protocol(skb, true)) {
99 	case htons(ETH_P_IP):
100 		wlen += sizeof(struct iphdr);
101 		if (!pskb_may_pull(skb, wlen))
102 			goto out;
103 
104 		proto = NFPROTO_IPV4;
105 		break;
106 	case htons(ETH_P_IPV6):
107 		wlen += sizeof(struct ipv6hdr);
108 		if (!pskb_may_pull(skb, wlen))
109 			goto out;
110 
111 		proto = NFPROTO_IPV6;
112 		break;
113 	default:
114 		goto out;
115 	}
116 
117 	ct = nf_ct_get(skb, &ctinfo);
118 	if (!ct) { /* look harder, usually ingress */
119 		if (!nf_ct_get_tuplepr(skb, skb_network_offset(skb),
120 				       proto, cp->net, &tuple))
121 			goto out;
122 		zone.id = cp->zone;
123 		zone.dir = NF_CT_DEFAULT_ZONE_DIR;
124 
125 		thash = nf_conntrack_find_get(cp->net, &zone, &tuple);
126 		if (!thash)
127 			goto out;
128 
129 		ct = nf_ct_tuplehash_to_ctrack(thash);
130 	}
131 
132 	if (cp->mode & CTINFO_MODE_DSCP)
133 		if (!cp->dscpstatemask || (ct->mark & cp->dscpstatemask))
134 			tcf_ctinfo_dscp_set(ct, ca, cp, skb, wlen, proto);
135 
136 	if (cp->mode & CTINFO_MODE_CPMARK)
137 		tcf_ctinfo_cpmark_set(ct, ca, cp, skb);
138 
139 	if (thash)
140 		nf_ct_put(ct);
141 out:
142 	return action;
143 }
144 
145 static const struct nla_policy ctinfo_policy[TCA_CTINFO_MAX + 1] = {
146 	[TCA_CTINFO_ACT]		  =
147 		NLA_POLICY_EXACT_LEN(sizeof(struct tc_ctinfo)),
148 	[TCA_CTINFO_ZONE]		  = { .type = NLA_U16 },
149 	[TCA_CTINFO_PARMS_DSCP_MASK]	  = { .type = NLA_U32 },
150 	[TCA_CTINFO_PARMS_DSCP_STATEMASK] = { .type = NLA_U32 },
151 	[TCA_CTINFO_PARMS_CPMARK_MASK]	  = { .type = NLA_U32 },
152 };
153 
154 static int tcf_ctinfo_init(struct net *net, struct nlattr *nla,
155 			   struct nlattr *est, struct tc_action **a,
156 			   struct tcf_proto *tp, u32 flags,
157 			   struct netlink_ext_ack *extack)
158 {
159 	struct tc_action_net *tn = net_generic(net, act_ctinfo_ops.net_id);
160 	bool bind = flags & TCA_ACT_FLAGS_BIND;
161 	u32 dscpmask = 0, dscpstatemask, index;
162 	struct nlattr *tb[TCA_CTINFO_MAX + 1];
163 	struct tcf_ctinfo_params *cp_new;
164 	struct tcf_chain *goto_ch = NULL;
165 	struct tc_ctinfo *actparm;
166 	struct tcf_ctinfo *ci;
167 	u8 dscpmaskshift;
168 	int ret = 0, err;
169 
170 	if (!nla) {
171 		NL_SET_ERR_MSG_MOD(extack, "ctinfo requires attributes to be passed");
172 		return -EINVAL;
173 	}
174 
175 	err = nla_parse_nested(tb, TCA_CTINFO_MAX, nla, ctinfo_policy, extack);
176 	if (err < 0)
177 		return err;
178 
179 	if (!tb[TCA_CTINFO_ACT]) {
180 		NL_SET_ERR_MSG_MOD(extack,
181 				   "Missing required TCA_CTINFO_ACT attribute");
182 		return -EINVAL;
183 	}
184 	actparm = nla_data(tb[TCA_CTINFO_ACT]);
185 
186 	/* do some basic validation here before dynamically allocating things */
187 	/* that we would otherwise have to clean up.			      */
188 	if (tb[TCA_CTINFO_PARMS_DSCP_MASK]) {
189 		dscpmask = nla_get_u32(tb[TCA_CTINFO_PARMS_DSCP_MASK]);
190 		/* need contiguous 6 bit mask */
191 		dscpmaskshift = dscpmask ? __ffs(dscpmask) : 0;
192 		if ((~0 & (dscpmask >> dscpmaskshift)) != 0x3f) {
193 			NL_SET_ERR_MSG_ATTR(extack,
194 					    tb[TCA_CTINFO_PARMS_DSCP_MASK],
195 					    "dscp mask must be 6 contiguous bits");
196 			return -EINVAL;
197 		}
198 		dscpstatemask = tb[TCA_CTINFO_PARMS_DSCP_STATEMASK] ?
199 			nla_get_u32(tb[TCA_CTINFO_PARMS_DSCP_STATEMASK]) : 0;
200 		/* mask & statemask must not overlap */
201 		if (dscpmask & dscpstatemask) {
202 			NL_SET_ERR_MSG_ATTR(extack,
203 					    tb[TCA_CTINFO_PARMS_DSCP_STATEMASK],
204 					    "dscp statemask must not overlap dscp mask");
205 			return -EINVAL;
206 		}
207 	}
208 
209 	/* done the validation:now to the actual action allocation */
210 	index = actparm->index;
211 	err = tcf_idr_check_alloc(tn, &index, a, bind);
212 	if (!err) {
213 		ret = tcf_idr_create(tn, index, est, a,
214 				     &act_ctinfo_ops, bind, false, flags);
215 		if (ret) {
216 			tcf_idr_cleanup(tn, index);
217 			return ret;
218 		}
219 		ret = ACT_P_CREATED;
220 	} else if (err > 0) {
221 		if (bind) /* don't override defaults */
222 			return 0;
223 		if (!(flags & TCA_ACT_FLAGS_REPLACE)) {
224 			tcf_idr_release(*a, bind);
225 			return -EEXIST;
226 		}
227 	} else {
228 		return err;
229 	}
230 
231 	err = tcf_action_check_ctrlact(actparm->action, tp, &goto_ch, extack);
232 	if (err < 0)
233 		goto release_idr;
234 
235 	ci = to_ctinfo(*a);
236 
237 	cp_new = kzalloc(sizeof(*cp_new), GFP_KERNEL);
238 	if (unlikely(!cp_new)) {
239 		err = -ENOMEM;
240 		goto put_chain;
241 	}
242 
243 	cp_new->net = net;
244 	cp_new->zone = tb[TCA_CTINFO_ZONE] ?
245 			nla_get_u16(tb[TCA_CTINFO_ZONE]) : 0;
246 	if (dscpmask) {
247 		cp_new->dscpmask = dscpmask;
248 		cp_new->dscpmaskshift = dscpmaskshift;
249 		cp_new->dscpstatemask = dscpstatemask;
250 		cp_new->mode |= CTINFO_MODE_DSCP;
251 	}
252 
253 	if (tb[TCA_CTINFO_PARMS_CPMARK_MASK]) {
254 		cp_new->cpmarkmask =
255 				nla_get_u32(tb[TCA_CTINFO_PARMS_CPMARK_MASK]);
256 		cp_new->mode |= CTINFO_MODE_CPMARK;
257 	}
258 
259 	spin_lock_bh(&ci->tcf_lock);
260 	goto_ch = tcf_action_set_ctrlact(*a, actparm->action, goto_ch);
261 	cp_new = rcu_replace_pointer(ci->params, cp_new,
262 				     lockdep_is_held(&ci->tcf_lock));
263 	spin_unlock_bh(&ci->tcf_lock);
264 
265 	if (goto_ch)
266 		tcf_chain_put_by_act(goto_ch);
267 	if (cp_new)
268 		kfree_rcu(cp_new, rcu);
269 
270 	return ret;
271 
272 put_chain:
273 	if (goto_ch)
274 		tcf_chain_put_by_act(goto_ch);
275 release_idr:
276 	tcf_idr_release(*a, bind);
277 	return err;
278 }
279 
280 static int tcf_ctinfo_dump(struct sk_buff *skb, struct tc_action *a,
281 			   int bind, int ref)
282 {
283 	struct tcf_ctinfo *ci = to_ctinfo(a);
284 	struct tc_ctinfo opt = {
285 		.index   = ci->tcf_index,
286 		.refcnt  = refcount_read(&ci->tcf_refcnt) - ref,
287 		.bindcnt = atomic_read(&ci->tcf_bindcnt) - bind,
288 	};
289 	unsigned char *b = skb_tail_pointer(skb);
290 	struct tcf_ctinfo_params *cp;
291 	struct tcf_t t;
292 
293 	spin_lock_bh(&ci->tcf_lock);
294 	cp = rcu_dereference_protected(ci->params,
295 				       lockdep_is_held(&ci->tcf_lock));
296 
297 	tcf_tm_dump(&t, &ci->tcf_tm);
298 	if (nla_put_64bit(skb, TCA_CTINFO_TM, sizeof(t), &t, TCA_CTINFO_PAD))
299 		goto nla_put_failure;
300 
301 	opt.action = ci->tcf_action;
302 	if (nla_put(skb, TCA_CTINFO_ACT, sizeof(opt), &opt))
303 		goto nla_put_failure;
304 
305 	if (nla_put_u16(skb, TCA_CTINFO_ZONE, cp->zone))
306 		goto nla_put_failure;
307 
308 	if (cp->mode & CTINFO_MODE_DSCP) {
309 		if (nla_put_u32(skb, TCA_CTINFO_PARMS_DSCP_MASK,
310 				cp->dscpmask))
311 			goto nla_put_failure;
312 		if (nla_put_u32(skb, TCA_CTINFO_PARMS_DSCP_STATEMASK,
313 				cp->dscpstatemask))
314 			goto nla_put_failure;
315 	}
316 
317 	if (cp->mode & CTINFO_MODE_CPMARK) {
318 		if (nla_put_u32(skb, TCA_CTINFO_PARMS_CPMARK_MASK,
319 				cp->cpmarkmask))
320 			goto nla_put_failure;
321 	}
322 
323 	if (nla_put_u64_64bit(skb, TCA_CTINFO_STATS_DSCP_SET,
324 			      ci->stats_dscp_set, TCA_CTINFO_PAD))
325 		goto nla_put_failure;
326 
327 	if (nla_put_u64_64bit(skb, TCA_CTINFO_STATS_DSCP_ERROR,
328 			      ci->stats_dscp_error, TCA_CTINFO_PAD))
329 		goto nla_put_failure;
330 
331 	if (nla_put_u64_64bit(skb, TCA_CTINFO_STATS_CPMARK_SET,
332 			      ci->stats_cpmark_set, TCA_CTINFO_PAD))
333 		goto nla_put_failure;
334 
335 	spin_unlock_bh(&ci->tcf_lock);
336 	return skb->len;
337 
338 nla_put_failure:
339 	spin_unlock_bh(&ci->tcf_lock);
340 	nlmsg_trim(skb, b);
341 	return -1;
342 }
343 
344 static void tcf_ctinfo_cleanup(struct tc_action *a)
345 {
346 	struct tcf_ctinfo *ci = to_ctinfo(a);
347 	struct tcf_ctinfo_params *cp;
348 
349 	cp = rcu_dereference_protected(ci->params, 1);
350 	if (cp)
351 		kfree_rcu(cp, rcu);
352 }
353 
354 static struct tc_action_ops act_ctinfo_ops = {
355 	.kind	= "ctinfo",
356 	.id	= TCA_ID_CTINFO,
357 	.owner	= THIS_MODULE,
358 	.act	= tcf_ctinfo_act,
359 	.dump	= tcf_ctinfo_dump,
360 	.init	= tcf_ctinfo_init,
361 	.cleanup= tcf_ctinfo_cleanup,
362 	.size	= sizeof(struct tcf_ctinfo),
363 };
364 
365 static __net_init int ctinfo_init_net(struct net *net)
366 {
367 	struct tc_action_net *tn = net_generic(net, act_ctinfo_ops.net_id);
368 
369 	return tc_action_net_init(net, tn, &act_ctinfo_ops);
370 }
371 
372 static void __net_exit ctinfo_exit_net(struct list_head *net_list)
373 {
374 	tc_action_net_exit(net_list, act_ctinfo_ops.net_id);
375 }
376 
377 static struct pernet_operations ctinfo_net_ops = {
378 	.init		= ctinfo_init_net,
379 	.exit_batch	= ctinfo_exit_net,
380 	.id		= &act_ctinfo_ops.net_id,
381 	.size		= sizeof(struct tc_action_net),
382 };
383 
384 static int __init ctinfo_init_module(void)
385 {
386 	return tcf_register_action(&act_ctinfo_ops, &ctinfo_net_ops);
387 }
388 
389 static void __exit ctinfo_cleanup_module(void)
390 {
391 	tcf_unregister_action(&act_ctinfo_ops, &ctinfo_net_ops);
392 }
393 
394 module_init(ctinfo_init_module);
395 module_exit(ctinfo_cleanup_module);
396 MODULE_AUTHOR("Kevin Darbyshire-Bryant <ldir@darbyshire-bryant.me.uk>");
397 MODULE_DESCRIPTION("Connection tracking mark actions");
398 MODULE_LICENSE("GPL");
399