1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * net/sched/act_police.c Input police filter 4 * 5 * Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru> 6 * J Hadi Salim (action changes) 7 */ 8 9 #include <linux/module.h> 10 #include <linux/types.h> 11 #include <linux/kernel.h> 12 #include <linux/string.h> 13 #include <linux/errno.h> 14 #include <linux/skbuff.h> 15 #include <linux/rtnetlink.h> 16 #include <linux/init.h> 17 #include <linux/slab.h> 18 #include <net/act_api.h> 19 #include <net/netlink.h> 20 #include <net/pkt_cls.h> 21 #include <net/tc_act/tc_police.h> 22 23 /* Each policer is serialized by its individual spinlock */ 24 25 static unsigned int police_net_id; 26 static struct tc_action_ops act_police_ops; 27 28 static int tcf_police_walker(struct net *net, struct sk_buff *skb, 29 struct netlink_callback *cb, int type, 30 const struct tc_action_ops *ops, 31 struct netlink_ext_ack *extack) 32 { 33 struct tc_action_net *tn = net_generic(net, police_net_id); 34 35 return tcf_generic_walker(tn, skb, cb, type, ops, extack); 36 } 37 38 static const struct nla_policy police_policy[TCA_POLICE_MAX + 1] = { 39 [TCA_POLICE_RATE] = { .len = TC_RTAB_SIZE }, 40 [TCA_POLICE_PEAKRATE] = { .len = TC_RTAB_SIZE }, 41 [TCA_POLICE_AVRATE] = { .type = NLA_U32 }, 42 [TCA_POLICE_RESULT] = { .type = NLA_U32 }, 43 [TCA_POLICE_RATE64] = { .type = NLA_U64 }, 44 [TCA_POLICE_PEAKRATE64] = { .type = NLA_U64 }, 45 }; 46 47 static int tcf_police_init(struct net *net, struct nlattr *nla, 48 struct nlattr *est, struct tc_action **a, 49 int ovr, int bind, bool rtnl_held, 50 struct tcf_proto *tp, u32 flags, 51 struct netlink_ext_ack *extack) 52 { 53 int ret = 0, tcfp_result = TC_ACT_OK, err, size; 54 struct nlattr *tb[TCA_POLICE_MAX + 1]; 55 struct tcf_chain *goto_ch = NULL; 56 struct tc_police *parm; 57 struct tcf_police *police; 58 struct qdisc_rate_table *R_tab = NULL, *P_tab = NULL; 59 struct tc_action_net *tn = net_generic(net, police_net_id); 60 struct tcf_police_params *new; 61 bool exists = false; 62 u32 index; 63 u64 rate64, prate64; 64 65 if (nla == NULL) 66 return -EINVAL; 67 68 err = nla_parse_nested_deprecated(tb, TCA_POLICE_MAX, nla, 69 police_policy, NULL); 70 if (err < 0) 71 return err; 72 73 if (tb[TCA_POLICE_TBF] == NULL) 74 return -EINVAL; 75 size = nla_len(tb[TCA_POLICE_TBF]); 76 if (size != sizeof(*parm) && size != sizeof(struct tc_police_compat)) 77 return -EINVAL; 78 79 parm = nla_data(tb[TCA_POLICE_TBF]); 80 index = parm->index; 81 err = tcf_idr_check_alloc(tn, &index, a, bind); 82 if (err < 0) 83 return err; 84 exists = err; 85 if (exists && bind) 86 return 0; 87 88 if (!exists) { 89 ret = tcf_idr_create(tn, index, NULL, a, 90 &act_police_ops, bind, true, 0); 91 if (ret) { 92 tcf_idr_cleanup(tn, index); 93 return ret; 94 } 95 ret = ACT_P_CREATED; 96 spin_lock_init(&(to_police(*a)->tcfp_lock)); 97 } else if (!ovr) { 98 tcf_idr_release(*a, bind); 99 return -EEXIST; 100 } 101 err = tcf_action_check_ctrlact(parm->action, tp, &goto_ch, extack); 102 if (err < 0) 103 goto release_idr; 104 105 police = to_police(*a); 106 if (parm->rate.rate) { 107 err = -ENOMEM; 108 R_tab = qdisc_get_rtab(&parm->rate, tb[TCA_POLICE_RATE], NULL); 109 if (R_tab == NULL) 110 goto failure; 111 112 if (parm->peakrate.rate) { 113 P_tab = qdisc_get_rtab(&parm->peakrate, 114 tb[TCA_POLICE_PEAKRATE], NULL); 115 if (P_tab == NULL) 116 goto failure; 117 } 118 } 119 120 if (est) { 121 err = gen_replace_estimator(&police->tcf_bstats, 122 police->common.cpu_bstats, 123 &police->tcf_rate_est, 124 &police->tcf_lock, 125 NULL, est); 126 if (err) 127 goto failure; 128 } else if (tb[TCA_POLICE_AVRATE] && 129 (ret == ACT_P_CREATED || 130 !gen_estimator_active(&police->tcf_rate_est))) { 131 err = -EINVAL; 132 goto failure; 133 } 134 135 if (tb[TCA_POLICE_RESULT]) { 136 tcfp_result = nla_get_u32(tb[TCA_POLICE_RESULT]); 137 if (TC_ACT_EXT_CMP(tcfp_result, TC_ACT_GOTO_CHAIN)) { 138 NL_SET_ERR_MSG(extack, 139 "goto chain not allowed on fallback"); 140 err = -EINVAL; 141 goto failure; 142 } 143 } 144 145 new = kzalloc(sizeof(*new), GFP_KERNEL); 146 if (unlikely(!new)) { 147 err = -ENOMEM; 148 goto failure; 149 } 150 151 /* No failure allowed after this point */ 152 new->tcfp_result = tcfp_result; 153 new->tcfp_mtu = parm->mtu; 154 if (!new->tcfp_mtu) { 155 new->tcfp_mtu = ~0; 156 if (R_tab) 157 new->tcfp_mtu = 255 << R_tab->rate.cell_log; 158 } 159 if (R_tab) { 160 new->rate_present = true; 161 rate64 = tb[TCA_POLICE_RATE64] ? 162 nla_get_u64(tb[TCA_POLICE_RATE64]) : 0; 163 psched_ratecfg_precompute(&new->rate, &R_tab->rate, rate64); 164 qdisc_put_rtab(R_tab); 165 } else { 166 new->rate_present = false; 167 } 168 if (P_tab) { 169 new->peak_present = true; 170 prate64 = tb[TCA_POLICE_PEAKRATE64] ? 171 nla_get_u64(tb[TCA_POLICE_PEAKRATE64]) : 0; 172 psched_ratecfg_precompute(&new->peak, &P_tab->rate, prate64); 173 qdisc_put_rtab(P_tab); 174 } else { 175 new->peak_present = false; 176 } 177 178 new->tcfp_burst = PSCHED_TICKS2NS(parm->burst); 179 if (new->peak_present) 180 new->tcfp_mtu_ptoks = (s64)psched_l2t_ns(&new->peak, 181 new->tcfp_mtu); 182 183 if (tb[TCA_POLICE_AVRATE]) 184 new->tcfp_ewma_rate = nla_get_u32(tb[TCA_POLICE_AVRATE]); 185 186 spin_lock_bh(&police->tcf_lock); 187 spin_lock_bh(&police->tcfp_lock); 188 police->tcfp_t_c = ktime_get_ns(); 189 police->tcfp_toks = new->tcfp_burst; 190 if (new->peak_present) 191 police->tcfp_ptoks = new->tcfp_mtu_ptoks; 192 spin_unlock_bh(&police->tcfp_lock); 193 goto_ch = tcf_action_set_ctrlact(*a, parm->action, goto_ch); 194 new = rcu_replace_pointer(police->params, 195 new, 196 lockdep_is_held(&police->tcf_lock)); 197 spin_unlock_bh(&police->tcf_lock); 198 199 if (goto_ch) 200 tcf_chain_put_by_act(goto_ch); 201 if (new) 202 kfree_rcu(new, rcu); 203 204 if (ret == ACT_P_CREATED) 205 tcf_idr_insert(tn, *a); 206 return ret; 207 208 failure: 209 qdisc_put_rtab(P_tab); 210 qdisc_put_rtab(R_tab); 211 if (goto_ch) 212 tcf_chain_put_by_act(goto_ch); 213 release_idr: 214 tcf_idr_release(*a, bind); 215 return err; 216 } 217 218 static int tcf_police_act(struct sk_buff *skb, const struct tc_action *a, 219 struct tcf_result *res) 220 { 221 struct tcf_police *police = to_police(a); 222 struct tcf_police_params *p; 223 s64 now, toks, ptoks = 0; 224 int ret; 225 226 tcf_lastuse_update(&police->tcf_tm); 227 bstats_cpu_update(this_cpu_ptr(police->common.cpu_bstats), skb); 228 229 ret = READ_ONCE(police->tcf_action); 230 p = rcu_dereference_bh(police->params); 231 232 if (p->tcfp_ewma_rate) { 233 struct gnet_stats_rate_est64 sample; 234 235 if (!gen_estimator_read(&police->tcf_rate_est, &sample) || 236 sample.bps >= p->tcfp_ewma_rate) 237 goto inc_overlimits; 238 } 239 240 if (qdisc_pkt_len(skb) <= p->tcfp_mtu) { 241 if (!p->rate_present) { 242 ret = p->tcfp_result; 243 goto end; 244 } 245 246 now = ktime_get_ns(); 247 spin_lock_bh(&police->tcfp_lock); 248 toks = min_t(s64, now - police->tcfp_t_c, p->tcfp_burst); 249 if (p->peak_present) { 250 ptoks = toks + police->tcfp_ptoks; 251 if (ptoks > p->tcfp_mtu_ptoks) 252 ptoks = p->tcfp_mtu_ptoks; 253 ptoks -= (s64)psched_l2t_ns(&p->peak, 254 qdisc_pkt_len(skb)); 255 } 256 toks += police->tcfp_toks; 257 if (toks > p->tcfp_burst) 258 toks = p->tcfp_burst; 259 toks -= (s64)psched_l2t_ns(&p->rate, qdisc_pkt_len(skb)); 260 if ((toks|ptoks) >= 0) { 261 police->tcfp_t_c = now; 262 police->tcfp_toks = toks; 263 police->tcfp_ptoks = ptoks; 264 spin_unlock_bh(&police->tcfp_lock); 265 ret = p->tcfp_result; 266 goto inc_drops; 267 } 268 spin_unlock_bh(&police->tcfp_lock); 269 } 270 271 inc_overlimits: 272 qstats_overlimit_inc(this_cpu_ptr(police->common.cpu_qstats)); 273 inc_drops: 274 if (ret == TC_ACT_SHOT) 275 qstats_drop_inc(this_cpu_ptr(police->common.cpu_qstats)); 276 end: 277 return ret; 278 } 279 280 static void tcf_police_cleanup(struct tc_action *a) 281 { 282 struct tcf_police *police = to_police(a); 283 struct tcf_police_params *p; 284 285 p = rcu_dereference_protected(police->params, 1); 286 if (p) 287 kfree_rcu(p, rcu); 288 } 289 290 static void tcf_police_stats_update(struct tc_action *a, 291 u64 bytes, u64 packets, u64 drops, 292 u64 lastuse, bool hw) 293 { 294 struct tcf_police *police = to_police(a); 295 struct tcf_t *tm = &police->tcf_tm; 296 297 tcf_action_update_stats(a, bytes, packets, drops, hw); 298 tm->lastuse = max_t(u64, tm->lastuse, lastuse); 299 } 300 301 static int tcf_police_dump(struct sk_buff *skb, struct tc_action *a, 302 int bind, int ref) 303 { 304 unsigned char *b = skb_tail_pointer(skb); 305 struct tcf_police *police = to_police(a); 306 struct tcf_police_params *p; 307 struct tc_police opt = { 308 .index = police->tcf_index, 309 .refcnt = refcount_read(&police->tcf_refcnt) - ref, 310 .bindcnt = atomic_read(&police->tcf_bindcnt) - bind, 311 }; 312 struct tcf_t t; 313 314 spin_lock_bh(&police->tcf_lock); 315 opt.action = police->tcf_action; 316 p = rcu_dereference_protected(police->params, 317 lockdep_is_held(&police->tcf_lock)); 318 opt.mtu = p->tcfp_mtu; 319 opt.burst = PSCHED_NS2TICKS(p->tcfp_burst); 320 if (p->rate_present) { 321 psched_ratecfg_getrate(&opt.rate, &p->rate); 322 if ((police->params->rate.rate_bytes_ps >= (1ULL << 32)) && 323 nla_put_u64_64bit(skb, TCA_POLICE_RATE64, 324 police->params->rate.rate_bytes_ps, 325 TCA_POLICE_PAD)) 326 goto nla_put_failure; 327 } 328 if (p->peak_present) { 329 psched_ratecfg_getrate(&opt.peakrate, &p->peak); 330 if ((police->params->peak.rate_bytes_ps >= (1ULL << 32)) && 331 nla_put_u64_64bit(skb, TCA_POLICE_PEAKRATE64, 332 police->params->peak.rate_bytes_ps, 333 TCA_POLICE_PAD)) 334 goto nla_put_failure; 335 } 336 if (nla_put(skb, TCA_POLICE_TBF, sizeof(opt), &opt)) 337 goto nla_put_failure; 338 if (p->tcfp_result && 339 nla_put_u32(skb, TCA_POLICE_RESULT, p->tcfp_result)) 340 goto nla_put_failure; 341 if (p->tcfp_ewma_rate && 342 nla_put_u32(skb, TCA_POLICE_AVRATE, p->tcfp_ewma_rate)) 343 goto nla_put_failure; 344 345 tcf_tm_dump(&t, &police->tcf_tm); 346 if (nla_put_64bit(skb, TCA_POLICE_TM, sizeof(t), &t, TCA_POLICE_PAD)) 347 goto nla_put_failure; 348 spin_unlock_bh(&police->tcf_lock); 349 350 return skb->len; 351 352 nla_put_failure: 353 spin_unlock_bh(&police->tcf_lock); 354 nlmsg_trim(skb, b); 355 return -1; 356 } 357 358 static int tcf_police_search(struct net *net, struct tc_action **a, u32 index) 359 { 360 struct tc_action_net *tn = net_generic(net, police_net_id); 361 362 return tcf_idr_search(tn, a, index); 363 } 364 365 MODULE_AUTHOR("Alexey Kuznetsov"); 366 MODULE_DESCRIPTION("Policing actions"); 367 MODULE_LICENSE("GPL"); 368 369 static struct tc_action_ops act_police_ops = { 370 .kind = "police", 371 .id = TCA_ID_POLICE, 372 .owner = THIS_MODULE, 373 .stats_update = tcf_police_stats_update, 374 .act = tcf_police_act, 375 .dump = tcf_police_dump, 376 .init = tcf_police_init, 377 .walk = tcf_police_walker, 378 .lookup = tcf_police_search, 379 .cleanup = tcf_police_cleanup, 380 .size = sizeof(struct tcf_police), 381 }; 382 383 static __net_init int police_init_net(struct net *net) 384 { 385 struct tc_action_net *tn = net_generic(net, police_net_id); 386 387 return tc_action_net_init(net, tn, &act_police_ops); 388 } 389 390 static void __net_exit police_exit_net(struct list_head *net_list) 391 { 392 tc_action_net_exit(net_list, police_net_id); 393 } 394 395 static struct pernet_operations police_net_ops = { 396 .init = police_init_net, 397 .exit_batch = police_exit_net, 398 .id = &police_net_id, 399 .size = sizeof(struct tc_action_net), 400 }; 401 402 static int __init police_init_module(void) 403 { 404 return tcf_register_action(&act_police_ops, &police_net_ops); 405 } 406 407 static void __exit police_cleanup_module(void) 408 { 409 tcf_unregister_action(&act_police_ops, &police_net_ops); 410 } 411 412 module_init(police_init_module); 413 module_exit(police_cleanup_module); 414