1 // SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0 2 /* Copyright (c) 2017-2020 Mellanox Technologies. All rights reserved */ 3 4 #include <linux/kernel.h> 5 #include <linux/errno.h> 6 #include <linux/netdevice.h> 7 #include <net/flow_offload.h> 8 9 #include "spectrum.h" 10 #include "spectrum_span.h" 11 #include "reg.h" 12 13 static struct mlxsw_sp_mall_entry * 14 mlxsw_sp_mall_entry_find(struct mlxsw_sp_flow_block *block, unsigned long cookie) 15 { 16 struct mlxsw_sp_mall_entry *mall_entry; 17 18 list_for_each_entry(mall_entry, &block->mall.list, list) 19 if (mall_entry->cookie == cookie) 20 return mall_entry; 21 22 return NULL; 23 } 24 25 static int 26 mlxsw_sp_mall_port_mirror_add(struct mlxsw_sp_port *mlxsw_sp_port, 27 struct mlxsw_sp_mall_entry *mall_entry, 28 struct netlink_ext_ack *extack) 29 { 30 struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; 31 struct mlxsw_sp_span_agent_parms agent_parms = {}; 32 struct mlxsw_sp_span_trigger_parms parms; 33 enum mlxsw_sp_span_trigger trigger; 34 int err; 35 36 if (!mall_entry->mirror.to_dev) { 37 NL_SET_ERR_MSG(extack, "Could not find requested device"); 38 return -EINVAL; 39 } 40 41 agent_parms.to_dev = mall_entry->mirror.to_dev; 42 err = mlxsw_sp_span_agent_get(mlxsw_sp, &mall_entry->mirror.span_id, 43 &agent_parms); 44 if (err) { 45 NL_SET_ERR_MSG(extack, "Failed to get SPAN agent"); 46 return err; 47 } 48 49 err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port, 50 mall_entry->ingress); 51 if (err) { 52 NL_SET_ERR_MSG(extack, "Failed to get analyzed port"); 53 goto err_analyzed_port_get; 54 } 55 56 trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : 57 MLXSW_SP_SPAN_TRIGGER_EGRESS; 58 parms.span_id = mall_entry->mirror.span_id; 59 parms.probability_rate = 1; 60 err = mlxsw_sp_span_agent_bind(mlxsw_sp, trigger, mlxsw_sp_port, 61 &parms); 62 if (err) { 63 NL_SET_ERR_MSG(extack, "Failed to bind SPAN agent"); 64 goto err_agent_bind; 65 } 66 67 return 0; 68 69 err_agent_bind: 70 mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); 71 err_analyzed_port_get: 72 mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->mirror.span_id); 73 return err; 74 } 75 76 static void 77 mlxsw_sp_mall_port_mirror_del(struct mlxsw_sp_port *mlxsw_sp_port, 78 struct mlxsw_sp_mall_entry *mall_entry) 79 { 80 struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; 81 struct mlxsw_sp_span_trigger_parms parms; 82 enum mlxsw_sp_span_trigger trigger; 83 84 trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : 85 MLXSW_SP_SPAN_TRIGGER_EGRESS; 86 parms.span_id = mall_entry->mirror.span_id; 87 mlxsw_sp_span_agent_unbind(mlxsw_sp, trigger, mlxsw_sp_port, &parms); 88 mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); 89 mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->mirror.span_id); 90 } 91 92 static int mlxsw_sp_mall_port_sample_set(struct mlxsw_sp_port *mlxsw_sp_port, 93 bool enable, u32 rate) 94 { 95 struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; 96 char mpsc_pl[MLXSW_REG_MPSC_LEN]; 97 98 mlxsw_reg_mpsc_pack(mpsc_pl, mlxsw_sp_port->local_port, enable, rate); 99 return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(mpsc), mpsc_pl); 100 } 101 102 static int 103 mlxsw_sp_mall_port_sample_add(struct mlxsw_sp_port *mlxsw_sp_port, 104 struct mlxsw_sp_mall_entry *mall_entry, 105 struct netlink_ext_ack *extack) 106 { 107 struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; 108 struct mlxsw_sp_sample_trigger trigger; 109 int err; 110 111 if (mall_entry->ingress) 112 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_INGRESS; 113 else 114 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_EGRESS; 115 trigger.local_port = mlxsw_sp_port->local_port; 116 err = mlxsw_sp_sample_trigger_params_set(mlxsw_sp, &trigger, 117 &mall_entry->sample.params, 118 extack); 119 if (err) 120 return err; 121 122 err = mlxsw_sp->mall_ops->sample_add(mlxsw_sp, mlxsw_sp_port, 123 mall_entry, extack); 124 if (err) 125 goto err_port_sample_set; 126 return 0; 127 128 err_port_sample_set: 129 mlxsw_sp_sample_trigger_params_unset(mlxsw_sp, &trigger); 130 return err; 131 } 132 133 static void 134 mlxsw_sp_mall_port_sample_del(struct mlxsw_sp_port *mlxsw_sp_port, 135 struct mlxsw_sp_mall_entry *mall_entry) 136 { 137 struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; 138 struct mlxsw_sp_sample_trigger trigger; 139 140 if (mall_entry->ingress) 141 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_INGRESS; 142 else 143 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_EGRESS; 144 trigger.local_port = mlxsw_sp_port->local_port; 145 146 mlxsw_sp->mall_ops->sample_del(mlxsw_sp, mlxsw_sp_port, mall_entry); 147 mlxsw_sp_sample_trigger_params_unset(mlxsw_sp, &trigger); 148 } 149 150 static int 151 mlxsw_sp_mall_port_rule_add(struct mlxsw_sp_port *mlxsw_sp_port, 152 struct mlxsw_sp_mall_entry *mall_entry, 153 struct netlink_ext_ack *extack) 154 { 155 switch (mall_entry->type) { 156 case MLXSW_SP_MALL_ACTION_TYPE_MIRROR: 157 return mlxsw_sp_mall_port_mirror_add(mlxsw_sp_port, mall_entry, 158 extack); 159 case MLXSW_SP_MALL_ACTION_TYPE_SAMPLE: 160 return mlxsw_sp_mall_port_sample_add(mlxsw_sp_port, mall_entry, 161 extack); 162 default: 163 WARN_ON(1); 164 return -EINVAL; 165 } 166 } 167 168 static void 169 mlxsw_sp_mall_port_rule_del(struct mlxsw_sp_port *mlxsw_sp_port, 170 struct mlxsw_sp_mall_entry *mall_entry) 171 { 172 switch (mall_entry->type) { 173 case MLXSW_SP_MALL_ACTION_TYPE_MIRROR: 174 mlxsw_sp_mall_port_mirror_del(mlxsw_sp_port, mall_entry); 175 break; 176 case MLXSW_SP_MALL_ACTION_TYPE_SAMPLE: 177 mlxsw_sp_mall_port_sample_del(mlxsw_sp_port, mall_entry); 178 break; 179 default: 180 WARN_ON(1); 181 } 182 } 183 184 static void mlxsw_sp_mall_prio_update(struct mlxsw_sp_flow_block *block) 185 { 186 struct mlxsw_sp_mall_entry *mall_entry; 187 188 if (list_empty(&block->mall.list)) 189 return; 190 block->mall.min_prio = UINT_MAX; 191 block->mall.max_prio = 0; 192 list_for_each_entry(mall_entry, &block->mall.list, list) { 193 if (mall_entry->priority < block->mall.min_prio) 194 block->mall.min_prio = mall_entry->priority; 195 if (mall_entry->priority > block->mall.max_prio) 196 block->mall.max_prio = mall_entry->priority; 197 } 198 } 199 200 int mlxsw_sp_mall_replace(struct mlxsw_sp *mlxsw_sp, 201 struct mlxsw_sp_flow_block *block, 202 struct tc_cls_matchall_offload *f) 203 { 204 struct mlxsw_sp_flow_block_binding *binding; 205 struct mlxsw_sp_mall_entry *mall_entry; 206 __be16 protocol = f->common.protocol; 207 struct flow_action_entry *act; 208 unsigned int flower_min_prio; 209 unsigned int flower_max_prio; 210 bool flower_prio_valid; 211 int err; 212 213 if (!flow_offload_has_one_action(&f->rule->action)) { 214 NL_SET_ERR_MSG(f->common.extack, "Only singular actions are supported"); 215 return -EOPNOTSUPP; 216 } 217 218 if (f->common.chain_index) { 219 NL_SET_ERR_MSG(f->common.extack, "Only chain 0 is supported"); 220 return -EOPNOTSUPP; 221 } 222 223 if (mlxsw_sp_flow_block_is_mixed_bound(block)) { 224 NL_SET_ERR_MSG(f->common.extack, "Only not mixed bound blocks are supported"); 225 return -EOPNOTSUPP; 226 } 227 228 err = mlxsw_sp_flower_prio_get(mlxsw_sp, block, f->common.chain_index, 229 &flower_min_prio, &flower_max_prio); 230 if (err) { 231 if (err != -ENOENT) { 232 NL_SET_ERR_MSG(f->common.extack, "Failed to get flower priorities"); 233 return err; 234 } 235 flower_prio_valid = false; 236 /* No flower filters are installed in specified chain. */ 237 } else { 238 flower_prio_valid = true; 239 } 240 241 if (protocol != htons(ETH_P_ALL)) { 242 NL_SET_ERR_MSG(f->common.extack, "matchall rules only supported with 'all' protocol"); 243 return -EOPNOTSUPP; 244 } 245 246 mall_entry = kzalloc(sizeof(*mall_entry), GFP_KERNEL); 247 if (!mall_entry) 248 return -ENOMEM; 249 mall_entry->cookie = f->cookie; 250 mall_entry->priority = f->common.prio; 251 mall_entry->ingress = mlxsw_sp_flow_block_is_ingress_bound(block); 252 253 if (flower_prio_valid && mall_entry->ingress && 254 mall_entry->priority >= flower_min_prio) { 255 NL_SET_ERR_MSG(f->common.extack, "Failed to add behind existing flower rules"); 256 err = -EOPNOTSUPP; 257 goto errout; 258 } 259 if (flower_prio_valid && !mall_entry->ingress && 260 mall_entry->priority <= flower_max_prio) { 261 NL_SET_ERR_MSG(f->common.extack, "Failed to add in front of existing flower rules"); 262 err = -EOPNOTSUPP; 263 goto errout; 264 } 265 266 act = &f->rule->action.entries[0]; 267 268 switch (act->id) { 269 case FLOW_ACTION_MIRRED: 270 mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_MIRROR; 271 mall_entry->mirror.to_dev = act->dev; 272 break; 273 case FLOW_ACTION_SAMPLE: 274 mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_SAMPLE; 275 mall_entry->sample.params.psample_group = act->sample.psample_group; 276 mall_entry->sample.params.truncate = act->sample.truncate; 277 mall_entry->sample.params.trunc_size = act->sample.trunc_size; 278 mall_entry->sample.params.rate = act->sample.rate; 279 break; 280 default: 281 err = -EOPNOTSUPP; 282 goto errout; 283 } 284 285 list_for_each_entry(binding, &block->binding_list, list) { 286 err = mlxsw_sp_mall_port_rule_add(binding->mlxsw_sp_port, 287 mall_entry, f->common.extack); 288 if (err) 289 goto rollback; 290 } 291 292 block->rule_count++; 293 if (mall_entry->ingress) 294 block->egress_blocker_rule_count++; 295 else 296 block->ingress_blocker_rule_count++; 297 list_add_tail(&mall_entry->list, &block->mall.list); 298 mlxsw_sp_mall_prio_update(block); 299 return 0; 300 301 rollback: 302 list_for_each_entry_continue_reverse(binding, &block->binding_list, 303 list) 304 mlxsw_sp_mall_port_rule_del(binding->mlxsw_sp_port, mall_entry); 305 errout: 306 kfree(mall_entry); 307 return err; 308 } 309 310 void mlxsw_sp_mall_destroy(struct mlxsw_sp_flow_block *block, 311 struct tc_cls_matchall_offload *f) 312 { 313 struct mlxsw_sp_flow_block_binding *binding; 314 struct mlxsw_sp_mall_entry *mall_entry; 315 316 mall_entry = mlxsw_sp_mall_entry_find(block, f->cookie); 317 if (!mall_entry) { 318 NL_SET_ERR_MSG(f->common.extack, "Entry not found"); 319 return; 320 } 321 322 list_del(&mall_entry->list); 323 if (mall_entry->ingress) 324 block->egress_blocker_rule_count--; 325 else 326 block->ingress_blocker_rule_count--; 327 block->rule_count--; 328 list_for_each_entry(binding, &block->binding_list, list) 329 mlxsw_sp_mall_port_rule_del(binding->mlxsw_sp_port, mall_entry); 330 kfree_rcu(mall_entry, rcu); /* sample RX packets may be in-flight */ 331 mlxsw_sp_mall_prio_update(block); 332 } 333 334 int mlxsw_sp_mall_port_bind(struct mlxsw_sp_flow_block *block, 335 struct mlxsw_sp_port *mlxsw_sp_port, 336 struct netlink_ext_ack *extack) 337 { 338 struct mlxsw_sp_mall_entry *mall_entry; 339 int err; 340 341 list_for_each_entry(mall_entry, &block->mall.list, list) { 342 err = mlxsw_sp_mall_port_rule_add(mlxsw_sp_port, mall_entry, 343 extack); 344 if (err) 345 goto rollback; 346 } 347 return 0; 348 349 rollback: 350 list_for_each_entry_continue_reverse(mall_entry, &block->mall.list, 351 list) 352 mlxsw_sp_mall_port_rule_del(mlxsw_sp_port, mall_entry); 353 return err; 354 } 355 356 void mlxsw_sp_mall_port_unbind(struct mlxsw_sp_flow_block *block, 357 struct mlxsw_sp_port *mlxsw_sp_port) 358 { 359 struct mlxsw_sp_mall_entry *mall_entry; 360 361 list_for_each_entry(mall_entry, &block->mall.list, list) 362 mlxsw_sp_mall_port_rule_del(mlxsw_sp_port, mall_entry); 363 } 364 365 int mlxsw_sp_mall_prio_get(struct mlxsw_sp_flow_block *block, u32 chain_index, 366 unsigned int *p_min_prio, unsigned int *p_max_prio) 367 { 368 if (chain_index || list_empty(&block->mall.list)) 369 /* In case there are no matchall rules, the caller 370 * receives -ENOENT to indicate there is no need 371 * to check the priorities. 372 */ 373 return -ENOENT; 374 *p_min_prio = block->mall.min_prio; 375 *p_max_prio = block->mall.max_prio; 376 return 0; 377 } 378 379 static int mlxsw_sp1_mall_sample_add(struct mlxsw_sp *mlxsw_sp, 380 struct mlxsw_sp_port *mlxsw_sp_port, 381 struct mlxsw_sp_mall_entry *mall_entry, 382 struct netlink_ext_ack *extack) 383 { 384 u32 rate = mall_entry->sample.params.rate; 385 386 if (!mall_entry->ingress) { 387 NL_SET_ERR_MSG(extack, "Sampling is not supported on egress"); 388 return -EOPNOTSUPP; 389 } 390 391 if (rate > MLXSW_REG_MPSC_RATE_MAX) { 392 NL_SET_ERR_MSG(extack, "Unsupported sampling rate"); 393 return -EOPNOTSUPP; 394 } 395 396 return mlxsw_sp_mall_port_sample_set(mlxsw_sp_port, true, rate); 397 } 398 399 static void mlxsw_sp1_mall_sample_del(struct mlxsw_sp *mlxsw_sp, 400 struct mlxsw_sp_port *mlxsw_sp_port, 401 struct mlxsw_sp_mall_entry *mall_entry) 402 { 403 mlxsw_sp_mall_port_sample_set(mlxsw_sp_port, false, 1); 404 } 405 406 const struct mlxsw_sp_mall_ops mlxsw_sp1_mall_ops = { 407 .sample_add = mlxsw_sp1_mall_sample_add, 408 .sample_del = mlxsw_sp1_mall_sample_del, 409 }; 410 411 static int mlxsw_sp2_mall_sample_add(struct mlxsw_sp *mlxsw_sp, 412 struct mlxsw_sp_port *mlxsw_sp_port, 413 struct mlxsw_sp_mall_entry *mall_entry, 414 struct netlink_ext_ack *extack) 415 { 416 struct mlxsw_sp_span_trigger_parms trigger_parms = {}; 417 struct mlxsw_sp_span_agent_parms agent_parms = { 418 .to_dev = NULL, /* Mirror to CPU. */ 419 .session_id = MLXSW_SP_SPAN_SESSION_ID_SAMPLING, 420 }; 421 u32 rate = mall_entry->sample.params.rate; 422 enum mlxsw_sp_span_trigger span_trigger; 423 int err; 424 425 err = mlxsw_sp_span_agent_get(mlxsw_sp, &mall_entry->sample.span_id, 426 &agent_parms); 427 if (err) { 428 NL_SET_ERR_MSG(extack, "Failed to get SPAN agent"); 429 return err; 430 } 431 432 err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port, 433 mall_entry->ingress); 434 if (err) { 435 NL_SET_ERR_MSG(extack, "Failed to get analyzed port"); 436 goto err_analyzed_port_get; 437 } 438 439 span_trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : 440 MLXSW_SP_SPAN_TRIGGER_EGRESS; 441 trigger_parms.span_id = mall_entry->sample.span_id; 442 trigger_parms.probability_rate = rate; 443 err = mlxsw_sp_span_agent_bind(mlxsw_sp, span_trigger, mlxsw_sp_port, 444 &trigger_parms); 445 if (err) { 446 NL_SET_ERR_MSG(extack, "Failed to bind SPAN agent"); 447 goto err_agent_bind; 448 } 449 450 return 0; 451 452 err_agent_bind: 453 mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); 454 err_analyzed_port_get: 455 mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->sample.span_id); 456 return err; 457 } 458 459 static void mlxsw_sp2_mall_sample_del(struct mlxsw_sp *mlxsw_sp, 460 struct mlxsw_sp_port *mlxsw_sp_port, 461 struct mlxsw_sp_mall_entry *mall_entry) 462 { 463 struct mlxsw_sp_span_trigger_parms trigger_parms = {}; 464 enum mlxsw_sp_span_trigger span_trigger; 465 466 span_trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : 467 MLXSW_SP_SPAN_TRIGGER_EGRESS; 468 trigger_parms.span_id = mall_entry->sample.span_id; 469 mlxsw_sp_span_agent_unbind(mlxsw_sp, span_trigger, mlxsw_sp_port, 470 &trigger_parms); 471 mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); 472 mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->sample.span_id); 473 } 474 475 const struct mlxsw_sp_mall_ops mlxsw_sp2_mall_ops = { 476 .sample_add = mlxsw_sp2_mall_sample_add, 477 .sample_del = mlxsw_sp2_mall_sample_del, 478 }; 479