1 /* 2 * Copyright (c) 2015, Mellanox Technologies. All rights reserved. 3 * 4 * This software is available to you under a choice of one of two 5 * licenses. You may choose to be licensed under the terms of the GNU 6 * General Public License (GPL) Version 2, available from the file 7 * COPYING in the main directory of this source tree, or the 8 * OpenIB.org BSD license below: 9 * 10 * Redistribution and use in source and binary forms, with or 11 * without modification, are permitted provided that the following 12 * conditions are met: 13 * 14 * - Redistributions of source code must retain the above 15 * copyright notice, this list of conditions and the following 16 * disclaimer. 17 * 18 * - Redistributions in binary form must reproduce the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer in the documentation and/or other materials 21 * provided with the distribution. 22 * 23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 30 * SOFTWARE. 31 */ 32 33 #include <linux/mlx5/driver.h> 34 #include <linux/mlx5/device.h> 35 #include <linux/mlx5/mlx5_ifc.h> 36 37 #include "fs_core.h" 38 #include "fs_cmd.h" 39 #include "mlx5_core.h" 40 #include "eswitch.h" 41 42 static int mlx5_cmd_stub_update_root_ft(struct mlx5_flow_root_namespace *ns, 43 struct mlx5_flow_table *ft, 44 u32 underlay_qpn, 45 bool disconnect) 46 { 47 return 0; 48 } 49 50 static int mlx5_cmd_stub_create_flow_table(struct mlx5_flow_root_namespace *ns, 51 struct mlx5_flow_table *ft, 52 unsigned int log_size, 53 struct mlx5_flow_table *next_ft) 54 { 55 return 0; 56 } 57 58 static int mlx5_cmd_stub_destroy_flow_table(struct mlx5_flow_root_namespace *ns, 59 struct mlx5_flow_table *ft) 60 { 61 return 0; 62 } 63 64 static int mlx5_cmd_stub_modify_flow_table(struct mlx5_flow_root_namespace *ns, 65 struct mlx5_flow_table *ft, 66 struct mlx5_flow_table *next_ft) 67 { 68 return 0; 69 } 70 71 static int mlx5_cmd_stub_create_flow_group(struct mlx5_flow_root_namespace *ns, 72 struct mlx5_flow_table *ft, 73 u32 *in, 74 struct mlx5_flow_group *fg) 75 { 76 return 0; 77 } 78 79 static int mlx5_cmd_stub_destroy_flow_group(struct mlx5_flow_root_namespace *ns, 80 struct mlx5_flow_table *ft, 81 struct mlx5_flow_group *fg) 82 { 83 return 0; 84 } 85 86 static int mlx5_cmd_stub_create_fte(struct mlx5_flow_root_namespace *ns, 87 struct mlx5_flow_table *ft, 88 struct mlx5_flow_group *group, 89 struct fs_fte *fte) 90 { 91 return 0; 92 } 93 94 static int mlx5_cmd_stub_update_fte(struct mlx5_flow_root_namespace *ns, 95 struct mlx5_flow_table *ft, 96 struct mlx5_flow_group *group, 97 int modify_mask, 98 struct fs_fte *fte) 99 { 100 return -EOPNOTSUPP; 101 } 102 103 static int mlx5_cmd_stub_delete_fte(struct mlx5_flow_root_namespace *ns, 104 struct mlx5_flow_table *ft, 105 struct fs_fte *fte) 106 { 107 return 0; 108 } 109 110 static int mlx5_cmd_stub_packet_reformat_alloc(struct mlx5_flow_root_namespace *ns, 111 int reformat_type, 112 size_t size, 113 void *reformat_data, 114 enum mlx5_flow_namespace_type namespace, 115 struct mlx5_pkt_reformat *pkt_reformat) 116 { 117 return 0; 118 } 119 120 static void mlx5_cmd_stub_packet_reformat_dealloc(struct mlx5_flow_root_namespace *ns, 121 struct mlx5_pkt_reformat *pkt_reformat) 122 { 123 } 124 125 static int mlx5_cmd_stub_modify_header_alloc(struct mlx5_flow_root_namespace *ns, 126 u8 namespace, u8 num_actions, 127 void *modify_actions, 128 struct mlx5_modify_hdr *modify_hdr) 129 { 130 return 0; 131 } 132 133 static void mlx5_cmd_stub_modify_header_dealloc(struct mlx5_flow_root_namespace *ns, 134 struct mlx5_modify_hdr *modify_hdr) 135 { 136 } 137 138 static int mlx5_cmd_stub_set_peer(struct mlx5_flow_root_namespace *ns, 139 struct mlx5_flow_root_namespace *peer_ns) 140 { 141 return 0; 142 } 143 144 static int mlx5_cmd_stub_create_ns(struct mlx5_flow_root_namespace *ns) 145 { 146 return 0; 147 } 148 149 static int mlx5_cmd_stub_destroy_ns(struct mlx5_flow_root_namespace *ns) 150 { 151 return 0; 152 } 153 154 static int mlx5_cmd_update_root_ft(struct mlx5_flow_root_namespace *ns, 155 struct mlx5_flow_table *ft, u32 underlay_qpn, 156 bool disconnect) 157 { 158 u32 in[MLX5_ST_SZ_DW(set_flow_table_root_in)] = {}; 159 struct mlx5_core_dev *dev = ns->dev; 160 161 if ((MLX5_CAP_GEN(dev, port_type) == MLX5_CAP_PORT_TYPE_IB) && 162 underlay_qpn == 0) 163 return 0; 164 165 MLX5_SET(set_flow_table_root_in, in, opcode, 166 MLX5_CMD_OP_SET_FLOW_TABLE_ROOT); 167 MLX5_SET(set_flow_table_root_in, in, table_type, ft->type); 168 169 if (disconnect) 170 MLX5_SET(set_flow_table_root_in, in, op_mod, 1); 171 else 172 MLX5_SET(set_flow_table_root_in, in, table_id, ft->id); 173 174 MLX5_SET(set_flow_table_root_in, in, underlay_qpn, underlay_qpn); 175 MLX5_SET(set_flow_table_root_in, in, vport_number, ft->vport); 176 MLX5_SET(set_flow_table_root_in, in, other_vport, 177 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 178 179 return mlx5_cmd_exec_in(dev, set_flow_table_root, in); 180 } 181 182 static int mlx5_cmd_create_flow_table(struct mlx5_flow_root_namespace *ns, 183 struct mlx5_flow_table *ft, 184 unsigned int log_size, 185 struct mlx5_flow_table *next_ft) 186 { 187 int en_encap = !!(ft->flags & MLX5_FLOW_TABLE_TUNNEL_EN_REFORMAT); 188 int en_decap = !!(ft->flags & MLX5_FLOW_TABLE_TUNNEL_EN_DECAP); 189 int term = !!(ft->flags & MLX5_FLOW_TABLE_TERMINATION); 190 u32 out[MLX5_ST_SZ_DW(create_flow_table_out)] = {}; 191 u32 in[MLX5_ST_SZ_DW(create_flow_table_in)] = {}; 192 struct mlx5_core_dev *dev = ns->dev; 193 int err; 194 195 MLX5_SET(create_flow_table_in, in, opcode, 196 MLX5_CMD_OP_CREATE_FLOW_TABLE); 197 198 MLX5_SET(create_flow_table_in, in, table_type, ft->type); 199 MLX5_SET(create_flow_table_in, in, flow_table_context.level, ft->level); 200 MLX5_SET(create_flow_table_in, in, flow_table_context.log_size, log_size); 201 MLX5_SET(create_flow_table_in, in, vport_number, ft->vport); 202 MLX5_SET(create_flow_table_in, in, other_vport, 203 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 204 205 MLX5_SET(create_flow_table_in, in, flow_table_context.decap_en, 206 en_decap); 207 MLX5_SET(create_flow_table_in, in, flow_table_context.reformat_en, 208 en_encap); 209 MLX5_SET(create_flow_table_in, in, flow_table_context.termination_table, 210 term); 211 212 switch (ft->op_mod) { 213 case FS_FT_OP_MOD_NORMAL: 214 if (next_ft) { 215 MLX5_SET(create_flow_table_in, in, 216 flow_table_context.table_miss_action, 217 MLX5_FLOW_TABLE_MISS_ACTION_FWD); 218 MLX5_SET(create_flow_table_in, in, 219 flow_table_context.table_miss_id, next_ft->id); 220 } else { 221 MLX5_SET(create_flow_table_in, in, 222 flow_table_context.table_miss_action, 223 ft->def_miss_action); 224 } 225 break; 226 227 case FS_FT_OP_MOD_LAG_DEMUX: 228 MLX5_SET(create_flow_table_in, in, op_mod, 0x1); 229 if (next_ft) 230 MLX5_SET(create_flow_table_in, in, 231 flow_table_context.lag_master_next_table_id, 232 next_ft->id); 233 break; 234 } 235 236 err = mlx5_cmd_exec_inout(dev, create_flow_table, in, out); 237 if (!err) 238 ft->id = MLX5_GET(create_flow_table_out, out, 239 table_id); 240 return err; 241 } 242 243 static int mlx5_cmd_destroy_flow_table(struct mlx5_flow_root_namespace *ns, 244 struct mlx5_flow_table *ft) 245 { 246 u32 in[MLX5_ST_SZ_DW(destroy_flow_table_in)] = {}; 247 struct mlx5_core_dev *dev = ns->dev; 248 249 MLX5_SET(destroy_flow_table_in, in, opcode, 250 MLX5_CMD_OP_DESTROY_FLOW_TABLE); 251 MLX5_SET(destroy_flow_table_in, in, table_type, ft->type); 252 MLX5_SET(destroy_flow_table_in, in, table_id, ft->id); 253 MLX5_SET(destroy_flow_table_in, in, vport_number, ft->vport); 254 MLX5_SET(destroy_flow_table_in, in, other_vport, 255 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 256 257 return mlx5_cmd_exec_in(dev, destroy_flow_table, in); 258 } 259 260 static int mlx5_cmd_modify_flow_table(struct mlx5_flow_root_namespace *ns, 261 struct mlx5_flow_table *ft, 262 struct mlx5_flow_table *next_ft) 263 { 264 u32 in[MLX5_ST_SZ_DW(modify_flow_table_in)] = {}; 265 struct mlx5_core_dev *dev = ns->dev; 266 267 MLX5_SET(modify_flow_table_in, in, opcode, 268 MLX5_CMD_OP_MODIFY_FLOW_TABLE); 269 MLX5_SET(modify_flow_table_in, in, table_type, ft->type); 270 MLX5_SET(modify_flow_table_in, in, table_id, ft->id); 271 272 if (ft->op_mod == FS_FT_OP_MOD_LAG_DEMUX) { 273 MLX5_SET(modify_flow_table_in, in, modify_field_select, 274 MLX5_MODIFY_FLOW_TABLE_LAG_NEXT_TABLE_ID); 275 if (next_ft) { 276 MLX5_SET(modify_flow_table_in, in, 277 flow_table_context.lag_master_next_table_id, next_ft->id); 278 } else { 279 MLX5_SET(modify_flow_table_in, in, 280 flow_table_context.lag_master_next_table_id, 0); 281 } 282 } else { 283 MLX5_SET(modify_flow_table_in, in, vport_number, ft->vport); 284 MLX5_SET(modify_flow_table_in, in, other_vport, 285 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 286 MLX5_SET(modify_flow_table_in, in, modify_field_select, 287 MLX5_MODIFY_FLOW_TABLE_MISS_TABLE_ID); 288 if (next_ft) { 289 MLX5_SET(modify_flow_table_in, in, 290 flow_table_context.table_miss_action, 291 MLX5_FLOW_TABLE_MISS_ACTION_FWD); 292 MLX5_SET(modify_flow_table_in, in, 293 flow_table_context.table_miss_id, 294 next_ft->id); 295 } else { 296 MLX5_SET(modify_flow_table_in, in, 297 flow_table_context.table_miss_action, 298 ft->def_miss_action); 299 } 300 } 301 302 return mlx5_cmd_exec_in(dev, modify_flow_table, in); 303 } 304 305 static int mlx5_cmd_create_flow_group(struct mlx5_flow_root_namespace *ns, 306 struct mlx5_flow_table *ft, 307 u32 *in, 308 struct mlx5_flow_group *fg) 309 { 310 u32 out[MLX5_ST_SZ_DW(create_flow_group_out)] = {}; 311 struct mlx5_core_dev *dev = ns->dev; 312 int err; 313 314 MLX5_SET(create_flow_group_in, in, opcode, 315 MLX5_CMD_OP_CREATE_FLOW_GROUP); 316 MLX5_SET(create_flow_group_in, in, table_type, ft->type); 317 MLX5_SET(create_flow_group_in, in, table_id, ft->id); 318 if (ft->vport) { 319 MLX5_SET(create_flow_group_in, in, vport_number, ft->vport); 320 MLX5_SET(create_flow_group_in, in, other_vport, 1); 321 } 322 323 MLX5_SET(create_flow_group_in, in, vport_number, ft->vport); 324 MLX5_SET(create_flow_group_in, in, other_vport, 325 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 326 err = mlx5_cmd_exec_inout(dev, create_flow_group, in, out); 327 if (!err) 328 fg->id = MLX5_GET(create_flow_group_out, out, 329 group_id); 330 return err; 331 } 332 333 static int mlx5_cmd_destroy_flow_group(struct mlx5_flow_root_namespace *ns, 334 struct mlx5_flow_table *ft, 335 struct mlx5_flow_group *fg) 336 { 337 u32 in[MLX5_ST_SZ_DW(destroy_flow_group_in)] = {}; 338 struct mlx5_core_dev *dev = ns->dev; 339 340 MLX5_SET(destroy_flow_group_in, in, opcode, 341 MLX5_CMD_OP_DESTROY_FLOW_GROUP); 342 MLX5_SET(destroy_flow_group_in, in, table_type, ft->type); 343 MLX5_SET(destroy_flow_group_in, in, table_id, ft->id); 344 MLX5_SET(destroy_flow_group_in, in, group_id, fg->id); 345 MLX5_SET(destroy_flow_group_in, in, vport_number, ft->vport); 346 MLX5_SET(destroy_flow_group_in, in, other_vport, 347 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 348 return mlx5_cmd_exec_in(dev, destroy_flow_group, in); 349 } 350 351 static int mlx5_set_extended_dest(struct mlx5_core_dev *dev, 352 struct fs_fte *fte, bool *extended_dest) 353 { 354 int fw_log_max_fdb_encap_uplink = 355 MLX5_CAP_ESW(dev, log_max_fdb_encap_uplink); 356 int num_fwd_destinations = 0; 357 struct mlx5_flow_rule *dst; 358 int num_encap = 0; 359 360 *extended_dest = false; 361 if (!(fte->action.action & MLX5_FLOW_CONTEXT_ACTION_FWD_DEST)) 362 return 0; 363 364 list_for_each_entry(dst, &fte->node.children, node.list) { 365 if (dst->dest_attr.type == MLX5_FLOW_DESTINATION_TYPE_COUNTER) 366 continue; 367 if (dst->dest_attr.type == MLX5_FLOW_DESTINATION_TYPE_VPORT && 368 dst->dest_attr.vport.flags & MLX5_FLOW_DEST_VPORT_REFORMAT_ID) 369 num_encap++; 370 num_fwd_destinations++; 371 } 372 if (num_fwd_destinations > 1 && num_encap > 0) 373 *extended_dest = true; 374 375 if (*extended_dest && !fw_log_max_fdb_encap_uplink) { 376 mlx5_core_warn(dev, "FW does not support extended destination"); 377 return -EOPNOTSUPP; 378 } 379 if (num_encap > (1 << fw_log_max_fdb_encap_uplink)) { 380 mlx5_core_warn(dev, "FW does not support more than %d encaps", 381 1 << fw_log_max_fdb_encap_uplink); 382 return -EOPNOTSUPP; 383 } 384 385 return 0; 386 } 387 static int mlx5_cmd_set_fte(struct mlx5_core_dev *dev, 388 int opmod, int modify_mask, 389 struct mlx5_flow_table *ft, 390 unsigned group_id, 391 struct fs_fte *fte) 392 { 393 u32 out[MLX5_ST_SZ_DW(set_fte_out)] = {0}; 394 bool extended_dest = false; 395 struct mlx5_flow_rule *dst; 396 void *in_flow_context, *vlan; 397 void *in_match_value; 398 unsigned int inlen; 399 int dst_cnt_size; 400 void *in_dests; 401 u32 *in; 402 int err; 403 404 if (mlx5_set_extended_dest(dev, fte, &extended_dest)) 405 return -EOPNOTSUPP; 406 407 if (!extended_dest) 408 dst_cnt_size = MLX5_ST_SZ_BYTES(dest_format_struct); 409 else 410 dst_cnt_size = MLX5_ST_SZ_BYTES(extended_dest_format); 411 412 inlen = MLX5_ST_SZ_BYTES(set_fte_in) + fte->dests_size * dst_cnt_size; 413 in = kvzalloc(inlen, GFP_KERNEL); 414 if (!in) 415 return -ENOMEM; 416 417 MLX5_SET(set_fte_in, in, opcode, MLX5_CMD_OP_SET_FLOW_TABLE_ENTRY); 418 MLX5_SET(set_fte_in, in, op_mod, opmod); 419 MLX5_SET(set_fte_in, in, modify_enable_mask, modify_mask); 420 MLX5_SET(set_fte_in, in, table_type, ft->type); 421 MLX5_SET(set_fte_in, in, table_id, ft->id); 422 MLX5_SET(set_fte_in, in, flow_index, fte->index); 423 MLX5_SET(set_fte_in, in, ignore_flow_level, 424 !!(fte->action.flags & FLOW_ACT_IGNORE_FLOW_LEVEL)); 425 426 MLX5_SET(set_fte_in, in, vport_number, ft->vport); 427 MLX5_SET(set_fte_in, in, other_vport, 428 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 429 430 in_flow_context = MLX5_ADDR_OF(set_fte_in, in, flow_context); 431 MLX5_SET(flow_context, in_flow_context, group_id, group_id); 432 433 MLX5_SET(flow_context, in_flow_context, flow_tag, 434 fte->flow_context.flow_tag); 435 MLX5_SET(flow_context, in_flow_context, flow_source, 436 fte->flow_context.flow_source); 437 438 MLX5_SET(flow_context, in_flow_context, extended_destination, 439 extended_dest); 440 if (extended_dest) { 441 u32 action; 442 443 action = fte->action.action & 444 ~MLX5_FLOW_CONTEXT_ACTION_PACKET_REFORMAT; 445 MLX5_SET(flow_context, in_flow_context, action, action); 446 } else { 447 MLX5_SET(flow_context, in_flow_context, action, 448 fte->action.action); 449 if (fte->action.pkt_reformat) 450 MLX5_SET(flow_context, in_flow_context, packet_reformat_id, 451 fte->action.pkt_reformat->id); 452 } 453 if (fte->action.modify_hdr) 454 MLX5_SET(flow_context, in_flow_context, modify_header_id, 455 fte->action.modify_hdr->id); 456 457 MLX5_SET(flow_context, in_flow_context, ipsec_obj_id, fte->action.ipsec_obj_id); 458 459 vlan = MLX5_ADDR_OF(flow_context, in_flow_context, push_vlan); 460 461 MLX5_SET(vlan, vlan, ethtype, fte->action.vlan[0].ethtype); 462 MLX5_SET(vlan, vlan, vid, fte->action.vlan[0].vid); 463 MLX5_SET(vlan, vlan, prio, fte->action.vlan[0].prio); 464 465 vlan = MLX5_ADDR_OF(flow_context, in_flow_context, push_vlan_2); 466 467 MLX5_SET(vlan, vlan, ethtype, fte->action.vlan[1].ethtype); 468 MLX5_SET(vlan, vlan, vid, fte->action.vlan[1].vid); 469 MLX5_SET(vlan, vlan, prio, fte->action.vlan[1].prio); 470 471 in_match_value = MLX5_ADDR_OF(flow_context, in_flow_context, 472 match_value); 473 memcpy(in_match_value, &fte->val, sizeof(fte->val)); 474 475 in_dests = MLX5_ADDR_OF(flow_context, in_flow_context, destination); 476 if (fte->action.action & MLX5_FLOW_CONTEXT_ACTION_FWD_DEST) { 477 int list_size = 0; 478 479 list_for_each_entry(dst, &fte->node.children, node.list) { 480 unsigned int id, type = dst->dest_attr.type; 481 482 if (type == MLX5_FLOW_DESTINATION_TYPE_COUNTER) 483 continue; 484 485 switch (type) { 486 case MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE_NUM: 487 id = dst->dest_attr.ft_num; 488 type = MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE; 489 break; 490 case MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE: 491 id = dst->dest_attr.ft->id; 492 break; 493 case MLX5_FLOW_DESTINATION_TYPE_VPORT: 494 id = dst->dest_attr.vport.num; 495 MLX5_SET(dest_format_struct, in_dests, 496 destination_eswitch_owner_vhca_id_valid, 497 !!(dst->dest_attr.vport.flags & 498 MLX5_FLOW_DEST_VPORT_VHCA_ID)); 499 MLX5_SET(dest_format_struct, in_dests, 500 destination_eswitch_owner_vhca_id, 501 dst->dest_attr.vport.vhca_id); 502 if (extended_dest && 503 dst->dest_attr.vport.pkt_reformat) { 504 MLX5_SET(dest_format_struct, in_dests, 505 packet_reformat, 506 !!(dst->dest_attr.vport.flags & 507 MLX5_FLOW_DEST_VPORT_REFORMAT_ID)); 508 MLX5_SET(extended_dest_format, in_dests, 509 packet_reformat_id, 510 dst->dest_attr.vport.pkt_reformat->id); 511 } 512 break; 513 case MLX5_FLOW_DESTINATION_TYPE_FLOW_SAMPLER: 514 id = dst->dest_attr.sampler_id; 515 break; 516 default: 517 id = dst->dest_attr.tir_num; 518 } 519 520 MLX5_SET(dest_format_struct, in_dests, destination_type, 521 type); 522 MLX5_SET(dest_format_struct, in_dests, destination_id, id); 523 in_dests += dst_cnt_size; 524 list_size++; 525 } 526 527 MLX5_SET(flow_context, in_flow_context, destination_list_size, 528 list_size); 529 } 530 531 if (fte->action.action & MLX5_FLOW_CONTEXT_ACTION_COUNT) { 532 int max_list_size = BIT(MLX5_CAP_FLOWTABLE_TYPE(dev, 533 log_max_flow_counter, 534 ft->type)); 535 int list_size = 0; 536 537 list_for_each_entry(dst, &fte->node.children, node.list) { 538 if (dst->dest_attr.type != 539 MLX5_FLOW_DESTINATION_TYPE_COUNTER) 540 continue; 541 542 MLX5_SET(flow_counter_list, in_dests, flow_counter_id, 543 dst->dest_attr.counter_id); 544 in_dests += dst_cnt_size; 545 list_size++; 546 } 547 if (list_size > max_list_size) { 548 err = -EINVAL; 549 goto err_out; 550 } 551 552 MLX5_SET(flow_context, in_flow_context, flow_counter_list_size, 553 list_size); 554 } 555 556 err = mlx5_cmd_exec(dev, in, inlen, out, sizeof(out)); 557 err_out: 558 kvfree(in); 559 return err; 560 } 561 562 static int mlx5_cmd_create_fte(struct mlx5_flow_root_namespace *ns, 563 struct mlx5_flow_table *ft, 564 struct mlx5_flow_group *group, 565 struct fs_fte *fte) 566 { 567 struct mlx5_core_dev *dev = ns->dev; 568 unsigned int group_id = group->id; 569 570 return mlx5_cmd_set_fte(dev, 0, 0, ft, group_id, fte); 571 } 572 573 static int mlx5_cmd_update_fte(struct mlx5_flow_root_namespace *ns, 574 struct mlx5_flow_table *ft, 575 struct mlx5_flow_group *fg, 576 int modify_mask, 577 struct fs_fte *fte) 578 { 579 int opmod; 580 struct mlx5_core_dev *dev = ns->dev; 581 int atomic_mod_cap = MLX5_CAP_FLOWTABLE(dev, 582 flow_table_properties_nic_receive. 583 flow_modify_en); 584 if (!atomic_mod_cap) 585 return -EOPNOTSUPP; 586 opmod = 1; 587 588 return mlx5_cmd_set_fte(dev, opmod, modify_mask, ft, fg->id, fte); 589 } 590 591 static int mlx5_cmd_delete_fte(struct mlx5_flow_root_namespace *ns, 592 struct mlx5_flow_table *ft, 593 struct fs_fte *fte) 594 { 595 u32 in[MLX5_ST_SZ_DW(delete_fte_in)] = {}; 596 struct mlx5_core_dev *dev = ns->dev; 597 598 MLX5_SET(delete_fte_in, in, opcode, MLX5_CMD_OP_DELETE_FLOW_TABLE_ENTRY); 599 MLX5_SET(delete_fte_in, in, table_type, ft->type); 600 MLX5_SET(delete_fte_in, in, table_id, ft->id); 601 MLX5_SET(delete_fte_in, in, flow_index, fte->index); 602 MLX5_SET(delete_fte_in, in, vport_number, ft->vport); 603 MLX5_SET(delete_fte_in, in, other_vport, 604 !!(ft->flags & MLX5_FLOW_TABLE_OTHER_VPORT)); 605 606 return mlx5_cmd_exec_in(dev, delete_fte, in); 607 } 608 609 int mlx5_cmd_fc_bulk_alloc(struct mlx5_core_dev *dev, 610 enum mlx5_fc_bulk_alloc_bitmask alloc_bitmask, 611 u32 *id) 612 { 613 u32 out[MLX5_ST_SZ_DW(alloc_flow_counter_out)] = {}; 614 u32 in[MLX5_ST_SZ_DW(alloc_flow_counter_in)] = {}; 615 int err; 616 617 MLX5_SET(alloc_flow_counter_in, in, opcode, 618 MLX5_CMD_OP_ALLOC_FLOW_COUNTER); 619 MLX5_SET(alloc_flow_counter_in, in, flow_counter_bulk, alloc_bitmask); 620 621 err = mlx5_cmd_exec_inout(dev, alloc_flow_counter, in, out); 622 if (!err) 623 *id = MLX5_GET(alloc_flow_counter_out, out, flow_counter_id); 624 return err; 625 } 626 627 int mlx5_cmd_fc_alloc(struct mlx5_core_dev *dev, u32 *id) 628 { 629 return mlx5_cmd_fc_bulk_alloc(dev, 0, id); 630 } 631 632 int mlx5_cmd_fc_free(struct mlx5_core_dev *dev, u32 id) 633 { 634 u32 in[MLX5_ST_SZ_DW(dealloc_flow_counter_in)] = {}; 635 636 MLX5_SET(dealloc_flow_counter_in, in, opcode, 637 MLX5_CMD_OP_DEALLOC_FLOW_COUNTER); 638 MLX5_SET(dealloc_flow_counter_in, in, flow_counter_id, id); 639 return mlx5_cmd_exec_in(dev, dealloc_flow_counter, in); 640 } 641 642 int mlx5_cmd_fc_query(struct mlx5_core_dev *dev, u32 id, 643 u64 *packets, u64 *bytes) 644 { 645 u32 out[MLX5_ST_SZ_BYTES(query_flow_counter_out) + 646 MLX5_ST_SZ_BYTES(traffic_counter)] = {}; 647 u32 in[MLX5_ST_SZ_DW(query_flow_counter_in)] = {}; 648 void *stats; 649 int err = 0; 650 651 MLX5_SET(query_flow_counter_in, in, opcode, 652 MLX5_CMD_OP_QUERY_FLOW_COUNTER); 653 MLX5_SET(query_flow_counter_in, in, op_mod, 0); 654 MLX5_SET(query_flow_counter_in, in, flow_counter_id, id); 655 err = mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out)); 656 if (err) 657 return err; 658 659 stats = MLX5_ADDR_OF(query_flow_counter_out, out, flow_statistics); 660 *packets = MLX5_GET64(traffic_counter, stats, packets); 661 *bytes = MLX5_GET64(traffic_counter, stats, octets); 662 return 0; 663 } 664 665 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len) 666 { 667 return MLX5_ST_SZ_BYTES(query_flow_counter_out) + 668 MLX5_ST_SZ_BYTES(traffic_counter) * bulk_len; 669 } 670 671 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len, 672 u32 *out) 673 { 674 int outlen = mlx5_cmd_fc_get_bulk_query_out_len(bulk_len); 675 u32 in[MLX5_ST_SZ_DW(query_flow_counter_in)] = {}; 676 677 MLX5_SET(query_flow_counter_in, in, opcode, 678 MLX5_CMD_OP_QUERY_FLOW_COUNTER); 679 MLX5_SET(query_flow_counter_in, in, flow_counter_id, base_id); 680 MLX5_SET(query_flow_counter_in, in, num_of_counters, bulk_len); 681 return mlx5_cmd_exec(dev, in, sizeof(in), out, outlen); 682 } 683 684 static int mlx5_cmd_packet_reformat_alloc(struct mlx5_flow_root_namespace *ns, 685 int reformat_type, 686 size_t size, 687 void *reformat_data, 688 enum mlx5_flow_namespace_type namespace, 689 struct mlx5_pkt_reformat *pkt_reformat) 690 { 691 u32 out[MLX5_ST_SZ_DW(alloc_packet_reformat_context_out)] = {}; 692 struct mlx5_core_dev *dev = ns->dev; 693 void *packet_reformat_context_in; 694 int max_encap_size; 695 void *reformat; 696 int inlen; 697 int err; 698 u32 *in; 699 700 if (namespace == MLX5_FLOW_NAMESPACE_FDB) 701 max_encap_size = MLX5_CAP_ESW(dev, max_encap_header_size); 702 else 703 max_encap_size = MLX5_CAP_FLOWTABLE(dev, max_encap_header_size); 704 705 if (size > max_encap_size) { 706 mlx5_core_warn(dev, "encap size %zd too big, max supported is %d\n", 707 size, max_encap_size); 708 return -EINVAL; 709 } 710 711 in = kzalloc(MLX5_ST_SZ_BYTES(alloc_packet_reformat_context_in) + size, 712 GFP_KERNEL); 713 if (!in) 714 return -ENOMEM; 715 716 packet_reformat_context_in = MLX5_ADDR_OF(alloc_packet_reformat_context_in, 717 in, packet_reformat_context); 718 reformat = MLX5_ADDR_OF(packet_reformat_context_in, 719 packet_reformat_context_in, 720 reformat_data); 721 inlen = reformat - (void *)in + size; 722 723 MLX5_SET(alloc_packet_reformat_context_in, in, opcode, 724 MLX5_CMD_OP_ALLOC_PACKET_REFORMAT_CONTEXT); 725 MLX5_SET(packet_reformat_context_in, packet_reformat_context_in, 726 reformat_data_size, size); 727 MLX5_SET(packet_reformat_context_in, packet_reformat_context_in, 728 reformat_type, reformat_type); 729 memcpy(reformat, reformat_data, size); 730 731 err = mlx5_cmd_exec(dev, in, inlen, out, sizeof(out)); 732 733 pkt_reformat->id = MLX5_GET(alloc_packet_reformat_context_out, 734 out, packet_reformat_id); 735 kfree(in); 736 return err; 737 } 738 739 static void mlx5_cmd_packet_reformat_dealloc(struct mlx5_flow_root_namespace *ns, 740 struct mlx5_pkt_reformat *pkt_reformat) 741 { 742 u32 in[MLX5_ST_SZ_DW(dealloc_packet_reformat_context_in)] = {}; 743 struct mlx5_core_dev *dev = ns->dev; 744 745 MLX5_SET(dealloc_packet_reformat_context_in, in, opcode, 746 MLX5_CMD_OP_DEALLOC_PACKET_REFORMAT_CONTEXT); 747 MLX5_SET(dealloc_packet_reformat_context_in, in, packet_reformat_id, 748 pkt_reformat->id); 749 750 mlx5_cmd_exec_in(dev, dealloc_packet_reformat_context, in); 751 } 752 753 static int mlx5_cmd_modify_header_alloc(struct mlx5_flow_root_namespace *ns, 754 u8 namespace, u8 num_actions, 755 void *modify_actions, 756 struct mlx5_modify_hdr *modify_hdr) 757 { 758 u32 out[MLX5_ST_SZ_DW(alloc_modify_header_context_out)] = {}; 759 int max_actions, actions_size, inlen, err; 760 struct mlx5_core_dev *dev = ns->dev; 761 void *actions_in; 762 u8 table_type; 763 u32 *in; 764 765 switch (namespace) { 766 case MLX5_FLOW_NAMESPACE_FDB: 767 max_actions = MLX5_CAP_ESW_FLOWTABLE_FDB(dev, max_modify_header_actions); 768 table_type = FS_FT_FDB; 769 break; 770 case MLX5_FLOW_NAMESPACE_KERNEL: 771 case MLX5_FLOW_NAMESPACE_BYPASS: 772 max_actions = MLX5_CAP_FLOWTABLE_NIC_RX(dev, max_modify_header_actions); 773 table_type = FS_FT_NIC_RX; 774 break; 775 case MLX5_FLOW_NAMESPACE_EGRESS: 776 #ifdef CONFIG_MLX5_IPSEC 777 case MLX5_FLOW_NAMESPACE_EGRESS_KERNEL: 778 #endif 779 max_actions = MLX5_CAP_FLOWTABLE_NIC_TX(dev, max_modify_header_actions); 780 table_type = FS_FT_NIC_TX; 781 break; 782 case MLX5_FLOW_NAMESPACE_ESW_INGRESS: 783 max_actions = MLX5_CAP_ESW_INGRESS_ACL(dev, max_modify_header_actions); 784 table_type = FS_FT_ESW_INGRESS_ACL; 785 break; 786 case MLX5_FLOW_NAMESPACE_RDMA_TX: 787 max_actions = MLX5_CAP_FLOWTABLE_RDMA_TX(dev, max_modify_header_actions); 788 table_type = FS_FT_RDMA_TX; 789 break; 790 default: 791 return -EOPNOTSUPP; 792 } 793 794 if (num_actions > max_actions) { 795 mlx5_core_warn(dev, "too many modify header actions %d, max supported %d\n", 796 num_actions, max_actions); 797 return -EOPNOTSUPP; 798 } 799 800 actions_size = MLX5_UN_SZ_BYTES(set_add_copy_action_in_auto) * num_actions; 801 inlen = MLX5_ST_SZ_BYTES(alloc_modify_header_context_in) + actions_size; 802 803 in = kzalloc(inlen, GFP_KERNEL); 804 if (!in) 805 return -ENOMEM; 806 807 MLX5_SET(alloc_modify_header_context_in, in, opcode, 808 MLX5_CMD_OP_ALLOC_MODIFY_HEADER_CONTEXT); 809 MLX5_SET(alloc_modify_header_context_in, in, table_type, table_type); 810 MLX5_SET(alloc_modify_header_context_in, in, num_of_actions, num_actions); 811 812 actions_in = MLX5_ADDR_OF(alloc_modify_header_context_in, in, actions); 813 memcpy(actions_in, modify_actions, actions_size); 814 815 err = mlx5_cmd_exec(dev, in, inlen, out, sizeof(out)); 816 817 modify_hdr->id = MLX5_GET(alloc_modify_header_context_out, out, modify_header_id); 818 kfree(in); 819 return err; 820 } 821 822 static void mlx5_cmd_modify_header_dealloc(struct mlx5_flow_root_namespace *ns, 823 struct mlx5_modify_hdr *modify_hdr) 824 { 825 u32 in[MLX5_ST_SZ_DW(dealloc_modify_header_context_in)] = {}; 826 struct mlx5_core_dev *dev = ns->dev; 827 828 MLX5_SET(dealloc_modify_header_context_in, in, opcode, 829 MLX5_CMD_OP_DEALLOC_MODIFY_HEADER_CONTEXT); 830 MLX5_SET(dealloc_modify_header_context_in, in, modify_header_id, 831 modify_hdr->id); 832 833 mlx5_cmd_exec_in(dev, dealloc_modify_header_context, in); 834 } 835 836 static const struct mlx5_flow_cmds mlx5_flow_cmds = { 837 .create_flow_table = mlx5_cmd_create_flow_table, 838 .destroy_flow_table = mlx5_cmd_destroy_flow_table, 839 .modify_flow_table = mlx5_cmd_modify_flow_table, 840 .create_flow_group = mlx5_cmd_create_flow_group, 841 .destroy_flow_group = mlx5_cmd_destroy_flow_group, 842 .create_fte = mlx5_cmd_create_fte, 843 .update_fte = mlx5_cmd_update_fte, 844 .delete_fte = mlx5_cmd_delete_fte, 845 .update_root_ft = mlx5_cmd_update_root_ft, 846 .packet_reformat_alloc = mlx5_cmd_packet_reformat_alloc, 847 .packet_reformat_dealloc = mlx5_cmd_packet_reformat_dealloc, 848 .modify_header_alloc = mlx5_cmd_modify_header_alloc, 849 .modify_header_dealloc = mlx5_cmd_modify_header_dealloc, 850 .set_peer = mlx5_cmd_stub_set_peer, 851 .create_ns = mlx5_cmd_stub_create_ns, 852 .destroy_ns = mlx5_cmd_stub_destroy_ns, 853 }; 854 855 static const struct mlx5_flow_cmds mlx5_flow_cmd_stubs = { 856 .create_flow_table = mlx5_cmd_stub_create_flow_table, 857 .destroy_flow_table = mlx5_cmd_stub_destroy_flow_table, 858 .modify_flow_table = mlx5_cmd_stub_modify_flow_table, 859 .create_flow_group = mlx5_cmd_stub_create_flow_group, 860 .destroy_flow_group = mlx5_cmd_stub_destroy_flow_group, 861 .create_fte = mlx5_cmd_stub_create_fte, 862 .update_fte = mlx5_cmd_stub_update_fte, 863 .delete_fte = mlx5_cmd_stub_delete_fte, 864 .update_root_ft = mlx5_cmd_stub_update_root_ft, 865 .packet_reformat_alloc = mlx5_cmd_stub_packet_reformat_alloc, 866 .packet_reformat_dealloc = mlx5_cmd_stub_packet_reformat_dealloc, 867 .modify_header_alloc = mlx5_cmd_stub_modify_header_alloc, 868 .modify_header_dealloc = mlx5_cmd_stub_modify_header_dealloc, 869 .set_peer = mlx5_cmd_stub_set_peer, 870 .create_ns = mlx5_cmd_stub_create_ns, 871 .destroy_ns = mlx5_cmd_stub_destroy_ns, 872 }; 873 874 const struct mlx5_flow_cmds *mlx5_fs_cmd_get_fw_cmds(void) 875 { 876 return &mlx5_flow_cmds; 877 } 878 879 static const struct mlx5_flow_cmds *mlx5_fs_cmd_get_stub_cmds(void) 880 { 881 return &mlx5_flow_cmd_stubs; 882 } 883 884 const struct mlx5_flow_cmds *mlx5_fs_cmd_get_default(enum fs_flow_table_type type) 885 { 886 switch (type) { 887 case FS_FT_NIC_RX: 888 case FS_FT_ESW_EGRESS_ACL: 889 case FS_FT_ESW_INGRESS_ACL: 890 case FS_FT_FDB: 891 case FS_FT_SNIFFER_RX: 892 case FS_FT_SNIFFER_TX: 893 case FS_FT_NIC_TX: 894 case FS_FT_RDMA_RX: 895 case FS_FT_RDMA_TX: 896 return mlx5_fs_cmd_get_fw_cmds(); 897 default: 898 return mlx5_fs_cmd_get_stub_cmds(); 899 } 900 } 901