1 /* 2 * This file is part of the Chelsio T4 Ethernet driver for Linux. 3 * 4 * Copyright (c) 2016 Chelsio Communications, Inc. All rights reserved. 5 * 6 * This software is available to you under a choice of one of two 7 * licenses. You may choose to be licensed under the terms of the GNU 8 * General Public License (GPL) Version 2, available from the file 9 * COPYING in the main directory of this source tree, or the 10 * OpenIB.org BSD license below: 11 * 12 * Redistribution and use in source and binary forms, with or 13 * without modification, are permitted provided that the following 14 * conditions are met: 15 * 16 * - Redistributions of source code must retain the above 17 * copyright notice, this list of conditions and the following 18 * disclaimer. 19 * 20 * - Redistributions in binary form must reproduce the above 21 * copyright notice, this list of conditions and the following 22 * disclaimer in the documentation and/or other materials 23 * provided with the distribution. 24 * 25 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 26 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 27 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 28 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 29 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 30 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 31 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 32 * SOFTWARE. 33 */ 34 35 #include <net/tc_act/tc_gact.h> 36 #include <net/tc_act/tc_mirred.h> 37 38 #include "cxgb4.h" 39 #include "cxgb4_filter.h" 40 #include "cxgb4_tc_u32_parse.h" 41 #include "cxgb4_tc_u32.h" 42 43 /* Fill ch_filter_specification with parsed match value/mask pair. */ 44 static int fill_match_fields(struct adapter *adap, 45 struct ch_filter_specification *fs, 46 struct tc_cls_u32_offload *cls, 47 const struct cxgb4_match_field *entry, 48 bool next_header) 49 { 50 unsigned int i, j; 51 __be32 val, mask; 52 int off, err; 53 bool found; 54 55 for (i = 0; i < cls->knode.sel->nkeys; i++) { 56 off = cls->knode.sel->keys[i].off; 57 val = cls->knode.sel->keys[i].val; 58 mask = cls->knode.sel->keys[i].mask; 59 60 if (next_header) { 61 /* For next headers, parse only keys with offmask */ 62 if (!cls->knode.sel->keys[i].offmask) 63 continue; 64 } else { 65 /* For the remaining, parse only keys without offmask */ 66 if (cls->knode.sel->keys[i].offmask) 67 continue; 68 } 69 70 found = false; 71 72 for (j = 0; entry[j].val; j++) { 73 if (off == entry[j].off) { 74 found = true; 75 err = entry[j].val(fs, val, mask); 76 if (err) 77 return err; 78 break; 79 } 80 } 81 82 if (!found) 83 return -EINVAL; 84 } 85 86 return 0; 87 } 88 89 /* Fill ch_filter_specification with parsed action. */ 90 static int fill_action_fields(struct adapter *adap, 91 struct ch_filter_specification *fs, 92 struct tc_cls_u32_offload *cls) 93 { 94 unsigned int num_actions = 0; 95 const struct tc_action *a; 96 struct tcf_exts *exts; 97 int i; 98 99 exts = cls->knode.exts; 100 if (!tcf_exts_has_actions(exts)) 101 return -EINVAL; 102 103 tcf_exts_for_each_action(i, a, exts) { 104 /* Don't allow more than one action per rule. */ 105 if (num_actions) 106 return -EINVAL; 107 108 /* Drop in hardware. */ 109 if (is_tcf_gact_shot(a)) { 110 fs->action = FILTER_DROP; 111 num_actions++; 112 continue; 113 } 114 115 /* Re-direct to specified port in hardware. */ 116 if (is_tcf_mirred_egress_redirect(a)) { 117 struct net_device *n_dev, *target_dev; 118 bool found = false; 119 unsigned int i; 120 121 target_dev = tcf_mirred_dev(a); 122 for_each_port(adap, i) { 123 n_dev = adap->port[i]; 124 if (target_dev == n_dev) { 125 fs->action = FILTER_SWITCH; 126 fs->eport = i; 127 found = true; 128 break; 129 } 130 } 131 132 /* Interface doesn't belong to any port of 133 * the underlying hardware. 134 */ 135 if (!found) 136 return -EINVAL; 137 138 num_actions++; 139 continue; 140 } 141 142 /* Un-supported action. */ 143 return -EINVAL; 144 } 145 146 return 0; 147 } 148 149 int cxgb4_config_knode(struct net_device *dev, struct tc_cls_u32_offload *cls) 150 { 151 const struct cxgb4_match_field *start, *link_start = NULL; 152 struct netlink_ext_ack *extack = cls->common.extack; 153 struct adapter *adapter = netdev2adap(dev); 154 __be16 protocol = cls->common.protocol; 155 struct ch_filter_specification fs; 156 struct cxgb4_tc_u32_table *t; 157 struct cxgb4_link *link; 158 u32 uhtid, link_uhtid; 159 bool is_ipv6 = false; 160 u8 inet_family; 161 int filter_id; 162 int ret; 163 164 if (!can_tc_u32_offload(dev)) 165 return -EOPNOTSUPP; 166 167 if (protocol != htons(ETH_P_IP) && protocol != htons(ETH_P_IPV6)) 168 return -EOPNOTSUPP; 169 170 inet_family = (protocol == htons(ETH_P_IPV6)) ? PF_INET6 : PF_INET; 171 172 /* Get a free filter entry TID, where we can insert this new 173 * rule. Only insert rule if its prio doesn't conflict with 174 * existing rules. 175 */ 176 filter_id = cxgb4_get_free_ftid(dev, inet_family, false, 177 TC_U32_NODE(cls->knode.handle)); 178 if (filter_id < 0) { 179 NL_SET_ERR_MSG_MOD(extack, 180 "No free LETCAM index available"); 181 return -ENOMEM; 182 } 183 184 t = adapter->tc_u32; 185 uhtid = TC_U32_USERHTID(cls->knode.handle); 186 link_uhtid = TC_U32_USERHTID(cls->knode.link_handle); 187 188 /* Ensure that uhtid is either root u32 (i.e. 0x800) 189 * or a a valid linked bucket. 190 */ 191 if (uhtid != 0x800 && uhtid >= t->size) 192 return -EINVAL; 193 194 /* Ensure link handle uhtid is sane, if specified. */ 195 if (link_uhtid >= t->size) 196 return -EINVAL; 197 198 memset(&fs, 0, sizeof(fs)); 199 200 if (filter_id < adapter->tids.nhpftids) 201 fs.prio = 1; 202 fs.tc_prio = cls->common.prio; 203 fs.tc_cookie = cls->knode.handle; 204 205 if (protocol == htons(ETH_P_IPV6)) { 206 start = cxgb4_ipv6_fields; 207 is_ipv6 = true; 208 } else { 209 start = cxgb4_ipv4_fields; 210 is_ipv6 = false; 211 } 212 213 if (uhtid != 0x800) { 214 /* Link must exist from root node before insertion. */ 215 if (!t->table[uhtid - 1].link_handle) 216 return -EINVAL; 217 218 /* Link must have a valid supported next header. */ 219 link_start = t->table[uhtid - 1].match_field; 220 if (!link_start) 221 return -EINVAL; 222 } 223 224 /* Parse links and record them for subsequent jumps to valid 225 * next headers. 226 */ 227 if (link_uhtid) { 228 const struct cxgb4_next_header *next; 229 bool found = false; 230 unsigned int i, j; 231 __be32 val, mask; 232 int off; 233 234 if (t->table[link_uhtid - 1].link_handle) { 235 dev_err(adapter->pdev_dev, 236 "Link handle exists for: 0x%x\n", 237 link_uhtid); 238 return -EINVAL; 239 } 240 241 next = is_ipv6 ? cxgb4_ipv6_jumps : cxgb4_ipv4_jumps; 242 243 /* Try to find matches that allow jumps to next header. */ 244 for (i = 0; next[i].jump; i++) { 245 if (next[i].sel.offoff != cls->knode.sel->offoff || 246 next[i].sel.offshift != cls->knode.sel->offshift || 247 next[i].sel.offmask != cls->knode.sel->offmask || 248 next[i].sel.off != cls->knode.sel->off) 249 continue; 250 251 /* Found a possible candidate. Find a key that 252 * matches the corresponding offset, value, and 253 * mask to jump to next header. 254 */ 255 for (j = 0; j < cls->knode.sel->nkeys; j++) { 256 off = cls->knode.sel->keys[j].off; 257 val = cls->knode.sel->keys[j].val; 258 mask = cls->knode.sel->keys[j].mask; 259 260 if (next[i].key.off == off && 261 next[i].key.val == val && 262 next[i].key.mask == mask) { 263 found = true; 264 break; 265 } 266 } 267 268 if (!found) 269 continue; /* Try next candidate. */ 270 271 /* Candidate to jump to next header found. 272 * Translate all keys to internal specification 273 * and store them in jump table. This spec is copied 274 * later to set the actual filters. 275 */ 276 ret = fill_match_fields(adapter, &fs, cls, 277 start, false); 278 if (ret) 279 goto out; 280 281 link = &t->table[link_uhtid - 1]; 282 link->match_field = next[i].jump; 283 link->link_handle = cls->knode.handle; 284 memcpy(&link->fs, &fs, sizeof(fs)); 285 break; 286 } 287 288 /* No candidate found to jump to next header. */ 289 if (!found) 290 return -EINVAL; 291 292 return 0; 293 } 294 295 /* Fill ch_filter_specification match fields to be shipped to hardware. 296 * Copy the linked spec (if any) first. And then update the spec as 297 * needed. 298 */ 299 if (uhtid != 0x800 && t->table[uhtid - 1].link_handle) { 300 /* Copy linked ch_filter_specification */ 301 memcpy(&fs, &t->table[uhtid - 1].fs, sizeof(fs)); 302 ret = fill_match_fields(adapter, &fs, cls, 303 link_start, true); 304 if (ret) 305 goto out; 306 } 307 308 ret = fill_match_fields(adapter, &fs, cls, start, false); 309 if (ret) 310 goto out; 311 312 /* Fill ch_filter_specification action fields to be shipped to 313 * hardware. 314 */ 315 ret = fill_action_fields(adapter, &fs, cls); 316 if (ret) 317 goto out; 318 319 /* The filter spec has been completely built from the info 320 * provided from u32. We now set some default fields in the 321 * spec for sanity. 322 */ 323 324 /* Match only packets coming from the ingress port where this 325 * filter will be created. 326 */ 327 fs.val.iport = netdev2pinfo(dev)->port_id; 328 fs.mask.iport = ~0; 329 330 /* Enable filter hit counts. */ 331 fs.hitcnts = 1; 332 333 /* Set type of filter - IPv6 or IPv4 */ 334 fs.type = is_ipv6 ? 1 : 0; 335 336 /* Set the filter */ 337 ret = cxgb4_set_filter(dev, filter_id, &fs); 338 if (ret) 339 goto out; 340 341 /* If this is a linked bucket, then set the corresponding 342 * entry in the bitmap to mark it as belonging to this linked 343 * bucket. 344 */ 345 if (uhtid != 0x800 && t->table[uhtid - 1].link_handle) 346 set_bit(filter_id, t->table[uhtid - 1].tid_map); 347 348 out: 349 return ret; 350 } 351 352 int cxgb4_delete_knode(struct net_device *dev, struct tc_cls_u32_offload *cls) 353 { 354 struct adapter *adapter = netdev2adap(dev); 355 unsigned int filter_id, max_tids, i, j; 356 struct cxgb4_link *link = NULL; 357 struct cxgb4_tc_u32_table *t; 358 struct filter_entry *f; 359 bool found = false; 360 u32 handle, uhtid; 361 u8 nslots; 362 int ret; 363 364 if (!can_tc_u32_offload(dev)) 365 return -EOPNOTSUPP; 366 367 /* Fetch the location to delete the filter. */ 368 max_tids = adapter->tids.nhpftids + adapter->tids.nftids; 369 370 spin_lock_bh(&adapter->tids.ftid_lock); 371 filter_id = 0; 372 while (filter_id < max_tids) { 373 if (filter_id < adapter->tids.nhpftids) { 374 i = filter_id; 375 f = &adapter->tids.hpftid_tab[i]; 376 if (f->valid && f->fs.tc_cookie == cls->knode.handle) { 377 found = true; 378 break; 379 } 380 381 i = find_next_bit(adapter->tids.hpftid_bmap, 382 adapter->tids.nhpftids, i + 1); 383 if (i >= adapter->tids.nhpftids) { 384 filter_id = adapter->tids.nhpftids; 385 continue; 386 } 387 388 filter_id = i; 389 } else { 390 i = filter_id - adapter->tids.nhpftids; 391 f = &adapter->tids.ftid_tab[i]; 392 if (f->valid && f->fs.tc_cookie == cls->knode.handle) { 393 found = true; 394 break; 395 } 396 397 i = find_next_bit(adapter->tids.ftid_bmap, 398 adapter->tids.nftids, i + 1); 399 if (i >= adapter->tids.nftids) 400 break; 401 402 filter_id = i + adapter->tids.nhpftids; 403 } 404 405 nslots = 0; 406 if (f->fs.type) { 407 nslots++; 408 if (CHELSIO_CHIP_VERSION(adapter->params.chip) < 409 CHELSIO_T6) 410 nslots += 2; 411 } 412 413 filter_id += nslots; 414 } 415 spin_unlock_bh(&adapter->tids.ftid_lock); 416 417 if (!found) 418 return -ERANGE; 419 420 t = adapter->tc_u32; 421 handle = cls->knode.handle; 422 uhtid = TC_U32_USERHTID(cls->knode.handle); 423 424 /* Ensure that uhtid is either root u32 (i.e. 0x800) 425 * or a a valid linked bucket. 426 */ 427 if (uhtid != 0x800 && uhtid >= t->size) 428 return -EINVAL; 429 430 /* Delete the specified filter */ 431 if (uhtid != 0x800) { 432 link = &t->table[uhtid - 1]; 433 if (!link->link_handle) 434 return -EINVAL; 435 436 if (!test_bit(filter_id, link->tid_map)) 437 return -EINVAL; 438 } 439 440 ret = cxgb4_del_filter(dev, filter_id, NULL); 441 if (ret) 442 goto out; 443 444 if (link) 445 clear_bit(filter_id, link->tid_map); 446 447 /* If a link is being deleted, then delete all filters 448 * associated with the link. 449 */ 450 for (i = 0; i < t->size; i++) { 451 link = &t->table[i]; 452 453 if (link->link_handle == handle) { 454 for (j = 0; j < max_tids; j++) { 455 if (!test_bit(j, link->tid_map)) 456 continue; 457 458 ret = __cxgb4_del_filter(dev, j, NULL, NULL); 459 if (ret) 460 goto out; 461 462 clear_bit(j, link->tid_map); 463 } 464 465 /* Clear the link state */ 466 link->match_field = NULL; 467 link->link_handle = 0; 468 memset(&link->fs, 0, sizeof(link->fs)); 469 break; 470 } 471 } 472 473 out: 474 return ret; 475 } 476 477 void cxgb4_cleanup_tc_u32(struct adapter *adap) 478 { 479 struct cxgb4_tc_u32_table *t; 480 unsigned int i; 481 482 if (!adap->tc_u32) 483 return; 484 485 /* Free up all allocated memory. */ 486 t = adap->tc_u32; 487 for (i = 0; i < t->size; i++) { 488 struct cxgb4_link *link = &t->table[i]; 489 490 kvfree(link->tid_map); 491 } 492 kvfree(adap->tc_u32); 493 } 494 495 struct cxgb4_tc_u32_table *cxgb4_init_tc_u32(struct adapter *adap) 496 { 497 unsigned int max_tids = adap->tids.nftids + adap->tids.nhpftids; 498 struct cxgb4_tc_u32_table *t; 499 unsigned int i; 500 501 if (!max_tids) 502 return NULL; 503 504 t = kvzalloc(struct_size(t, table, max_tids), GFP_KERNEL); 505 if (!t) 506 return NULL; 507 508 t->size = max_tids; 509 510 for (i = 0; i < t->size; i++) { 511 struct cxgb4_link *link = &t->table[i]; 512 unsigned int bmap_size; 513 514 bmap_size = BITS_TO_LONGS(max_tids); 515 link->tid_map = kvcalloc(bmap_size, sizeof(unsigned long), 516 GFP_KERNEL); 517 if (!link->tid_map) 518 goto out_no_mem; 519 bitmap_zero(link->tid_map, max_tids); 520 } 521 522 return t; 523 524 out_no_mem: 525 for (i = 0; i < t->size; i++) { 526 struct cxgb4_link *link = &t->table[i]; 527 528 if (link->tid_map) 529 kvfree(link->tid_map); 530 } 531 532 if (t) 533 kvfree(t); 534 535 return NULL; 536 } 537