1 /* 2 * Copyright (C) 2016 Netronome Systems, Inc. 3 * 4 * This software is dual licensed under the GNU General License Version 2, 5 * June 1991 as shown in the file COPYING in the top-level directory of this 6 * source tree or the BSD 2-Clause License provided below. You have the 7 * option to license this software under the complete terms of either license. 8 * 9 * The BSD 2-Clause License: 10 * 11 * Redistribution and use in source and binary forms, with or 12 * without modification, are permitted provided that the following 13 * conditions are met: 14 * 15 * 1. Redistributions of source code must retain the above 16 * copyright notice, this list of conditions and the following 17 * disclaimer. 18 * 19 * 2. Redistributions in binary form must reproduce the above 20 * copyright notice, this list of conditions and the following 21 * disclaimer in the documentation and/or other materials 22 * provided with the distribution. 23 * 24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 31 * SOFTWARE. 32 */ 33 34 #define pr_fmt(fmt) "NFP net bpf: " fmt 35 36 #include <linux/kernel.h> 37 #include <linux/bpf.h> 38 #include <linux/filter.h> 39 #include <linux/pkt_cls.h> 40 #include <linux/unistd.h> 41 42 #include "main.h" 43 #include "../nfp_asm.h" 44 45 /* --- NFP prog --- */ 46 /* Foreach "multiple" entries macros provide pos and next<n> pointers. 47 * It's safe to modify the next pointers (but not pos). 48 */ 49 #define nfp_for_each_insn_walk2(nfp_prog, pos, next) \ 50 for (pos = list_first_entry(&(nfp_prog)->insns, typeof(*pos), l), \ 51 next = list_next_entry(pos, l); \ 52 &(nfp_prog)->insns != &pos->l && \ 53 &(nfp_prog)->insns != &next->l; \ 54 pos = nfp_meta_next(pos), \ 55 next = nfp_meta_next(pos)) 56 57 #define nfp_for_each_insn_walk3(nfp_prog, pos, next, next2) \ 58 for (pos = list_first_entry(&(nfp_prog)->insns, typeof(*pos), l), \ 59 next = list_next_entry(pos, l), \ 60 next2 = list_next_entry(next, l); \ 61 &(nfp_prog)->insns != &pos->l && \ 62 &(nfp_prog)->insns != &next->l && \ 63 &(nfp_prog)->insns != &next2->l; \ 64 pos = nfp_meta_next(pos), \ 65 next = nfp_meta_next(pos), \ 66 next2 = nfp_meta_next(next)) 67 68 static bool 69 nfp_meta_has_next(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 70 { 71 return meta->l.next != &nfp_prog->insns; 72 } 73 74 static bool 75 nfp_meta_has_prev(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 76 { 77 return meta->l.prev != &nfp_prog->insns; 78 } 79 80 static void nfp_prog_free(struct nfp_prog *nfp_prog) 81 { 82 struct nfp_insn_meta *meta, *tmp; 83 84 list_for_each_entry_safe(meta, tmp, &nfp_prog->insns, l) { 85 list_del(&meta->l); 86 kfree(meta); 87 } 88 kfree(nfp_prog); 89 } 90 91 static void nfp_prog_push(struct nfp_prog *nfp_prog, u64 insn) 92 { 93 if (nfp_prog->__prog_alloc_len == nfp_prog->prog_len) { 94 nfp_prog->error = -ENOSPC; 95 return; 96 } 97 98 nfp_prog->prog[nfp_prog->prog_len] = insn; 99 nfp_prog->prog_len++; 100 } 101 102 static unsigned int nfp_prog_current_offset(struct nfp_prog *nfp_prog) 103 { 104 return nfp_prog->start_off + nfp_prog->prog_len; 105 } 106 107 static unsigned int 108 nfp_prog_offset_to_index(struct nfp_prog *nfp_prog, unsigned int offset) 109 { 110 return offset - nfp_prog->start_off; 111 } 112 113 /* --- Emitters --- */ 114 static void 115 __emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, 116 u8 mode, u8 xfer, u8 areg, u8 breg, u8 size, bool sync) 117 { 118 enum cmd_ctx_swap ctx; 119 u64 insn; 120 121 if (sync) 122 ctx = CMD_CTX_SWAP; 123 else 124 ctx = CMD_CTX_NO_SWAP; 125 126 insn = FIELD_PREP(OP_CMD_A_SRC, areg) | 127 FIELD_PREP(OP_CMD_CTX, ctx) | 128 FIELD_PREP(OP_CMD_B_SRC, breg) | 129 FIELD_PREP(OP_CMD_TOKEN, cmd_tgt_act[op].token) | 130 FIELD_PREP(OP_CMD_XFER, xfer) | 131 FIELD_PREP(OP_CMD_CNT, size) | 132 FIELD_PREP(OP_CMD_SIG, sync) | 133 FIELD_PREP(OP_CMD_TGT_CMD, cmd_tgt_act[op].tgt_cmd) | 134 FIELD_PREP(OP_CMD_MODE, mode); 135 136 nfp_prog_push(nfp_prog, insn); 137 } 138 139 static void 140 emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, 141 u8 mode, u8 xfer, swreg lreg, swreg rreg, u8 size, bool sync) 142 { 143 struct nfp_insn_re_regs reg; 144 int err; 145 146 err = swreg_to_restricted(reg_none(), lreg, rreg, ®, false); 147 if (err) { 148 nfp_prog->error = err; 149 return; 150 } 151 if (reg.swap) { 152 pr_err("cmd can't swap arguments\n"); 153 nfp_prog->error = -EFAULT; 154 return; 155 } 156 if (reg.dst_lmextn || reg.src_lmextn) { 157 pr_err("cmd can't use LMextn\n"); 158 nfp_prog->error = -EFAULT; 159 return; 160 } 161 162 __emit_cmd(nfp_prog, op, mode, xfer, reg.areg, reg.breg, size, sync); 163 } 164 165 static void 166 __emit_br(struct nfp_prog *nfp_prog, enum br_mask mask, enum br_ev_pip ev_pip, 167 enum br_ctx_signal_state css, u16 addr, u8 defer) 168 { 169 u16 addr_lo, addr_hi; 170 u64 insn; 171 172 addr_lo = addr & (OP_BR_ADDR_LO >> __bf_shf(OP_BR_ADDR_LO)); 173 addr_hi = addr != addr_lo; 174 175 insn = OP_BR_BASE | 176 FIELD_PREP(OP_BR_MASK, mask) | 177 FIELD_PREP(OP_BR_EV_PIP, ev_pip) | 178 FIELD_PREP(OP_BR_CSS, css) | 179 FIELD_PREP(OP_BR_DEFBR, defer) | 180 FIELD_PREP(OP_BR_ADDR_LO, addr_lo) | 181 FIELD_PREP(OP_BR_ADDR_HI, addr_hi); 182 183 nfp_prog_push(nfp_prog, insn); 184 } 185 186 static void emit_br_def(struct nfp_prog *nfp_prog, u16 addr, u8 defer) 187 { 188 if (defer > 2) { 189 pr_err("BUG: branch defer out of bounds %d\n", defer); 190 nfp_prog->error = -EFAULT; 191 return; 192 } 193 __emit_br(nfp_prog, BR_UNC, BR_EV_PIP_UNCOND, BR_CSS_NONE, addr, defer); 194 } 195 196 static void 197 emit_br(struct nfp_prog *nfp_prog, enum br_mask mask, u16 addr, u8 defer) 198 { 199 __emit_br(nfp_prog, mask, 200 mask != BR_UNC ? BR_EV_PIP_COND : BR_EV_PIP_UNCOND, 201 BR_CSS_NONE, addr, defer); 202 } 203 204 static void 205 __emit_br_byte(struct nfp_prog *nfp_prog, u8 areg, u8 breg, bool imm8, 206 u8 byte, bool equal, u16 addr, u8 defer, bool src_lmextn) 207 { 208 u16 addr_lo, addr_hi; 209 u64 insn; 210 211 addr_lo = addr & (OP_BB_ADDR_LO >> __bf_shf(OP_BB_ADDR_LO)); 212 addr_hi = addr != addr_lo; 213 214 insn = OP_BBYTE_BASE | 215 FIELD_PREP(OP_BB_A_SRC, areg) | 216 FIELD_PREP(OP_BB_BYTE, byte) | 217 FIELD_PREP(OP_BB_B_SRC, breg) | 218 FIELD_PREP(OP_BB_I8, imm8) | 219 FIELD_PREP(OP_BB_EQ, equal) | 220 FIELD_PREP(OP_BB_DEFBR, defer) | 221 FIELD_PREP(OP_BB_ADDR_LO, addr_lo) | 222 FIELD_PREP(OP_BB_ADDR_HI, addr_hi) | 223 FIELD_PREP(OP_BB_SRC_LMEXTN, src_lmextn); 224 225 nfp_prog_push(nfp_prog, insn); 226 } 227 228 static void 229 emit_br_byte_neq(struct nfp_prog *nfp_prog, 230 swreg src, u8 imm, u8 byte, u16 addr, u8 defer) 231 { 232 struct nfp_insn_re_regs reg; 233 int err; 234 235 err = swreg_to_restricted(reg_none(), src, reg_imm(imm), ®, true); 236 if (err) { 237 nfp_prog->error = err; 238 return; 239 } 240 241 __emit_br_byte(nfp_prog, reg.areg, reg.breg, reg.i8, byte, false, addr, 242 defer, reg.src_lmextn); 243 } 244 245 static void 246 __emit_immed(struct nfp_prog *nfp_prog, u16 areg, u16 breg, u16 imm_hi, 247 enum immed_width width, bool invert, 248 enum immed_shift shift, bool wr_both, 249 bool dst_lmextn, bool src_lmextn) 250 { 251 u64 insn; 252 253 insn = OP_IMMED_BASE | 254 FIELD_PREP(OP_IMMED_A_SRC, areg) | 255 FIELD_PREP(OP_IMMED_B_SRC, breg) | 256 FIELD_PREP(OP_IMMED_IMM, imm_hi) | 257 FIELD_PREP(OP_IMMED_WIDTH, width) | 258 FIELD_PREP(OP_IMMED_INV, invert) | 259 FIELD_PREP(OP_IMMED_SHIFT, shift) | 260 FIELD_PREP(OP_IMMED_WR_AB, wr_both) | 261 FIELD_PREP(OP_IMMED_SRC_LMEXTN, src_lmextn) | 262 FIELD_PREP(OP_IMMED_DST_LMEXTN, dst_lmextn); 263 264 nfp_prog_push(nfp_prog, insn); 265 } 266 267 static void 268 emit_immed(struct nfp_prog *nfp_prog, swreg dst, u16 imm, 269 enum immed_width width, bool invert, enum immed_shift shift) 270 { 271 struct nfp_insn_ur_regs reg; 272 int err; 273 274 if (swreg_type(dst) == NN_REG_IMM) { 275 nfp_prog->error = -EFAULT; 276 return; 277 } 278 279 err = swreg_to_unrestricted(dst, dst, reg_imm(imm & 0xff), ®); 280 if (err) { 281 nfp_prog->error = err; 282 return; 283 } 284 285 __emit_immed(nfp_prog, reg.areg, reg.breg, imm >> 8, width, 286 invert, shift, reg.wr_both, 287 reg.dst_lmextn, reg.src_lmextn); 288 } 289 290 static void 291 __emit_shf(struct nfp_prog *nfp_prog, u16 dst, enum alu_dst_ab dst_ab, 292 enum shf_sc sc, u8 shift, 293 u16 areg, enum shf_op op, u16 breg, bool i8, bool sw, bool wr_both, 294 bool dst_lmextn, bool src_lmextn) 295 { 296 u64 insn; 297 298 if (!FIELD_FIT(OP_SHF_SHIFT, shift)) { 299 nfp_prog->error = -EFAULT; 300 return; 301 } 302 303 if (sc == SHF_SC_L_SHF) 304 shift = 32 - shift; 305 306 insn = OP_SHF_BASE | 307 FIELD_PREP(OP_SHF_A_SRC, areg) | 308 FIELD_PREP(OP_SHF_SC, sc) | 309 FIELD_PREP(OP_SHF_B_SRC, breg) | 310 FIELD_PREP(OP_SHF_I8, i8) | 311 FIELD_PREP(OP_SHF_SW, sw) | 312 FIELD_PREP(OP_SHF_DST, dst) | 313 FIELD_PREP(OP_SHF_SHIFT, shift) | 314 FIELD_PREP(OP_SHF_OP, op) | 315 FIELD_PREP(OP_SHF_DST_AB, dst_ab) | 316 FIELD_PREP(OP_SHF_WR_AB, wr_both) | 317 FIELD_PREP(OP_SHF_SRC_LMEXTN, src_lmextn) | 318 FIELD_PREP(OP_SHF_DST_LMEXTN, dst_lmextn); 319 320 nfp_prog_push(nfp_prog, insn); 321 } 322 323 static void 324 emit_shf(struct nfp_prog *nfp_prog, swreg dst, 325 swreg lreg, enum shf_op op, swreg rreg, enum shf_sc sc, u8 shift) 326 { 327 struct nfp_insn_re_regs reg; 328 int err; 329 330 err = swreg_to_restricted(dst, lreg, rreg, ®, true); 331 if (err) { 332 nfp_prog->error = err; 333 return; 334 } 335 336 __emit_shf(nfp_prog, reg.dst, reg.dst_ab, sc, shift, 337 reg.areg, op, reg.breg, reg.i8, reg.swap, reg.wr_both, 338 reg.dst_lmextn, reg.src_lmextn); 339 } 340 341 static void 342 __emit_alu(struct nfp_prog *nfp_prog, u16 dst, enum alu_dst_ab dst_ab, 343 u16 areg, enum alu_op op, u16 breg, bool swap, bool wr_both, 344 bool dst_lmextn, bool src_lmextn) 345 { 346 u64 insn; 347 348 insn = OP_ALU_BASE | 349 FIELD_PREP(OP_ALU_A_SRC, areg) | 350 FIELD_PREP(OP_ALU_B_SRC, breg) | 351 FIELD_PREP(OP_ALU_DST, dst) | 352 FIELD_PREP(OP_ALU_SW, swap) | 353 FIELD_PREP(OP_ALU_OP, op) | 354 FIELD_PREP(OP_ALU_DST_AB, dst_ab) | 355 FIELD_PREP(OP_ALU_WR_AB, wr_both) | 356 FIELD_PREP(OP_ALU_SRC_LMEXTN, src_lmextn) | 357 FIELD_PREP(OP_ALU_DST_LMEXTN, dst_lmextn); 358 359 nfp_prog_push(nfp_prog, insn); 360 } 361 362 static void 363 emit_alu(struct nfp_prog *nfp_prog, swreg dst, 364 swreg lreg, enum alu_op op, swreg rreg) 365 { 366 struct nfp_insn_ur_regs reg; 367 int err; 368 369 err = swreg_to_unrestricted(dst, lreg, rreg, ®); 370 if (err) { 371 nfp_prog->error = err; 372 return; 373 } 374 375 __emit_alu(nfp_prog, reg.dst, reg.dst_ab, 376 reg.areg, op, reg.breg, reg.swap, reg.wr_both, 377 reg.dst_lmextn, reg.src_lmextn); 378 } 379 380 static void 381 __emit_ld_field(struct nfp_prog *nfp_prog, enum shf_sc sc, 382 u8 areg, u8 bmask, u8 breg, u8 shift, bool imm8, 383 bool zero, bool swap, bool wr_both, 384 bool dst_lmextn, bool src_lmextn) 385 { 386 u64 insn; 387 388 insn = OP_LDF_BASE | 389 FIELD_PREP(OP_LDF_A_SRC, areg) | 390 FIELD_PREP(OP_LDF_SC, sc) | 391 FIELD_PREP(OP_LDF_B_SRC, breg) | 392 FIELD_PREP(OP_LDF_I8, imm8) | 393 FIELD_PREP(OP_LDF_SW, swap) | 394 FIELD_PREP(OP_LDF_ZF, zero) | 395 FIELD_PREP(OP_LDF_BMASK, bmask) | 396 FIELD_PREP(OP_LDF_SHF, shift) | 397 FIELD_PREP(OP_LDF_WR_AB, wr_both) | 398 FIELD_PREP(OP_LDF_SRC_LMEXTN, src_lmextn) | 399 FIELD_PREP(OP_LDF_DST_LMEXTN, dst_lmextn); 400 401 nfp_prog_push(nfp_prog, insn); 402 } 403 404 static void 405 emit_ld_field_any(struct nfp_prog *nfp_prog, swreg dst, u8 bmask, swreg src, 406 enum shf_sc sc, u8 shift, bool zero) 407 { 408 struct nfp_insn_re_regs reg; 409 int err; 410 411 /* Note: ld_field is special as it uses one of the src regs as dst */ 412 err = swreg_to_restricted(dst, dst, src, ®, true); 413 if (err) { 414 nfp_prog->error = err; 415 return; 416 } 417 418 __emit_ld_field(nfp_prog, sc, reg.areg, bmask, reg.breg, shift, 419 reg.i8, zero, reg.swap, reg.wr_both, 420 reg.dst_lmextn, reg.src_lmextn); 421 } 422 423 static void 424 emit_ld_field(struct nfp_prog *nfp_prog, swreg dst, u8 bmask, swreg src, 425 enum shf_sc sc, u8 shift) 426 { 427 emit_ld_field_any(nfp_prog, dst, bmask, src, sc, shift, false); 428 } 429 430 static void emit_nop(struct nfp_prog *nfp_prog) 431 { 432 __emit_immed(nfp_prog, UR_REG_IMM, UR_REG_IMM, 0, 0, 0, 0, 0, 0, 0); 433 } 434 435 /* --- Wrappers --- */ 436 static bool pack_immed(u32 imm, u16 *val, enum immed_shift *shift) 437 { 438 if (!(imm & 0xffff0000)) { 439 *val = imm; 440 *shift = IMMED_SHIFT_0B; 441 } else if (!(imm & 0xff0000ff)) { 442 *val = imm >> 8; 443 *shift = IMMED_SHIFT_1B; 444 } else if (!(imm & 0x0000ffff)) { 445 *val = imm >> 16; 446 *shift = IMMED_SHIFT_2B; 447 } else { 448 return false; 449 } 450 451 return true; 452 } 453 454 static void wrp_immed(struct nfp_prog *nfp_prog, swreg dst, u32 imm) 455 { 456 enum immed_shift shift; 457 u16 val; 458 459 if (pack_immed(imm, &val, &shift)) { 460 emit_immed(nfp_prog, dst, val, IMMED_WIDTH_ALL, false, shift); 461 } else if (pack_immed(~imm, &val, &shift)) { 462 emit_immed(nfp_prog, dst, val, IMMED_WIDTH_ALL, true, shift); 463 } else { 464 emit_immed(nfp_prog, dst, imm & 0xffff, IMMED_WIDTH_ALL, 465 false, IMMED_SHIFT_0B); 466 emit_immed(nfp_prog, dst, imm >> 16, IMMED_WIDTH_WORD, 467 false, IMMED_SHIFT_2B); 468 } 469 } 470 471 /* ur_load_imm_any() - encode immediate or use tmp register (unrestricted) 472 * If the @imm is small enough encode it directly in operand and return 473 * otherwise load @imm to a spare register and return its encoding. 474 */ 475 static swreg ur_load_imm_any(struct nfp_prog *nfp_prog, u32 imm, swreg tmp_reg) 476 { 477 if (FIELD_FIT(UR_REG_IMM_MAX, imm)) 478 return reg_imm(imm); 479 480 wrp_immed(nfp_prog, tmp_reg, imm); 481 return tmp_reg; 482 } 483 484 /* re_load_imm_any() - encode immediate or use tmp register (restricted) 485 * If the @imm is small enough encode it directly in operand and return 486 * otherwise load @imm to a spare register and return its encoding. 487 */ 488 static swreg re_load_imm_any(struct nfp_prog *nfp_prog, u32 imm, swreg tmp_reg) 489 { 490 if (FIELD_FIT(RE_REG_IMM_MAX, imm)) 491 return reg_imm(imm); 492 493 wrp_immed(nfp_prog, tmp_reg, imm); 494 return tmp_reg; 495 } 496 497 static void wrp_nops(struct nfp_prog *nfp_prog, unsigned int count) 498 { 499 while (count--) 500 emit_nop(nfp_prog); 501 } 502 503 static void 504 wrp_br_special(struct nfp_prog *nfp_prog, enum br_mask mask, 505 enum br_special special) 506 { 507 emit_br(nfp_prog, mask, 0, 0); 508 509 nfp_prog->prog[nfp_prog->prog_len - 1] |= 510 FIELD_PREP(OP_BR_SPECIAL, special); 511 } 512 513 static void wrp_mov(struct nfp_prog *nfp_prog, swreg dst, swreg src) 514 { 515 emit_alu(nfp_prog, dst, reg_none(), ALU_OP_NONE, src); 516 } 517 518 static void wrp_reg_mov(struct nfp_prog *nfp_prog, u16 dst, u16 src) 519 { 520 wrp_mov(nfp_prog, reg_both(dst), reg_b(src)); 521 } 522 523 static int 524 data_ld(struct nfp_prog *nfp_prog, swreg offset, u8 dst_gpr, int size) 525 { 526 unsigned int i; 527 u16 shift, sz; 528 529 /* We load the value from the address indicated in @offset and then 530 * shift out the data we don't need. Note: this is big endian! 531 */ 532 sz = max(size, 4); 533 shift = size < 4 ? 4 - size : 0; 534 535 emit_cmd(nfp_prog, CMD_TGT_READ8, CMD_MODE_32b, 0, 536 pptr_reg(nfp_prog), offset, sz - 1, true); 537 538 i = 0; 539 if (shift) 540 emit_shf(nfp_prog, reg_both(dst_gpr), reg_none(), SHF_OP_NONE, 541 reg_xfer(0), SHF_SC_R_SHF, shift * 8); 542 else 543 for (; i * 4 < size; i++) 544 wrp_mov(nfp_prog, reg_both(dst_gpr + i), reg_xfer(i)); 545 546 if (i < 2) 547 wrp_immed(nfp_prog, reg_both(dst_gpr + 1), 0); 548 549 return 0; 550 } 551 552 static int 553 data_ld_host_order(struct nfp_prog *nfp_prog, u8 src_gpr, swreg offset, 554 u8 dst_gpr, int size) 555 { 556 unsigned int i; 557 u8 mask, sz; 558 559 /* We load the value from the address indicated in @offset and then 560 * mask out the data we don't need. Note: this is little endian! 561 */ 562 sz = max(size, 4); 563 mask = size < 4 ? GENMASK(size - 1, 0) : 0; 564 565 emit_cmd(nfp_prog, CMD_TGT_READ32_SWAP, CMD_MODE_32b, 0, 566 reg_a(src_gpr), offset, sz / 4 - 1, true); 567 568 i = 0; 569 if (mask) 570 emit_ld_field_any(nfp_prog, reg_both(dst_gpr), mask, 571 reg_xfer(0), SHF_SC_NONE, 0, true); 572 else 573 for (; i * 4 < size; i++) 574 wrp_mov(nfp_prog, reg_both(dst_gpr + i), reg_xfer(i)); 575 576 if (i < 2) 577 wrp_immed(nfp_prog, reg_both(dst_gpr + 1), 0); 578 579 return 0; 580 } 581 582 static int 583 construct_data_ind_ld(struct nfp_prog *nfp_prog, u16 offset, u16 src, u8 size) 584 { 585 swreg tmp_reg; 586 587 /* Calculate the true offset (src_reg + imm) */ 588 tmp_reg = ur_load_imm_any(nfp_prog, offset, imm_b(nfp_prog)); 589 emit_alu(nfp_prog, imm_both(nfp_prog), reg_a(src), ALU_OP_ADD, tmp_reg); 590 591 /* Check packet length (size guaranteed to fit b/c it's u8) */ 592 emit_alu(nfp_prog, imm_a(nfp_prog), 593 imm_a(nfp_prog), ALU_OP_ADD, reg_imm(size)); 594 emit_alu(nfp_prog, reg_none(), 595 plen_reg(nfp_prog), ALU_OP_SUB, imm_a(nfp_prog)); 596 wrp_br_special(nfp_prog, BR_BLO, OP_BR_GO_ABORT); 597 598 /* Load data */ 599 return data_ld(nfp_prog, imm_b(nfp_prog), 0, size); 600 } 601 602 static int construct_data_ld(struct nfp_prog *nfp_prog, u16 offset, u8 size) 603 { 604 swreg tmp_reg; 605 606 /* Check packet length */ 607 tmp_reg = ur_load_imm_any(nfp_prog, offset + size, imm_a(nfp_prog)); 608 emit_alu(nfp_prog, reg_none(), plen_reg(nfp_prog), ALU_OP_SUB, tmp_reg); 609 wrp_br_special(nfp_prog, BR_BLO, OP_BR_GO_ABORT); 610 611 /* Load data */ 612 tmp_reg = re_load_imm_any(nfp_prog, offset, imm_b(nfp_prog)); 613 return data_ld(nfp_prog, tmp_reg, 0, size); 614 } 615 616 static int 617 data_stx_host_order(struct nfp_prog *nfp_prog, u8 dst_gpr, swreg offset, 618 u8 src_gpr, u8 size) 619 { 620 unsigned int i; 621 622 for (i = 0; i * 4 < size; i++) 623 wrp_mov(nfp_prog, reg_xfer(i), reg_a(src_gpr + i)); 624 625 emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 626 reg_a(dst_gpr), offset, size - 1, true); 627 628 return 0; 629 } 630 631 static int 632 data_st_host_order(struct nfp_prog *nfp_prog, u8 dst_gpr, swreg offset, 633 u64 imm, u8 size) 634 { 635 wrp_immed(nfp_prog, reg_xfer(0), imm); 636 if (size == 8) 637 wrp_immed(nfp_prog, reg_xfer(1), imm >> 32); 638 639 emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 640 reg_a(dst_gpr), offset, size - 1, true); 641 642 return 0; 643 } 644 645 typedef int 646 (*lmem_step)(struct nfp_prog *nfp_prog, u8 gpr, u8 gpr_byte, s32 off, 647 unsigned int size, bool first, bool new_gpr, bool last); 648 649 static int 650 wrp_lmem_load(struct nfp_prog *nfp_prog, u8 dst, u8 dst_byte, s32 off, 651 unsigned int size, bool first, bool new_gpr, bool last) 652 { 653 u32 idx, src_byte; 654 enum shf_sc sc; 655 swreg reg; 656 int shf; 657 u8 mask; 658 659 if (WARN_ON_ONCE(dst_byte + size > 4 || off % 4 + size > 4)) 660 return -EOPNOTSUPP; 661 662 idx = off / 4; 663 664 /* Move the entire word */ 665 if (size == 4) { 666 wrp_mov(nfp_prog, reg_both(dst), reg_lm(0, idx)); 667 return 0; 668 } 669 670 src_byte = off % 4; 671 672 mask = (1 << size) - 1; 673 mask <<= dst_byte; 674 675 if (WARN_ON_ONCE(mask > 0xf)) 676 return -EOPNOTSUPP; 677 678 shf = abs(src_byte - dst_byte) * 8; 679 if (src_byte == dst_byte) { 680 sc = SHF_SC_NONE; 681 } else if (src_byte < dst_byte) { 682 shf = 32 - shf; 683 sc = SHF_SC_L_SHF; 684 } else { 685 sc = SHF_SC_R_SHF; 686 } 687 688 /* ld_field can address fewer indexes, if offset too large do RMW. 689 * Because we RMV twice we waste 2 cycles on unaligned 8 byte writes. 690 */ 691 if (idx <= RE_REG_LM_IDX_MAX) { 692 reg = reg_lm(0, idx); 693 } else { 694 reg = imm_a(nfp_prog); 695 /* If it's not the first part of the load and we start a new GPR 696 * that means we are loading a second part of the LMEM word into 697 * a new GPR. IOW we've already looked that LMEM word and 698 * therefore it has been loaded into imm_a(). 699 */ 700 if (first || !new_gpr) 701 wrp_mov(nfp_prog, reg, reg_lm(0, idx)); 702 } 703 704 emit_ld_field_any(nfp_prog, reg_both(dst), mask, reg, sc, shf, new_gpr); 705 706 return 0; 707 } 708 709 static int 710 wrp_lmem_store(struct nfp_prog *nfp_prog, u8 src, u8 src_byte, s32 off, 711 unsigned int size, bool first, bool new_gpr, bool last) 712 { 713 u32 idx, dst_byte; 714 enum shf_sc sc; 715 swreg reg; 716 int shf; 717 u8 mask; 718 719 if (WARN_ON_ONCE(src_byte + size > 4 || off % 4 + size > 4)) 720 return -EOPNOTSUPP; 721 722 idx = off / 4; 723 724 /* Move the entire word */ 725 if (size == 4) { 726 wrp_mov(nfp_prog, reg_lm(0, idx), reg_b(src)); 727 return 0; 728 } 729 730 dst_byte = off % 4; 731 732 mask = (1 << size) - 1; 733 mask <<= dst_byte; 734 735 if (WARN_ON_ONCE(mask > 0xf)) 736 return -EOPNOTSUPP; 737 738 shf = abs(src_byte - dst_byte) * 8; 739 if (src_byte == dst_byte) { 740 sc = SHF_SC_NONE; 741 } else if (src_byte < dst_byte) { 742 shf = 32 - shf; 743 sc = SHF_SC_L_SHF; 744 } else { 745 sc = SHF_SC_R_SHF; 746 } 747 748 /* ld_field can address fewer indexes, if offset too large do RMW. 749 * Because we RMV twice we waste 2 cycles on unaligned 8 byte writes. 750 */ 751 if (idx <= RE_REG_LM_IDX_MAX) { 752 reg = reg_lm(0, idx); 753 } else { 754 reg = imm_a(nfp_prog); 755 /* Only first and last LMEM locations are going to need RMW, 756 * the middle location will be overwritten fully. 757 */ 758 if (first || last) 759 wrp_mov(nfp_prog, reg, reg_lm(0, idx)); 760 } 761 762 emit_ld_field(nfp_prog, reg, mask, reg_b(src), sc, shf); 763 764 if (new_gpr || last) { 765 if (idx > RE_REG_LM_IDX_MAX) 766 wrp_mov(nfp_prog, reg_lm(0, idx), reg); 767 } 768 769 return 0; 770 } 771 772 static int 773 mem_op_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 774 unsigned int size, u8 gpr, bool clr_gpr, lmem_step step) 775 { 776 s32 off = nfp_prog->stack_depth + meta->insn.off; 777 bool first = true, last; 778 u8 prev_gpr = 255; 779 u32 gpr_byte = 0; 780 int ret; 781 782 if (clr_gpr && size < 8) 783 wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 784 785 while (size) { 786 u32 slice_end; 787 u8 slice_size; 788 789 slice_size = min(size, 4 - gpr_byte); 790 slice_end = min(off + slice_size, round_up(off + 1, 4)); 791 slice_size = slice_end - off; 792 793 last = slice_size == size; 794 795 ret = step(nfp_prog, gpr, gpr_byte, off, slice_size, 796 first, gpr != prev_gpr, last); 797 if (ret) 798 return ret; 799 800 prev_gpr = gpr; 801 first = false; 802 803 gpr_byte += slice_size; 804 if (gpr_byte >= 4) { 805 gpr_byte -= 4; 806 gpr++; 807 } 808 809 size -= slice_size; 810 off += slice_size; 811 } 812 813 return 0; 814 } 815 816 static void 817 wrp_alu_imm(struct nfp_prog *nfp_prog, u8 dst, enum alu_op alu_op, u32 imm) 818 { 819 swreg tmp_reg; 820 821 if (alu_op == ALU_OP_AND) { 822 if (!imm) 823 wrp_immed(nfp_prog, reg_both(dst), 0); 824 if (!imm || !~imm) 825 return; 826 } 827 if (alu_op == ALU_OP_OR) { 828 if (!~imm) 829 wrp_immed(nfp_prog, reg_both(dst), ~0U); 830 if (!imm || !~imm) 831 return; 832 } 833 if (alu_op == ALU_OP_XOR) { 834 if (!~imm) 835 emit_alu(nfp_prog, reg_both(dst), reg_none(), 836 ALU_OP_NEG, reg_b(dst)); 837 if (!imm || !~imm) 838 return; 839 } 840 841 tmp_reg = ur_load_imm_any(nfp_prog, imm, imm_b(nfp_prog)); 842 emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, tmp_reg); 843 } 844 845 static int 846 wrp_alu64_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 847 enum alu_op alu_op, bool skip) 848 { 849 const struct bpf_insn *insn = &meta->insn; 850 u64 imm = insn->imm; /* sign extend */ 851 852 if (skip) { 853 meta->skip = true; 854 return 0; 855 } 856 857 wrp_alu_imm(nfp_prog, insn->dst_reg * 2, alu_op, imm & ~0U); 858 wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, alu_op, imm >> 32); 859 860 return 0; 861 } 862 863 static int 864 wrp_alu64_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 865 enum alu_op alu_op) 866 { 867 u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2; 868 869 emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, reg_b(src)); 870 emit_alu(nfp_prog, reg_both(dst + 1), 871 reg_a(dst + 1), alu_op, reg_b(src + 1)); 872 873 return 0; 874 } 875 876 static int 877 wrp_alu32_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 878 enum alu_op alu_op, bool skip) 879 { 880 const struct bpf_insn *insn = &meta->insn; 881 882 if (skip) { 883 meta->skip = true; 884 return 0; 885 } 886 887 wrp_alu_imm(nfp_prog, insn->dst_reg * 2, alu_op, insn->imm); 888 wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 889 890 return 0; 891 } 892 893 static int 894 wrp_alu32_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 895 enum alu_op alu_op) 896 { 897 u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2; 898 899 emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, reg_b(src)); 900 wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 901 902 return 0; 903 } 904 905 static void 906 wrp_test_reg_one(struct nfp_prog *nfp_prog, u8 dst, enum alu_op alu_op, u8 src, 907 enum br_mask br_mask, u16 off) 908 { 909 emit_alu(nfp_prog, reg_none(), reg_a(dst), alu_op, reg_b(src)); 910 emit_br(nfp_prog, br_mask, off, 0); 911 } 912 913 static int 914 wrp_test_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 915 enum alu_op alu_op, enum br_mask br_mask) 916 { 917 const struct bpf_insn *insn = &meta->insn; 918 919 if (insn->off < 0) /* TODO */ 920 return -EOPNOTSUPP; 921 922 wrp_test_reg_one(nfp_prog, insn->dst_reg * 2, alu_op, 923 insn->src_reg * 2, br_mask, insn->off); 924 wrp_test_reg_one(nfp_prog, insn->dst_reg * 2 + 1, alu_op, 925 insn->src_reg * 2 + 1, br_mask, insn->off); 926 927 return 0; 928 } 929 930 static int 931 wrp_cmp_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 932 enum br_mask br_mask, bool swap) 933 { 934 const struct bpf_insn *insn = &meta->insn; 935 u64 imm = insn->imm; /* sign extend */ 936 u8 reg = insn->dst_reg * 2; 937 swreg tmp_reg; 938 939 if (insn->off < 0) /* TODO */ 940 return -EOPNOTSUPP; 941 942 tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 943 if (!swap) 944 emit_alu(nfp_prog, reg_none(), reg_a(reg), ALU_OP_SUB, tmp_reg); 945 else 946 emit_alu(nfp_prog, reg_none(), tmp_reg, ALU_OP_SUB, reg_a(reg)); 947 948 tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 949 if (!swap) 950 emit_alu(nfp_prog, reg_none(), 951 reg_a(reg + 1), ALU_OP_SUB_C, tmp_reg); 952 else 953 emit_alu(nfp_prog, reg_none(), 954 tmp_reg, ALU_OP_SUB_C, reg_a(reg + 1)); 955 956 emit_br(nfp_prog, br_mask, insn->off, 0); 957 958 return 0; 959 } 960 961 static int 962 wrp_cmp_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 963 enum br_mask br_mask, bool swap) 964 { 965 const struct bpf_insn *insn = &meta->insn; 966 u8 areg, breg; 967 968 areg = insn->dst_reg * 2; 969 breg = insn->src_reg * 2; 970 971 if (insn->off < 0) /* TODO */ 972 return -EOPNOTSUPP; 973 974 if (swap) { 975 areg ^= breg; 976 breg ^= areg; 977 areg ^= breg; 978 } 979 980 emit_alu(nfp_prog, reg_none(), reg_a(areg), ALU_OP_SUB, reg_b(breg)); 981 emit_alu(nfp_prog, reg_none(), 982 reg_a(areg + 1), ALU_OP_SUB_C, reg_b(breg + 1)); 983 emit_br(nfp_prog, br_mask, insn->off, 0); 984 985 return 0; 986 } 987 988 static void wrp_end32(struct nfp_prog *nfp_prog, swreg reg_in, u8 gpr_out) 989 { 990 emit_ld_field(nfp_prog, reg_both(gpr_out), 0xf, reg_in, 991 SHF_SC_R_ROT, 8); 992 emit_ld_field(nfp_prog, reg_both(gpr_out), 0x5, reg_a(gpr_out), 993 SHF_SC_R_ROT, 16); 994 } 995 996 /* --- Callbacks --- */ 997 static int mov_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 998 { 999 const struct bpf_insn *insn = &meta->insn; 1000 1001 wrp_reg_mov(nfp_prog, insn->dst_reg * 2, insn->src_reg * 2); 1002 wrp_reg_mov(nfp_prog, insn->dst_reg * 2 + 1, insn->src_reg * 2 + 1); 1003 1004 return 0; 1005 } 1006 1007 static int mov_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1008 { 1009 u64 imm = meta->insn.imm; /* sign extend */ 1010 1011 wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2), imm & ~0U); 1012 wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), imm >> 32); 1013 1014 return 0; 1015 } 1016 1017 static int xor_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1018 { 1019 return wrp_alu64_reg(nfp_prog, meta, ALU_OP_XOR); 1020 } 1021 1022 static int xor_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1023 { 1024 return wrp_alu64_imm(nfp_prog, meta, ALU_OP_XOR, !meta->insn.imm); 1025 } 1026 1027 static int and_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1028 { 1029 return wrp_alu64_reg(nfp_prog, meta, ALU_OP_AND); 1030 } 1031 1032 static int and_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1033 { 1034 return wrp_alu64_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm); 1035 } 1036 1037 static int or_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1038 { 1039 return wrp_alu64_reg(nfp_prog, meta, ALU_OP_OR); 1040 } 1041 1042 static int or_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1043 { 1044 return wrp_alu64_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm); 1045 } 1046 1047 static int add_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1048 { 1049 const struct bpf_insn *insn = &meta->insn; 1050 1051 emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), 1052 reg_a(insn->dst_reg * 2), ALU_OP_ADD, 1053 reg_b(insn->src_reg * 2)); 1054 emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 1055 reg_a(insn->dst_reg * 2 + 1), ALU_OP_ADD_C, 1056 reg_b(insn->src_reg * 2 + 1)); 1057 1058 return 0; 1059 } 1060 1061 static int add_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1062 { 1063 const struct bpf_insn *insn = &meta->insn; 1064 u64 imm = insn->imm; /* sign extend */ 1065 1066 wrp_alu_imm(nfp_prog, insn->dst_reg * 2, ALU_OP_ADD, imm & ~0U); 1067 wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, ALU_OP_ADD_C, imm >> 32); 1068 1069 return 0; 1070 } 1071 1072 static int sub_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1073 { 1074 const struct bpf_insn *insn = &meta->insn; 1075 1076 emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), 1077 reg_a(insn->dst_reg * 2), ALU_OP_SUB, 1078 reg_b(insn->src_reg * 2)); 1079 emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 1080 reg_a(insn->dst_reg * 2 + 1), ALU_OP_SUB_C, 1081 reg_b(insn->src_reg * 2 + 1)); 1082 1083 return 0; 1084 } 1085 1086 static int sub_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1087 { 1088 const struct bpf_insn *insn = &meta->insn; 1089 u64 imm = insn->imm; /* sign extend */ 1090 1091 wrp_alu_imm(nfp_prog, insn->dst_reg * 2, ALU_OP_SUB, imm & ~0U); 1092 wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, ALU_OP_SUB_C, imm >> 32); 1093 1094 return 0; 1095 } 1096 1097 static int shl_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1098 { 1099 const struct bpf_insn *insn = &meta->insn; 1100 u8 dst = insn->dst_reg * 2; 1101 1102 if (insn->imm < 32) { 1103 emit_shf(nfp_prog, reg_both(dst + 1), 1104 reg_a(dst + 1), SHF_OP_NONE, reg_b(dst), 1105 SHF_SC_R_DSHF, 32 - insn->imm); 1106 emit_shf(nfp_prog, reg_both(dst), 1107 reg_none(), SHF_OP_NONE, reg_b(dst), 1108 SHF_SC_L_SHF, insn->imm); 1109 } else if (insn->imm == 32) { 1110 wrp_reg_mov(nfp_prog, dst + 1, dst); 1111 wrp_immed(nfp_prog, reg_both(dst), 0); 1112 } else if (insn->imm > 32) { 1113 emit_shf(nfp_prog, reg_both(dst + 1), 1114 reg_none(), SHF_OP_NONE, reg_b(dst), 1115 SHF_SC_L_SHF, insn->imm - 32); 1116 wrp_immed(nfp_prog, reg_both(dst), 0); 1117 } 1118 1119 return 0; 1120 } 1121 1122 static int shr_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1123 { 1124 const struct bpf_insn *insn = &meta->insn; 1125 u8 dst = insn->dst_reg * 2; 1126 1127 if (insn->imm < 32) { 1128 emit_shf(nfp_prog, reg_both(dst), 1129 reg_a(dst + 1), SHF_OP_NONE, reg_b(dst), 1130 SHF_SC_R_DSHF, insn->imm); 1131 emit_shf(nfp_prog, reg_both(dst + 1), 1132 reg_none(), SHF_OP_NONE, reg_b(dst + 1), 1133 SHF_SC_R_SHF, insn->imm); 1134 } else if (insn->imm == 32) { 1135 wrp_reg_mov(nfp_prog, dst, dst + 1); 1136 wrp_immed(nfp_prog, reg_both(dst + 1), 0); 1137 } else if (insn->imm > 32) { 1138 emit_shf(nfp_prog, reg_both(dst), 1139 reg_none(), SHF_OP_NONE, reg_b(dst + 1), 1140 SHF_SC_R_SHF, insn->imm - 32); 1141 wrp_immed(nfp_prog, reg_both(dst + 1), 0); 1142 } 1143 1144 return 0; 1145 } 1146 1147 static int mov_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1148 { 1149 const struct bpf_insn *insn = &meta->insn; 1150 1151 wrp_reg_mov(nfp_prog, insn->dst_reg * 2, insn->src_reg * 2); 1152 wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1153 1154 return 0; 1155 } 1156 1157 static int mov_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1158 { 1159 const struct bpf_insn *insn = &meta->insn; 1160 1161 wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2), insn->imm); 1162 wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1163 1164 return 0; 1165 } 1166 1167 static int xor_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1168 { 1169 return wrp_alu32_reg(nfp_prog, meta, ALU_OP_XOR); 1170 } 1171 1172 static int xor_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1173 { 1174 return wrp_alu32_imm(nfp_prog, meta, ALU_OP_XOR, !~meta->insn.imm); 1175 } 1176 1177 static int and_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1178 { 1179 return wrp_alu32_reg(nfp_prog, meta, ALU_OP_AND); 1180 } 1181 1182 static int and_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1183 { 1184 return wrp_alu32_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm); 1185 } 1186 1187 static int or_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1188 { 1189 return wrp_alu32_reg(nfp_prog, meta, ALU_OP_OR); 1190 } 1191 1192 static int or_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1193 { 1194 return wrp_alu32_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm); 1195 } 1196 1197 static int add_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1198 { 1199 return wrp_alu32_reg(nfp_prog, meta, ALU_OP_ADD); 1200 } 1201 1202 static int add_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1203 { 1204 return wrp_alu32_imm(nfp_prog, meta, ALU_OP_ADD, !meta->insn.imm); 1205 } 1206 1207 static int sub_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1208 { 1209 return wrp_alu32_reg(nfp_prog, meta, ALU_OP_SUB); 1210 } 1211 1212 static int sub_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1213 { 1214 return wrp_alu32_imm(nfp_prog, meta, ALU_OP_SUB, !meta->insn.imm); 1215 } 1216 1217 static int shl_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1218 { 1219 const struct bpf_insn *insn = &meta->insn; 1220 1221 if (!insn->imm) 1222 return 1; /* TODO: zero shift means indirect */ 1223 1224 emit_shf(nfp_prog, reg_both(insn->dst_reg * 2), 1225 reg_none(), SHF_OP_NONE, reg_b(insn->dst_reg * 2), 1226 SHF_SC_L_SHF, insn->imm); 1227 wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1228 1229 return 0; 1230 } 1231 1232 static int end_reg32(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1233 { 1234 const struct bpf_insn *insn = &meta->insn; 1235 u8 gpr = insn->dst_reg * 2; 1236 1237 switch (insn->imm) { 1238 case 16: 1239 emit_ld_field(nfp_prog, reg_both(gpr), 0x9, reg_b(gpr), 1240 SHF_SC_R_ROT, 8); 1241 emit_ld_field(nfp_prog, reg_both(gpr), 0xe, reg_a(gpr), 1242 SHF_SC_R_SHF, 16); 1243 1244 wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 1245 break; 1246 case 32: 1247 wrp_end32(nfp_prog, reg_a(gpr), gpr); 1248 wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 1249 break; 1250 case 64: 1251 wrp_mov(nfp_prog, imm_a(nfp_prog), reg_b(gpr + 1)); 1252 1253 wrp_end32(nfp_prog, reg_a(gpr), gpr + 1); 1254 wrp_end32(nfp_prog, imm_a(nfp_prog), gpr); 1255 break; 1256 } 1257 1258 return 0; 1259 } 1260 1261 static int imm_ld8_part2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1262 { 1263 wrp_immed(nfp_prog, reg_both(nfp_meta_prev(meta)->insn.dst_reg * 2 + 1), 1264 meta->insn.imm); 1265 1266 return 0; 1267 } 1268 1269 static int imm_ld8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1270 { 1271 const struct bpf_insn *insn = &meta->insn; 1272 1273 meta->double_cb = imm_ld8_part2; 1274 wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2), insn->imm); 1275 1276 return 0; 1277 } 1278 1279 static int data_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1280 { 1281 return construct_data_ld(nfp_prog, meta->insn.imm, 1); 1282 } 1283 1284 static int data_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1285 { 1286 return construct_data_ld(nfp_prog, meta->insn.imm, 2); 1287 } 1288 1289 static int data_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1290 { 1291 return construct_data_ld(nfp_prog, meta->insn.imm, 4); 1292 } 1293 1294 static int data_ind_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1295 { 1296 return construct_data_ind_ld(nfp_prog, meta->insn.imm, 1297 meta->insn.src_reg * 2, 1); 1298 } 1299 1300 static int data_ind_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1301 { 1302 return construct_data_ind_ld(nfp_prog, meta->insn.imm, 1303 meta->insn.src_reg * 2, 2); 1304 } 1305 1306 static int data_ind_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1307 { 1308 return construct_data_ind_ld(nfp_prog, meta->insn.imm, 1309 meta->insn.src_reg * 2, 4); 1310 } 1311 1312 static int 1313 mem_ldx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1314 unsigned int size) 1315 { 1316 return mem_op_stack(nfp_prog, meta, size, meta->insn.dst_reg * 2, true, 1317 wrp_lmem_load); 1318 } 1319 1320 static int mem_ldx_skb(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1321 u8 size) 1322 { 1323 swreg dst = reg_both(meta->insn.dst_reg * 2); 1324 1325 switch (meta->insn.off) { 1326 case offsetof(struct sk_buff, len): 1327 if (size != FIELD_SIZEOF(struct sk_buff, len)) 1328 return -EOPNOTSUPP; 1329 wrp_mov(nfp_prog, dst, plen_reg(nfp_prog)); 1330 break; 1331 case offsetof(struct sk_buff, data): 1332 if (size != sizeof(void *)) 1333 return -EOPNOTSUPP; 1334 wrp_mov(nfp_prog, dst, pptr_reg(nfp_prog)); 1335 break; 1336 case offsetof(struct sk_buff, cb) + 1337 offsetof(struct bpf_skb_data_end, data_end): 1338 if (size != sizeof(void *)) 1339 return -EOPNOTSUPP; 1340 emit_alu(nfp_prog, dst, 1341 plen_reg(nfp_prog), ALU_OP_ADD, pptr_reg(nfp_prog)); 1342 break; 1343 default: 1344 return -EOPNOTSUPP; 1345 } 1346 1347 wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1348 1349 return 0; 1350 } 1351 1352 static int mem_ldx_xdp(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1353 u8 size) 1354 { 1355 swreg dst = reg_both(meta->insn.dst_reg * 2); 1356 1357 if (size != sizeof(void *)) 1358 return -EINVAL; 1359 1360 switch (meta->insn.off) { 1361 case offsetof(struct xdp_buff, data): 1362 wrp_mov(nfp_prog, dst, pptr_reg(nfp_prog)); 1363 break; 1364 case offsetof(struct xdp_buff, data_end): 1365 emit_alu(nfp_prog, dst, 1366 plen_reg(nfp_prog), ALU_OP_ADD, pptr_reg(nfp_prog)); 1367 break; 1368 default: 1369 return -EOPNOTSUPP; 1370 } 1371 1372 wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1373 1374 return 0; 1375 } 1376 1377 static int 1378 mem_ldx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1379 unsigned int size) 1380 { 1381 swreg tmp_reg; 1382 1383 tmp_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1384 1385 return data_ld_host_order(nfp_prog, meta->insn.src_reg * 2, tmp_reg, 1386 meta->insn.dst_reg * 2, size); 1387 } 1388 1389 static int 1390 mem_ldx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1391 unsigned int size) 1392 { 1393 if (meta->ptr.type == PTR_TO_CTX) { 1394 if (nfp_prog->act == NN_ACT_XDP) 1395 return mem_ldx_xdp(nfp_prog, meta, size); 1396 else 1397 return mem_ldx_skb(nfp_prog, meta, size); 1398 } 1399 1400 if (meta->ptr.type == PTR_TO_PACKET) 1401 return mem_ldx_data(nfp_prog, meta, size); 1402 1403 if (meta->ptr.type == PTR_TO_STACK) 1404 return mem_ldx_stack(nfp_prog, meta, size); 1405 1406 return -EOPNOTSUPP; 1407 } 1408 1409 static int mem_ldx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1410 { 1411 return mem_ldx(nfp_prog, meta, 1); 1412 } 1413 1414 static int mem_ldx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1415 { 1416 return mem_ldx(nfp_prog, meta, 2); 1417 } 1418 1419 static int mem_ldx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1420 { 1421 return mem_ldx(nfp_prog, meta, 4); 1422 } 1423 1424 static int mem_ldx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1425 { 1426 return mem_ldx(nfp_prog, meta, 8); 1427 } 1428 1429 static int 1430 mem_st_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1431 unsigned int size) 1432 { 1433 u64 imm = meta->insn.imm; /* sign extend */ 1434 swreg off_reg; 1435 1436 off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1437 1438 return data_st_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg, 1439 imm, size); 1440 } 1441 1442 static int mem_st(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1443 unsigned int size) 1444 { 1445 if (meta->ptr.type == PTR_TO_PACKET) 1446 return mem_st_data(nfp_prog, meta, size); 1447 1448 return -EOPNOTSUPP; 1449 } 1450 1451 static int mem_st1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1452 { 1453 return mem_st(nfp_prog, meta, 1); 1454 } 1455 1456 static int mem_st2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1457 { 1458 return mem_st(nfp_prog, meta, 2); 1459 } 1460 1461 static int mem_st4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1462 { 1463 return mem_st(nfp_prog, meta, 4); 1464 } 1465 1466 static int mem_st8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1467 { 1468 return mem_st(nfp_prog, meta, 8); 1469 } 1470 1471 static int 1472 mem_stx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1473 unsigned int size) 1474 { 1475 swreg off_reg; 1476 1477 off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1478 1479 return data_stx_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg, 1480 meta->insn.src_reg * 2, size); 1481 } 1482 1483 static int 1484 mem_stx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1485 unsigned int size) 1486 { 1487 return mem_op_stack(nfp_prog, meta, size, meta->insn.src_reg * 2, false, 1488 wrp_lmem_store); 1489 } 1490 1491 static int 1492 mem_stx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1493 unsigned int size) 1494 { 1495 if (meta->ptr.type == PTR_TO_PACKET) 1496 return mem_stx_data(nfp_prog, meta, size); 1497 1498 if (meta->ptr.type == PTR_TO_STACK) 1499 return mem_stx_stack(nfp_prog, meta, size); 1500 1501 return -EOPNOTSUPP; 1502 } 1503 1504 static int mem_stx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1505 { 1506 return mem_stx(nfp_prog, meta, 1); 1507 } 1508 1509 static int mem_stx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1510 { 1511 return mem_stx(nfp_prog, meta, 2); 1512 } 1513 1514 static int mem_stx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1515 { 1516 return mem_stx(nfp_prog, meta, 4); 1517 } 1518 1519 static int mem_stx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1520 { 1521 return mem_stx(nfp_prog, meta, 8); 1522 } 1523 1524 static int jump(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1525 { 1526 if (meta->insn.off < 0) /* TODO */ 1527 return -EOPNOTSUPP; 1528 emit_br(nfp_prog, BR_UNC, meta->insn.off, 0); 1529 1530 return 0; 1531 } 1532 1533 static int jeq_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1534 { 1535 const struct bpf_insn *insn = &meta->insn; 1536 u64 imm = insn->imm; /* sign extend */ 1537 swreg or1, or2, tmp_reg; 1538 1539 or1 = reg_a(insn->dst_reg * 2); 1540 or2 = reg_b(insn->dst_reg * 2 + 1); 1541 1542 if (insn->off < 0) /* TODO */ 1543 return -EOPNOTSUPP; 1544 1545 if (imm & ~0U) { 1546 tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1547 emit_alu(nfp_prog, imm_a(nfp_prog), 1548 reg_a(insn->dst_reg * 2), ALU_OP_XOR, tmp_reg); 1549 or1 = imm_a(nfp_prog); 1550 } 1551 1552 if (imm >> 32) { 1553 tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1554 emit_alu(nfp_prog, imm_b(nfp_prog), 1555 reg_a(insn->dst_reg * 2 + 1), ALU_OP_XOR, tmp_reg); 1556 or2 = imm_b(nfp_prog); 1557 } 1558 1559 emit_alu(nfp_prog, reg_none(), or1, ALU_OP_OR, or2); 1560 emit_br(nfp_prog, BR_BEQ, insn->off, 0); 1561 1562 return 0; 1563 } 1564 1565 static int jgt_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1566 { 1567 return wrp_cmp_imm(nfp_prog, meta, BR_BLO, true); 1568 } 1569 1570 static int jge_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1571 { 1572 return wrp_cmp_imm(nfp_prog, meta, BR_BHS, false); 1573 } 1574 1575 static int jlt_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1576 { 1577 return wrp_cmp_imm(nfp_prog, meta, BR_BLO, false); 1578 } 1579 1580 static int jle_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1581 { 1582 return wrp_cmp_imm(nfp_prog, meta, BR_BHS, true); 1583 } 1584 1585 static int jset_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1586 { 1587 const struct bpf_insn *insn = &meta->insn; 1588 u64 imm = insn->imm; /* sign extend */ 1589 swreg tmp_reg; 1590 1591 if (insn->off < 0) /* TODO */ 1592 return -EOPNOTSUPP; 1593 1594 if (!imm) { 1595 meta->skip = true; 1596 return 0; 1597 } 1598 1599 if (imm & ~0U) { 1600 tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1601 emit_alu(nfp_prog, reg_none(), 1602 reg_a(insn->dst_reg * 2), ALU_OP_AND, tmp_reg); 1603 emit_br(nfp_prog, BR_BNE, insn->off, 0); 1604 } 1605 1606 if (imm >> 32) { 1607 tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1608 emit_alu(nfp_prog, reg_none(), 1609 reg_a(insn->dst_reg * 2 + 1), ALU_OP_AND, tmp_reg); 1610 emit_br(nfp_prog, BR_BNE, insn->off, 0); 1611 } 1612 1613 return 0; 1614 } 1615 1616 static int jne_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1617 { 1618 const struct bpf_insn *insn = &meta->insn; 1619 u64 imm = insn->imm; /* sign extend */ 1620 swreg tmp_reg; 1621 1622 if (insn->off < 0) /* TODO */ 1623 return -EOPNOTSUPP; 1624 1625 if (!imm) { 1626 emit_alu(nfp_prog, reg_none(), reg_a(insn->dst_reg * 2), 1627 ALU_OP_OR, reg_b(insn->dst_reg * 2 + 1)); 1628 emit_br(nfp_prog, BR_BNE, insn->off, 0); 1629 return 0; 1630 } 1631 1632 tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1633 emit_alu(nfp_prog, reg_none(), 1634 reg_a(insn->dst_reg * 2), ALU_OP_XOR, tmp_reg); 1635 emit_br(nfp_prog, BR_BNE, insn->off, 0); 1636 1637 tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1638 emit_alu(nfp_prog, reg_none(), 1639 reg_a(insn->dst_reg * 2 + 1), ALU_OP_XOR, tmp_reg); 1640 emit_br(nfp_prog, BR_BNE, insn->off, 0); 1641 1642 return 0; 1643 } 1644 1645 static int jeq_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1646 { 1647 const struct bpf_insn *insn = &meta->insn; 1648 1649 if (insn->off < 0) /* TODO */ 1650 return -EOPNOTSUPP; 1651 1652 emit_alu(nfp_prog, imm_a(nfp_prog), reg_a(insn->dst_reg * 2), 1653 ALU_OP_XOR, reg_b(insn->src_reg * 2)); 1654 emit_alu(nfp_prog, imm_b(nfp_prog), reg_a(insn->dst_reg * 2 + 1), 1655 ALU_OP_XOR, reg_b(insn->src_reg * 2 + 1)); 1656 emit_alu(nfp_prog, reg_none(), 1657 imm_a(nfp_prog), ALU_OP_OR, imm_b(nfp_prog)); 1658 emit_br(nfp_prog, BR_BEQ, insn->off, 0); 1659 1660 return 0; 1661 } 1662 1663 static int jgt_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1664 { 1665 return wrp_cmp_reg(nfp_prog, meta, BR_BLO, true); 1666 } 1667 1668 static int jge_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1669 { 1670 return wrp_cmp_reg(nfp_prog, meta, BR_BHS, false); 1671 } 1672 1673 static int jlt_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1674 { 1675 return wrp_cmp_reg(nfp_prog, meta, BR_BLO, false); 1676 } 1677 1678 static int jle_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1679 { 1680 return wrp_cmp_reg(nfp_prog, meta, BR_BHS, true); 1681 } 1682 1683 static int jset_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1684 { 1685 return wrp_test_reg(nfp_prog, meta, ALU_OP_AND, BR_BNE); 1686 } 1687 1688 static int jne_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1689 { 1690 return wrp_test_reg(nfp_prog, meta, ALU_OP_XOR, BR_BNE); 1691 } 1692 1693 static int goto_out(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1694 { 1695 wrp_br_special(nfp_prog, BR_UNC, OP_BR_GO_OUT); 1696 1697 return 0; 1698 } 1699 1700 static const instr_cb_t instr_cb[256] = { 1701 [BPF_ALU64 | BPF_MOV | BPF_X] = mov_reg64, 1702 [BPF_ALU64 | BPF_MOV | BPF_K] = mov_imm64, 1703 [BPF_ALU64 | BPF_XOR | BPF_X] = xor_reg64, 1704 [BPF_ALU64 | BPF_XOR | BPF_K] = xor_imm64, 1705 [BPF_ALU64 | BPF_AND | BPF_X] = and_reg64, 1706 [BPF_ALU64 | BPF_AND | BPF_K] = and_imm64, 1707 [BPF_ALU64 | BPF_OR | BPF_X] = or_reg64, 1708 [BPF_ALU64 | BPF_OR | BPF_K] = or_imm64, 1709 [BPF_ALU64 | BPF_ADD | BPF_X] = add_reg64, 1710 [BPF_ALU64 | BPF_ADD | BPF_K] = add_imm64, 1711 [BPF_ALU64 | BPF_SUB | BPF_X] = sub_reg64, 1712 [BPF_ALU64 | BPF_SUB | BPF_K] = sub_imm64, 1713 [BPF_ALU64 | BPF_LSH | BPF_K] = shl_imm64, 1714 [BPF_ALU64 | BPF_RSH | BPF_K] = shr_imm64, 1715 [BPF_ALU | BPF_MOV | BPF_X] = mov_reg, 1716 [BPF_ALU | BPF_MOV | BPF_K] = mov_imm, 1717 [BPF_ALU | BPF_XOR | BPF_X] = xor_reg, 1718 [BPF_ALU | BPF_XOR | BPF_K] = xor_imm, 1719 [BPF_ALU | BPF_AND | BPF_X] = and_reg, 1720 [BPF_ALU | BPF_AND | BPF_K] = and_imm, 1721 [BPF_ALU | BPF_OR | BPF_X] = or_reg, 1722 [BPF_ALU | BPF_OR | BPF_K] = or_imm, 1723 [BPF_ALU | BPF_ADD | BPF_X] = add_reg, 1724 [BPF_ALU | BPF_ADD | BPF_K] = add_imm, 1725 [BPF_ALU | BPF_SUB | BPF_X] = sub_reg, 1726 [BPF_ALU | BPF_SUB | BPF_K] = sub_imm, 1727 [BPF_ALU | BPF_LSH | BPF_K] = shl_imm, 1728 [BPF_ALU | BPF_END | BPF_X] = end_reg32, 1729 [BPF_LD | BPF_IMM | BPF_DW] = imm_ld8, 1730 [BPF_LD | BPF_ABS | BPF_B] = data_ld1, 1731 [BPF_LD | BPF_ABS | BPF_H] = data_ld2, 1732 [BPF_LD | BPF_ABS | BPF_W] = data_ld4, 1733 [BPF_LD | BPF_IND | BPF_B] = data_ind_ld1, 1734 [BPF_LD | BPF_IND | BPF_H] = data_ind_ld2, 1735 [BPF_LD | BPF_IND | BPF_W] = data_ind_ld4, 1736 [BPF_LDX | BPF_MEM | BPF_B] = mem_ldx1, 1737 [BPF_LDX | BPF_MEM | BPF_H] = mem_ldx2, 1738 [BPF_LDX | BPF_MEM | BPF_W] = mem_ldx4, 1739 [BPF_LDX | BPF_MEM | BPF_DW] = mem_ldx8, 1740 [BPF_STX | BPF_MEM | BPF_B] = mem_stx1, 1741 [BPF_STX | BPF_MEM | BPF_H] = mem_stx2, 1742 [BPF_STX | BPF_MEM | BPF_W] = mem_stx4, 1743 [BPF_STX | BPF_MEM | BPF_DW] = mem_stx8, 1744 [BPF_ST | BPF_MEM | BPF_B] = mem_st1, 1745 [BPF_ST | BPF_MEM | BPF_H] = mem_st2, 1746 [BPF_ST | BPF_MEM | BPF_W] = mem_st4, 1747 [BPF_ST | BPF_MEM | BPF_DW] = mem_st8, 1748 [BPF_JMP | BPF_JA | BPF_K] = jump, 1749 [BPF_JMP | BPF_JEQ | BPF_K] = jeq_imm, 1750 [BPF_JMP | BPF_JGT | BPF_K] = jgt_imm, 1751 [BPF_JMP | BPF_JGE | BPF_K] = jge_imm, 1752 [BPF_JMP | BPF_JLT | BPF_K] = jlt_imm, 1753 [BPF_JMP | BPF_JLE | BPF_K] = jle_imm, 1754 [BPF_JMP | BPF_JSET | BPF_K] = jset_imm, 1755 [BPF_JMP | BPF_JNE | BPF_K] = jne_imm, 1756 [BPF_JMP | BPF_JEQ | BPF_X] = jeq_reg, 1757 [BPF_JMP | BPF_JGT | BPF_X] = jgt_reg, 1758 [BPF_JMP | BPF_JGE | BPF_X] = jge_reg, 1759 [BPF_JMP | BPF_JLT | BPF_X] = jlt_reg, 1760 [BPF_JMP | BPF_JLE | BPF_X] = jle_reg, 1761 [BPF_JMP | BPF_JSET | BPF_X] = jset_reg, 1762 [BPF_JMP | BPF_JNE | BPF_X] = jne_reg, 1763 [BPF_JMP | BPF_EXIT] = goto_out, 1764 }; 1765 1766 /* --- Misc code --- */ 1767 static void br_set_offset(u64 *instr, u16 offset) 1768 { 1769 u16 addr_lo, addr_hi; 1770 1771 addr_lo = offset & (OP_BR_ADDR_LO >> __bf_shf(OP_BR_ADDR_LO)); 1772 addr_hi = offset != addr_lo; 1773 *instr &= ~(OP_BR_ADDR_HI | OP_BR_ADDR_LO); 1774 *instr |= FIELD_PREP(OP_BR_ADDR_HI, addr_hi); 1775 *instr |= FIELD_PREP(OP_BR_ADDR_LO, addr_lo); 1776 } 1777 1778 /* --- Assembler logic --- */ 1779 static int nfp_fixup_branches(struct nfp_prog *nfp_prog) 1780 { 1781 struct nfp_insn_meta *meta, *next; 1782 u32 off, br_idx; 1783 u32 idx; 1784 1785 nfp_for_each_insn_walk2(nfp_prog, meta, next) { 1786 if (meta->skip) 1787 continue; 1788 if (BPF_CLASS(meta->insn.code) != BPF_JMP) 1789 continue; 1790 1791 br_idx = nfp_prog_offset_to_index(nfp_prog, next->off) - 1; 1792 if (!nfp_is_br(nfp_prog->prog[br_idx])) { 1793 pr_err("Fixup found block not ending in branch %d %02x %016llx!!\n", 1794 br_idx, meta->insn.code, nfp_prog->prog[br_idx]); 1795 return -ELOOP; 1796 } 1797 /* Leave special branches for later */ 1798 if (FIELD_GET(OP_BR_SPECIAL, nfp_prog->prog[br_idx])) 1799 continue; 1800 1801 /* Find the target offset in assembler realm */ 1802 off = meta->insn.off; 1803 if (!off) { 1804 pr_err("Fixup found zero offset!!\n"); 1805 return -ELOOP; 1806 } 1807 1808 while (off && nfp_meta_has_next(nfp_prog, next)) { 1809 next = nfp_meta_next(next); 1810 off--; 1811 } 1812 if (off) { 1813 pr_err("Fixup found too large jump!! %d\n", off); 1814 return -ELOOP; 1815 } 1816 1817 if (next->skip) { 1818 pr_err("Branch landing on removed instruction!!\n"); 1819 return -ELOOP; 1820 } 1821 1822 for (idx = nfp_prog_offset_to_index(nfp_prog, meta->off); 1823 idx <= br_idx; idx++) { 1824 if (!nfp_is_br(nfp_prog->prog[idx])) 1825 continue; 1826 br_set_offset(&nfp_prog->prog[idx], next->off); 1827 } 1828 } 1829 1830 /* Fixup 'goto out's separately, they can be scattered around */ 1831 for (br_idx = 0; br_idx < nfp_prog->prog_len; br_idx++) { 1832 enum br_special special; 1833 1834 if ((nfp_prog->prog[br_idx] & OP_BR_BASE_MASK) != OP_BR_BASE) 1835 continue; 1836 1837 special = FIELD_GET(OP_BR_SPECIAL, nfp_prog->prog[br_idx]); 1838 switch (special) { 1839 case OP_BR_NORMAL: 1840 break; 1841 case OP_BR_GO_OUT: 1842 br_set_offset(&nfp_prog->prog[br_idx], 1843 nfp_prog->tgt_out); 1844 break; 1845 case OP_BR_GO_ABORT: 1846 br_set_offset(&nfp_prog->prog[br_idx], 1847 nfp_prog->tgt_abort); 1848 break; 1849 } 1850 1851 nfp_prog->prog[br_idx] &= ~OP_BR_SPECIAL; 1852 } 1853 1854 return 0; 1855 } 1856 1857 static void nfp_intro(struct nfp_prog *nfp_prog) 1858 { 1859 wrp_immed(nfp_prog, plen_reg(nfp_prog), GENMASK(13, 0)); 1860 emit_alu(nfp_prog, plen_reg(nfp_prog), 1861 plen_reg(nfp_prog), ALU_OP_AND, pv_len(nfp_prog)); 1862 } 1863 1864 static void nfp_outro_tc_legacy(struct nfp_prog *nfp_prog) 1865 { 1866 const u8 act2code[] = { 1867 [NN_ACT_TC_DROP] = 0x22, 1868 [NN_ACT_TC_REDIR] = 0x24 1869 }; 1870 /* Target for aborts */ 1871 nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 1872 wrp_immed(nfp_prog, reg_both(0), 0); 1873 1874 /* Target for normal exits */ 1875 nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 1876 /* Legacy TC mode: 1877 * 0 0x11 -> pass, count as stat0 1878 * -1 drop 0x22 -> drop, count as stat1 1879 * redir 0x24 -> redir, count as stat1 1880 * ife mark 0x21 -> pass, count as stat1 1881 * ife + tx 0x24 -> redir, count as stat1 1882 */ 1883 emit_br_byte_neq(nfp_prog, reg_b(0), 0xff, 0, nfp_prog->tgt_done, 2); 1884 wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 1885 emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x11), SHF_SC_L_SHF, 16); 1886 1887 emit_br(nfp_prog, BR_UNC, nfp_prog->tgt_done, 1); 1888 emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(act2code[nfp_prog->act]), 1889 SHF_SC_L_SHF, 16); 1890 } 1891 1892 static void nfp_outro_tc_da(struct nfp_prog *nfp_prog) 1893 { 1894 /* TC direct-action mode: 1895 * 0,1 ok NOT SUPPORTED[1] 1896 * 2 drop 0x22 -> drop, count as stat1 1897 * 4,5 nuke 0x02 -> drop 1898 * 7 redir 0x44 -> redir, count as stat2 1899 * * unspec 0x11 -> pass, count as stat0 1900 * 1901 * [1] We can't support OK and RECLASSIFY because we can't tell TC 1902 * the exact decision made. We are forced to support UNSPEC 1903 * to handle aborts so that's the only one we handle for passing 1904 * packets up the stack. 1905 */ 1906 /* Target for aborts */ 1907 nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 1908 1909 emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 1910 1911 wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 1912 emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x11), SHF_SC_L_SHF, 16); 1913 1914 /* Target for normal exits */ 1915 nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 1916 1917 /* if R0 > 7 jump to abort */ 1918 emit_alu(nfp_prog, reg_none(), reg_imm(7), ALU_OP_SUB, reg_b(0)); 1919 emit_br(nfp_prog, BR_BLO, nfp_prog->tgt_abort, 0); 1920 wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 1921 1922 wrp_immed(nfp_prog, reg_b(2), 0x41221211); 1923 wrp_immed(nfp_prog, reg_b(3), 0x41001211); 1924 1925 emit_shf(nfp_prog, reg_a(1), 1926 reg_none(), SHF_OP_NONE, reg_b(0), SHF_SC_L_SHF, 2); 1927 1928 emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 1929 emit_shf(nfp_prog, reg_a(2), 1930 reg_imm(0xf), SHF_OP_AND, reg_b(2), SHF_SC_R_SHF, 0); 1931 1932 emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 1933 emit_shf(nfp_prog, reg_b(2), 1934 reg_imm(0xf), SHF_OP_AND, reg_b(3), SHF_SC_R_SHF, 0); 1935 1936 emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 1937 1938 emit_shf(nfp_prog, reg_b(2), 1939 reg_a(2), SHF_OP_OR, reg_b(2), SHF_SC_L_SHF, 4); 1940 emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_b(2), SHF_SC_L_SHF, 16); 1941 } 1942 1943 static void nfp_outro_xdp(struct nfp_prog *nfp_prog) 1944 { 1945 /* XDP return codes: 1946 * 0 aborted 0x82 -> drop, count as stat3 1947 * 1 drop 0x22 -> drop, count as stat1 1948 * 2 pass 0x11 -> pass, count as stat0 1949 * 3 tx 0x44 -> redir, count as stat2 1950 * * unknown 0x82 -> drop, count as stat3 1951 */ 1952 /* Target for aborts */ 1953 nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 1954 1955 emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 1956 1957 wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 1958 emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x82), SHF_SC_L_SHF, 16); 1959 1960 /* Target for normal exits */ 1961 nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 1962 1963 /* if R0 > 3 jump to abort */ 1964 emit_alu(nfp_prog, reg_none(), reg_imm(3), ALU_OP_SUB, reg_b(0)); 1965 emit_br(nfp_prog, BR_BLO, nfp_prog->tgt_abort, 0); 1966 1967 wrp_immed(nfp_prog, reg_b(2), 0x44112282); 1968 1969 emit_shf(nfp_prog, reg_a(1), 1970 reg_none(), SHF_OP_NONE, reg_b(0), SHF_SC_L_SHF, 3); 1971 1972 emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 1973 emit_shf(nfp_prog, reg_b(2), 1974 reg_imm(0xff), SHF_OP_AND, reg_b(2), SHF_SC_R_SHF, 0); 1975 1976 emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 1977 1978 wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 1979 emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_b(2), SHF_SC_L_SHF, 16); 1980 } 1981 1982 static void nfp_outro(struct nfp_prog *nfp_prog) 1983 { 1984 switch (nfp_prog->act) { 1985 case NN_ACT_DIRECT: 1986 nfp_outro_tc_da(nfp_prog); 1987 break; 1988 case NN_ACT_TC_DROP: 1989 case NN_ACT_TC_REDIR: 1990 nfp_outro_tc_legacy(nfp_prog); 1991 break; 1992 case NN_ACT_XDP: 1993 nfp_outro_xdp(nfp_prog); 1994 break; 1995 } 1996 } 1997 1998 static int nfp_translate(struct nfp_prog *nfp_prog) 1999 { 2000 struct nfp_insn_meta *meta; 2001 int err; 2002 2003 nfp_intro(nfp_prog); 2004 if (nfp_prog->error) 2005 return nfp_prog->error; 2006 2007 list_for_each_entry(meta, &nfp_prog->insns, l) { 2008 instr_cb_t cb = instr_cb[meta->insn.code]; 2009 2010 meta->off = nfp_prog_current_offset(nfp_prog); 2011 2012 if (meta->skip) { 2013 nfp_prog->n_translated++; 2014 continue; 2015 } 2016 2017 if (nfp_meta_has_prev(nfp_prog, meta) && 2018 nfp_meta_prev(meta)->double_cb) 2019 cb = nfp_meta_prev(meta)->double_cb; 2020 if (!cb) 2021 return -ENOENT; 2022 err = cb(nfp_prog, meta); 2023 if (err) 2024 return err; 2025 2026 nfp_prog->n_translated++; 2027 } 2028 2029 nfp_outro(nfp_prog); 2030 if (nfp_prog->error) 2031 return nfp_prog->error; 2032 2033 wrp_nops(nfp_prog, NFP_USTORE_PREFETCH_WINDOW); 2034 if (nfp_prog->error) 2035 return nfp_prog->error; 2036 2037 return nfp_fixup_branches(nfp_prog); 2038 } 2039 2040 static int 2041 nfp_prog_prepare(struct nfp_prog *nfp_prog, const struct bpf_insn *prog, 2042 unsigned int cnt) 2043 { 2044 unsigned int i; 2045 2046 for (i = 0; i < cnt; i++) { 2047 struct nfp_insn_meta *meta; 2048 2049 meta = kzalloc(sizeof(*meta), GFP_KERNEL); 2050 if (!meta) 2051 return -ENOMEM; 2052 2053 meta->insn = prog[i]; 2054 meta->n = i; 2055 2056 list_add_tail(&meta->l, &nfp_prog->insns); 2057 } 2058 2059 return 0; 2060 } 2061 2062 /* --- Optimizations --- */ 2063 static void nfp_bpf_opt_reg_init(struct nfp_prog *nfp_prog) 2064 { 2065 struct nfp_insn_meta *meta; 2066 2067 list_for_each_entry(meta, &nfp_prog->insns, l) { 2068 struct bpf_insn insn = meta->insn; 2069 2070 /* Programs converted from cBPF start with register xoring */ 2071 if (insn.code == (BPF_ALU64 | BPF_XOR | BPF_X) && 2072 insn.src_reg == insn.dst_reg) 2073 continue; 2074 2075 /* Programs start with R6 = R1 but we ignore the skb pointer */ 2076 if (insn.code == (BPF_ALU64 | BPF_MOV | BPF_X) && 2077 insn.src_reg == 1 && insn.dst_reg == 6) 2078 meta->skip = true; 2079 2080 /* Return as soon as something doesn't match */ 2081 if (!meta->skip) 2082 return; 2083 } 2084 } 2085 2086 /* Remove masking after load since our load guarantees this is not needed */ 2087 static void nfp_bpf_opt_ld_mask(struct nfp_prog *nfp_prog) 2088 { 2089 struct nfp_insn_meta *meta1, *meta2; 2090 const s32 exp_mask[] = { 2091 [BPF_B] = 0x000000ffU, 2092 [BPF_H] = 0x0000ffffU, 2093 [BPF_W] = 0xffffffffU, 2094 }; 2095 2096 nfp_for_each_insn_walk2(nfp_prog, meta1, meta2) { 2097 struct bpf_insn insn, next; 2098 2099 insn = meta1->insn; 2100 next = meta2->insn; 2101 2102 if (BPF_CLASS(insn.code) != BPF_LD) 2103 continue; 2104 if (BPF_MODE(insn.code) != BPF_ABS && 2105 BPF_MODE(insn.code) != BPF_IND) 2106 continue; 2107 2108 if (next.code != (BPF_ALU64 | BPF_AND | BPF_K)) 2109 continue; 2110 2111 if (!exp_mask[BPF_SIZE(insn.code)]) 2112 continue; 2113 if (exp_mask[BPF_SIZE(insn.code)] != next.imm) 2114 continue; 2115 2116 if (next.src_reg || next.dst_reg) 2117 continue; 2118 2119 meta2->skip = true; 2120 } 2121 } 2122 2123 static void nfp_bpf_opt_ld_shift(struct nfp_prog *nfp_prog) 2124 { 2125 struct nfp_insn_meta *meta1, *meta2, *meta3; 2126 2127 nfp_for_each_insn_walk3(nfp_prog, meta1, meta2, meta3) { 2128 struct bpf_insn insn, next1, next2; 2129 2130 insn = meta1->insn; 2131 next1 = meta2->insn; 2132 next2 = meta3->insn; 2133 2134 if (BPF_CLASS(insn.code) != BPF_LD) 2135 continue; 2136 if (BPF_MODE(insn.code) != BPF_ABS && 2137 BPF_MODE(insn.code) != BPF_IND) 2138 continue; 2139 if (BPF_SIZE(insn.code) != BPF_W) 2140 continue; 2141 2142 if (!(next1.code == (BPF_LSH | BPF_K | BPF_ALU64) && 2143 next2.code == (BPF_RSH | BPF_K | BPF_ALU64)) && 2144 !(next1.code == (BPF_RSH | BPF_K | BPF_ALU64) && 2145 next2.code == (BPF_LSH | BPF_K | BPF_ALU64))) 2146 continue; 2147 2148 if (next1.src_reg || next1.dst_reg || 2149 next2.src_reg || next2.dst_reg) 2150 continue; 2151 2152 if (next1.imm != 0x20 || next2.imm != 0x20) 2153 continue; 2154 2155 meta2->skip = true; 2156 meta3->skip = true; 2157 } 2158 } 2159 2160 static int nfp_bpf_optimize(struct nfp_prog *nfp_prog) 2161 { 2162 nfp_bpf_opt_reg_init(nfp_prog); 2163 2164 nfp_bpf_opt_ld_mask(nfp_prog); 2165 nfp_bpf_opt_ld_shift(nfp_prog); 2166 2167 return 0; 2168 } 2169 2170 static int nfp_bpf_ustore_calc(struct nfp_prog *nfp_prog, __le64 *ustore) 2171 { 2172 int i; 2173 2174 for (i = 0; i < nfp_prog->prog_len; i++) { 2175 int err; 2176 2177 err = nfp_ustore_check_valid_no_ecc(nfp_prog->prog[i]); 2178 if (err) 2179 return err; 2180 2181 nfp_prog->prog[i] = nfp_ustore_calc_ecc_insn(nfp_prog->prog[i]); 2182 2183 ustore[i] = cpu_to_le64(nfp_prog->prog[i]); 2184 } 2185 2186 return 0; 2187 } 2188 2189 /** 2190 * nfp_bpf_jit() - translate BPF code into NFP assembly 2191 * @filter: kernel BPF filter struct 2192 * @prog_mem: memory to store assembler instructions 2193 * @act: action attached to this eBPF program 2194 * @prog_start: offset of the first instruction when loaded 2195 * @prog_done: where to jump on exit 2196 * @prog_sz: size of @prog_mem in instructions 2197 * @res: achieved parameters of translation results 2198 */ 2199 int 2200 nfp_bpf_jit(struct bpf_prog *filter, void *prog_mem, 2201 enum nfp_bpf_action_type act, 2202 unsigned int prog_start, unsigned int prog_done, 2203 unsigned int prog_sz, struct nfp_bpf_result *res) 2204 { 2205 struct nfp_prog *nfp_prog; 2206 int ret; 2207 2208 nfp_prog = kzalloc(sizeof(*nfp_prog), GFP_KERNEL); 2209 if (!nfp_prog) 2210 return -ENOMEM; 2211 2212 INIT_LIST_HEAD(&nfp_prog->insns); 2213 nfp_prog->act = act; 2214 nfp_prog->start_off = prog_start; 2215 nfp_prog->tgt_done = prog_done; 2216 2217 ret = nfp_prog_prepare(nfp_prog, filter->insnsi, filter->len); 2218 if (ret) 2219 goto out; 2220 2221 ret = nfp_prog_verify(nfp_prog, filter); 2222 if (ret) 2223 goto out; 2224 2225 ret = nfp_bpf_optimize(nfp_prog); 2226 if (ret) 2227 goto out; 2228 2229 nfp_prog->num_regs = MAX_BPF_REG; 2230 nfp_prog->regs_per_thread = 32; 2231 2232 nfp_prog->prog = prog_mem; 2233 nfp_prog->__prog_alloc_len = prog_sz; 2234 2235 ret = nfp_translate(nfp_prog); 2236 if (ret) { 2237 pr_err("Translation failed with error %d (translated: %u)\n", 2238 ret, nfp_prog->n_translated); 2239 ret = -EINVAL; 2240 goto out; 2241 } 2242 2243 ret = nfp_bpf_ustore_calc(nfp_prog, (__force __le64 *)prog_mem); 2244 2245 res->n_instr = nfp_prog->prog_len; 2246 res->dense_mode = false; 2247 out: 2248 nfp_prog_free(nfp_prog); 2249 2250 return ret; 2251 } 2252