1d9ae7f2bSJakub Kicinski /* 2854dc87dSJiong Wang * Copyright (C) 2016-2017 Netronome Systems, Inc. 3d9ae7f2bSJakub Kicinski * 4d9ae7f2bSJakub Kicinski * This software is dual licensed under the GNU General License Version 2, 5d9ae7f2bSJakub Kicinski * June 1991 as shown in the file COPYING in the top-level directory of this 6d9ae7f2bSJakub Kicinski * source tree or the BSD 2-Clause License provided below. You have the 7d9ae7f2bSJakub Kicinski * option to license this software under the complete terms of either license. 8d9ae7f2bSJakub Kicinski * 9d9ae7f2bSJakub Kicinski * The BSD 2-Clause License: 10d9ae7f2bSJakub Kicinski * 11d9ae7f2bSJakub Kicinski * Redistribution and use in source and binary forms, with or 12d9ae7f2bSJakub Kicinski * without modification, are permitted provided that the following 13d9ae7f2bSJakub Kicinski * conditions are met: 14d9ae7f2bSJakub Kicinski * 15d9ae7f2bSJakub Kicinski * 1. Redistributions of source code must retain the above 16d9ae7f2bSJakub Kicinski * copyright notice, this list of conditions and the following 17d9ae7f2bSJakub Kicinski * disclaimer. 18d9ae7f2bSJakub Kicinski * 19d9ae7f2bSJakub Kicinski * 2. Redistributions in binary form must reproduce the above 20d9ae7f2bSJakub Kicinski * copyright notice, this list of conditions and the following 21d9ae7f2bSJakub Kicinski * disclaimer in the documentation and/or other materials 22d9ae7f2bSJakub Kicinski * provided with the distribution. 23d9ae7f2bSJakub Kicinski * 24d9ae7f2bSJakub Kicinski * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 25d9ae7f2bSJakub Kicinski * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 26d9ae7f2bSJakub Kicinski * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 27d9ae7f2bSJakub Kicinski * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 28d9ae7f2bSJakub Kicinski * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 29d9ae7f2bSJakub Kicinski * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 30d9ae7f2bSJakub Kicinski * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 31d9ae7f2bSJakub Kicinski * SOFTWARE. 32d9ae7f2bSJakub Kicinski */ 33d9ae7f2bSJakub Kicinski 34d9ae7f2bSJakub Kicinski #define pr_fmt(fmt) "NFP net bpf: " fmt 35d9ae7f2bSJakub Kicinski 360d49eaf4SJakub Kicinski #include <linux/bug.h> 37d9ae7f2bSJakub Kicinski #include <linux/kernel.h> 38d9ae7f2bSJakub Kicinski #include <linux/bpf.h> 39d9ae7f2bSJakub Kicinski #include <linux/filter.h> 40d9ae7f2bSJakub Kicinski #include <linux/pkt_cls.h> 41d9ae7f2bSJakub Kicinski #include <linux/unistd.h> 42d9ae7f2bSJakub Kicinski 43d9ae7f2bSJakub Kicinski #include "main.h" 44d9ae7f2bSJakub Kicinski #include "../nfp_asm.h" 45d9ae7f2bSJakub Kicinski 46d9ae7f2bSJakub Kicinski /* --- NFP prog --- */ 47d9ae7f2bSJakub Kicinski /* Foreach "multiple" entries macros provide pos and next<n> pointers. 48d9ae7f2bSJakub Kicinski * It's safe to modify the next pointers (but not pos). 49d9ae7f2bSJakub Kicinski */ 50d9ae7f2bSJakub Kicinski #define nfp_for_each_insn_walk2(nfp_prog, pos, next) \ 51d9ae7f2bSJakub Kicinski for (pos = list_first_entry(&(nfp_prog)->insns, typeof(*pos), l), \ 52d9ae7f2bSJakub Kicinski next = list_next_entry(pos, l); \ 53d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &pos->l && \ 54d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &next->l; \ 55d9ae7f2bSJakub Kicinski pos = nfp_meta_next(pos), \ 56d9ae7f2bSJakub Kicinski next = nfp_meta_next(pos)) 57d9ae7f2bSJakub Kicinski 58d9ae7f2bSJakub Kicinski #define nfp_for_each_insn_walk3(nfp_prog, pos, next, next2) \ 59d9ae7f2bSJakub Kicinski for (pos = list_first_entry(&(nfp_prog)->insns, typeof(*pos), l), \ 60d9ae7f2bSJakub Kicinski next = list_next_entry(pos, l), \ 61d9ae7f2bSJakub Kicinski next2 = list_next_entry(next, l); \ 62d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &pos->l && \ 63d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &next->l && \ 64d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &next2->l; \ 65d9ae7f2bSJakub Kicinski pos = nfp_meta_next(pos), \ 66d9ae7f2bSJakub Kicinski next = nfp_meta_next(pos), \ 67d9ae7f2bSJakub Kicinski next2 = nfp_meta_next(next)) 68d9ae7f2bSJakub Kicinski 69d9ae7f2bSJakub Kicinski static bool 70d9ae7f2bSJakub Kicinski nfp_meta_has_prev(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 71d9ae7f2bSJakub Kicinski { 72d9ae7f2bSJakub Kicinski return meta->l.prev != &nfp_prog->insns; 73d9ae7f2bSJakub Kicinski } 74d9ae7f2bSJakub Kicinski 75d9ae7f2bSJakub Kicinski static void nfp_prog_push(struct nfp_prog *nfp_prog, u64 insn) 76d9ae7f2bSJakub Kicinski { 77d9ae7f2bSJakub Kicinski if (nfp_prog->__prog_alloc_len == nfp_prog->prog_len) { 78d9ae7f2bSJakub Kicinski nfp_prog->error = -ENOSPC; 79d9ae7f2bSJakub Kicinski return; 80d9ae7f2bSJakub Kicinski } 81d9ae7f2bSJakub Kicinski 82d9ae7f2bSJakub Kicinski nfp_prog->prog[nfp_prog->prog_len] = insn; 83d9ae7f2bSJakub Kicinski nfp_prog->prog_len++; 84d9ae7f2bSJakub Kicinski } 85d9ae7f2bSJakub Kicinski 86d9ae7f2bSJakub Kicinski static unsigned int nfp_prog_current_offset(struct nfp_prog *nfp_prog) 87d9ae7f2bSJakub Kicinski { 88d9ae7f2bSJakub Kicinski return nfp_prog->start_off + nfp_prog->prog_len; 89d9ae7f2bSJakub Kicinski } 90d9ae7f2bSJakub Kicinski 910d49eaf4SJakub Kicinski static bool 920d49eaf4SJakub Kicinski nfp_prog_confirm_current_offset(struct nfp_prog *nfp_prog, unsigned int off) 930d49eaf4SJakub Kicinski { 940d49eaf4SJakub Kicinski /* If there is a recorded error we may have dropped instructions; 950d49eaf4SJakub Kicinski * that doesn't have to be due to translator bug, and the translation 960d49eaf4SJakub Kicinski * will fail anyway, so just return OK. 970d49eaf4SJakub Kicinski */ 980d49eaf4SJakub Kicinski if (nfp_prog->error) 990d49eaf4SJakub Kicinski return true; 1000d49eaf4SJakub Kicinski return !WARN_ON_ONCE(nfp_prog_current_offset(nfp_prog) != off); 1010d49eaf4SJakub Kicinski } 1020d49eaf4SJakub Kicinski 103d9ae7f2bSJakub Kicinski static unsigned int 104d9ae7f2bSJakub Kicinski nfp_prog_offset_to_index(struct nfp_prog *nfp_prog, unsigned int offset) 105d9ae7f2bSJakub Kicinski { 106d9ae7f2bSJakub Kicinski return offset - nfp_prog->start_off; 107d9ae7f2bSJakub Kicinski } 108d9ae7f2bSJakub Kicinski 109d9ae7f2bSJakub Kicinski /* --- Emitters --- */ 110d9ae7f2bSJakub Kicinski static void 111d9ae7f2bSJakub Kicinski __emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, 1125468a8b9SJakub Kicinski u8 mode, u8 xfer, u8 areg, u8 breg, u8 size, bool sync, bool indir) 113d9ae7f2bSJakub Kicinski { 114d9ae7f2bSJakub Kicinski enum cmd_ctx_swap ctx; 115d9ae7f2bSJakub Kicinski u64 insn; 116d9ae7f2bSJakub Kicinski 117d9ae7f2bSJakub Kicinski if (sync) 118d9ae7f2bSJakub Kicinski ctx = CMD_CTX_SWAP; 119d9ae7f2bSJakub Kicinski else 120d9ae7f2bSJakub Kicinski ctx = CMD_CTX_NO_SWAP; 121d9ae7f2bSJakub Kicinski 122d9ae7f2bSJakub Kicinski insn = FIELD_PREP(OP_CMD_A_SRC, areg) | 123d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_CTX, ctx) | 124d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_B_SRC, breg) | 125d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_TOKEN, cmd_tgt_act[op].token) | 126d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_XFER, xfer) | 127d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_CNT, size) | 128d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_SIG, sync) | 129d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_TGT_CMD, cmd_tgt_act[op].tgt_cmd) | 1305468a8b9SJakub Kicinski FIELD_PREP(OP_CMD_INDIR, indir) | 131d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_MODE, mode); 132d9ae7f2bSJakub Kicinski 133d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 134d9ae7f2bSJakub Kicinski } 135d9ae7f2bSJakub Kicinski 136d9ae7f2bSJakub Kicinski static void 1375468a8b9SJakub Kicinski emit_cmd_any(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer, 1385468a8b9SJakub Kicinski swreg lreg, swreg rreg, u8 size, bool sync, bool indir) 139d9ae7f2bSJakub Kicinski { 140d9ae7f2bSJakub Kicinski struct nfp_insn_re_regs reg; 141d9ae7f2bSJakub Kicinski int err; 142d9ae7f2bSJakub Kicinski 143d9ae7f2bSJakub Kicinski err = swreg_to_restricted(reg_none(), lreg, rreg, ®, false); 144d9ae7f2bSJakub Kicinski if (err) { 145d9ae7f2bSJakub Kicinski nfp_prog->error = err; 146d9ae7f2bSJakub Kicinski return; 147d9ae7f2bSJakub Kicinski } 148d9ae7f2bSJakub Kicinski if (reg.swap) { 149d9ae7f2bSJakub Kicinski pr_err("cmd can't swap arguments\n"); 150d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 151d9ae7f2bSJakub Kicinski return; 152d9ae7f2bSJakub Kicinski } 153995e101fSJakub Kicinski if (reg.dst_lmextn || reg.src_lmextn) { 154995e101fSJakub Kicinski pr_err("cmd can't use LMextn\n"); 155995e101fSJakub Kicinski nfp_prog->error = -EFAULT; 156995e101fSJakub Kicinski return; 157995e101fSJakub Kicinski } 158d9ae7f2bSJakub Kicinski 1595468a8b9SJakub Kicinski __emit_cmd(nfp_prog, op, mode, xfer, reg.areg, reg.breg, size, sync, 1605468a8b9SJakub Kicinski indir); 1615468a8b9SJakub Kicinski } 1625468a8b9SJakub Kicinski 1635468a8b9SJakub Kicinski static void 1645468a8b9SJakub Kicinski emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer, 1655468a8b9SJakub Kicinski swreg lreg, swreg rreg, u8 size, bool sync) 1665468a8b9SJakub Kicinski { 1675468a8b9SJakub Kicinski emit_cmd_any(nfp_prog, op, mode, xfer, lreg, rreg, size, sync, false); 168d9ae7f2bSJakub Kicinski } 169d9ae7f2bSJakub Kicinski 170d9ae7f2bSJakub Kicinski static void 1719879a381SJiong Wang emit_cmd_indir(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer, 1729879a381SJiong Wang swreg lreg, swreg rreg, u8 size, bool sync) 1739879a381SJiong Wang { 1749879a381SJiong Wang emit_cmd_any(nfp_prog, op, mode, xfer, lreg, rreg, size, sync, true); 1759879a381SJiong Wang } 1769879a381SJiong Wang 1779879a381SJiong Wang static void 178d9ae7f2bSJakub Kicinski __emit_br(struct nfp_prog *nfp_prog, enum br_mask mask, enum br_ev_pip ev_pip, 179d9ae7f2bSJakub Kicinski enum br_ctx_signal_state css, u16 addr, u8 defer) 180d9ae7f2bSJakub Kicinski { 181d9ae7f2bSJakub Kicinski u16 addr_lo, addr_hi; 182d9ae7f2bSJakub Kicinski u64 insn; 183d9ae7f2bSJakub Kicinski 184d9ae7f2bSJakub Kicinski addr_lo = addr & (OP_BR_ADDR_LO >> __bf_shf(OP_BR_ADDR_LO)); 185d9ae7f2bSJakub Kicinski addr_hi = addr != addr_lo; 186d9ae7f2bSJakub Kicinski 187d9ae7f2bSJakub Kicinski insn = OP_BR_BASE | 188d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_MASK, mask) | 189d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_EV_PIP, ev_pip) | 190d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_CSS, css) | 191d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_DEFBR, defer) | 192d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_ADDR_LO, addr_lo) | 193d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_ADDR_HI, addr_hi); 194d9ae7f2bSJakub Kicinski 195d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 196d9ae7f2bSJakub Kicinski } 197d9ae7f2bSJakub Kicinski 198d9ae7f2bSJakub Kicinski static void emit_br_def(struct nfp_prog *nfp_prog, u16 addr, u8 defer) 199d9ae7f2bSJakub Kicinski { 200d9ae7f2bSJakub Kicinski if (defer > 2) { 201d9ae7f2bSJakub Kicinski pr_err("BUG: branch defer out of bounds %d\n", defer); 202d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 203d9ae7f2bSJakub Kicinski return; 204d9ae7f2bSJakub Kicinski } 205d9ae7f2bSJakub Kicinski __emit_br(nfp_prog, BR_UNC, BR_EV_PIP_UNCOND, BR_CSS_NONE, addr, defer); 206d9ae7f2bSJakub Kicinski } 207d9ae7f2bSJakub Kicinski 208d9ae7f2bSJakub Kicinski static void 209d9ae7f2bSJakub Kicinski emit_br(struct nfp_prog *nfp_prog, enum br_mask mask, u16 addr, u8 defer) 210d9ae7f2bSJakub Kicinski { 211d9ae7f2bSJakub Kicinski __emit_br(nfp_prog, mask, 212d9ae7f2bSJakub Kicinski mask != BR_UNC ? BR_EV_PIP_COND : BR_EV_PIP_UNCOND, 213d9ae7f2bSJakub Kicinski BR_CSS_NONE, addr, defer); 214d9ae7f2bSJakub Kicinski } 215d9ae7f2bSJakub Kicinski 216d9ae7f2bSJakub Kicinski static void 217d9ae7f2bSJakub Kicinski __emit_immed(struct nfp_prog *nfp_prog, u16 areg, u16 breg, u16 imm_hi, 218d9ae7f2bSJakub Kicinski enum immed_width width, bool invert, 219995e101fSJakub Kicinski enum immed_shift shift, bool wr_both, 220995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 221d9ae7f2bSJakub Kicinski { 222d9ae7f2bSJakub Kicinski u64 insn; 223d9ae7f2bSJakub Kicinski 224d9ae7f2bSJakub Kicinski insn = OP_IMMED_BASE | 225d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_A_SRC, areg) | 226d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_B_SRC, breg) | 227d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_IMM, imm_hi) | 228d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_WIDTH, width) | 229d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_INV, invert) | 230d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_SHIFT, shift) | 231995e101fSJakub Kicinski FIELD_PREP(OP_IMMED_WR_AB, wr_both) | 232995e101fSJakub Kicinski FIELD_PREP(OP_IMMED_SRC_LMEXTN, src_lmextn) | 233995e101fSJakub Kicinski FIELD_PREP(OP_IMMED_DST_LMEXTN, dst_lmextn); 234d9ae7f2bSJakub Kicinski 235d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 236d9ae7f2bSJakub Kicinski } 237d9ae7f2bSJakub Kicinski 238d9ae7f2bSJakub Kicinski static void 239b3f868dfSJakub Kicinski emit_immed(struct nfp_prog *nfp_prog, swreg dst, u16 imm, 240d9ae7f2bSJakub Kicinski enum immed_width width, bool invert, enum immed_shift shift) 241d9ae7f2bSJakub Kicinski { 242d9ae7f2bSJakub Kicinski struct nfp_insn_ur_regs reg; 243d9ae7f2bSJakub Kicinski int err; 244d9ae7f2bSJakub Kicinski 245b3f868dfSJakub Kicinski if (swreg_type(dst) == NN_REG_IMM) { 246d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 247d9ae7f2bSJakub Kicinski return; 248d9ae7f2bSJakub Kicinski } 249d9ae7f2bSJakub Kicinski 250d9ae7f2bSJakub Kicinski err = swreg_to_unrestricted(dst, dst, reg_imm(imm & 0xff), ®); 251d9ae7f2bSJakub Kicinski if (err) { 252d9ae7f2bSJakub Kicinski nfp_prog->error = err; 253d9ae7f2bSJakub Kicinski return; 254d9ae7f2bSJakub Kicinski } 255d9ae7f2bSJakub Kicinski 2563239e7bbSJiong Wang /* Use reg.dst when destination is No-Dest. */ 2573239e7bbSJiong Wang __emit_immed(nfp_prog, 2583239e7bbSJiong Wang swreg_type(dst) == NN_REG_NONE ? reg.dst : reg.areg, 2593239e7bbSJiong Wang reg.breg, imm >> 8, width, invert, shift, 2603239e7bbSJiong Wang reg.wr_both, reg.dst_lmextn, reg.src_lmextn); 261d9ae7f2bSJakub Kicinski } 262d9ae7f2bSJakub Kicinski 263d9ae7f2bSJakub Kicinski static void 264d9ae7f2bSJakub Kicinski __emit_shf(struct nfp_prog *nfp_prog, u16 dst, enum alu_dst_ab dst_ab, 265d9ae7f2bSJakub Kicinski enum shf_sc sc, u8 shift, 266995e101fSJakub Kicinski u16 areg, enum shf_op op, u16 breg, bool i8, bool sw, bool wr_both, 267995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 268d9ae7f2bSJakub Kicinski { 269d9ae7f2bSJakub Kicinski u64 insn; 270d9ae7f2bSJakub Kicinski 271d9ae7f2bSJakub Kicinski if (!FIELD_FIT(OP_SHF_SHIFT, shift)) { 272d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 273d9ae7f2bSJakub Kicinski return; 274d9ae7f2bSJakub Kicinski } 275d9ae7f2bSJakub Kicinski 276d9ae7f2bSJakub Kicinski if (sc == SHF_SC_L_SHF) 277d9ae7f2bSJakub Kicinski shift = 32 - shift; 278d9ae7f2bSJakub Kicinski 279d9ae7f2bSJakub Kicinski insn = OP_SHF_BASE | 280d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_A_SRC, areg) | 281d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_SC, sc) | 282d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_B_SRC, breg) | 283d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_I8, i8) | 284d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_SW, sw) | 285d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_DST, dst) | 286d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_SHIFT, shift) | 287d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_OP, op) | 288d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_DST_AB, dst_ab) | 289995e101fSJakub Kicinski FIELD_PREP(OP_SHF_WR_AB, wr_both) | 290995e101fSJakub Kicinski FIELD_PREP(OP_SHF_SRC_LMEXTN, src_lmextn) | 291995e101fSJakub Kicinski FIELD_PREP(OP_SHF_DST_LMEXTN, dst_lmextn); 292d9ae7f2bSJakub Kicinski 293d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 294d9ae7f2bSJakub Kicinski } 295d9ae7f2bSJakub Kicinski 296d9ae7f2bSJakub Kicinski static void 297b3f868dfSJakub Kicinski emit_shf(struct nfp_prog *nfp_prog, swreg dst, 298b3f868dfSJakub Kicinski swreg lreg, enum shf_op op, swreg rreg, enum shf_sc sc, u8 shift) 299d9ae7f2bSJakub Kicinski { 300d9ae7f2bSJakub Kicinski struct nfp_insn_re_regs reg; 301d9ae7f2bSJakub Kicinski int err; 302d9ae7f2bSJakub Kicinski 303d9ae7f2bSJakub Kicinski err = swreg_to_restricted(dst, lreg, rreg, ®, true); 304d9ae7f2bSJakub Kicinski if (err) { 305d9ae7f2bSJakub Kicinski nfp_prog->error = err; 306d9ae7f2bSJakub Kicinski return; 307d9ae7f2bSJakub Kicinski } 308d9ae7f2bSJakub Kicinski 309d9ae7f2bSJakub Kicinski __emit_shf(nfp_prog, reg.dst, reg.dst_ab, sc, shift, 310995e101fSJakub Kicinski reg.areg, op, reg.breg, reg.i8, reg.swap, reg.wr_both, 311995e101fSJakub Kicinski reg.dst_lmextn, reg.src_lmextn); 312d9ae7f2bSJakub Kicinski } 313d9ae7f2bSJakub Kicinski 314d9ae7f2bSJakub Kicinski static void 315d9ae7f2bSJakub Kicinski __emit_alu(struct nfp_prog *nfp_prog, u16 dst, enum alu_dst_ab dst_ab, 316995e101fSJakub Kicinski u16 areg, enum alu_op op, u16 breg, bool swap, bool wr_both, 317995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 318d9ae7f2bSJakub Kicinski { 319d9ae7f2bSJakub Kicinski u64 insn; 320d9ae7f2bSJakub Kicinski 321d9ae7f2bSJakub Kicinski insn = OP_ALU_BASE | 322d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_A_SRC, areg) | 323d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_B_SRC, breg) | 324d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_DST, dst) | 325d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_SW, swap) | 326d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_OP, op) | 327d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_DST_AB, dst_ab) | 328995e101fSJakub Kicinski FIELD_PREP(OP_ALU_WR_AB, wr_both) | 329995e101fSJakub Kicinski FIELD_PREP(OP_ALU_SRC_LMEXTN, src_lmextn) | 330995e101fSJakub Kicinski FIELD_PREP(OP_ALU_DST_LMEXTN, dst_lmextn); 331d9ae7f2bSJakub Kicinski 332d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 333d9ae7f2bSJakub Kicinski } 334d9ae7f2bSJakub Kicinski 335d9ae7f2bSJakub Kicinski static void 336b3f868dfSJakub Kicinski emit_alu(struct nfp_prog *nfp_prog, swreg dst, 337b3f868dfSJakub Kicinski swreg lreg, enum alu_op op, swreg rreg) 338d9ae7f2bSJakub Kicinski { 339d9ae7f2bSJakub Kicinski struct nfp_insn_ur_regs reg; 340d9ae7f2bSJakub Kicinski int err; 341d9ae7f2bSJakub Kicinski 342d9ae7f2bSJakub Kicinski err = swreg_to_unrestricted(dst, lreg, rreg, ®); 343d9ae7f2bSJakub Kicinski if (err) { 344d9ae7f2bSJakub Kicinski nfp_prog->error = err; 345d9ae7f2bSJakub Kicinski return; 346d9ae7f2bSJakub Kicinski } 347d9ae7f2bSJakub Kicinski 348d9ae7f2bSJakub Kicinski __emit_alu(nfp_prog, reg.dst, reg.dst_ab, 349995e101fSJakub Kicinski reg.areg, op, reg.breg, reg.swap, reg.wr_both, 350995e101fSJakub Kicinski reg.dst_lmextn, reg.src_lmextn); 351d9ae7f2bSJakub Kicinski } 352d9ae7f2bSJakub Kicinski 353d9ae7f2bSJakub Kicinski static void 354d9ae7f2bSJakub Kicinski __emit_ld_field(struct nfp_prog *nfp_prog, enum shf_sc sc, 355d9ae7f2bSJakub Kicinski u8 areg, u8 bmask, u8 breg, u8 shift, bool imm8, 356995e101fSJakub Kicinski bool zero, bool swap, bool wr_both, 357995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 358d9ae7f2bSJakub Kicinski { 359d9ae7f2bSJakub Kicinski u64 insn; 360d9ae7f2bSJakub Kicinski 361d9ae7f2bSJakub Kicinski insn = OP_LDF_BASE | 362d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_A_SRC, areg) | 363d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_SC, sc) | 364d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_B_SRC, breg) | 365d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_I8, imm8) | 366d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_SW, swap) | 367d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_ZF, zero) | 368d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_BMASK, bmask) | 369d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_SHF, shift) | 370995e101fSJakub Kicinski FIELD_PREP(OP_LDF_WR_AB, wr_both) | 371995e101fSJakub Kicinski FIELD_PREP(OP_LDF_SRC_LMEXTN, src_lmextn) | 372995e101fSJakub Kicinski FIELD_PREP(OP_LDF_DST_LMEXTN, dst_lmextn); 373d9ae7f2bSJakub Kicinski 374d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 375d9ae7f2bSJakub Kicinski } 376d9ae7f2bSJakub Kicinski 377d9ae7f2bSJakub Kicinski static void 378bc8c80a8SJakub Kicinski emit_ld_field_any(struct nfp_prog *nfp_prog, swreg dst, u8 bmask, swreg src, 379bc8c80a8SJakub Kicinski enum shf_sc sc, u8 shift, bool zero) 380d9ae7f2bSJakub Kicinski { 381d9ae7f2bSJakub Kicinski struct nfp_insn_re_regs reg; 382d9ae7f2bSJakub Kicinski int err; 383d9ae7f2bSJakub Kicinski 3842de1be1dSJakub Kicinski /* Note: ld_field is special as it uses one of the src regs as dst */ 3852de1be1dSJakub Kicinski err = swreg_to_restricted(dst, dst, src, ®, true); 386d9ae7f2bSJakub Kicinski if (err) { 387d9ae7f2bSJakub Kicinski nfp_prog->error = err; 388d9ae7f2bSJakub Kicinski return; 389d9ae7f2bSJakub Kicinski } 390d9ae7f2bSJakub Kicinski 391d9ae7f2bSJakub Kicinski __emit_ld_field(nfp_prog, sc, reg.areg, bmask, reg.breg, shift, 392995e101fSJakub Kicinski reg.i8, zero, reg.swap, reg.wr_both, 393995e101fSJakub Kicinski reg.dst_lmextn, reg.src_lmextn); 394d9ae7f2bSJakub Kicinski } 395d9ae7f2bSJakub Kicinski 396d9ae7f2bSJakub Kicinski static void 397b3f868dfSJakub Kicinski emit_ld_field(struct nfp_prog *nfp_prog, swreg dst, u8 bmask, swreg src, 398d9ae7f2bSJakub Kicinski enum shf_sc sc, u8 shift) 399d9ae7f2bSJakub Kicinski { 400bc8c80a8SJakub Kicinski emit_ld_field_any(nfp_prog, dst, bmask, src, sc, shift, false); 401d9ae7f2bSJakub Kicinski } 402d9ae7f2bSJakub Kicinski 4032df03a50SJakub Kicinski static void 4042df03a50SJakub Kicinski __emit_lcsr(struct nfp_prog *nfp_prog, u16 areg, u16 breg, bool wr, u16 addr, 4052df03a50SJakub Kicinski bool dst_lmextn, bool src_lmextn) 4062df03a50SJakub Kicinski { 4072df03a50SJakub Kicinski u64 insn; 4082df03a50SJakub Kicinski 4092df03a50SJakub Kicinski insn = OP_LCSR_BASE | 4102df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_A_SRC, areg) | 4112df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_B_SRC, breg) | 4122df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_WRITE, wr) | 4132df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_ADDR, addr) | 4142df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_SRC_LMEXTN, src_lmextn) | 4152df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_DST_LMEXTN, dst_lmextn); 4162df03a50SJakub Kicinski 4172df03a50SJakub Kicinski nfp_prog_push(nfp_prog, insn); 4182df03a50SJakub Kicinski } 4192df03a50SJakub Kicinski 4202df03a50SJakub Kicinski static void emit_csr_wr(struct nfp_prog *nfp_prog, swreg src, u16 addr) 4212df03a50SJakub Kicinski { 4222df03a50SJakub Kicinski struct nfp_insn_ur_regs reg; 4232df03a50SJakub Kicinski int err; 4242df03a50SJakub Kicinski 4252df03a50SJakub Kicinski /* This instruction takes immeds instead of reg_none() for the ignored 4262df03a50SJakub Kicinski * operand, but we can't encode 2 immeds in one instr with our normal 4272df03a50SJakub Kicinski * swreg infra so if param is an immed, we encode as reg_none() and 4282df03a50SJakub Kicinski * copy the immed to both operands. 4292df03a50SJakub Kicinski */ 4302df03a50SJakub Kicinski if (swreg_type(src) == NN_REG_IMM) { 4312df03a50SJakub Kicinski err = swreg_to_unrestricted(reg_none(), src, reg_none(), ®); 4322df03a50SJakub Kicinski reg.breg = reg.areg; 4332df03a50SJakub Kicinski } else { 4342df03a50SJakub Kicinski err = swreg_to_unrestricted(reg_none(), src, reg_imm(0), ®); 4352df03a50SJakub Kicinski } 4362df03a50SJakub Kicinski if (err) { 4372df03a50SJakub Kicinski nfp_prog->error = err; 4382df03a50SJakub Kicinski return; 4392df03a50SJakub Kicinski } 4402df03a50SJakub Kicinski 4412df03a50SJakub Kicinski __emit_lcsr(nfp_prog, reg.areg, reg.breg, true, addr / 4, 4422df03a50SJakub Kicinski false, reg.src_lmextn); 4432df03a50SJakub Kicinski } 4442df03a50SJakub Kicinski 4451c03e03fSJakub Kicinski static void emit_nop(struct nfp_prog *nfp_prog) 4461c03e03fSJakub Kicinski { 4471c03e03fSJakub Kicinski __emit_immed(nfp_prog, UR_REG_IMM, UR_REG_IMM, 0, 0, 0, 0, 0, 0, 0); 4481c03e03fSJakub Kicinski } 4491c03e03fSJakub Kicinski 450d9ae7f2bSJakub Kicinski /* --- Wrappers --- */ 451d9ae7f2bSJakub Kicinski static bool pack_immed(u32 imm, u16 *val, enum immed_shift *shift) 452d9ae7f2bSJakub Kicinski { 453d9ae7f2bSJakub Kicinski if (!(imm & 0xffff0000)) { 454d9ae7f2bSJakub Kicinski *val = imm; 455d9ae7f2bSJakub Kicinski *shift = IMMED_SHIFT_0B; 456d9ae7f2bSJakub Kicinski } else if (!(imm & 0xff0000ff)) { 457d9ae7f2bSJakub Kicinski *val = imm >> 8; 458d9ae7f2bSJakub Kicinski *shift = IMMED_SHIFT_1B; 459d9ae7f2bSJakub Kicinski } else if (!(imm & 0x0000ffff)) { 460d9ae7f2bSJakub Kicinski *val = imm >> 16; 461d9ae7f2bSJakub Kicinski *shift = IMMED_SHIFT_2B; 462d9ae7f2bSJakub Kicinski } else { 463d9ae7f2bSJakub Kicinski return false; 464d9ae7f2bSJakub Kicinski } 465d9ae7f2bSJakub Kicinski 466d9ae7f2bSJakub Kicinski return true; 467d9ae7f2bSJakub Kicinski } 468d9ae7f2bSJakub Kicinski 469b3f868dfSJakub Kicinski static void wrp_immed(struct nfp_prog *nfp_prog, swreg dst, u32 imm) 470d9ae7f2bSJakub Kicinski { 471d9ae7f2bSJakub Kicinski enum immed_shift shift; 472d9ae7f2bSJakub Kicinski u16 val; 473d9ae7f2bSJakub Kicinski 474d9ae7f2bSJakub Kicinski if (pack_immed(imm, &val, &shift)) { 475d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, val, IMMED_WIDTH_ALL, false, shift); 476d9ae7f2bSJakub Kicinski } else if (pack_immed(~imm, &val, &shift)) { 477d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, val, IMMED_WIDTH_ALL, true, shift); 478d9ae7f2bSJakub Kicinski } else { 479d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, imm & 0xffff, IMMED_WIDTH_ALL, 480d9ae7f2bSJakub Kicinski false, IMMED_SHIFT_0B); 481d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, imm >> 16, IMMED_WIDTH_WORD, 482d9ae7f2bSJakub Kicinski false, IMMED_SHIFT_2B); 483d9ae7f2bSJakub Kicinski } 484d9ae7f2bSJakub Kicinski } 485d9ae7f2bSJakub Kicinski 486d9ae7f2bSJakub Kicinski /* ur_load_imm_any() - encode immediate or use tmp register (unrestricted) 487d9ae7f2bSJakub Kicinski * If the @imm is small enough encode it directly in operand and return 488d9ae7f2bSJakub Kicinski * otherwise load @imm to a spare register and return its encoding. 489d9ae7f2bSJakub Kicinski */ 490b3f868dfSJakub Kicinski static swreg ur_load_imm_any(struct nfp_prog *nfp_prog, u32 imm, swreg tmp_reg) 491d9ae7f2bSJakub Kicinski { 492d9ae7f2bSJakub Kicinski if (FIELD_FIT(UR_REG_IMM_MAX, imm)) 493d9ae7f2bSJakub Kicinski return reg_imm(imm); 494d9ae7f2bSJakub Kicinski 495d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, tmp_reg, imm); 496d9ae7f2bSJakub Kicinski return tmp_reg; 497d9ae7f2bSJakub Kicinski } 498d9ae7f2bSJakub Kicinski 499d9ae7f2bSJakub Kicinski /* re_load_imm_any() - encode immediate or use tmp register (restricted) 500d9ae7f2bSJakub Kicinski * If the @imm is small enough encode it directly in operand and return 501d9ae7f2bSJakub Kicinski * otherwise load @imm to a spare register and return its encoding. 502d9ae7f2bSJakub Kicinski */ 503b3f868dfSJakub Kicinski static swreg re_load_imm_any(struct nfp_prog *nfp_prog, u32 imm, swreg tmp_reg) 504d9ae7f2bSJakub Kicinski { 505d9ae7f2bSJakub Kicinski if (FIELD_FIT(RE_REG_IMM_MAX, imm)) 506d9ae7f2bSJakub Kicinski return reg_imm(imm); 507d9ae7f2bSJakub Kicinski 508d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, tmp_reg, imm); 509d9ae7f2bSJakub Kicinski return tmp_reg; 510d9ae7f2bSJakub Kicinski } 511d9ae7f2bSJakub Kicinski 512ff42bb9fSJakub Kicinski static void wrp_nops(struct nfp_prog *nfp_prog, unsigned int count) 513ff42bb9fSJakub Kicinski { 514ff42bb9fSJakub Kicinski while (count--) 515ff42bb9fSJakub Kicinski emit_nop(nfp_prog); 516ff42bb9fSJakub Kicinski } 517ff42bb9fSJakub Kicinski 518d9ae7f2bSJakub Kicinski static void 519d9ae7f2bSJakub Kicinski wrp_br_special(struct nfp_prog *nfp_prog, enum br_mask mask, 520d9ae7f2bSJakub Kicinski enum br_special special) 521d9ae7f2bSJakub Kicinski { 522d9ae7f2bSJakub Kicinski emit_br(nfp_prog, mask, 0, 0); 523d9ae7f2bSJakub Kicinski 524d9ae7f2bSJakub Kicinski nfp_prog->prog[nfp_prog->prog_len - 1] |= 525d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_SPECIAL, special); 526d9ae7f2bSJakub Kicinski } 527d9ae7f2bSJakub Kicinski 528c000dfb5SJakub Kicinski static void wrp_mov(struct nfp_prog *nfp_prog, swreg dst, swreg src) 529c000dfb5SJakub Kicinski { 530c000dfb5SJakub Kicinski emit_alu(nfp_prog, dst, reg_none(), ALU_OP_NONE, src); 531c000dfb5SJakub Kicinski } 532c000dfb5SJakub Kicinski 533d9ae7f2bSJakub Kicinski static void wrp_reg_mov(struct nfp_prog *nfp_prog, u16 dst, u16 src) 534d9ae7f2bSJakub Kicinski { 535c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst), reg_b(src)); 536d9ae7f2bSJakub Kicinski } 537d9ae7f2bSJakub Kicinski 5389879a381SJiong Wang /* wrp_reg_subpart() - load @field_len bytes from @offset of @src, write the 5399879a381SJiong Wang * result to @dst from low end. 5409879a381SJiong Wang */ 5419879a381SJiong Wang static void 5429879a381SJiong Wang wrp_reg_subpart(struct nfp_prog *nfp_prog, swreg dst, swreg src, u8 field_len, 5439879a381SJiong Wang u8 offset) 5449879a381SJiong Wang { 5459879a381SJiong Wang enum shf_sc sc = offset ? SHF_SC_R_SHF : SHF_SC_NONE; 5469879a381SJiong Wang u8 mask = (1 << field_len) - 1; 5479879a381SJiong Wang 5489879a381SJiong Wang emit_ld_field_any(nfp_prog, dst, mask, src, sc, offset * 8, true); 5499879a381SJiong Wang } 5509879a381SJiong Wang 5519879a381SJiong Wang /* NFP has Command Push Pull bus which supports bluk memory operations. */ 5529879a381SJiong Wang static int nfp_cpp_memcpy(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 5539879a381SJiong Wang { 5549879a381SJiong Wang bool descending_seq = meta->ldst_gather_len < 0; 5559879a381SJiong Wang s16 len = abs(meta->ldst_gather_len); 5569879a381SJiong Wang swreg src_base, off; 5579879a381SJiong Wang unsigned int i; 5589879a381SJiong Wang u8 xfer_num; 5599879a381SJiong Wang 5609879a381SJiong Wang off = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 5619879a381SJiong Wang src_base = reg_a(meta->insn.src_reg * 2); 5629879a381SJiong Wang xfer_num = round_up(len, 4) / 4; 5639879a381SJiong Wang 5648c900538SJiong Wang /* Setup PREV_ALU fields to override memory read length. */ 5658c900538SJiong Wang if (len > 32) 5668c900538SJiong Wang wrp_immed(nfp_prog, reg_none(), 5678c900538SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, xfer_num - 1)); 5688c900538SJiong Wang 5699879a381SJiong Wang /* Memory read from source addr into transfer-in registers. */ 5708c900538SJiong Wang emit_cmd_any(nfp_prog, CMD_TGT_READ32_SWAP, CMD_MODE_32b, 0, src_base, 5718c900538SJiong Wang off, xfer_num - 1, true, len > 32); 5729879a381SJiong Wang 5739879a381SJiong Wang /* Move from transfer-in to transfer-out. */ 5749879a381SJiong Wang for (i = 0; i < xfer_num; i++) 5759879a381SJiong Wang wrp_mov(nfp_prog, reg_xfer(i), reg_xfer(i)); 5769879a381SJiong Wang 5779879a381SJiong Wang off = re_load_imm_any(nfp_prog, meta->paired_st->off, imm_b(nfp_prog)); 5789879a381SJiong Wang 5799879a381SJiong Wang if (len <= 8) { 5809879a381SJiong Wang /* Use single direct_ref write8. */ 5819879a381SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 5829879a381SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, len - 1, 5839879a381SJiong Wang true); 5848c900538SJiong Wang } else if (len <= 32 && IS_ALIGNED(len, 4)) { 5859879a381SJiong Wang /* Use single direct_ref write32. */ 5869879a381SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 5879879a381SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, xfer_num - 1, 5889879a381SJiong Wang true); 5898c900538SJiong Wang } else if (len <= 32) { 5909879a381SJiong Wang /* Use single indirect_ref write8. */ 5919879a381SJiong Wang wrp_immed(nfp_prog, reg_none(), 5929879a381SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, len - 1)); 5939879a381SJiong Wang emit_cmd_indir(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 5949879a381SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 5959879a381SJiong Wang len - 1, true); 5968c900538SJiong Wang } else if (IS_ALIGNED(len, 4)) { 5978c900538SJiong Wang /* Use single indirect_ref write32. */ 5988c900538SJiong Wang wrp_immed(nfp_prog, reg_none(), 5998c900538SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, xfer_num - 1)); 6008c900538SJiong Wang emit_cmd_indir(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 6018c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 6028c900538SJiong Wang xfer_num - 1, true); 6038c900538SJiong Wang } else if (len <= 40) { 6048c900538SJiong Wang /* Use one direct_ref write32 to write the first 32-bytes, then 6058c900538SJiong Wang * another direct_ref write8 to write the remaining bytes. 6068c900538SJiong Wang */ 6078c900538SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 6088c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 7, 6098c900538SJiong Wang true); 6108c900538SJiong Wang 6118c900538SJiong Wang off = re_load_imm_any(nfp_prog, meta->paired_st->off + 32, 6128c900538SJiong Wang imm_b(nfp_prog)); 6138c900538SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 8, 6148c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, len - 33, 6158c900538SJiong Wang true); 6168c900538SJiong Wang } else { 6178c900538SJiong Wang /* Use one indirect_ref write32 to write 4-bytes aligned length, 6188c900538SJiong Wang * then another direct_ref write8 to write the remaining bytes. 6198c900538SJiong Wang */ 6208c900538SJiong Wang u8 new_off; 6218c900538SJiong Wang 6228c900538SJiong Wang wrp_immed(nfp_prog, reg_none(), 6238c900538SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, xfer_num - 2)); 6248c900538SJiong Wang emit_cmd_indir(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 6258c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 6268c900538SJiong Wang xfer_num - 2, true); 6278c900538SJiong Wang new_off = meta->paired_st->off + (xfer_num - 1) * 4; 6288c900538SJiong Wang off = re_load_imm_any(nfp_prog, new_off, imm_b(nfp_prog)); 6298c900538SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 6308c900538SJiong Wang xfer_num - 1, reg_a(meta->paired_st->dst_reg * 2), off, 6318c900538SJiong Wang (len & 0x3) - 1, true); 6329879a381SJiong Wang } 6339879a381SJiong Wang 6349879a381SJiong Wang /* TODO: The following extra load is to make sure data flow be identical 6359879a381SJiong Wang * before and after we do memory copy optimization. 6369879a381SJiong Wang * 6379879a381SJiong Wang * The load destination register is not guaranteed to be dead, so we 6389879a381SJiong Wang * need to make sure it is loaded with the value the same as before 6399879a381SJiong Wang * this transformation. 6409879a381SJiong Wang * 6419879a381SJiong Wang * These extra loads could be removed once we have accurate register 6429879a381SJiong Wang * usage information. 6439879a381SJiong Wang */ 6449879a381SJiong Wang if (descending_seq) 6459879a381SJiong Wang xfer_num = 0; 6469879a381SJiong Wang else if (BPF_SIZE(meta->insn.code) != BPF_DW) 6479879a381SJiong Wang xfer_num = xfer_num - 1; 6489879a381SJiong Wang else 6499879a381SJiong Wang xfer_num = xfer_num - 2; 6509879a381SJiong Wang 6519879a381SJiong Wang switch (BPF_SIZE(meta->insn.code)) { 6529879a381SJiong Wang case BPF_B: 6539879a381SJiong Wang wrp_reg_subpart(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6549879a381SJiong Wang reg_xfer(xfer_num), 1, 6559879a381SJiong Wang IS_ALIGNED(len, 4) ? 3 : (len & 3) - 1); 6569879a381SJiong Wang break; 6579879a381SJiong Wang case BPF_H: 6589879a381SJiong Wang wrp_reg_subpart(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6599879a381SJiong Wang reg_xfer(xfer_num), 2, (len & 3) ^ 2); 6609879a381SJiong Wang break; 6619879a381SJiong Wang case BPF_W: 6629879a381SJiong Wang wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6639879a381SJiong Wang reg_xfer(0)); 6649879a381SJiong Wang break; 6659879a381SJiong Wang case BPF_DW: 6669879a381SJiong Wang wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6679879a381SJiong Wang reg_xfer(xfer_num)); 6689879a381SJiong Wang wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 6699879a381SJiong Wang reg_xfer(xfer_num + 1)); 6709879a381SJiong Wang break; 6719879a381SJiong Wang } 6729879a381SJiong Wang 6739879a381SJiong Wang if (BPF_SIZE(meta->insn.code) != BPF_DW) 6749879a381SJiong Wang wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 6759879a381SJiong Wang 6769879a381SJiong Wang return 0; 6779879a381SJiong Wang } 6789879a381SJiong Wang 679d9ae7f2bSJakub Kicinski static int 6800a793977SJakub Kicinski data_ld(struct nfp_prog *nfp_prog, swreg offset, u8 dst_gpr, int size) 681d9ae7f2bSJakub Kicinski { 682d9ae7f2bSJakub Kicinski unsigned int i; 683d9ae7f2bSJakub Kicinski u16 shift, sz; 684d9ae7f2bSJakub Kicinski 685d9ae7f2bSJakub Kicinski /* We load the value from the address indicated in @offset and then 686d9ae7f2bSJakub Kicinski * shift out the data we don't need. Note: this is big endian! 687d9ae7f2bSJakub Kicinski */ 6880a793977SJakub Kicinski sz = max(size, 4); 689d9ae7f2bSJakub Kicinski shift = size < 4 ? 4 - size : 0; 690d9ae7f2bSJakub Kicinski 6910a793977SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_READ8, CMD_MODE_32b, 0, 6920a793977SJakub Kicinski pptr_reg(nfp_prog), offset, sz - 1, true); 6930a793977SJakub Kicinski 6940a793977SJakub Kicinski i = 0; 6950a793977SJakub Kicinski if (shift) 6960a793977SJakub Kicinski emit_shf(nfp_prog, reg_both(dst_gpr), reg_none(), SHF_OP_NONE, 6970a793977SJakub Kicinski reg_xfer(0), SHF_SC_R_SHF, shift * 8); 6980a793977SJakub Kicinski else 6990a793977SJakub Kicinski for (; i * 4 < size; i++) 7000a793977SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst_gpr + i), reg_xfer(i)); 7010a793977SJakub Kicinski 7020a793977SJakub Kicinski if (i < 2) 7030a793977SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst_gpr + 1), 0); 7040a793977SJakub Kicinski 7050a793977SJakub Kicinski return 0; 7060a793977SJakub Kicinski } 7070a793977SJakub Kicinski 7080a793977SJakub Kicinski static int 7092ca71441SJakub Kicinski data_ld_host_order(struct nfp_prog *nfp_prog, u8 src_gpr, swreg offset, 7102ca71441SJakub Kicinski u8 dst_gpr, int size) 7112ca71441SJakub Kicinski { 7122ca71441SJakub Kicinski unsigned int i; 7132ca71441SJakub Kicinski u8 mask, sz; 7142ca71441SJakub Kicinski 7152ca71441SJakub Kicinski /* We load the value from the address indicated in @offset and then 7162ca71441SJakub Kicinski * mask out the data we don't need. Note: this is little endian! 7172ca71441SJakub Kicinski */ 7182ca71441SJakub Kicinski sz = max(size, 4); 7192ca71441SJakub Kicinski mask = size < 4 ? GENMASK(size - 1, 0) : 0; 7202ca71441SJakub Kicinski 7212ca71441SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_READ32_SWAP, CMD_MODE_32b, 0, 7222ca71441SJakub Kicinski reg_a(src_gpr), offset, sz / 4 - 1, true); 7232ca71441SJakub Kicinski 7242ca71441SJakub Kicinski i = 0; 7252ca71441SJakub Kicinski if (mask) 7262ca71441SJakub Kicinski emit_ld_field_any(nfp_prog, reg_both(dst_gpr), mask, 7272ca71441SJakub Kicinski reg_xfer(0), SHF_SC_NONE, 0, true); 7282ca71441SJakub Kicinski else 7292ca71441SJakub Kicinski for (; i * 4 < size; i++) 7302ca71441SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst_gpr + i), reg_xfer(i)); 7312ca71441SJakub Kicinski 7322ca71441SJakub Kicinski if (i < 2) 7332ca71441SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst_gpr + 1), 0); 7342ca71441SJakub Kicinski 7352ca71441SJakub Kicinski return 0; 7362ca71441SJakub Kicinski } 7372ca71441SJakub Kicinski 7382ca71441SJakub Kicinski static int 7390a793977SJakub Kicinski construct_data_ind_ld(struct nfp_prog *nfp_prog, u16 offset, u16 src, u8 size) 7400a793977SJakub Kicinski { 7410a793977SJakub Kicinski swreg tmp_reg; 7420a793977SJakub Kicinski 743d9ae7f2bSJakub Kicinski /* Calculate the true offset (src_reg + imm) */ 744d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, offset, imm_b(nfp_prog)); 7450a793977SJakub Kicinski emit_alu(nfp_prog, imm_both(nfp_prog), reg_a(src), ALU_OP_ADD, tmp_reg); 7460a793977SJakub Kicinski 747d9ae7f2bSJakub Kicinski /* Check packet length (size guaranteed to fit b/c it's u8) */ 748d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_a(nfp_prog), 749d9ae7f2bSJakub Kicinski imm_a(nfp_prog), ALU_OP_ADD, reg_imm(size)); 750d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 75118e53b6cSJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, imm_a(nfp_prog)); 752d9ae7f2bSJakub Kicinski wrp_br_special(nfp_prog, BR_BLO, OP_BR_GO_ABORT); 7530a793977SJakub Kicinski 754d9ae7f2bSJakub Kicinski /* Load data */ 7550a793977SJakub Kicinski return data_ld(nfp_prog, imm_b(nfp_prog), 0, size); 756d9ae7f2bSJakub Kicinski } 757d9ae7f2bSJakub Kicinski 758d9ae7f2bSJakub Kicinski static int construct_data_ld(struct nfp_prog *nfp_prog, u16 offset, u8 size) 759d9ae7f2bSJakub Kicinski { 7600a793977SJakub Kicinski swreg tmp_reg; 7610a793977SJakub Kicinski 7620a793977SJakub Kicinski /* Check packet length */ 7630a793977SJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, offset + size, imm_a(nfp_prog)); 7640a793977SJakub Kicinski emit_alu(nfp_prog, reg_none(), plen_reg(nfp_prog), ALU_OP_SUB, tmp_reg); 7650a793977SJakub Kicinski wrp_br_special(nfp_prog, BR_BLO, OP_BR_GO_ABORT); 7660a793977SJakub Kicinski 7670a793977SJakub Kicinski /* Load data */ 7680a793977SJakub Kicinski tmp_reg = re_load_imm_any(nfp_prog, offset, imm_b(nfp_prog)); 7690a793977SJakub Kicinski return data_ld(nfp_prog, tmp_reg, 0, size); 770d9ae7f2bSJakub Kicinski } 771d9ae7f2bSJakub Kicinski 772e663fe38SJakub Kicinski static int 773e663fe38SJakub Kicinski data_stx_host_order(struct nfp_prog *nfp_prog, u8 dst_gpr, swreg offset, 774e663fe38SJakub Kicinski u8 src_gpr, u8 size) 775e663fe38SJakub Kicinski { 776e663fe38SJakub Kicinski unsigned int i; 777e663fe38SJakub Kicinski 778e663fe38SJakub Kicinski for (i = 0; i * 4 < size; i++) 779e663fe38SJakub Kicinski wrp_mov(nfp_prog, reg_xfer(i), reg_a(src_gpr + i)); 780e663fe38SJakub Kicinski 781e663fe38SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 782e663fe38SJakub Kicinski reg_a(dst_gpr), offset, size - 1, true); 783e663fe38SJakub Kicinski 784e663fe38SJakub Kicinski return 0; 785e663fe38SJakub Kicinski } 786e663fe38SJakub Kicinski 787e663fe38SJakub Kicinski static int 788e663fe38SJakub Kicinski data_st_host_order(struct nfp_prog *nfp_prog, u8 dst_gpr, swreg offset, 789e663fe38SJakub Kicinski u64 imm, u8 size) 790e663fe38SJakub Kicinski { 791e663fe38SJakub Kicinski wrp_immed(nfp_prog, reg_xfer(0), imm); 792e663fe38SJakub Kicinski if (size == 8) 793e663fe38SJakub Kicinski wrp_immed(nfp_prog, reg_xfer(1), imm >> 32); 794e663fe38SJakub Kicinski 795e663fe38SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 796e663fe38SJakub Kicinski reg_a(dst_gpr), offset, size - 1, true); 797e663fe38SJakub Kicinski 798e663fe38SJakub Kicinski return 0; 799e663fe38SJakub Kicinski } 800e663fe38SJakub Kicinski 801ee9133a8SJakub Kicinski typedef int 802ee9133a8SJakub Kicinski (*lmem_step)(struct nfp_prog *nfp_prog, u8 gpr, u8 gpr_byte, s32 off, 8032df03a50SJakub Kicinski unsigned int size, bool first, bool new_gpr, bool last, bool lm3, 8042df03a50SJakub Kicinski bool needs_inc); 805a82b23fbSJakub Kicinski 806a82b23fbSJakub Kicinski static int 807a82b23fbSJakub Kicinski wrp_lmem_load(struct nfp_prog *nfp_prog, u8 dst, u8 dst_byte, s32 off, 8082df03a50SJakub Kicinski unsigned int size, bool first, bool new_gpr, bool last, bool lm3, 8092df03a50SJakub Kicinski bool needs_inc) 810a82b23fbSJakub Kicinski { 8112df03a50SJakub Kicinski bool should_inc = needs_inc && new_gpr && !last; 812a82b23fbSJakub Kicinski u32 idx, src_byte; 813a82b23fbSJakub Kicinski enum shf_sc sc; 814a82b23fbSJakub Kicinski swreg reg; 815a82b23fbSJakub Kicinski int shf; 816a82b23fbSJakub Kicinski u8 mask; 817a82b23fbSJakub Kicinski 818a82b23fbSJakub Kicinski if (WARN_ON_ONCE(dst_byte + size > 4 || off % 4 + size > 4)) 819a82b23fbSJakub Kicinski return -EOPNOTSUPP; 820a82b23fbSJakub Kicinski 821a82b23fbSJakub Kicinski idx = off / 4; 822a82b23fbSJakub Kicinski 823a82b23fbSJakub Kicinski /* Move the entire word */ 824a82b23fbSJakub Kicinski if (size == 4) { 8252df03a50SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst), 8262df03a50SJakub Kicinski should_inc ? reg_lm_inc(3) : reg_lm(lm3 ? 3 : 0, idx)); 827a82b23fbSJakub Kicinski return 0; 828a82b23fbSJakub Kicinski } 829a82b23fbSJakub Kicinski 8302df03a50SJakub Kicinski if (WARN_ON_ONCE(lm3 && idx > RE_REG_LM_IDX_MAX)) 8312df03a50SJakub Kicinski return -EOPNOTSUPP; 8322df03a50SJakub Kicinski 833a82b23fbSJakub Kicinski src_byte = off % 4; 834a82b23fbSJakub Kicinski 835a82b23fbSJakub Kicinski mask = (1 << size) - 1; 836a82b23fbSJakub Kicinski mask <<= dst_byte; 837a82b23fbSJakub Kicinski 838a82b23fbSJakub Kicinski if (WARN_ON_ONCE(mask > 0xf)) 839a82b23fbSJakub Kicinski return -EOPNOTSUPP; 840a82b23fbSJakub Kicinski 841a82b23fbSJakub Kicinski shf = abs(src_byte - dst_byte) * 8; 842a82b23fbSJakub Kicinski if (src_byte == dst_byte) { 843a82b23fbSJakub Kicinski sc = SHF_SC_NONE; 844a82b23fbSJakub Kicinski } else if (src_byte < dst_byte) { 845a82b23fbSJakub Kicinski shf = 32 - shf; 846a82b23fbSJakub Kicinski sc = SHF_SC_L_SHF; 847a82b23fbSJakub Kicinski } else { 848a82b23fbSJakub Kicinski sc = SHF_SC_R_SHF; 849a82b23fbSJakub Kicinski } 850a82b23fbSJakub Kicinski 851a82b23fbSJakub Kicinski /* ld_field can address fewer indexes, if offset too large do RMW. 852a82b23fbSJakub Kicinski * Because we RMV twice we waste 2 cycles on unaligned 8 byte writes. 853a82b23fbSJakub Kicinski */ 854a82b23fbSJakub Kicinski if (idx <= RE_REG_LM_IDX_MAX) { 8552df03a50SJakub Kicinski reg = reg_lm(lm3 ? 3 : 0, idx); 856a82b23fbSJakub Kicinski } else { 857a82b23fbSJakub Kicinski reg = imm_a(nfp_prog); 8589a90c83cSJakub Kicinski /* If it's not the first part of the load and we start a new GPR 8599a90c83cSJakub Kicinski * that means we are loading a second part of the LMEM word into 8609a90c83cSJakub Kicinski * a new GPR. IOW we've already looked that LMEM word and 8619a90c83cSJakub Kicinski * therefore it has been loaded into imm_a(). 8629a90c83cSJakub Kicinski */ 8639a90c83cSJakub Kicinski if (first || !new_gpr) 864a82b23fbSJakub Kicinski wrp_mov(nfp_prog, reg, reg_lm(0, idx)); 865a82b23fbSJakub Kicinski } 866a82b23fbSJakub Kicinski 867a82b23fbSJakub Kicinski emit_ld_field_any(nfp_prog, reg_both(dst), mask, reg, sc, shf, new_gpr); 868a82b23fbSJakub Kicinski 8692df03a50SJakub Kicinski if (should_inc) 8702df03a50SJakub Kicinski wrp_mov(nfp_prog, reg_none(), reg_lm_inc(3)); 8712df03a50SJakub Kicinski 872a82b23fbSJakub Kicinski return 0; 873a82b23fbSJakub Kicinski } 874ee9133a8SJakub Kicinski 875ee9133a8SJakub Kicinski static int 876ee9133a8SJakub Kicinski wrp_lmem_store(struct nfp_prog *nfp_prog, u8 src, u8 src_byte, s32 off, 8772df03a50SJakub Kicinski unsigned int size, bool first, bool new_gpr, bool last, bool lm3, 8782df03a50SJakub Kicinski bool needs_inc) 879ee9133a8SJakub Kicinski { 8802df03a50SJakub Kicinski bool should_inc = needs_inc && new_gpr && !last; 881ee9133a8SJakub Kicinski u32 idx, dst_byte; 882ee9133a8SJakub Kicinski enum shf_sc sc; 883ee9133a8SJakub Kicinski swreg reg; 884ee9133a8SJakub Kicinski int shf; 885ee9133a8SJakub Kicinski u8 mask; 886ee9133a8SJakub Kicinski 887ee9133a8SJakub Kicinski if (WARN_ON_ONCE(src_byte + size > 4 || off % 4 + size > 4)) 888ee9133a8SJakub Kicinski return -EOPNOTSUPP; 889ee9133a8SJakub Kicinski 890ee9133a8SJakub Kicinski idx = off / 4; 891ee9133a8SJakub Kicinski 892ee9133a8SJakub Kicinski /* Move the entire word */ 893ee9133a8SJakub Kicinski if (size == 4) { 8942df03a50SJakub Kicinski wrp_mov(nfp_prog, 8952df03a50SJakub Kicinski should_inc ? reg_lm_inc(3) : reg_lm(lm3 ? 3 : 0, idx), 8962df03a50SJakub Kicinski reg_b(src)); 897ee9133a8SJakub Kicinski return 0; 898ee9133a8SJakub Kicinski } 899ee9133a8SJakub Kicinski 9002df03a50SJakub Kicinski if (WARN_ON_ONCE(lm3 && idx > RE_REG_LM_IDX_MAX)) 9012df03a50SJakub Kicinski return -EOPNOTSUPP; 9022df03a50SJakub Kicinski 903ee9133a8SJakub Kicinski dst_byte = off % 4; 904ee9133a8SJakub Kicinski 905ee9133a8SJakub Kicinski mask = (1 << size) - 1; 906ee9133a8SJakub Kicinski mask <<= dst_byte; 907ee9133a8SJakub Kicinski 908ee9133a8SJakub Kicinski if (WARN_ON_ONCE(mask > 0xf)) 909ee9133a8SJakub Kicinski return -EOPNOTSUPP; 910ee9133a8SJakub Kicinski 911ee9133a8SJakub Kicinski shf = abs(src_byte - dst_byte) * 8; 912ee9133a8SJakub Kicinski if (src_byte == dst_byte) { 913ee9133a8SJakub Kicinski sc = SHF_SC_NONE; 914ee9133a8SJakub Kicinski } else if (src_byte < dst_byte) { 915ee9133a8SJakub Kicinski shf = 32 - shf; 916ee9133a8SJakub Kicinski sc = SHF_SC_L_SHF; 917ee9133a8SJakub Kicinski } else { 918ee9133a8SJakub Kicinski sc = SHF_SC_R_SHF; 919ee9133a8SJakub Kicinski } 920ee9133a8SJakub Kicinski 921ee9133a8SJakub Kicinski /* ld_field can address fewer indexes, if offset too large do RMW. 922ee9133a8SJakub Kicinski * Because we RMV twice we waste 2 cycles on unaligned 8 byte writes. 923ee9133a8SJakub Kicinski */ 924ee9133a8SJakub Kicinski if (idx <= RE_REG_LM_IDX_MAX) { 9252df03a50SJakub Kicinski reg = reg_lm(lm3 ? 3 : 0, idx); 926ee9133a8SJakub Kicinski } else { 927ee9133a8SJakub Kicinski reg = imm_a(nfp_prog); 9289a90c83cSJakub Kicinski /* Only first and last LMEM locations are going to need RMW, 9299a90c83cSJakub Kicinski * the middle location will be overwritten fully. 9309a90c83cSJakub Kicinski */ 9319a90c83cSJakub Kicinski if (first || last) 932ee9133a8SJakub Kicinski wrp_mov(nfp_prog, reg, reg_lm(0, idx)); 933ee9133a8SJakub Kicinski } 934ee9133a8SJakub Kicinski 935ee9133a8SJakub Kicinski emit_ld_field(nfp_prog, reg, mask, reg_b(src), sc, shf); 936ee9133a8SJakub Kicinski 9379a90c83cSJakub Kicinski if (new_gpr || last) { 938ee9133a8SJakub Kicinski if (idx > RE_REG_LM_IDX_MAX) 939ee9133a8SJakub Kicinski wrp_mov(nfp_prog, reg_lm(0, idx), reg); 9402df03a50SJakub Kicinski if (should_inc) 9412df03a50SJakub Kicinski wrp_mov(nfp_prog, reg_none(), reg_lm_inc(3)); 9429a90c83cSJakub Kicinski } 943ee9133a8SJakub Kicinski 944ee9133a8SJakub Kicinski return 0; 945ee9133a8SJakub Kicinski } 946ee9133a8SJakub Kicinski 947ee9133a8SJakub Kicinski static int 948ee9133a8SJakub Kicinski mem_op_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 949b14157eeSJakub Kicinski unsigned int size, unsigned int ptr_off, u8 gpr, u8 ptr_gpr, 950b14157eeSJakub Kicinski bool clr_gpr, lmem_step step) 951ee9133a8SJakub Kicinski { 952d3488480SJakub Kicinski s32 off = nfp_prog->stack_depth + meta->insn.off + ptr_off; 9539a90c83cSJakub Kicinski bool first = true, last; 9542df03a50SJakub Kicinski bool needs_inc = false; 9552df03a50SJakub Kicinski swreg stack_off_reg; 956a82b23fbSJakub Kicinski u8 prev_gpr = 255; 957ee9133a8SJakub Kicinski u32 gpr_byte = 0; 9582df03a50SJakub Kicinski bool lm3 = true; 959ee9133a8SJakub Kicinski int ret; 960ee9133a8SJakub Kicinski 961b14157eeSJakub Kicinski if (meta->ptr_not_const) { 962b14157eeSJakub Kicinski /* Use of the last encountered ptr_off is OK, they all have 963b14157eeSJakub Kicinski * the same alignment. Depend on low bits of value being 964b14157eeSJakub Kicinski * discarded when written to LMaddr register. 965b14157eeSJakub Kicinski */ 966b14157eeSJakub Kicinski stack_off_reg = ur_load_imm_any(nfp_prog, meta->insn.off, 967b14157eeSJakub Kicinski stack_imm(nfp_prog)); 968b14157eeSJakub Kicinski 969b14157eeSJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 970b14157eeSJakub Kicinski reg_a(ptr_gpr), ALU_OP_ADD, stack_off_reg); 971b14157eeSJakub Kicinski 972b14157eeSJakub Kicinski needs_inc = true; 973b14157eeSJakub Kicinski } else if (off + size <= 64) { 9742df03a50SJakub Kicinski /* We can reach bottom 64B with LMaddr0 */ 9752df03a50SJakub Kicinski lm3 = false; 9762df03a50SJakub Kicinski } else if (round_down(off, 32) == round_down(off + size - 1, 32)) { 9772df03a50SJakub Kicinski /* We have to set up a new pointer. If we know the offset 9782df03a50SJakub Kicinski * and the entire access falls into a single 32 byte aligned 9792df03a50SJakub Kicinski * window we won't have to increment the LM pointer. 9802df03a50SJakub Kicinski * The 32 byte alignment is imporant because offset is ORed in 9812df03a50SJakub Kicinski * not added when doing *l$indexN[off]. 9822df03a50SJakub Kicinski */ 9832df03a50SJakub Kicinski stack_off_reg = ur_load_imm_any(nfp_prog, round_down(off, 32), 9842df03a50SJakub Kicinski stack_imm(nfp_prog)); 9852df03a50SJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 9862df03a50SJakub Kicinski stack_reg(nfp_prog), ALU_OP_ADD, stack_off_reg); 9872df03a50SJakub Kicinski 9882df03a50SJakub Kicinski off %= 32; 9892df03a50SJakub Kicinski } else { 9902df03a50SJakub Kicinski stack_off_reg = ur_load_imm_any(nfp_prog, round_down(off, 4), 9912df03a50SJakub Kicinski stack_imm(nfp_prog)); 9922df03a50SJakub Kicinski 9932df03a50SJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 9942df03a50SJakub Kicinski stack_reg(nfp_prog), ALU_OP_ADD, stack_off_reg); 9952df03a50SJakub Kicinski 9962df03a50SJakub Kicinski needs_inc = true; 9972df03a50SJakub Kicinski } 9982df03a50SJakub Kicinski if (lm3) { 9992df03a50SJakub Kicinski emit_csr_wr(nfp_prog, imm_b(nfp_prog), NFP_CSR_ACT_LM_ADDR3); 10002df03a50SJakub Kicinski /* For size < 4 one slot will be filled by zeroing of upper. */ 10012df03a50SJakub Kicinski wrp_nops(nfp_prog, clr_gpr && size < 8 ? 2 : 3); 10022df03a50SJakub Kicinski } 10032df03a50SJakub Kicinski 1004a82b23fbSJakub Kicinski if (clr_gpr && size < 8) 1005a82b23fbSJakub Kicinski wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 1006a82b23fbSJakub Kicinski 1007ee9133a8SJakub Kicinski while (size) { 1008ee9133a8SJakub Kicinski u32 slice_end; 1009ee9133a8SJakub Kicinski u8 slice_size; 1010ee9133a8SJakub Kicinski 1011ee9133a8SJakub Kicinski slice_size = min(size, 4 - gpr_byte); 1012ee9133a8SJakub Kicinski slice_end = min(off + slice_size, round_up(off + 1, 4)); 1013ee9133a8SJakub Kicinski slice_size = slice_end - off; 1014ee9133a8SJakub Kicinski 10159a90c83cSJakub Kicinski last = slice_size == size; 10169a90c83cSJakub Kicinski 10172df03a50SJakub Kicinski if (needs_inc) 10182df03a50SJakub Kicinski off %= 4; 10192df03a50SJakub Kicinski 1020a82b23fbSJakub Kicinski ret = step(nfp_prog, gpr, gpr_byte, off, slice_size, 10212df03a50SJakub Kicinski first, gpr != prev_gpr, last, lm3, needs_inc); 1022ee9133a8SJakub Kicinski if (ret) 1023ee9133a8SJakub Kicinski return ret; 1024ee9133a8SJakub Kicinski 1025a82b23fbSJakub Kicinski prev_gpr = gpr; 10269a90c83cSJakub Kicinski first = false; 10279a90c83cSJakub Kicinski 1028ee9133a8SJakub Kicinski gpr_byte += slice_size; 1029ee9133a8SJakub Kicinski if (gpr_byte >= 4) { 1030ee9133a8SJakub Kicinski gpr_byte -= 4; 1031ee9133a8SJakub Kicinski gpr++; 1032ee9133a8SJakub Kicinski } 1033ee9133a8SJakub Kicinski 1034ee9133a8SJakub Kicinski size -= slice_size; 1035ee9133a8SJakub Kicinski off += slice_size; 1036ee9133a8SJakub Kicinski } 1037ee9133a8SJakub Kicinski 1038ee9133a8SJakub Kicinski return 0; 1039ee9133a8SJakub Kicinski } 1040ee9133a8SJakub Kicinski 1041d9ae7f2bSJakub Kicinski static void 1042d9ae7f2bSJakub Kicinski wrp_alu_imm(struct nfp_prog *nfp_prog, u8 dst, enum alu_op alu_op, u32 imm) 1043d9ae7f2bSJakub Kicinski { 1044b3f868dfSJakub Kicinski swreg tmp_reg; 1045d9ae7f2bSJakub Kicinski 1046d9ae7f2bSJakub Kicinski if (alu_op == ALU_OP_AND) { 1047d9ae7f2bSJakub Kicinski if (!imm) 1048d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), 0); 1049d9ae7f2bSJakub Kicinski if (!imm || !~imm) 1050d9ae7f2bSJakub Kicinski return; 1051d9ae7f2bSJakub Kicinski } 1052d9ae7f2bSJakub Kicinski if (alu_op == ALU_OP_OR) { 1053d9ae7f2bSJakub Kicinski if (!~imm) 1054d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), ~0U); 1055d9ae7f2bSJakub Kicinski if (!imm || !~imm) 1056d9ae7f2bSJakub Kicinski return; 1057d9ae7f2bSJakub Kicinski } 1058d9ae7f2bSJakub Kicinski if (alu_op == ALU_OP_XOR) { 1059d9ae7f2bSJakub Kicinski if (!~imm) 1060d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_none(), 10615d42ced1SJiong Wang ALU_OP_NOT, reg_b(dst)); 1062d9ae7f2bSJakub Kicinski if (!imm || !~imm) 1063d9ae7f2bSJakub Kicinski return; 1064d9ae7f2bSJakub Kicinski } 1065d9ae7f2bSJakub Kicinski 1066d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm, imm_b(nfp_prog)); 1067d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, tmp_reg); 1068d9ae7f2bSJakub Kicinski } 1069d9ae7f2bSJakub Kicinski 1070d9ae7f2bSJakub Kicinski static int 1071d9ae7f2bSJakub Kicinski wrp_alu64_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1072d9ae7f2bSJakub Kicinski enum alu_op alu_op, bool skip) 1073d9ae7f2bSJakub Kicinski { 1074d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1075d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1076d9ae7f2bSJakub Kicinski 1077d9ae7f2bSJakub Kicinski if (skip) { 1078d9ae7f2bSJakub Kicinski meta->skip = true; 1079d9ae7f2bSJakub Kicinski return 0; 1080d9ae7f2bSJakub Kicinski } 1081d9ae7f2bSJakub Kicinski 1082d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, alu_op, imm & ~0U); 1083d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, alu_op, imm >> 32); 1084d9ae7f2bSJakub Kicinski 1085d9ae7f2bSJakub Kicinski return 0; 1086d9ae7f2bSJakub Kicinski } 1087d9ae7f2bSJakub Kicinski 1088d9ae7f2bSJakub Kicinski static int 1089d9ae7f2bSJakub Kicinski wrp_alu64_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1090d9ae7f2bSJakub Kicinski enum alu_op alu_op) 1091d9ae7f2bSJakub Kicinski { 1092d9ae7f2bSJakub Kicinski u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2; 1093d9ae7f2bSJakub Kicinski 1094d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, reg_b(src)); 1095d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst + 1), 1096d9ae7f2bSJakub Kicinski reg_a(dst + 1), alu_op, reg_b(src + 1)); 1097d9ae7f2bSJakub Kicinski 1098d9ae7f2bSJakub Kicinski return 0; 1099d9ae7f2bSJakub Kicinski } 1100d9ae7f2bSJakub Kicinski 1101d9ae7f2bSJakub Kicinski static int 1102d9ae7f2bSJakub Kicinski wrp_alu32_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1103d9ae7f2bSJakub Kicinski enum alu_op alu_op, bool skip) 1104d9ae7f2bSJakub Kicinski { 1105d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1106d9ae7f2bSJakub Kicinski 1107d9ae7f2bSJakub Kicinski if (skip) { 1108d9ae7f2bSJakub Kicinski meta->skip = true; 1109d9ae7f2bSJakub Kicinski return 0; 1110d9ae7f2bSJakub Kicinski } 1111d9ae7f2bSJakub Kicinski 1112d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, alu_op, insn->imm); 1113d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1114d9ae7f2bSJakub Kicinski 1115d9ae7f2bSJakub Kicinski return 0; 1116d9ae7f2bSJakub Kicinski } 1117d9ae7f2bSJakub Kicinski 1118d9ae7f2bSJakub Kicinski static int 1119d9ae7f2bSJakub Kicinski wrp_alu32_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1120d9ae7f2bSJakub Kicinski enum alu_op alu_op) 1121d9ae7f2bSJakub Kicinski { 1122d9ae7f2bSJakub Kicinski u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2; 1123d9ae7f2bSJakub Kicinski 1124d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, reg_b(src)); 1125d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1126d9ae7f2bSJakub Kicinski 1127d9ae7f2bSJakub Kicinski return 0; 1128d9ae7f2bSJakub Kicinski } 1129d9ae7f2bSJakub Kicinski 1130d9ae7f2bSJakub Kicinski static void 1131d9ae7f2bSJakub Kicinski wrp_test_reg_one(struct nfp_prog *nfp_prog, u8 dst, enum alu_op alu_op, u8 src, 1132d9ae7f2bSJakub Kicinski enum br_mask br_mask, u16 off) 1133d9ae7f2bSJakub Kicinski { 1134d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(dst), alu_op, reg_b(src)); 1135d9ae7f2bSJakub Kicinski emit_br(nfp_prog, br_mask, off, 0); 1136d9ae7f2bSJakub Kicinski } 1137d9ae7f2bSJakub Kicinski 1138d9ae7f2bSJakub Kicinski static int 1139d9ae7f2bSJakub Kicinski wrp_test_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1140d9ae7f2bSJakub Kicinski enum alu_op alu_op, enum br_mask br_mask) 1141d9ae7f2bSJakub Kicinski { 1142d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1143d9ae7f2bSJakub Kicinski 1144d9ae7f2bSJakub Kicinski wrp_test_reg_one(nfp_prog, insn->dst_reg * 2, alu_op, 1145d9ae7f2bSJakub Kicinski insn->src_reg * 2, br_mask, insn->off); 1146d9ae7f2bSJakub Kicinski wrp_test_reg_one(nfp_prog, insn->dst_reg * 2 + 1, alu_op, 1147d9ae7f2bSJakub Kicinski insn->src_reg * 2 + 1, br_mask, insn->off); 1148d9ae7f2bSJakub Kicinski 1149d9ae7f2bSJakub Kicinski return 0; 1150d9ae7f2bSJakub Kicinski } 1151d9ae7f2bSJakub Kicinski 1152d9ae7f2bSJakub Kicinski static int 1153d9ae7f2bSJakub Kicinski wrp_cmp_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1154d9ae7f2bSJakub Kicinski enum br_mask br_mask, bool swap) 1155d9ae7f2bSJakub Kicinski { 1156d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1157d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1158d9ae7f2bSJakub Kicinski u8 reg = insn->dst_reg * 2; 1159b3f868dfSJakub Kicinski swreg tmp_reg; 1160d9ae7f2bSJakub Kicinski 1161d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1162d9ae7f2bSJakub Kicinski if (!swap) 1163d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(reg), ALU_OP_SUB, tmp_reg); 1164d9ae7f2bSJakub Kicinski else 1165d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), tmp_reg, ALU_OP_SUB, reg_a(reg)); 1166d9ae7f2bSJakub Kicinski 1167d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1168d9ae7f2bSJakub Kicinski if (!swap) 1169d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1170d9ae7f2bSJakub Kicinski reg_a(reg + 1), ALU_OP_SUB_C, tmp_reg); 1171d9ae7f2bSJakub Kicinski else 1172d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1173d9ae7f2bSJakub Kicinski tmp_reg, ALU_OP_SUB_C, reg_a(reg + 1)); 1174d9ae7f2bSJakub Kicinski 1175d9ae7f2bSJakub Kicinski emit_br(nfp_prog, br_mask, insn->off, 0); 1176d9ae7f2bSJakub Kicinski 1177d9ae7f2bSJakub Kicinski return 0; 1178d9ae7f2bSJakub Kicinski } 1179d9ae7f2bSJakub Kicinski 1180d9ae7f2bSJakub Kicinski static int 1181d9ae7f2bSJakub Kicinski wrp_cmp_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1182d9ae7f2bSJakub Kicinski enum br_mask br_mask, bool swap) 1183d9ae7f2bSJakub Kicinski { 1184d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 118526fa818dSJakub Kicinski u8 areg, breg; 118626fa818dSJakub Kicinski 118726fa818dSJakub Kicinski areg = insn->dst_reg * 2; 118826fa818dSJakub Kicinski breg = insn->src_reg * 2; 1189d9ae7f2bSJakub Kicinski 1190d9ae7f2bSJakub Kicinski if (swap) { 1191d9ae7f2bSJakub Kicinski areg ^= breg; 1192d9ae7f2bSJakub Kicinski breg ^= areg; 1193d9ae7f2bSJakub Kicinski areg ^= breg; 1194d9ae7f2bSJakub Kicinski } 1195d9ae7f2bSJakub Kicinski 1196d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(areg), ALU_OP_SUB, reg_b(breg)); 1197d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1198d9ae7f2bSJakub Kicinski reg_a(areg + 1), ALU_OP_SUB_C, reg_b(breg + 1)); 1199d9ae7f2bSJakub Kicinski emit_br(nfp_prog, br_mask, insn->off, 0); 1200d9ae7f2bSJakub Kicinski 1201d9ae7f2bSJakub Kicinski return 0; 1202d9ae7f2bSJakub Kicinski } 1203d9ae7f2bSJakub Kicinski 12043119d1fdSJakub Kicinski static void wrp_end32(struct nfp_prog *nfp_prog, swreg reg_in, u8 gpr_out) 12053119d1fdSJakub Kicinski { 12063119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr_out), 0xf, reg_in, 12073119d1fdSJakub Kicinski SHF_SC_R_ROT, 8); 12083119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr_out), 0x5, reg_a(gpr_out), 12093119d1fdSJakub Kicinski SHF_SC_R_ROT, 16); 12103119d1fdSJakub Kicinski } 12113119d1fdSJakub Kicinski 12120d49eaf4SJakub Kicinski static int adjust_head(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 12130d49eaf4SJakub Kicinski { 12140d49eaf4SJakub Kicinski swreg tmp = imm_a(nfp_prog), tmp_len = imm_b(nfp_prog); 12150d49eaf4SJakub Kicinski struct nfp_bpf_cap_adjust_head *adjust_head; 12160d49eaf4SJakub Kicinski u32 ret_einval, end; 12170d49eaf4SJakub Kicinski 12180d49eaf4SJakub Kicinski adjust_head = &nfp_prog->bpf->adjust_head; 12190d49eaf4SJakub Kicinski 12208231f844SJakub Kicinski /* Optimized version - 5 vs 14 cycles */ 12218231f844SJakub Kicinski if (nfp_prog->adjust_head_location != UINT_MAX) { 12228231f844SJakub Kicinski if (WARN_ON_ONCE(nfp_prog->adjust_head_location != meta->n)) 12238231f844SJakub Kicinski return -EINVAL; 12248231f844SJakub Kicinski 12258231f844SJakub Kicinski emit_alu(nfp_prog, pptr_reg(nfp_prog), 12268231f844SJakub Kicinski reg_a(2 * 2), ALU_OP_ADD, pptr_reg(nfp_prog)); 12278231f844SJakub Kicinski emit_alu(nfp_prog, plen_reg(nfp_prog), 12288231f844SJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12298231f844SJakub Kicinski emit_alu(nfp_prog, pv_len(nfp_prog), 12308231f844SJakub Kicinski pv_len(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12318231f844SJakub Kicinski 12328231f844SJakub Kicinski wrp_immed(nfp_prog, reg_both(0), 0); 12338231f844SJakub Kicinski wrp_immed(nfp_prog, reg_both(1), 0); 12348231f844SJakub Kicinski 12358231f844SJakub Kicinski /* TODO: when adjust head is guaranteed to succeed we can 12368231f844SJakub Kicinski * also eliminate the following if (r0 == 0) branch. 12378231f844SJakub Kicinski */ 12388231f844SJakub Kicinski 12398231f844SJakub Kicinski return 0; 12408231f844SJakub Kicinski } 12418231f844SJakub Kicinski 12420d49eaf4SJakub Kicinski ret_einval = nfp_prog_current_offset(nfp_prog) + 14; 12430d49eaf4SJakub Kicinski end = ret_einval + 2; 12440d49eaf4SJakub Kicinski 12450d49eaf4SJakub Kicinski /* We need to use a temp because offset is just a part of the pkt ptr */ 12460d49eaf4SJakub Kicinski emit_alu(nfp_prog, tmp, 12470d49eaf4SJakub Kicinski reg_a(2 * 2), ALU_OP_ADD_2B, pptr_reg(nfp_prog)); 12480d49eaf4SJakub Kicinski 12490d49eaf4SJakub Kicinski /* Validate result will fit within FW datapath constraints */ 12500d49eaf4SJakub Kicinski emit_alu(nfp_prog, reg_none(), 12510d49eaf4SJakub Kicinski tmp, ALU_OP_SUB, reg_imm(adjust_head->off_min)); 12520d49eaf4SJakub Kicinski emit_br(nfp_prog, BR_BLO, ret_einval, 0); 12530d49eaf4SJakub Kicinski emit_alu(nfp_prog, reg_none(), 12540d49eaf4SJakub Kicinski reg_imm(adjust_head->off_max), ALU_OP_SUB, tmp); 12550d49eaf4SJakub Kicinski emit_br(nfp_prog, BR_BLO, ret_einval, 0); 12560d49eaf4SJakub Kicinski 12570d49eaf4SJakub Kicinski /* Validate the length is at least ETH_HLEN */ 12580d49eaf4SJakub Kicinski emit_alu(nfp_prog, tmp_len, 12590d49eaf4SJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12600d49eaf4SJakub Kicinski emit_alu(nfp_prog, reg_none(), 12610d49eaf4SJakub Kicinski tmp_len, ALU_OP_SUB, reg_imm(ETH_HLEN)); 12620d49eaf4SJakub Kicinski emit_br(nfp_prog, BR_BMI, ret_einval, 0); 12630d49eaf4SJakub Kicinski 12640d49eaf4SJakub Kicinski /* Load the ret code */ 12650d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(0), 0); 12660d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(1), 0); 12670d49eaf4SJakub Kicinski 12680d49eaf4SJakub Kicinski /* Modify the packet metadata */ 12690d49eaf4SJakub Kicinski emit_ld_field(nfp_prog, pptr_reg(nfp_prog), 0x3, tmp, SHF_SC_NONE, 0); 12700d49eaf4SJakub Kicinski 12710d49eaf4SJakub Kicinski /* Skip over the -EINVAL ret code (defer 2) */ 12720d49eaf4SJakub Kicinski emit_br_def(nfp_prog, end, 2); 12730d49eaf4SJakub Kicinski 12740d49eaf4SJakub Kicinski emit_alu(nfp_prog, plen_reg(nfp_prog), 12750d49eaf4SJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12760d49eaf4SJakub Kicinski emit_alu(nfp_prog, pv_len(nfp_prog), 12770d49eaf4SJakub Kicinski pv_len(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12780d49eaf4SJakub Kicinski 12790d49eaf4SJakub Kicinski /* return -EINVAL target */ 12800d49eaf4SJakub Kicinski if (!nfp_prog_confirm_current_offset(nfp_prog, ret_einval)) 12810d49eaf4SJakub Kicinski return -EINVAL; 12820d49eaf4SJakub Kicinski 12830d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(0), -22); 12840d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(1), ~0); 12850d49eaf4SJakub Kicinski 12860d49eaf4SJakub Kicinski if (!nfp_prog_confirm_current_offset(nfp_prog, end)) 12870d49eaf4SJakub Kicinski return -EINVAL; 12880d49eaf4SJakub Kicinski 12890d49eaf4SJakub Kicinski return 0; 12900d49eaf4SJakub Kicinski } 12910d49eaf4SJakub Kicinski 1292d9ae7f2bSJakub Kicinski /* --- Callbacks --- */ 1293d9ae7f2bSJakub Kicinski static int mov_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1294d9ae7f2bSJakub Kicinski { 1295d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1296b14157eeSJakub Kicinski u8 dst = insn->dst_reg * 2; 1297b14157eeSJakub Kicinski u8 src = insn->src_reg * 2; 1298d9ae7f2bSJakub Kicinski 1299b14157eeSJakub Kicinski if (insn->src_reg == BPF_REG_10) { 1300b14157eeSJakub Kicinski swreg stack_depth_reg; 1301b14157eeSJakub Kicinski 1302b14157eeSJakub Kicinski stack_depth_reg = ur_load_imm_any(nfp_prog, 1303b14157eeSJakub Kicinski nfp_prog->stack_depth, 1304b14157eeSJakub Kicinski stack_imm(nfp_prog)); 1305b14157eeSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), 1306b14157eeSJakub Kicinski stack_reg(nfp_prog), ALU_OP_ADD, stack_depth_reg); 1307b14157eeSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), 0); 1308b14157eeSJakub Kicinski } else { 1309b14157eeSJakub Kicinski wrp_reg_mov(nfp_prog, dst, src); 1310b14157eeSJakub Kicinski wrp_reg_mov(nfp_prog, dst + 1, src + 1); 1311b14157eeSJakub Kicinski } 1312d9ae7f2bSJakub Kicinski 1313d9ae7f2bSJakub Kicinski return 0; 1314d9ae7f2bSJakub Kicinski } 1315d9ae7f2bSJakub Kicinski 1316d9ae7f2bSJakub Kicinski static int mov_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1317d9ae7f2bSJakub Kicinski { 1318d9ae7f2bSJakub Kicinski u64 imm = meta->insn.imm; /* sign extend */ 1319d9ae7f2bSJakub Kicinski 1320d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2), imm & ~0U); 1321d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), imm >> 32); 1322d9ae7f2bSJakub Kicinski 1323d9ae7f2bSJakub Kicinski return 0; 1324d9ae7f2bSJakub Kicinski } 1325d9ae7f2bSJakub Kicinski 1326d9ae7f2bSJakub Kicinski static int xor_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1327d9ae7f2bSJakub Kicinski { 1328d9ae7f2bSJakub Kicinski return wrp_alu64_reg(nfp_prog, meta, ALU_OP_XOR); 1329d9ae7f2bSJakub Kicinski } 1330d9ae7f2bSJakub Kicinski 1331d9ae7f2bSJakub Kicinski static int xor_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1332d9ae7f2bSJakub Kicinski { 1333d9ae7f2bSJakub Kicinski return wrp_alu64_imm(nfp_prog, meta, ALU_OP_XOR, !meta->insn.imm); 1334d9ae7f2bSJakub Kicinski } 1335d9ae7f2bSJakub Kicinski 1336d9ae7f2bSJakub Kicinski static int and_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1337d9ae7f2bSJakub Kicinski { 1338d9ae7f2bSJakub Kicinski return wrp_alu64_reg(nfp_prog, meta, ALU_OP_AND); 1339d9ae7f2bSJakub Kicinski } 1340d9ae7f2bSJakub Kicinski 1341d9ae7f2bSJakub Kicinski static int and_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1342d9ae7f2bSJakub Kicinski { 1343d9ae7f2bSJakub Kicinski return wrp_alu64_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm); 1344d9ae7f2bSJakub Kicinski } 1345d9ae7f2bSJakub Kicinski 1346d9ae7f2bSJakub Kicinski static int or_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1347d9ae7f2bSJakub Kicinski { 1348d9ae7f2bSJakub Kicinski return wrp_alu64_reg(nfp_prog, meta, ALU_OP_OR); 1349d9ae7f2bSJakub Kicinski } 1350d9ae7f2bSJakub Kicinski 1351d9ae7f2bSJakub Kicinski static int or_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1352d9ae7f2bSJakub Kicinski { 1353d9ae7f2bSJakub Kicinski return wrp_alu64_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm); 1354d9ae7f2bSJakub Kicinski } 1355d9ae7f2bSJakub Kicinski 1356d9ae7f2bSJakub Kicinski static int add_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1357d9ae7f2bSJakub Kicinski { 1358d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1359d9ae7f2bSJakub Kicinski 1360d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), 1361d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_ADD, 1362d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2)); 1363d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 1364d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_ADD_C, 1365d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2 + 1)); 1366d9ae7f2bSJakub Kicinski 1367d9ae7f2bSJakub Kicinski return 0; 1368d9ae7f2bSJakub Kicinski } 1369d9ae7f2bSJakub Kicinski 1370d9ae7f2bSJakub Kicinski static int add_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1371d9ae7f2bSJakub Kicinski { 1372d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1373d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1374d9ae7f2bSJakub Kicinski 1375d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, ALU_OP_ADD, imm & ~0U); 1376d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, ALU_OP_ADD_C, imm >> 32); 1377d9ae7f2bSJakub Kicinski 1378d9ae7f2bSJakub Kicinski return 0; 1379d9ae7f2bSJakub Kicinski } 1380d9ae7f2bSJakub Kicinski 1381d9ae7f2bSJakub Kicinski static int sub_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1382d9ae7f2bSJakub Kicinski { 1383d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1384d9ae7f2bSJakub Kicinski 1385d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), 1386d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_SUB, 1387d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2)); 1388d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 1389d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_SUB_C, 1390d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2 + 1)); 1391d9ae7f2bSJakub Kicinski 1392d9ae7f2bSJakub Kicinski return 0; 1393d9ae7f2bSJakub Kicinski } 1394d9ae7f2bSJakub Kicinski 1395d9ae7f2bSJakub Kicinski static int sub_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1396d9ae7f2bSJakub Kicinski { 1397d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1398d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1399d9ae7f2bSJakub Kicinski 1400d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, ALU_OP_SUB, imm & ~0U); 1401d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, ALU_OP_SUB_C, imm >> 32); 1402d9ae7f2bSJakub Kicinski 1403d9ae7f2bSJakub Kicinski return 0; 1404d9ae7f2bSJakub Kicinski } 1405d9ae7f2bSJakub Kicinski 1406254ef4d7SJiong Wang static int neg_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1407254ef4d7SJiong Wang { 1408254ef4d7SJiong Wang const struct bpf_insn *insn = &meta->insn; 1409254ef4d7SJiong Wang 1410254ef4d7SJiong Wang emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), reg_imm(0), 1411254ef4d7SJiong Wang ALU_OP_SUB, reg_b(insn->dst_reg * 2)); 1412254ef4d7SJiong Wang emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), reg_imm(0), 1413254ef4d7SJiong Wang ALU_OP_SUB_C, reg_b(insn->dst_reg * 2 + 1)); 1414254ef4d7SJiong Wang 1415254ef4d7SJiong Wang return 0; 1416254ef4d7SJiong Wang } 1417254ef4d7SJiong Wang 1418d9ae7f2bSJakub Kicinski static int shl_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1419d9ae7f2bSJakub Kicinski { 1420d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 14213cae1319SJakub Kicinski u8 dst = insn->dst_reg * 2; 1422d9ae7f2bSJakub Kicinski 14233cae1319SJakub Kicinski if (insn->imm < 32) { 14243cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst + 1), 14253cae1319SJakub Kicinski reg_a(dst + 1), SHF_OP_NONE, reg_b(dst), 14263cae1319SJakub Kicinski SHF_SC_R_DSHF, 32 - insn->imm); 14273cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst), 14283cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst), 14293cae1319SJakub Kicinski SHF_SC_L_SHF, insn->imm); 14303cae1319SJakub Kicinski } else if (insn->imm == 32) { 14313cae1319SJakub Kicinski wrp_reg_mov(nfp_prog, dst + 1, dst); 14323cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), 0); 14333cae1319SJakub Kicinski } else if (insn->imm > 32) { 14343cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst + 1), 14353cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst), 14363cae1319SJakub Kicinski SHF_SC_L_SHF, insn->imm - 32); 14373cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), 0); 14383cae1319SJakub Kicinski } 1439d9ae7f2bSJakub Kicinski 1440d9ae7f2bSJakub Kicinski return 0; 1441d9ae7f2bSJakub Kicinski } 1442d9ae7f2bSJakub Kicinski 1443d9ae7f2bSJakub Kicinski static int shr_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1444d9ae7f2bSJakub Kicinski { 1445d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 14463cae1319SJakub Kicinski u8 dst = insn->dst_reg * 2; 1447d9ae7f2bSJakub Kicinski 14483cae1319SJakub Kicinski if (insn->imm < 32) { 14493cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst), 14503cae1319SJakub Kicinski reg_a(dst + 1), SHF_OP_NONE, reg_b(dst), 14513cae1319SJakub Kicinski SHF_SC_R_DSHF, insn->imm); 14523cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst + 1), 14533cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst + 1), 14543cae1319SJakub Kicinski SHF_SC_R_SHF, insn->imm); 14553cae1319SJakub Kicinski } else if (insn->imm == 32) { 14563cae1319SJakub Kicinski wrp_reg_mov(nfp_prog, dst, dst + 1); 14573cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), 0); 14583cae1319SJakub Kicinski } else if (insn->imm > 32) { 14593cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst), 14603cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst + 1), 14613cae1319SJakub Kicinski SHF_SC_R_SHF, insn->imm - 32); 14623cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), 0); 14633cae1319SJakub Kicinski } 1464d9ae7f2bSJakub Kicinski 1465d9ae7f2bSJakub Kicinski return 0; 1466d9ae7f2bSJakub Kicinski } 1467d9ae7f2bSJakub Kicinski 1468d9ae7f2bSJakub Kicinski static int mov_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1469d9ae7f2bSJakub Kicinski { 1470d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1471d9ae7f2bSJakub Kicinski 1472d9ae7f2bSJakub Kicinski wrp_reg_mov(nfp_prog, insn->dst_reg * 2, insn->src_reg * 2); 1473d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1474d9ae7f2bSJakub Kicinski 1475d9ae7f2bSJakub Kicinski return 0; 1476d9ae7f2bSJakub Kicinski } 1477d9ae7f2bSJakub Kicinski 1478d9ae7f2bSJakub Kicinski static int mov_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1479d9ae7f2bSJakub Kicinski { 1480d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1481d9ae7f2bSJakub Kicinski 1482d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2), insn->imm); 1483d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1484d9ae7f2bSJakub Kicinski 1485d9ae7f2bSJakub Kicinski return 0; 1486d9ae7f2bSJakub Kicinski } 1487d9ae7f2bSJakub Kicinski 1488d9ae7f2bSJakub Kicinski static int xor_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1489d9ae7f2bSJakub Kicinski { 1490d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_XOR); 1491d9ae7f2bSJakub Kicinski } 1492d9ae7f2bSJakub Kicinski 1493d9ae7f2bSJakub Kicinski static int xor_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1494d9ae7f2bSJakub Kicinski { 1495d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_XOR, !~meta->insn.imm); 1496d9ae7f2bSJakub Kicinski } 1497d9ae7f2bSJakub Kicinski 1498d9ae7f2bSJakub Kicinski static int and_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1499d9ae7f2bSJakub Kicinski { 1500d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_AND); 1501d9ae7f2bSJakub Kicinski } 1502d9ae7f2bSJakub Kicinski 1503d9ae7f2bSJakub Kicinski static int and_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1504d9ae7f2bSJakub Kicinski { 1505d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm); 1506d9ae7f2bSJakub Kicinski } 1507d9ae7f2bSJakub Kicinski 1508d9ae7f2bSJakub Kicinski static int or_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1509d9ae7f2bSJakub Kicinski { 1510d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_OR); 1511d9ae7f2bSJakub Kicinski } 1512d9ae7f2bSJakub Kicinski 1513d9ae7f2bSJakub Kicinski static int or_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1514d9ae7f2bSJakub Kicinski { 1515d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm); 1516d9ae7f2bSJakub Kicinski } 1517d9ae7f2bSJakub Kicinski 1518d9ae7f2bSJakub Kicinski static int add_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1519d9ae7f2bSJakub Kicinski { 1520d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_ADD); 1521d9ae7f2bSJakub Kicinski } 1522d9ae7f2bSJakub Kicinski 1523d9ae7f2bSJakub Kicinski static int add_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1524d9ae7f2bSJakub Kicinski { 1525d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_ADD, !meta->insn.imm); 1526d9ae7f2bSJakub Kicinski } 1527d9ae7f2bSJakub Kicinski 1528d9ae7f2bSJakub Kicinski static int sub_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1529d9ae7f2bSJakub Kicinski { 1530d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_SUB); 1531d9ae7f2bSJakub Kicinski } 1532d9ae7f2bSJakub Kicinski 1533d9ae7f2bSJakub Kicinski static int sub_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1534d9ae7f2bSJakub Kicinski { 1535d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_SUB, !meta->insn.imm); 1536d9ae7f2bSJakub Kicinski } 1537d9ae7f2bSJakub Kicinski 1538254ef4d7SJiong Wang static int neg_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1539254ef4d7SJiong Wang { 1540254ef4d7SJiong Wang u8 dst = meta->insn.dst_reg * 2; 1541254ef4d7SJiong Wang 1542254ef4d7SJiong Wang emit_alu(nfp_prog, reg_both(dst), reg_imm(0), ALU_OP_SUB, reg_b(dst)); 1543254ef4d7SJiong Wang wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1544254ef4d7SJiong Wang 1545254ef4d7SJiong Wang return 0; 1546254ef4d7SJiong Wang } 1547254ef4d7SJiong Wang 1548d9ae7f2bSJakub Kicinski static int shl_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1549d9ae7f2bSJakub Kicinski { 1550d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1551d9ae7f2bSJakub Kicinski 1552d9ae7f2bSJakub Kicinski if (!insn->imm) 1553d9ae7f2bSJakub Kicinski return 1; /* TODO: zero shift means indirect */ 1554d9ae7f2bSJakub Kicinski 1555d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_both(insn->dst_reg * 2), 1556d9ae7f2bSJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(insn->dst_reg * 2), 1557d9ae7f2bSJakub Kicinski SHF_SC_L_SHF, insn->imm); 1558d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1559d9ae7f2bSJakub Kicinski 1560d9ae7f2bSJakub Kicinski return 0; 1561d9ae7f2bSJakub Kicinski } 1562d9ae7f2bSJakub Kicinski 15633119d1fdSJakub Kicinski static int end_reg32(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 15643119d1fdSJakub Kicinski { 15653119d1fdSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 15663119d1fdSJakub Kicinski u8 gpr = insn->dst_reg * 2; 15673119d1fdSJakub Kicinski 15683119d1fdSJakub Kicinski switch (insn->imm) { 15693119d1fdSJakub Kicinski case 16: 15703119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr), 0x9, reg_b(gpr), 15713119d1fdSJakub Kicinski SHF_SC_R_ROT, 8); 15723119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr), 0xe, reg_a(gpr), 15733119d1fdSJakub Kicinski SHF_SC_R_SHF, 16); 15743119d1fdSJakub Kicinski 15753119d1fdSJakub Kicinski wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 15763119d1fdSJakub Kicinski break; 15773119d1fdSJakub Kicinski case 32: 15783119d1fdSJakub Kicinski wrp_end32(nfp_prog, reg_a(gpr), gpr); 15793119d1fdSJakub Kicinski wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 15803119d1fdSJakub Kicinski break; 15813119d1fdSJakub Kicinski case 64: 15823119d1fdSJakub Kicinski wrp_mov(nfp_prog, imm_a(nfp_prog), reg_b(gpr + 1)); 15833119d1fdSJakub Kicinski 15843119d1fdSJakub Kicinski wrp_end32(nfp_prog, reg_a(gpr), gpr + 1); 15853119d1fdSJakub Kicinski wrp_end32(nfp_prog, imm_a(nfp_prog), gpr); 15863119d1fdSJakub Kicinski break; 15873119d1fdSJakub Kicinski } 15883119d1fdSJakub Kicinski 15893119d1fdSJakub Kicinski return 0; 15903119d1fdSJakub Kicinski } 15913119d1fdSJakub Kicinski 1592d9ae7f2bSJakub Kicinski static int imm_ld8_part2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1593d9ae7f2bSJakub Kicinski { 15949f16c8abSJakub Kicinski struct nfp_insn_meta *prev = nfp_meta_prev(meta); 15959f16c8abSJakub Kicinski u32 imm_lo, imm_hi; 15969f16c8abSJakub Kicinski u8 dst; 15979f16c8abSJakub Kicinski 15989f16c8abSJakub Kicinski dst = prev->insn.dst_reg * 2; 15999f16c8abSJakub Kicinski imm_lo = prev->insn.imm; 16009f16c8abSJakub Kicinski imm_hi = meta->insn.imm; 16019f16c8abSJakub Kicinski 16029f16c8abSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), imm_lo); 16039f16c8abSJakub Kicinski 16049f16c8abSJakub Kicinski /* mov is always 1 insn, load imm may be two, so try to use mov */ 16059f16c8abSJakub Kicinski if (imm_hi == imm_lo) 16069f16c8abSJakub Kicinski wrp_mov(nfp_prog, reg_both(dst + 1), reg_a(dst)); 16079f16c8abSJakub Kicinski else 16089f16c8abSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), imm_hi); 1609d9ae7f2bSJakub Kicinski 1610d9ae7f2bSJakub Kicinski return 0; 1611d9ae7f2bSJakub Kicinski } 1612d9ae7f2bSJakub Kicinski 1613d9ae7f2bSJakub Kicinski static int imm_ld8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1614d9ae7f2bSJakub Kicinski { 1615d9ae7f2bSJakub Kicinski meta->double_cb = imm_ld8_part2; 1616d9ae7f2bSJakub Kicinski return 0; 1617d9ae7f2bSJakub Kicinski } 1618d9ae7f2bSJakub Kicinski 1619d9ae7f2bSJakub Kicinski static int data_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1620d9ae7f2bSJakub Kicinski { 1621d9ae7f2bSJakub Kicinski return construct_data_ld(nfp_prog, meta->insn.imm, 1); 1622d9ae7f2bSJakub Kicinski } 1623d9ae7f2bSJakub Kicinski 1624d9ae7f2bSJakub Kicinski static int data_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1625d9ae7f2bSJakub Kicinski { 1626d9ae7f2bSJakub Kicinski return construct_data_ld(nfp_prog, meta->insn.imm, 2); 1627d9ae7f2bSJakub Kicinski } 1628d9ae7f2bSJakub Kicinski 1629d9ae7f2bSJakub Kicinski static int data_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1630d9ae7f2bSJakub Kicinski { 1631d9ae7f2bSJakub Kicinski return construct_data_ld(nfp_prog, meta->insn.imm, 4); 1632d9ae7f2bSJakub Kicinski } 1633d9ae7f2bSJakub Kicinski 1634d9ae7f2bSJakub Kicinski static int data_ind_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1635d9ae7f2bSJakub Kicinski { 1636d9ae7f2bSJakub Kicinski return construct_data_ind_ld(nfp_prog, meta->insn.imm, 16370a793977SJakub Kicinski meta->insn.src_reg * 2, 1); 1638d9ae7f2bSJakub Kicinski } 1639d9ae7f2bSJakub Kicinski 1640d9ae7f2bSJakub Kicinski static int data_ind_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1641d9ae7f2bSJakub Kicinski { 1642d9ae7f2bSJakub Kicinski return construct_data_ind_ld(nfp_prog, meta->insn.imm, 16430a793977SJakub Kicinski meta->insn.src_reg * 2, 2); 1644d9ae7f2bSJakub Kicinski } 1645d9ae7f2bSJakub Kicinski 1646d9ae7f2bSJakub Kicinski static int data_ind_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1647d9ae7f2bSJakub Kicinski { 1648d9ae7f2bSJakub Kicinski return construct_data_ind_ld(nfp_prog, meta->insn.imm, 16490a793977SJakub Kicinski meta->insn.src_reg * 2, 4); 1650d9ae7f2bSJakub Kicinski } 1651d9ae7f2bSJakub Kicinski 1652a82b23fbSJakub Kicinski static int 1653a82b23fbSJakub Kicinski mem_ldx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1654d3488480SJakub Kicinski unsigned int size, unsigned int ptr_off) 1655a82b23fbSJakub Kicinski { 1656d3488480SJakub Kicinski return mem_op_stack(nfp_prog, meta, size, ptr_off, 1657b14157eeSJakub Kicinski meta->insn.dst_reg * 2, meta->insn.src_reg * 2, 1658b14157eeSJakub Kicinski true, wrp_lmem_load); 1659a82b23fbSJakub Kicinski } 1660a82b23fbSJakub Kicinski 1661943c57b9SJakub Kicinski static int mem_ldx_skb(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1662943c57b9SJakub Kicinski u8 size) 1663d9ae7f2bSJakub Kicinski { 1664bfddbc8aSJakub Kicinski swreg dst = reg_both(meta->insn.dst_reg * 2); 1665bfddbc8aSJakub Kicinski 1666943c57b9SJakub Kicinski switch (meta->insn.off) { 1667c6c580d7SJakub Kicinski case offsetof(struct __sk_buff, len): 1668c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct __sk_buff, len)) 1669943c57b9SJakub Kicinski return -EOPNOTSUPP; 1670bfddbc8aSJakub Kicinski wrp_mov(nfp_prog, dst, plen_reg(nfp_prog)); 1671bfddbc8aSJakub Kicinski break; 1672c6c580d7SJakub Kicinski case offsetof(struct __sk_buff, data): 1673c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct __sk_buff, data)) 1674bfddbc8aSJakub Kicinski return -EOPNOTSUPP; 1675bfddbc8aSJakub Kicinski wrp_mov(nfp_prog, dst, pptr_reg(nfp_prog)); 1676bfddbc8aSJakub Kicinski break; 1677c6c580d7SJakub Kicinski case offsetof(struct __sk_buff, data_end): 1678c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct __sk_buff, data_end)) 1679bfddbc8aSJakub Kicinski return -EOPNOTSUPP; 1680bfddbc8aSJakub Kicinski emit_alu(nfp_prog, dst, 1681bfddbc8aSJakub Kicinski plen_reg(nfp_prog), ALU_OP_ADD, pptr_reg(nfp_prog)); 1682943c57b9SJakub Kicinski break; 1683943c57b9SJakub Kicinski default: 1684d9ae7f2bSJakub Kicinski return -EOPNOTSUPP; 1685943c57b9SJakub Kicinski } 1686943c57b9SJakub Kicinski 1687943c57b9SJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1688d9ae7f2bSJakub Kicinski 1689d9ae7f2bSJakub Kicinski return 0; 1690d9ae7f2bSJakub Kicinski } 1691d9ae7f2bSJakub Kicinski 1692943c57b9SJakub Kicinski static int mem_ldx_xdp(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1693943c57b9SJakub Kicinski u8 size) 1694d9ae7f2bSJakub Kicinski { 1695b3f868dfSJakub Kicinski swreg dst = reg_both(meta->insn.dst_reg * 2); 1696d9ae7f2bSJakub Kicinski 1697943c57b9SJakub Kicinski switch (meta->insn.off) { 1698c6c580d7SJakub Kicinski case offsetof(struct xdp_md, data): 1699c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct xdp_md, data)) 1700c6c580d7SJakub Kicinski return -EOPNOTSUPP; 1701c000dfb5SJakub Kicinski wrp_mov(nfp_prog, dst, pptr_reg(nfp_prog)); 1702943c57b9SJakub Kicinski break; 1703c6c580d7SJakub Kicinski case offsetof(struct xdp_md, data_end): 1704c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct xdp_md, data_end)) 1705c6c580d7SJakub Kicinski return -EOPNOTSUPP; 1706943c57b9SJakub Kicinski emit_alu(nfp_prog, dst, 1707943c57b9SJakub Kicinski plen_reg(nfp_prog), ALU_OP_ADD, pptr_reg(nfp_prog)); 1708943c57b9SJakub Kicinski break; 1709943c57b9SJakub Kicinski default: 1710943c57b9SJakub Kicinski return -EOPNOTSUPP; 1711943c57b9SJakub Kicinski } 1712d9ae7f2bSJakub Kicinski 1713943c57b9SJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1714d9ae7f2bSJakub Kicinski 1715d9ae7f2bSJakub Kicinski return 0; 1716d9ae7f2bSJakub Kicinski } 1717d9ae7f2bSJakub Kicinski 17182ca71441SJakub Kicinski static int 17192ca71441SJakub Kicinski mem_ldx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 17202ca71441SJakub Kicinski unsigned int size) 17212ca71441SJakub Kicinski { 17222ca71441SJakub Kicinski swreg tmp_reg; 17232ca71441SJakub Kicinski 17242ca71441SJakub Kicinski tmp_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 17252ca71441SJakub Kicinski 17262ca71441SJakub Kicinski return data_ld_host_order(nfp_prog, meta->insn.src_reg * 2, tmp_reg, 17272ca71441SJakub Kicinski meta->insn.dst_reg * 2, size); 17282ca71441SJakub Kicinski } 17292ca71441SJakub Kicinski 17302ca71441SJakub Kicinski static int 17312ca71441SJakub Kicinski mem_ldx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 17322ca71441SJakub Kicinski unsigned int size) 17332ca71441SJakub Kicinski { 17349879a381SJiong Wang if (meta->ldst_gather_len) 17359879a381SJiong Wang return nfp_cpp_memcpy(nfp_prog, meta); 17369879a381SJiong Wang 17372ca71441SJakub Kicinski if (meta->ptr.type == PTR_TO_CTX) { 1738012bb8a8SJakub Kicinski if (nfp_prog->type == BPF_PROG_TYPE_XDP) 17392ca71441SJakub Kicinski return mem_ldx_xdp(nfp_prog, meta, size); 17402ca71441SJakub Kicinski else 17412ca71441SJakub Kicinski return mem_ldx_skb(nfp_prog, meta, size); 17422ca71441SJakub Kicinski } 17432ca71441SJakub Kicinski 17442ca71441SJakub Kicinski if (meta->ptr.type == PTR_TO_PACKET) 17452ca71441SJakub Kicinski return mem_ldx_data(nfp_prog, meta, size); 17462ca71441SJakub Kicinski 1747a82b23fbSJakub Kicinski if (meta->ptr.type == PTR_TO_STACK) 1748d3488480SJakub Kicinski return mem_ldx_stack(nfp_prog, meta, size, 1749d3488480SJakub Kicinski meta->ptr.off + meta->ptr.var_off.value); 1750a82b23fbSJakub Kicinski 17512ca71441SJakub Kicinski return -EOPNOTSUPP; 17522ca71441SJakub Kicinski } 17532ca71441SJakub Kicinski 17542ca71441SJakub Kicinski static int mem_ldx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 17552ca71441SJakub Kicinski { 17562ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 1); 17572ca71441SJakub Kicinski } 17582ca71441SJakub Kicinski 17592ca71441SJakub Kicinski static int mem_ldx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 17602ca71441SJakub Kicinski { 17612ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 2); 17622ca71441SJakub Kicinski } 17632ca71441SJakub Kicinski 1764d9ae7f2bSJakub Kicinski static int mem_ldx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1765d9ae7f2bSJakub Kicinski { 17662ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 4); 17672ca71441SJakub Kicinski } 17682ca71441SJakub Kicinski 17692ca71441SJakub Kicinski static int mem_ldx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 17702ca71441SJakub Kicinski { 17712ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 8); 1772d9ae7f2bSJakub Kicinski } 1773d9ae7f2bSJakub Kicinski 1774e663fe38SJakub Kicinski static int 1775e663fe38SJakub Kicinski mem_st_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1776e663fe38SJakub Kicinski unsigned int size) 1777d9ae7f2bSJakub Kicinski { 1778e663fe38SJakub Kicinski u64 imm = meta->insn.imm; /* sign extend */ 1779e663fe38SJakub Kicinski swreg off_reg; 1780e663fe38SJakub Kicinski 1781e663fe38SJakub Kicinski off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1782e663fe38SJakub Kicinski 1783e663fe38SJakub Kicinski return data_st_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg, 1784e663fe38SJakub Kicinski imm, size); 1785e663fe38SJakub Kicinski } 1786e663fe38SJakub Kicinski 1787e663fe38SJakub Kicinski static int mem_st(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1788e663fe38SJakub Kicinski unsigned int size) 1789e663fe38SJakub Kicinski { 1790e663fe38SJakub Kicinski if (meta->ptr.type == PTR_TO_PACKET) 1791e663fe38SJakub Kicinski return mem_st_data(nfp_prog, meta, size); 1792e663fe38SJakub Kicinski 1793d9ae7f2bSJakub Kicinski return -EOPNOTSUPP; 1794d9ae7f2bSJakub Kicinski } 1795d9ae7f2bSJakub Kicinski 1796e663fe38SJakub Kicinski static int mem_st1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1797d9ae7f2bSJakub Kicinski { 1798e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 1); 1799e663fe38SJakub Kicinski } 1800e663fe38SJakub Kicinski 1801e663fe38SJakub Kicinski static int mem_st2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1802e663fe38SJakub Kicinski { 1803e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 2); 1804e663fe38SJakub Kicinski } 1805e663fe38SJakub Kicinski 1806e663fe38SJakub Kicinski static int mem_st4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1807e663fe38SJakub Kicinski { 1808e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 4); 1809e663fe38SJakub Kicinski } 1810e663fe38SJakub Kicinski 1811e663fe38SJakub Kicinski static int mem_st8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1812e663fe38SJakub Kicinski { 1813e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 8); 1814e663fe38SJakub Kicinski } 1815e663fe38SJakub Kicinski 1816e663fe38SJakub Kicinski static int 1817e663fe38SJakub Kicinski mem_stx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1818e663fe38SJakub Kicinski unsigned int size) 1819e663fe38SJakub Kicinski { 1820e663fe38SJakub Kicinski swreg off_reg; 1821e663fe38SJakub Kicinski 1822e663fe38SJakub Kicinski off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1823e663fe38SJakub Kicinski 1824e663fe38SJakub Kicinski return data_stx_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg, 1825e663fe38SJakub Kicinski meta->insn.src_reg * 2, size); 1826e663fe38SJakub Kicinski } 1827e663fe38SJakub Kicinski 1828e663fe38SJakub Kicinski static int 1829ee9133a8SJakub Kicinski mem_stx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1830d3488480SJakub Kicinski unsigned int size, unsigned int ptr_off) 1831ee9133a8SJakub Kicinski { 1832d3488480SJakub Kicinski return mem_op_stack(nfp_prog, meta, size, ptr_off, 1833b14157eeSJakub Kicinski meta->insn.src_reg * 2, meta->insn.dst_reg * 2, 1834b14157eeSJakub Kicinski false, wrp_lmem_store); 1835ee9133a8SJakub Kicinski } 1836ee9133a8SJakub Kicinski 1837ee9133a8SJakub Kicinski static int 1838e663fe38SJakub Kicinski mem_stx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1839e663fe38SJakub Kicinski unsigned int size) 1840e663fe38SJakub Kicinski { 1841e663fe38SJakub Kicinski if (meta->ptr.type == PTR_TO_PACKET) 1842e663fe38SJakub Kicinski return mem_stx_data(nfp_prog, meta, size); 1843e663fe38SJakub Kicinski 1844ee9133a8SJakub Kicinski if (meta->ptr.type == PTR_TO_STACK) 1845d3488480SJakub Kicinski return mem_stx_stack(nfp_prog, meta, size, 1846d3488480SJakub Kicinski meta->ptr.off + meta->ptr.var_off.value); 1847ee9133a8SJakub Kicinski 1848d9ae7f2bSJakub Kicinski return -EOPNOTSUPP; 1849d9ae7f2bSJakub Kicinski } 1850d9ae7f2bSJakub Kicinski 1851e663fe38SJakub Kicinski static int mem_stx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1852e663fe38SJakub Kicinski { 1853e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 1); 1854e663fe38SJakub Kicinski } 1855e663fe38SJakub Kicinski 1856e663fe38SJakub Kicinski static int mem_stx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1857e663fe38SJakub Kicinski { 1858e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 2); 1859e663fe38SJakub Kicinski } 1860e663fe38SJakub Kicinski 1861d9ae7f2bSJakub Kicinski static int mem_stx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1862d9ae7f2bSJakub Kicinski { 1863e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 4); 1864e663fe38SJakub Kicinski } 18652ca71441SJakub Kicinski 1866e663fe38SJakub Kicinski static int mem_stx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1867e663fe38SJakub Kicinski { 1868e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 8); 1869d9ae7f2bSJakub Kicinski } 1870d9ae7f2bSJakub Kicinski 1871d9ae7f2bSJakub Kicinski static int jump(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1872d9ae7f2bSJakub Kicinski { 1873d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_UNC, meta->insn.off, 0); 1874d9ae7f2bSJakub Kicinski 1875d9ae7f2bSJakub Kicinski return 0; 1876d9ae7f2bSJakub Kicinski } 1877d9ae7f2bSJakub Kicinski 1878d9ae7f2bSJakub Kicinski static int jeq_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1879d9ae7f2bSJakub Kicinski { 1880d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1881d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1882b3f868dfSJakub Kicinski swreg or1, or2, tmp_reg; 1883b3f868dfSJakub Kicinski 1884b3f868dfSJakub Kicinski or1 = reg_a(insn->dst_reg * 2); 1885b3f868dfSJakub Kicinski or2 = reg_b(insn->dst_reg * 2 + 1); 1886d9ae7f2bSJakub Kicinski 1887d9ae7f2bSJakub Kicinski if (imm & ~0U) { 1888d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1889d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_a(nfp_prog), 1890d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_XOR, tmp_reg); 1891d9ae7f2bSJakub Kicinski or1 = imm_a(nfp_prog); 1892d9ae7f2bSJakub Kicinski } 1893d9ae7f2bSJakub Kicinski 1894d9ae7f2bSJakub Kicinski if (imm >> 32) { 1895d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1896d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 1897d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_XOR, tmp_reg); 1898d9ae7f2bSJakub Kicinski or2 = imm_b(nfp_prog); 1899d9ae7f2bSJakub Kicinski } 1900d9ae7f2bSJakub Kicinski 1901d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), or1, ALU_OP_OR, or2); 1902d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BEQ, insn->off, 0); 1903d9ae7f2bSJakub Kicinski 1904d9ae7f2bSJakub Kicinski return 0; 1905d9ae7f2bSJakub Kicinski } 1906d9ae7f2bSJakub Kicinski 1907d9ae7f2bSJakub Kicinski static int jgt_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1908d9ae7f2bSJakub Kicinski { 190926fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BLO, true); 1910d9ae7f2bSJakub Kicinski } 1911d9ae7f2bSJakub Kicinski 1912d9ae7f2bSJakub Kicinski static int jge_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1913d9ae7f2bSJakub Kicinski { 191426fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BHS, false); 1915d9ae7f2bSJakub Kicinski } 1916d9ae7f2bSJakub Kicinski 19175dd294d4SDaniel Borkmann static int jlt_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 19185dd294d4SDaniel Borkmann { 191926fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BLO, false); 19205dd294d4SDaniel Borkmann } 19215dd294d4SDaniel Borkmann 19225dd294d4SDaniel Borkmann static int jle_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 19235dd294d4SDaniel Borkmann { 192426fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BHS, true); 19255dd294d4SDaniel Borkmann } 19265dd294d4SDaniel Borkmann 1927d9ae7f2bSJakub Kicinski static int jset_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1928d9ae7f2bSJakub Kicinski { 1929d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1930d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1931b3f868dfSJakub Kicinski swreg tmp_reg; 1932d9ae7f2bSJakub Kicinski 1933d9ae7f2bSJakub Kicinski if (!imm) { 1934d9ae7f2bSJakub Kicinski meta->skip = true; 1935d9ae7f2bSJakub Kicinski return 0; 1936d9ae7f2bSJakub Kicinski } 1937d9ae7f2bSJakub Kicinski 1938d9ae7f2bSJakub Kicinski if (imm & ~0U) { 1939d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1940d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1941d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_AND, tmp_reg); 1942d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1943d9ae7f2bSJakub Kicinski } 1944d9ae7f2bSJakub Kicinski 1945d9ae7f2bSJakub Kicinski if (imm >> 32) { 1946d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1947d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1948d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_AND, tmp_reg); 1949d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1950d9ae7f2bSJakub Kicinski } 1951d9ae7f2bSJakub Kicinski 1952d9ae7f2bSJakub Kicinski return 0; 1953d9ae7f2bSJakub Kicinski } 1954d9ae7f2bSJakub Kicinski 1955d9ae7f2bSJakub Kicinski static int jne_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1956d9ae7f2bSJakub Kicinski { 1957d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1958d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1959b3f868dfSJakub Kicinski swreg tmp_reg; 1960d9ae7f2bSJakub Kicinski 1961d9ae7f2bSJakub Kicinski if (!imm) { 1962d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(insn->dst_reg * 2), 1963d9ae7f2bSJakub Kicinski ALU_OP_OR, reg_b(insn->dst_reg * 2 + 1)); 1964d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 196582837370SJakub Kicinski return 0; 1966d9ae7f2bSJakub Kicinski } 1967d9ae7f2bSJakub Kicinski 1968d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1969d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1970d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_XOR, tmp_reg); 1971d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1972d9ae7f2bSJakub Kicinski 1973d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1974d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1975d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_XOR, tmp_reg); 1976d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1977d9ae7f2bSJakub Kicinski 1978d9ae7f2bSJakub Kicinski return 0; 1979d9ae7f2bSJakub Kicinski } 1980d9ae7f2bSJakub Kicinski 1981d9ae7f2bSJakub Kicinski static int jeq_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1982d9ae7f2bSJakub Kicinski { 1983d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1984d9ae7f2bSJakub Kicinski 1985d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_a(nfp_prog), reg_a(insn->dst_reg * 2), 1986d9ae7f2bSJakub Kicinski ALU_OP_XOR, reg_b(insn->src_reg * 2)); 1987d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), reg_a(insn->dst_reg * 2 + 1), 1988d9ae7f2bSJakub Kicinski ALU_OP_XOR, reg_b(insn->src_reg * 2 + 1)); 1989d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1990d9ae7f2bSJakub Kicinski imm_a(nfp_prog), ALU_OP_OR, imm_b(nfp_prog)); 1991d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BEQ, insn->off, 0); 1992d9ae7f2bSJakub Kicinski 1993d9ae7f2bSJakub Kicinski return 0; 1994d9ae7f2bSJakub Kicinski } 1995d9ae7f2bSJakub Kicinski 1996d9ae7f2bSJakub Kicinski static int jgt_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1997d9ae7f2bSJakub Kicinski { 199826fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BLO, true); 1999d9ae7f2bSJakub Kicinski } 2000d9ae7f2bSJakub Kicinski 2001d9ae7f2bSJakub Kicinski static int jge_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 2002d9ae7f2bSJakub Kicinski { 200326fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BHS, false); 2004d9ae7f2bSJakub Kicinski } 2005d9ae7f2bSJakub Kicinski 20065dd294d4SDaniel Borkmann static int jlt_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 20075dd294d4SDaniel Borkmann { 200826fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BLO, false); 20095dd294d4SDaniel Borkmann } 20105dd294d4SDaniel Borkmann 20115dd294d4SDaniel Borkmann static int jle_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 20125dd294d4SDaniel Borkmann { 201326fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BHS, true); 20145dd294d4SDaniel Borkmann } 20155dd294d4SDaniel Borkmann 2016d9ae7f2bSJakub Kicinski static int jset_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 2017d9ae7f2bSJakub Kicinski { 2018d9ae7f2bSJakub Kicinski return wrp_test_reg(nfp_prog, meta, ALU_OP_AND, BR_BNE); 2019d9ae7f2bSJakub Kicinski } 2020d9ae7f2bSJakub Kicinski 2021d9ae7f2bSJakub Kicinski static int jne_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 2022d9ae7f2bSJakub Kicinski { 2023d9ae7f2bSJakub Kicinski return wrp_test_reg(nfp_prog, meta, ALU_OP_XOR, BR_BNE); 2024d9ae7f2bSJakub Kicinski } 2025d9ae7f2bSJakub Kicinski 20262cb230bdSJakub Kicinski static int call(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 20272cb230bdSJakub Kicinski { 20282cb230bdSJakub Kicinski switch (meta->insn.imm) { 20290d49eaf4SJakub Kicinski case BPF_FUNC_xdp_adjust_head: 20300d49eaf4SJakub Kicinski return adjust_head(nfp_prog, meta); 20312cb230bdSJakub Kicinski default: 20322cb230bdSJakub Kicinski WARN_ONCE(1, "verifier allowed unsupported function\n"); 20332cb230bdSJakub Kicinski return -EOPNOTSUPP; 20342cb230bdSJakub Kicinski } 20352cb230bdSJakub Kicinski } 20362cb230bdSJakub Kicinski 2037d9ae7f2bSJakub Kicinski static int goto_out(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 2038d9ae7f2bSJakub Kicinski { 2039d9ae7f2bSJakub Kicinski wrp_br_special(nfp_prog, BR_UNC, OP_BR_GO_OUT); 2040d9ae7f2bSJakub Kicinski 2041d9ae7f2bSJakub Kicinski return 0; 2042d9ae7f2bSJakub Kicinski } 2043d9ae7f2bSJakub Kicinski 2044d9ae7f2bSJakub Kicinski static const instr_cb_t instr_cb[256] = { 2045d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_MOV | BPF_X] = mov_reg64, 2046d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_MOV | BPF_K] = mov_imm64, 2047d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_XOR | BPF_X] = xor_reg64, 2048d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_XOR | BPF_K] = xor_imm64, 2049d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_AND | BPF_X] = and_reg64, 2050d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_AND | BPF_K] = and_imm64, 2051d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_OR | BPF_X] = or_reg64, 2052d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_OR | BPF_K] = or_imm64, 2053d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_ADD | BPF_X] = add_reg64, 2054d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_ADD | BPF_K] = add_imm64, 2055d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_SUB | BPF_X] = sub_reg64, 2056d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_SUB | BPF_K] = sub_imm64, 2057254ef4d7SJiong Wang [BPF_ALU64 | BPF_NEG] = neg_reg64, 2058d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_LSH | BPF_K] = shl_imm64, 2059d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_RSH | BPF_K] = shr_imm64, 2060d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_MOV | BPF_X] = mov_reg, 2061d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_MOV | BPF_K] = mov_imm, 2062d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_XOR | BPF_X] = xor_reg, 2063d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_XOR | BPF_K] = xor_imm, 2064d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_AND | BPF_X] = and_reg, 2065d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_AND | BPF_K] = and_imm, 2066d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_OR | BPF_X] = or_reg, 2067d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_OR | BPF_K] = or_imm, 2068d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_ADD | BPF_X] = add_reg, 2069d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_ADD | BPF_K] = add_imm, 2070d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_SUB | BPF_X] = sub_reg, 2071d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_SUB | BPF_K] = sub_imm, 2072254ef4d7SJiong Wang [BPF_ALU | BPF_NEG] = neg_reg, 2073d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_LSH | BPF_K] = shl_imm, 20743119d1fdSJakub Kicinski [BPF_ALU | BPF_END | BPF_X] = end_reg32, 2075d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IMM | BPF_DW] = imm_ld8, 2076d9ae7f2bSJakub Kicinski [BPF_LD | BPF_ABS | BPF_B] = data_ld1, 2077d9ae7f2bSJakub Kicinski [BPF_LD | BPF_ABS | BPF_H] = data_ld2, 2078d9ae7f2bSJakub Kicinski [BPF_LD | BPF_ABS | BPF_W] = data_ld4, 2079d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IND | BPF_B] = data_ind_ld1, 2080d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IND | BPF_H] = data_ind_ld2, 2081d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IND | BPF_W] = data_ind_ld4, 20822ca71441SJakub Kicinski [BPF_LDX | BPF_MEM | BPF_B] = mem_ldx1, 20832ca71441SJakub Kicinski [BPF_LDX | BPF_MEM | BPF_H] = mem_ldx2, 2084d9ae7f2bSJakub Kicinski [BPF_LDX | BPF_MEM | BPF_W] = mem_ldx4, 20852ca71441SJakub Kicinski [BPF_LDX | BPF_MEM | BPF_DW] = mem_ldx8, 2086e663fe38SJakub Kicinski [BPF_STX | BPF_MEM | BPF_B] = mem_stx1, 2087e663fe38SJakub Kicinski [BPF_STX | BPF_MEM | BPF_H] = mem_stx2, 2088d9ae7f2bSJakub Kicinski [BPF_STX | BPF_MEM | BPF_W] = mem_stx4, 2089e663fe38SJakub Kicinski [BPF_STX | BPF_MEM | BPF_DW] = mem_stx8, 2090e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_B] = mem_st1, 2091e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_H] = mem_st2, 2092e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_W] = mem_st4, 2093e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_DW] = mem_st8, 2094d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JA | BPF_K] = jump, 2095d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JEQ | BPF_K] = jeq_imm, 2096d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGT | BPF_K] = jgt_imm, 2097d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGE | BPF_K] = jge_imm, 20985dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLT | BPF_K] = jlt_imm, 20995dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLE | BPF_K] = jle_imm, 2100d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JSET | BPF_K] = jset_imm, 2101d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JNE | BPF_K] = jne_imm, 2102d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JEQ | BPF_X] = jeq_reg, 2103d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGT | BPF_X] = jgt_reg, 2104d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGE | BPF_X] = jge_reg, 21055dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLT | BPF_X] = jlt_reg, 21065dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLE | BPF_X] = jle_reg, 2107d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JSET | BPF_X] = jset_reg, 2108d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JNE | BPF_X] = jne_reg, 21092cb230bdSJakub Kicinski [BPF_JMP | BPF_CALL] = call, 2110d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_EXIT] = goto_out, 2111d9ae7f2bSJakub Kicinski }; 2112d9ae7f2bSJakub Kicinski 2113d9ae7f2bSJakub Kicinski /* --- Misc code --- */ 2114d9ae7f2bSJakub Kicinski static void br_set_offset(u64 *instr, u16 offset) 2115d9ae7f2bSJakub Kicinski { 2116d9ae7f2bSJakub Kicinski u16 addr_lo, addr_hi; 2117d9ae7f2bSJakub Kicinski 2118d9ae7f2bSJakub Kicinski addr_lo = offset & (OP_BR_ADDR_LO >> __bf_shf(OP_BR_ADDR_LO)); 2119d9ae7f2bSJakub Kicinski addr_hi = offset != addr_lo; 2120d9ae7f2bSJakub Kicinski *instr &= ~(OP_BR_ADDR_HI | OP_BR_ADDR_LO); 2121d9ae7f2bSJakub Kicinski *instr |= FIELD_PREP(OP_BR_ADDR_HI, addr_hi); 2122d9ae7f2bSJakub Kicinski *instr |= FIELD_PREP(OP_BR_ADDR_LO, addr_lo); 2123d9ae7f2bSJakub Kicinski } 2124d9ae7f2bSJakub Kicinski 2125d9ae7f2bSJakub Kicinski /* --- Assembler logic --- */ 2126d9ae7f2bSJakub Kicinski static int nfp_fixup_branches(struct nfp_prog *nfp_prog) 2127d9ae7f2bSJakub Kicinski { 21285b674140SJiong Wang struct nfp_insn_meta *meta, *jmp_dst; 2129854dc87dSJiong Wang u32 idx, br_idx; 2130d9ae7f2bSJakub Kicinski 2131854dc87dSJiong Wang list_for_each_entry(meta, &nfp_prog->insns, l) { 2132d9ae7f2bSJakub Kicinski if (meta->skip) 2133d9ae7f2bSJakub Kicinski continue; 21342cb230bdSJakub Kicinski if (meta->insn.code == (BPF_JMP | BPF_CALL)) 21352cb230bdSJakub Kicinski continue; 2136d9ae7f2bSJakub Kicinski if (BPF_CLASS(meta->insn.code) != BPF_JMP) 2137d9ae7f2bSJakub Kicinski continue; 2138d9ae7f2bSJakub Kicinski 21395b674140SJiong Wang if (list_is_last(&meta->l, &nfp_prog->insns)) 2140854dc87dSJiong Wang idx = nfp_prog->last_bpf_off; 21415b674140SJiong Wang else 21425b674140SJiong Wang idx = list_next_entry(meta, l)->off - 1; 2143854dc87dSJiong Wang 2144854dc87dSJiong Wang br_idx = nfp_prog_offset_to_index(nfp_prog, idx); 2145854dc87dSJiong Wang 2146d9ae7f2bSJakub Kicinski if (!nfp_is_br(nfp_prog->prog[br_idx])) { 2147d9ae7f2bSJakub Kicinski pr_err("Fixup found block not ending in branch %d %02x %016llx!!\n", 2148d9ae7f2bSJakub Kicinski br_idx, meta->insn.code, nfp_prog->prog[br_idx]); 2149d9ae7f2bSJakub Kicinski return -ELOOP; 2150d9ae7f2bSJakub Kicinski } 2151d9ae7f2bSJakub Kicinski /* Leave special branches for later */ 2152d9ae7f2bSJakub Kicinski if (FIELD_GET(OP_BR_SPECIAL, nfp_prog->prog[br_idx])) 2153d9ae7f2bSJakub Kicinski continue; 2154d9ae7f2bSJakub Kicinski 21555b674140SJiong Wang if (!meta->jmp_dst) { 21565b674140SJiong Wang pr_err("Non-exit jump doesn't have destination info recorded!!\n"); 2157d9ae7f2bSJakub Kicinski return -ELOOP; 2158d9ae7f2bSJakub Kicinski } 2159d9ae7f2bSJakub Kicinski 21605b674140SJiong Wang jmp_dst = meta->jmp_dst; 2161854dc87dSJiong Wang 21625b674140SJiong Wang if (jmp_dst->skip) { 2163d9ae7f2bSJakub Kicinski pr_err("Branch landing on removed instruction!!\n"); 2164d9ae7f2bSJakub Kicinski return -ELOOP; 2165d9ae7f2bSJakub Kicinski } 2166d9ae7f2bSJakub Kicinski 2167d9ae7f2bSJakub Kicinski for (idx = nfp_prog_offset_to_index(nfp_prog, meta->off); 2168d9ae7f2bSJakub Kicinski idx <= br_idx; idx++) { 2169d9ae7f2bSJakub Kicinski if (!nfp_is_br(nfp_prog->prog[idx])) 2170d9ae7f2bSJakub Kicinski continue; 21715b674140SJiong Wang br_set_offset(&nfp_prog->prog[idx], jmp_dst->off); 2172d9ae7f2bSJakub Kicinski } 2173d9ae7f2bSJakub Kicinski } 2174d9ae7f2bSJakub Kicinski 2175d9ae7f2bSJakub Kicinski /* Fixup 'goto out's separately, they can be scattered around */ 2176d9ae7f2bSJakub Kicinski for (br_idx = 0; br_idx < nfp_prog->prog_len; br_idx++) { 2177d9ae7f2bSJakub Kicinski enum br_special special; 2178d9ae7f2bSJakub Kicinski 2179d9ae7f2bSJakub Kicinski if ((nfp_prog->prog[br_idx] & OP_BR_BASE_MASK) != OP_BR_BASE) 2180d9ae7f2bSJakub Kicinski continue; 2181d9ae7f2bSJakub Kicinski 2182d9ae7f2bSJakub Kicinski special = FIELD_GET(OP_BR_SPECIAL, nfp_prog->prog[br_idx]); 2183d9ae7f2bSJakub Kicinski switch (special) { 2184d9ae7f2bSJakub Kicinski case OP_BR_NORMAL: 2185d9ae7f2bSJakub Kicinski break; 2186d9ae7f2bSJakub Kicinski case OP_BR_GO_OUT: 2187d9ae7f2bSJakub Kicinski br_set_offset(&nfp_prog->prog[br_idx], 2188d9ae7f2bSJakub Kicinski nfp_prog->tgt_out); 2189d9ae7f2bSJakub Kicinski break; 2190d9ae7f2bSJakub Kicinski case OP_BR_GO_ABORT: 2191d9ae7f2bSJakub Kicinski br_set_offset(&nfp_prog->prog[br_idx], 2192d9ae7f2bSJakub Kicinski nfp_prog->tgt_abort); 2193d9ae7f2bSJakub Kicinski break; 2194d9ae7f2bSJakub Kicinski } 2195d9ae7f2bSJakub Kicinski 2196d9ae7f2bSJakub Kicinski nfp_prog->prog[br_idx] &= ~OP_BR_SPECIAL; 2197d9ae7f2bSJakub Kicinski } 2198d9ae7f2bSJakub Kicinski 2199d9ae7f2bSJakub Kicinski return 0; 2200d9ae7f2bSJakub Kicinski } 2201d9ae7f2bSJakub Kicinski 2202d9ae7f2bSJakub Kicinski static void nfp_intro(struct nfp_prog *nfp_prog) 2203d9ae7f2bSJakub Kicinski { 220418e53b6cSJakub Kicinski wrp_immed(nfp_prog, plen_reg(nfp_prog), GENMASK(13, 0)); 220518e53b6cSJakub Kicinski emit_alu(nfp_prog, plen_reg(nfp_prog), 220618e53b6cSJakub Kicinski plen_reg(nfp_prog), ALU_OP_AND, pv_len(nfp_prog)); 2207d9ae7f2bSJakub Kicinski } 2208d9ae7f2bSJakub Kicinski 2209d9ae7f2bSJakub Kicinski static void nfp_outro_tc_da(struct nfp_prog *nfp_prog) 2210d9ae7f2bSJakub Kicinski { 2211d9ae7f2bSJakub Kicinski /* TC direct-action mode: 2212d9ae7f2bSJakub Kicinski * 0,1 ok NOT SUPPORTED[1] 2213d9ae7f2bSJakub Kicinski * 2 drop 0x22 -> drop, count as stat1 2214d9ae7f2bSJakub Kicinski * 4,5 nuke 0x02 -> drop 2215d9ae7f2bSJakub Kicinski * 7 redir 0x44 -> redir, count as stat2 2216d9ae7f2bSJakub Kicinski * * unspec 0x11 -> pass, count as stat0 2217d9ae7f2bSJakub Kicinski * 2218d9ae7f2bSJakub Kicinski * [1] We can't support OK and RECLASSIFY because we can't tell TC 2219d9ae7f2bSJakub Kicinski * the exact decision made. We are forced to support UNSPEC 2220d9ae7f2bSJakub Kicinski * to handle aborts so that's the only one we handle for passing 2221d9ae7f2bSJakub Kicinski * packets up the stack. 2222d9ae7f2bSJakub Kicinski */ 2223d9ae7f2bSJakub Kicinski /* Target for aborts */ 2224d9ae7f2bSJakub Kicinski nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 2225d9ae7f2bSJakub Kicinski 2226d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2227d9ae7f2bSJakub Kicinski 2228c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2229d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x11), SHF_SC_L_SHF, 16); 2230d9ae7f2bSJakub Kicinski 2231d9ae7f2bSJakub Kicinski /* Target for normal exits */ 2232d9ae7f2bSJakub Kicinski nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 2233d9ae7f2bSJakub Kicinski 2234d9ae7f2bSJakub Kicinski /* if R0 > 7 jump to abort */ 2235d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_imm(7), ALU_OP_SUB, reg_b(0)); 2236d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BLO, nfp_prog->tgt_abort, 0); 2237c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2238d9ae7f2bSJakub Kicinski 2239d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_b(2), 0x41221211); 2240d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_b(3), 0x41001211); 2241d9ae7f2bSJakub Kicinski 2242d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_a(1), 2243d9ae7f2bSJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(0), SHF_SC_L_SHF, 2); 2244d9ae7f2bSJakub Kicinski 2245d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 2246d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_a(2), 2247d9ae7f2bSJakub Kicinski reg_imm(0xf), SHF_OP_AND, reg_b(2), SHF_SC_R_SHF, 0); 2248d9ae7f2bSJakub Kicinski 2249d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 2250d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_b(2), 2251d9ae7f2bSJakub Kicinski reg_imm(0xf), SHF_OP_AND, reg_b(3), SHF_SC_R_SHF, 0); 2252d9ae7f2bSJakub Kicinski 2253d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2254d9ae7f2bSJakub Kicinski 2255d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_b(2), 2256d9ae7f2bSJakub Kicinski reg_a(2), SHF_OP_OR, reg_b(2), SHF_SC_L_SHF, 4); 2257d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_b(2), SHF_SC_L_SHF, 16); 2258d9ae7f2bSJakub Kicinski } 2259d9ae7f2bSJakub Kicinski 2260d9ae7f2bSJakub Kicinski static void nfp_outro_xdp(struct nfp_prog *nfp_prog) 2261d9ae7f2bSJakub Kicinski { 2262d9ae7f2bSJakub Kicinski /* XDP return codes: 2263d9ae7f2bSJakub Kicinski * 0 aborted 0x82 -> drop, count as stat3 2264d9ae7f2bSJakub Kicinski * 1 drop 0x22 -> drop, count as stat1 2265d9ae7f2bSJakub Kicinski * 2 pass 0x11 -> pass, count as stat0 2266d9ae7f2bSJakub Kicinski * 3 tx 0x44 -> redir, count as stat2 2267d9ae7f2bSJakub Kicinski * * unknown 0x82 -> drop, count as stat3 2268d9ae7f2bSJakub Kicinski */ 2269d9ae7f2bSJakub Kicinski /* Target for aborts */ 2270d9ae7f2bSJakub Kicinski nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 2271d9ae7f2bSJakub Kicinski 2272d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2273d9ae7f2bSJakub Kicinski 2274c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2275d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x82), SHF_SC_L_SHF, 16); 2276d9ae7f2bSJakub Kicinski 2277d9ae7f2bSJakub Kicinski /* Target for normal exits */ 2278d9ae7f2bSJakub Kicinski nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 2279d9ae7f2bSJakub Kicinski 2280d9ae7f2bSJakub Kicinski /* if R0 > 3 jump to abort */ 2281d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_imm(3), ALU_OP_SUB, reg_b(0)); 2282d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BLO, nfp_prog->tgt_abort, 0); 2283d9ae7f2bSJakub Kicinski 2284d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_b(2), 0x44112282); 2285d9ae7f2bSJakub Kicinski 2286d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_a(1), 2287d9ae7f2bSJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(0), SHF_SC_L_SHF, 3); 2288d9ae7f2bSJakub Kicinski 2289d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 2290d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_b(2), 2291d9ae7f2bSJakub Kicinski reg_imm(0xff), SHF_OP_AND, reg_b(2), SHF_SC_R_SHF, 0); 2292d9ae7f2bSJakub Kicinski 2293d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2294d9ae7f2bSJakub Kicinski 2295c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2296d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_b(2), SHF_SC_L_SHF, 16); 2297d9ae7f2bSJakub Kicinski } 2298d9ae7f2bSJakub Kicinski 2299d9ae7f2bSJakub Kicinski static void nfp_outro(struct nfp_prog *nfp_prog) 2300d9ae7f2bSJakub Kicinski { 2301012bb8a8SJakub Kicinski switch (nfp_prog->type) { 2302012bb8a8SJakub Kicinski case BPF_PROG_TYPE_SCHED_CLS: 2303d9ae7f2bSJakub Kicinski nfp_outro_tc_da(nfp_prog); 2304d9ae7f2bSJakub Kicinski break; 2305012bb8a8SJakub Kicinski case BPF_PROG_TYPE_XDP: 2306d9ae7f2bSJakub Kicinski nfp_outro_xdp(nfp_prog); 2307d9ae7f2bSJakub Kicinski break; 2308012bb8a8SJakub Kicinski default: 2309012bb8a8SJakub Kicinski WARN_ON(1); 2310d9ae7f2bSJakub Kicinski } 2311d9ae7f2bSJakub Kicinski } 2312d9ae7f2bSJakub Kicinski 2313d9ae7f2bSJakub Kicinski static int nfp_translate(struct nfp_prog *nfp_prog) 2314d9ae7f2bSJakub Kicinski { 2315d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta; 2316ff42bb9fSJakub Kicinski int err; 2317d9ae7f2bSJakub Kicinski 2318d9ae7f2bSJakub Kicinski nfp_intro(nfp_prog); 2319d9ae7f2bSJakub Kicinski if (nfp_prog->error) 2320d9ae7f2bSJakub Kicinski return nfp_prog->error; 2321d9ae7f2bSJakub Kicinski 2322d9ae7f2bSJakub Kicinski list_for_each_entry(meta, &nfp_prog->insns, l) { 2323d9ae7f2bSJakub Kicinski instr_cb_t cb = instr_cb[meta->insn.code]; 2324d9ae7f2bSJakub Kicinski 2325d9ae7f2bSJakub Kicinski meta->off = nfp_prog_current_offset(nfp_prog); 2326d9ae7f2bSJakub Kicinski 2327d9ae7f2bSJakub Kicinski if (meta->skip) { 2328d9ae7f2bSJakub Kicinski nfp_prog->n_translated++; 2329d9ae7f2bSJakub Kicinski continue; 2330d9ae7f2bSJakub Kicinski } 2331d9ae7f2bSJakub Kicinski 2332d9ae7f2bSJakub Kicinski if (nfp_meta_has_prev(nfp_prog, meta) && 2333d9ae7f2bSJakub Kicinski nfp_meta_prev(meta)->double_cb) 2334d9ae7f2bSJakub Kicinski cb = nfp_meta_prev(meta)->double_cb; 2335d9ae7f2bSJakub Kicinski if (!cb) 2336d9ae7f2bSJakub Kicinski return -ENOENT; 2337d9ae7f2bSJakub Kicinski err = cb(nfp_prog, meta); 2338d9ae7f2bSJakub Kicinski if (err) 2339d9ae7f2bSJakub Kicinski return err; 2340d9ae7f2bSJakub Kicinski 2341d9ae7f2bSJakub Kicinski nfp_prog->n_translated++; 2342d9ae7f2bSJakub Kicinski } 2343d9ae7f2bSJakub Kicinski 2344854dc87dSJiong Wang nfp_prog->last_bpf_off = nfp_prog_current_offset(nfp_prog) - 1; 2345854dc87dSJiong Wang 2346d9ae7f2bSJakub Kicinski nfp_outro(nfp_prog); 2347d9ae7f2bSJakub Kicinski if (nfp_prog->error) 2348d9ae7f2bSJakub Kicinski return nfp_prog->error; 2349d9ae7f2bSJakub Kicinski 2350ff42bb9fSJakub Kicinski wrp_nops(nfp_prog, NFP_USTORE_PREFETCH_WINDOW); 23511c03e03fSJakub Kicinski if (nfp_prog->error) 23521c03e03fSJakub Kicinski return nfp_prog->error; 23531c03e03fSJakub Kicinski 2354d9ae7f2bSJakub Kicinski return nfp_fixup_branches(nfp_prog); 2355d9ae7f2bSJakub Kicinski } 2356d9ae7f2bSJakub Kicinski 2357d9ae7f2bSJakub Kicinski /* --- Optimizations --- */ 2358d9ae7f2bSJakub Kicinski static void nfp_bpf_opt_reg_init(struct nfp_prog *nfp_prog) 2359d9ae7f2bSJakub Kicinski { 2360d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta; 2361d9ae7f2bSJakub Kicinski 2362d9ae7f2bSJakub Kicinski list_for_each_entry(meta, &nfp_prog->insns, l) { 2363d9ae7f2bSJakub Kicinski struct bpf_insn insn = meta->insn; 2364d9ae7f2bSJakub Kicinski 2365d9ae7f2bSJakub Kicinski /* Programs converted from cBPF start with register xoring */ 2366d9ae7f2bSJakub Kicinski if (insn.code == (BPF_ALU64 | BPF_XOR | BPF_X) && 2367d9ae7f2bSJakub Kicinski insn.src_reg == insn.dst_reg) 2368d9ae7f2bSJakub Kicinski continue; 2369d9ae7f2bSJakub Kicinski 2370d9ae7f2bSJakub Kicinski /* Programs start with R6 = R1 but we ignore the skb pointer */ 2371d9ae7f2bSJakub Kicinski if (insn.code == (BPF_ALU64 | BPF_MOV | BPF_X) && 2372d9ae7f2bSJakub Kicinski insn.src_reg == 1 && insn.dst_reg == 6) 2373d9ae7f2bSJakub Kicinski meta->skip = true; 2374d9ae7f2bSJakub Kicinski 2375d9ae7f2bSJakub Kicinski /* Return as soon as something doesn't match */ 2376d9ae7f2bSJakub Kicinski if (!meta->skip) 2377d9ae7f2bSJakub Kicinski return; 2378d9ae7f2bSJakub Kicinski } 2379d9ae7f2bSJakub Kicinski } 2380d9ae7f2bSJakub Kicinski 2381d9ae7f2bSJakub Kicinski /* Remove masking after load since our load guarantees this is not needed */ 2382d9ae7f2bSJakub Kicinski static void nfp_bpf_opt_ld_mask(struct nfp_prog *nfp_prog) 2383d9ae7f2bSJakub Kicinski { 2384d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta1, *meta2; 2385d9ae7f2bSJakub Kicinski const s32 exp_mask[] = { 2386d9ae7f2bSJakub Kicinski [BPF_B] = 0x000000ffU, 2387d9ae7f2bSJakub Kicinski [BPF_H] = 0x0000ffffU, 2388d9ae7f2bSJakub Kicinski [BPF_W] = 0xffffffffU, 2389d9ae7f2bSJakub Kicinski }; 2390d9ae7f2bSJakub Kicinski 2391d9ae7f2bSJakub Kicinski nfp_for_each_insn_walk2(nfp_prog, meta1, meta2) { 2392d9ae7f2bSJakub Kicinski struct bpf_insn insn, next; 2393d9ae7f2bSJakub Kicinski 2394d9ae7f2bSJakub Kicinski insn = meta1->insn; 2395d9ae7f2bSJakub Kicinski next = meta2->insn; 2396d9ae7f2bSJakub Kicinski 2397d9ae7f2bSJakub Kicinski if (BPF_CLASS(insn.code) != BPF_LD) 2398d9ae7f2bSJakub Kicinski continue; 2399d9ae7f2bSJakub Kicinski if (BPF_MODE(insn.code) != BPF_ABS && 2400d9ae7f2bSJakub Kicinski BPF_MODE(insn.code) != BPF_IND) 2401d9ae7f2bSJakub Kicinski continue; 2402d9ae7f2bSJakub Kicinski 2403d9ae7f2bSJakub Kicinski if (next.code != (BPF_ALU64 | BPF_AND | BPF_K)) 2404d9ae7f2bSJakub Kicinski continue; 2405d9ae7f2bSJakub Kicinski 2406d9ae7f2bSJakub Kicinski if (!exp_mask[BPF_SIZE(insn.code)]) 2407d9ae7f2bSJakub Kicinski continue; 2408d9ae7f2bSJakub Kicinski if (exp_mask[BPF_SIZE(insn.code)] != next.imm) 2409d9ae7f2bSJakub Kicinski continue; 2410d9ae7f2bSJakub Kicinski 2411d9ae7f2bSJakub Kicinski if (next.src_reg || next.dst_reg) 2412d9ae7f2bSJakub Kicinski continue; 2413d9ae7f2bSJakub Kicinski 24141266f5d6SJiong Wang if (meta2->flags & FLAG_INSN_IS_JUMP_DST) 24151266f5d6SJiong Wang continue; 24161266f5d6SJiong Wang 2417d9ae7f2bSJakub Kicinski meta2->skip = true; 2418d9ae7f2bSJakub Kicinski } 2419d9ae7f2bSJakub Kicinski } 2420d9ae7f2bSJakub Kicinski 2421d9ae7f2bSJakub Kicinski static void nfp_bpf_opt_ld_shift(struct nfp_prog *nfp_prog) 2422d9ae7f2bSJakub Kicinski { 2423d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta1, *meta2, *meta3; 2424d9ae7f2bSJakub Kicinski 2425d9ae7f2bSJakub Kicinski nfp_for_each_insn_walk3(nfp_prog, meta1, meta2, meta3) { 2426d9ae7f2bSJakub Kicinski struct bpf_insn insn, next1, next2; 2427d9ae7f2bSJakub Kicinski 2428d9ae7f2bSJakub Kicinski insn = meta1->insn; 2429d9ae7f2bSJakub Kicinski next1 = meta2->insn; 2430d9ae7f2bSJakub Kicinski next2 = meta3->insn; 2431d9ae7f2bSJakub Kicinski 2432d9ae7f2bSJakub Kicinski if (BPF_CLASS(insn.code) != BPF_LD) 2433d9ae7f2bSJakub Kicinski continue; 2434d9ae7f2bSJakub Kicinski if (BPF_MODE(insn.code) != BPF_ABS && 2435d9ae7f2bSJakub Kicinski BPF_MODE(insn.code) != BPF_IND) 2436d9ae7f2bSJakub Kicinski continue; 2437d9ae7f2bSJakub Kicinski if (BPF_SIZE(insn.code) != BPF_W) 2438d9ae7f2bSJakub Kicinski continue; 2439d9ae7f2bSJakub Kicinski 2440d9ae7f2bSJakub Kicinski if (!(next1.code == (BPF_LSH | BPF_K | BPF_ALU64) && 2441d9ae7f2bSJakub Kicinski next2.code == (BPF_RSH | BPF_K | BPF_ALU64)) && 2442d9ae7f2bSJakub Kicinski !(next1.code == (BPF_RSH | BPF_K | BPF_ALU64) && 2443d9ae7f2bSJakub Kicinski next2.code == (BPF_LSH | BPF_K | BPF_ALU64))) 2444d9ae7f2bSJakub Kicinski continue; 2445d9ae7f2bSJakub Kicinski 2446d9ae7f2bSJakub Kicinski if (next1.src_reg || next1.dst_reg || 2447d9ae7f2bSJakub Kicinski next2.src_reg || next2.dst_reg) 2448d9ae7f2bSJakub Kicinski continue; 2449d9ae7f2bSJakub Kicinski 2450d9ae7f2bSJakub Kicinski if (next1.imm != 0x20 || next2.imm != 0x20) 2451d9ae7f2bSJakub Kicinski continue; 2452d9ae7f2bSJakub Kicinski 245329fe46efSJiong Wang if (meta2->flags & FLAG_INSN_IS_JUMP_DST || 245429fe46efSJiong Wang meta3->flags & FLAG_INSN_IS_JUMP_DST) 245529fe46efSJiong Wang continue; 245629fe46efSJiong Wang 2457d9ae7f2bSJakub Kicinski meta2->skip = true; 2458d9ae7f2bSJakub Kicinski meta3->skip = true; 2459d9ae7f2bSJakub Kicinski } 2460d9ae7f2bSJakub Kicinski } 2461d9ae7f2bSJakub Kicinski 24626bc7103cSJiong Wang /* load/store pair that forms memory copy sould look like the following: 24636bc7103cSJiong Wang * 24646bc7103cSJiong Wang * ld_width R, [addr_src + offset_src] 24656bc7103cSJiong Wang * st_width [addr_dest + offset_dest], R 24666bc7103cSJiong Wang * 24676bc7103cSJiong Wang * The destination register of load and source register of store should 24686bc7103cSJiong Wang * be the same, load and store should also perform at the same width. 24696bc7103cSJiong Wang * If either of addr_src or addr_dest is stack pointer, we don't do the 24706bc7103cSJiong Wang * CPP optimization as stack is modelled by registers on NFP. 24716bc7103cSJiong Wang */ 24726bc7103cSJiong Wang static bool 24736bc7103cSJiong Wang curr_pair_is_memcpy(struct nfp_insn_meta *ld_meta, 24746bc7103cSJiong Wang struct nfp_insn_meta *st_meta) 24756bc7103cSJiong Wang { 24766bc7103cSJiong Wang struct bpf_insn *ld = &ld_meta->insn; 24776bc7103cSJiong Wang struct bpf_insn *st = &st_meta->insn; 24786bc7103cSJiong Wang 24796bc7103cSJiong Wang if (!is_mbpf_load(ld_meta) || !is_mbpf_store(st_meta)) 24806bc7103cSJiong Wang return false; 24816bc7103cSJiong Wang 24826bc7103cSJiong Wang if (ld_meta->ptr.type != PTR_TO_PACKET) 24836bc7103cSJiong Wang return false; 24846bc7103cSJiong Wang 24856bc7103cSJiong Wang if (st_meta->ptr.type != PTR_TO_PACKET) 24866bc7103cSJiong Wang return false; 24876bc7103cSJiong Wang 24886bc7103cSJiong Wang if (BPF_SIZE(ld->code) != BPF_SIZE(st->code)) 24896bc7103cSJiong Wang return false; 24906bc7103cSJiong Wang 24916bc7103cSJiong Wang if (ld->dst_reg != st->src_reg) 24926bc7103cSJiong Wang return false; 24936bc7103cSJiong Wang 24946bc7103cSJiong Wang /* There is jump to the store insn in this pair. */ 24956bc7103cSJiong Wang if (st_meta->flags & FLAG_INSN_IS_JUMP_DST) 24966bc7103cSJiong Wang return false; 24976bc7103cSJiong Wang 24986bc7103cSJiong Wang return true; 24996bc7103cSJiong Wang } 25006bc7103cSJiong Wang 25016bc7103cSJiong Wang /* Currently, we only support chaining load/store pairs if: 25026bc7103cSJiong Wang * 25036bc7103cSJiong Wang * - Their address base registers are the same. 25046bc7103cSJiong Wang * - Their address offsets are in the same order. 25056bc7103cSJiong Wang * - They operate at the same memory width. 25066bc7103cSJiong Wang * - There is no jump into the middle of them. 25076bc7103cSJiong Wang */ 25086bc7103cSJiong Wang static bool 25096bc7103cSJiong Wang curr_pair_chain_with_previous(struct nfp_insn_meta *ld_meta, 25106bc7103cSJiong Wang struct nfp_insn_meta *st_meta, 25116bc7103cSJiong Wang struct bpf_insn *prev_ld, 25126bc7103cSJiong Wang struct bpf_insn *prev_st) 25136bc7103cSJiong Wang { 25146bc7103cSJiong Wang u8 prev_size, curr_size, prev_ld_base, prev_st_base, prev_ld_dst; 25156bc7103cSJiong Wang struct bpf_insn *ld = &ld_meta->insn; 25166bc7103cSJiong Wang struct bpf_insn *st = &st_meta->insn; 25176bc7103cSJiong Wang s16 prev_ld_off, prev_st_off; 25186bc7103cSJiong Wang 25196bc7103cSJiong Wang /* This pair is the start pair. */ 25206bc7103cSJiong Wang if (!prev_ld) 25216bc7103cSJiong Wang return true; 25226bc7103cSJiong Wang 25236bc7103cSJiong Wang prev_size = BPF_LDST_BYTES(prev_ld); 25246bc7103cSJiong Wang curr_size = BPF_LDST_BYTES(ld); 25256bc7103cSJiong Wang prev_ld_base = prev_ld->src_reg; 25266bc7103cSJiong Wang prev_st_base = prev_st->dst_reg; 25276bc7103cSJiong Wang prev_ld_dst = prev_ld->dst_reg; 25286bc7103cSJiong Wang prev_ld_off = prev_ld->off; 25296bc7103cSJiong Wang prev_st_off = prev_st->off; 25306bc7103cSJiong Wang 25316bc7103cSJiong Wang if (ld->dst_reg != prev_ld_dst) 25326bc7103cSJiong Wang return false; 25336bc7103cSJiong Wang 25346bc7103cSJiong Wang if (ld->src_reg != prev_ld_base || st->dst_reg != prev_st_base) 25356bc7103cSJiong Wang return false; 25366bc7103cSJiong Wang 25376bc7103cSJiong Wang if (curr_size != prev_size) 25386bc7103cSJiong Wang return false; 25396bc7103cSJiong Wang 25406bc7103cSJiong Wang /* There is jump to the head of this pair. */ 25416bc7103cSJiong Wang if (ld_meta->flags & FLAG_INSN_IS_JUMP_DST) 25426bc7103cSJiong Wang return false; 25436bc7103cSJiong Wang 25446bc7103cSJiong Wang /* Both in ascending order. */ 25456bc7103cSJiong Wang if (prev_ld_off + prev_size == ld->off && 25466bc7103cSJiong Wang prev_st_off + prev_size == st->off) 25476bc7103cSJiong Wang return true; 25486bc7103cSJiong Wang 25496bc7103cSJiong Wang /* Both in descending order. */ 25506bc7103cSJiong Wang if (ld->off + curr_size == prev_ld_off && 25516bc7103cSJiong Wang st->off + curr_size == prev_st_off) 25526bc7103cSJiong Wang return true; 25536bc7103cSJiong Wang 25546bc7103cSJiong Wang return false; 25556bc7103cSJiong Wang } 25566bc7103cSJiong Wang 25576bc7103cSJiong Wang /* Return TRUE if cross memory access happens. Cross memory access means 25586bc7103cSJiong Wang * store area is overlapping with load area that a later load might load 25596bc7103cSJiong Wang * the value from previous store, for this case we can't treat the sequence 25606bc7103cSJiong Wang * as an memory copy. 25616bc7103cSJiong Wang */ 25626bc7103cSJiong Wang static bool 25636bc7103cSJiong Wang cross_mem_access(struct bpf_insn *ld, struct nfp_insn_meta *head_ld_meta, 25646bc7103cSJiong Wang struct nfp_insn_meta *head_st_meta) 25656bc7103cSJiong Wang { 25666bc7103cSJiong Wang s16 head_ld_off, head_st_off, ld_off; 25676bc7103cSJiong Wang 25686bc7103cSJiong Wang /* Different pointer types does not overlap. */ 25696bc7103cSJiong Wang if (head_ld_meta->ptr.type != head_st_meta->ptr.type) 25706bc7103cSJiong Wang return false; 25716bc7103cSJiong Wang 25726bc7103cSJiong Wang /* load and store are both PTR_TO_PACKET, check ID info. */ 25736bc7103cSJiong Wang if (head_ld_meta->ptr.id != head_st_meta->ptr.id) 25746bc7103cSJiong Wang return true; 25756bc7103cSJiong Wang 25766bc7103cSJiong Wang /* Canonicalize the offsets. Turn all of them against the original 25776bc7103cSJiong Wang * base register. 25786bc7103cSJiong Wang */ 25796bc7103cSJiong Wang head_ld_off = head_ld_meta->insn.off + head_ld_meta->ptr.off; 25806bc7103cSJiong Wang head_st_off = head_st_meta->insn.off + head_st_meta->ptr.off; 25816bc7103cSJiong Wang ld_off = ld->off + head_ld_meta->ptr.off; 25826bc7103cSJiong Wang 25836bc7103cSJiong Wang /* Ascending order cross. */ 25846bc7103cSJiong Wang if (ld_off > head_ld_off && 25856bc7103cSJiong Wang head_ld_off < head_st_off && ld_off >= head_st_off) 25866bc7103cSJiong Wang return true; 25876bc7103cSJiong Wang 25886bc7103cSJiong Wang /* Descending order cross. */ 25896bc7103cSJiong Wang if (ld_off < head_ld_off && 25906bc7103cSJiong Wang head_ld_off > head_st_off && ld_off <= head_st_off) 25916bc7103cSJiong Wang return true; 25926bc7103cSJiong Wang 25936bc7103cSJiong Wang return false; 25946bc7103cSJiong Wang } 25956bc7103cSJiong Wang 25966bc7103cSJiong Wang /* This pass try to identify the following instructoin sequences. 25976bc7103cSJiong Wang * 25986bc7103cSJiong Wang * load R, [regA + offA] 25996bc7103cSJiong Wang * store [regB + offB], R 26006bc7103cSJiong Wang * load R, [regA + offA + const_imm_A] 26016bc7103cSJiong Wang * store [regB + offB + const_imm_A], R 26026bc7103cSJiong Wang * load R, [regA + offA + 2 * const_imm_A] 26036bc7103cSJiong Wang * store [regB + offB + 2 * const_imm_A], R 26046bc7103cSJiong Wang * ... 26056bc7103cSJiong Wang * 26066bc7103cSJiong Wang * Above sequence is typically generated by compiler when lowering 26076bc7103cSJiong Wang * memcpy. NFP prefer using CPP instructions to accelerate it. 26086bc7103cSJiong Wang */ 26096bc7103cSJiong Wang static void nfp_bpf_opt_ldst_gather(struct nfp_prog *nfp_prog) 26106bc7103cSJiong Wang { 26116bc7103cSJiong Wang struct nfp_insn_meta *head_ld_meta = NULL; 26126bc7103cSJiong Wang struct nfp_insn_meta *head_st_meta = NULL; 26136bc7103cSJiong Wang struct nfp_insn_meta *meta1, *meta2; 26146bc7103cSJiong Wang struct bpf_insn *prev_ld = NULL; 26156bc7103cSJiong Wang struct bpf_insn *prev_st = NULL; 26166bc7103cSJiong Wang u8 count = 0; 26176bc7103cSJiong Wang 26186bc7103cSJiong Wang nfp_for_each_insn_walk2(nfp_prog, meta1, meta2) { 26196bc7103cSJiong Wang struct bpf_insn *ld = &meta1->insn; 26206bc7103cSJiong Wang struct bpf_insn *st = &meta2->insn; 26216bc7103cSJiong Wang 26226bc7103cSJiong Wang /* Reset record status if any of the following if true: 26236bc7103cSJiong Wang * - The current insn pair is not load/store. 26246bc7103cSJiong Wang * - The load/store pair doesn't chain with previous one. 26256bc7103cSJiong Wang * - The chained load/store pair crossed with previous pair. 26266bc7103cSJiong Wang * - The chained load/store pair has a total size of memory 26276bc7103cSJiong Wang * copy beyond 128 bytes which is the maximum length a 26286bc7103cSJiong Wang * single NFP CPP command can transfer. 26296bc7103cSJiong Wang */ 26306bc7103cSJiong Wang if (!curr_pair_is_memcpy(meta1, meta2) || 26316bc7103cSJiong Wang !curr_pair_chain_with_previous(meta1, meta2, prev_ld, 26326bc7103cSJiong Wang prev_st) || 26336bc7103cSJiong Wang (head_ld_meta && (cross_mem_access(ld, head_ld_meta, 26346bc7103cSJiong Wang head_st_meta) || 26356bc7103cSJiong Wang head_ld_meta->ldst_gather_len >= 128))) { 26366bc7103cSJiong Wang if (!count) 26376bc7103cSJiong Wang continue; 26386bc7103cSJiong Wang 26396bc7103cSJiong Wang if (count > 1) { 26406bc7103cSJiong Wang s16 prev_ld_off = prev_ld->off; 26416bc7103cSJiong Wang s16 prev_st_off = prev_st->off; 26426bc7103cSJiong Wang s16 head_ld_off = head_ld_meta->insn.off; 26436bc7103cSJiong Wang 26446bc7103cSJiong Wang if (prev_ld_off < head_ld_off) { 26456bc7103cSJiong Wang head_ld_meta->insn.off = prev_ld_off; 26466bc7103cSJiong Wang head_st_meta->insn.off = prev_st_off; 26476bc7103cSJiong Wang head_ld_meta->ldst_gather_len = 26486bc7103cSJiong Wang -head_ld_meta->ldst_gather_len; 26496bc7103cSJiong Wang } 26506bc7103cSJiong Wang 26516bc7103cSJiong Wang head_ld_meta->paired_st = &head_st_meta->insn; 26526bc7103cSJiong Wang head_st_meta->skip = true; 26536bc7103cSJiong Wang } else { 26546bc7103cSJiong Wang head_ld_meta->ldst_gather_len = 0; 26556bc7103cSJiong Wang } 26566bc7103cSJiong Wang 26576bc7103cSJiong Wang /* If the chain is ended by an load/store pair then this 26586bc7103cSJiong Wang * could serve as the new head of the the next chain. 26596bc7103cSJiong Wang */ 26606bc7103cSJiong Wang if (curr_pair_is_memcpy(meta1, meta2)) { 26616bc7103cSJiong Wang head_ld_meta = meta1; 26626bc7103cSJiong Wang head_st_meta = meta2; 26636bc7103cSJiong Wang head_ld_meta->ldst_gather_len = 26646bc7103cSJiong Wang BPF_LDST_BYTES(ld); 26656bc7103cSJiong Wang meta1 = nfp_meta_next(meta1); 26666bc7103cSJiong Wang meta2 = nfp_meta_next(meta2); 26676bc7103cSJiong Wang prev_ld = ld; 26686bc7103cSJiong Wang prev_st = st; 26696bc7103cSJiong Wang count = 1; 26706bc7103cSJiong Wang } else { 26716bc7103cSJiong Wang head_ld_meta = NULL; 26726bc7103cSJiong Wang head_st_meta = NULL; 26736bc7103cSJiong Wang prev_ld = NULL; 26746bc7103cSJiong Wang prev_st = NULL; 26756bc7103cSJiong Wang count = 0; 26766bc7103cSJiong Wang } 26776bc7103cSJiong Wang 26786bc7103cSJiong Wang continue; 26796bc7103cSJiong Wang } 26806bc7103cSJiong Wang 26816bc7103cSJiong Wang if (!head_ld_meta) { 26826bc7103cSJiong Wang head_ld_meta = meta1; 26836bc7103cSJiong Wang head_st_meta = meta2; 26846bc7103cSJiong Wang } else { 26856bc7103cSJiong Wang meta1->skip = true; 26866bc7103cSJiong Wang meta2->skip = true; 26876bc7103cSJiong Wang } 26886bc7103cSJiong Wang 26896bc7103cSJiong Wang head_ld_meta->ldst_gather_len += BPF_LDST_BYTES(ld); 26906bc7103cSJiong Wang meta1 = nfp_meta_next(meta1); 26916bc7103cSJiong Wang meta2 = nfp_meta_next(meta2); 26926bc7103cSJiong Wang prev_ld = ld; 26936bc7103cSJiong Wang prev_st = st; 26946bc7103cSJiong Wang count++; 26956bc7103cSJiong Wang } 26966bc7103cSJiong Wang } 26976bc7103cSJiong Wang 2698d9ae7f2bSJakub Kicinski static int nfp_bpf_optimize(struct nfp_prog *nfp_prog) 2699d9ae7f2bSJakub Kicinski { 2700d9ae7f2bSJakub Kicinski nfp_bpf_opt_reg_init(nfp_prog); 2701d9ae7f2bSJakub Kicinski 2702d9ae7f2bSJakub Kicinski nfp_bpf_opt_ld_mask(nfp_prog); 2703d9ae7f2bSJakub Kicinski nfp_bpf_opt_ld_shift(nfp_prog); 27046bc7103cSJiong Wang nfp_bpf_opt_ldst_gather(nfp_prog); 2705d9ae7f2bSJakub Kicinski 2706d9ae7f2bSJakub Kicinski return 0; 2707d9ae7f2bSJakub Kicinski } 2708d9ae7f2bSJakub Kicinski 27092e85d388SJakub Kicinski static int nfp_bpf_ustore_calc(struct nfp_prog *nfp_prog, __le64 *ustore) 2710fd068ddcSJakub Kicinski { 2711fd068ddcSJakub Kicinski int i; 2712fd068ddcSJakub Kicinski 2713fd068ddcSJakub Kicinski for (i = 0; i < nfp_prog->prog_len; i++) { 2714fd068ddcSJakub Kicinski int err; 2715fd068ddcSJakub Kicinski 2716fd068ddcSJakub Kicinski err = nfp_ustore_check_valid_no_ecc(nfp_prog->prog[i]); 2717fd068ddcSJakub Kicinski if (err) 2718fd068ddcSJakub Kicinski return err; 2719fd068ddcSJakub Kicinski 2720fd068ddcSJakub Kicinski nfp_prog->prog[i] = nfp_ustore_calc_ecc_insn(nfp_prog->prog[i]); 27212e85d388SJakub Kicinski 27222e85d388SJakub Kicinski ustore[i] = cpu_to_le64(nfp_prog->prog[i]); 2723fd068ddcSJakub Kicinski } 2724fd068ddcSJakub Kicinski 2725fd068ddcSJakub Kicinski return 0; 2726fd068ddcSJakub Kicinski } 2727fd068ddcSJakub Kicinski 2728c6c580d7SJakub Kicinski int nfp_bpf_jit(struct nfp_prog *nfp_prog) 2729d9ae7f2bSJakub Kicinski { 2730d9ae7f2bSJakub Kicinski int ret; 2731d9ae7f2bSJakub Kicinski 2732d9ae7f2bSJakub Kicinski ret = nfp_bpf_optimize(nfp_prog); 2733d9ae7f2bSJakub Kicinski if (ret) 27349314c442SJakub Kicinski return ret; 2735d9ae7f2bSJakub Kicinski 2736d9ae7f2bSJakub Kicinski ret = nfp_translate(nfp_prog); 2737d9ae7f2bSJakub Kicinski if (ret) { 2738d9ae7f2bSJakub Kicinski pr_err("Translation failed with error %d (translated: %u)\n", 2739d9ae7f2bSJakub Kicinski ret, nfp_prog->n_translated); 27409314c442SJakub Kicinski return -EINVAL; 2741d9ae7f2bSJakub Kicinski } 2742d9ae7f2bSJakub Kicinski 27439314c442SJakub Kicinski return nfp_bpf_ustore_calc(nfp_prog, (__force __le64 *)nfp_prog->prog); 2744d9ae7f2bSJakub Kicinski } 2745