1d9ae7f2bSJakub Kicinski /* 2854dc87dSJiong Wang * Copyright (C) 2016-2017 Netronome Systems, Inc. 3d9ae7f2bSJakub Kicinski * 4d9ae7f2bSJakub Kicinski * This software is dual licensed under the GNU General License Version 2, 5d9ae7f2bSJakub Kicinski * June 1991 as shown in the file COPYING in the top-level directory of this 6d9ae7f2bSJakub Kicinski * source tree or the BSD 2-Clause License provided below. You have the 7d9ae7f2bSJakub Kicinski * option to license this software under the complete terms of either license. 8d9ae7f2bSJakub Kicinski * 9d9ae7f2bSJakub Kicinski * The BSD 2-Clause License: 10d9ae7f2bSJakub Kicinski * 11d9ae7f2bSJakub Kicinski * Redistribution and use in source and binary forms, with or 12d9ae7f2bSJakub Kicinski * without modification, are permitted provided that the following 13d9ae7f2bSJakub Kicinski * conditions are met: 14d9ae7f2bSJakub Kicinski * 15d9ae7f2bSJakub Kicinski * 1. Redistributions of source code must retain the above 16d9ae7f2bSJakub Kicinski * copyright notice, this list of conditions and the following 17d9ae7f2bSJakub Kicinski * disclaimer. 18d9ae7f2bSJakub Kicinski * 19d9ae7f2bSJakub Kicinski * 2. Redistributions in binary form must reproduce the above 20d9ae7f2bSJakub Kicinski * copyright notice, this list of conditions and the following 21d9ae7f2bSJakub Kicinski * disclaimer in the documentation and/or other materials 22d9ae7f2bSJakub Kicinski * provided with the distribution. 23d9ae7f2bSJakub Kicinski * 24d9ae7f2bSJakub Kicinski * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 25d9ae7f2bSJakub Kicinski * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 26d9ae7f2bSJakub Kicinski * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 27d9ae7f2bSJakub Kicinski * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 28d9ae7f2bSJakub Kicinski * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 29d9ae7f2bSJakub Kicinski * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 30d9ae7f2bSJakub Kicinski * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 31d9ae7f2bSJakub Kicinski * SOFTWARE. 32d9ae7f2bSJakub Kicinski */ 33d9ae7f2bSJakub Kicinski 34d9ae7f2bSJakub Kicinski #define pr_fmt(fmt) "NFP net bpf: " fmt 35d9ae7f2bSJakub Kicinski 360d49eaf4SJakub Kicinski #include <linux/bug.h> 37d9ae7f2bSJakub Kicinski #include <linux/kernel.h> 38d9ae7f2bSJakub Kicinski #include <linux/bpf.h> 39d9ae7f2bSJakub Kicinski #include <linux/filter.h> 40d9ae7f2bSJakub Kicinski #include <linux/pkt_cls.h> 41d9ae7f2bSJakub Kicinski #include <linux/unistd.h> 42d9ae7f2bSJakub Kicinski 43d9ae7f2bSJakub Kicinski #include "main.h" 44d9ae7f2bSJakub Kicinski #include "../nfp_asm.h" 45d9ae7f2bSJakub Kicinski 46d9ae7f2bSJakub Kicinski /* --- NFP prog --- */ 47d9ae7f2bSJakub Kicinski /* Foreach "multiple" entries macros provide pos and next<n> pointers. 48d9ae7f2bSJakub Kicinski * It's safe to modify the next pointers (but not pos). 49d9ae7f2bSJakub Kicinski */ 50d9ae7f2bSJakub Kicinski #define nfp_for_each_insn_walk2(nfp_prog, pos, next) \ 51d9ae7f2bSJakub Kicinski for (pos = list_first_entry(&(nfp_prog)->insns, typeof(*pos), l), \ 52d9ae7f2bSJakub Kicinski next = list_next_entry(pos, l); \ 53d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &pos->l && \ 54d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &next->l; \ 55d9ae7f2bSJakub Kicinski pos = nfp_meta_next(pos), \ 56d9ae7f2bSJakub Kicinski next = nfp_meta_next(pos)) 57d9ae7f2bSJakub Kicinski 58d9ae7f2bSJakub Kicinski #define nfp_for_each_insn_walk3(nfp_prog, pos, next, next2) \ 59d9ae7f2bSJakub Kicinski for (pos = list_first_entry(&(nfp_prog)->insns, typeof(*pos), l), \ 60d9ae7f2bSJakub Kicinski next = list_next_entry(pos, l), \ 61d9ae7f2bSJakub Kicinski next2 = list_next_entry(next, l); \ 62d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &pos->l && \ 63d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &next->l && \ 64d9ae7f2bSJakub Kicinski &(nfp_prog)->insns != &next2->l; \ 65d9ae7f2bSJakub Kicinski pos = nfp_meta_next(pos), \ 66d9ae7f2bSJakub Kicinski next = nfp_meta_next(pos), \ 67d9ae7f2bSJakub Kicinski next2 = nfp_meta_next(next)) 68d9ae7f2bSJakub Kicinski 69d9ae7f2bSJakub Kicinski static bool 70d9ae7f2bSJakub Kicinski nfp_meta_has_prev(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 71d9ae7f2bSJakub Kicinski { 72d9ae7f2bSJakub Kicinski return meta->l.prev != &nfp_prog->insns; 73d9ae7f2bSJakub Kicinski } 74d9ae7f2bSJakub Kicinski 75d9ae7f2bSJakub Kicinski static void nfp_prog_push(struct nfp_prog *nfp_prog, u64 insn) 76d9ae7f2bSJakub Kicinski { 77d9ae7f2bSJakub Kicinski if (nfp_prog->__prog_alloc_len == nfp_prog->prog_len) { 78d9ae7f2bSJakub Kicinski nfp_prog->error = -ENOSPC; 79d9ae7f2bSJakub Kicinski return; 80d9ae7f2bSJakub Kicinski } 81d9ae7f2bSJakub Kicinski 82d9ae7f2bSJakub Kicinski nfp_prog->prog[nfp_prog->prog_len] = insn; 83d9ae7f2bSJakub Kicinski nfp_prog->prog_len++; 84d9ae7f2bSJakub Kicinski } 85d9ae7f2bSJakub Kicinski 86d9ae7f2bSJakub Kicinski static unsigned int nfp_prog_current_offset(struct nfp_prog *nfp_prog) 87d9ae7f2bSJakub Kicinski { 88d9ae7f2bSJakub Kicinski return nfp_prog->start_off + nfp_prog->prog_len; 89d9ae7f2bSJakub Kicinski } 90d9ae7f2bSJakub Kicinski 910d49eaf4SJakub Kicinski static bool 920d49eaf4SJakub Kicinski nfp_prog_confirm_current_offset(struct nfp_prog *nfp_prog, unsigned int off) 930d49eaf4SJakub Kicinski { 940d49eaf4SJakub Kicinski /* If there is a recorded error we may have dropped instructions; 950d49eaf4SJakub Kicinski * that doesn't have to be due to translator bug, and the translation 960d49eaf4SJakub Kicinski * will fail anyway, so just return OK. 970d49eaf4SJakub Kicinski */ 980d49eaf4SJakub Kicinski if (nfp_prog->error) 990d49eaf4SJakub Kicinski return true; 1000d49eaf4SJakub Kicinski return !WARN_ON_ONCE(nfp_prog_current_offset(nfp_prog) != off); 1010d49eaf4SJakub Kicinski } 1020d49eaf4SJakub Kicinski 103d9ae7f2bSJakub Kicinski static unsigned int 104d9ae7f2bSJakub Kicinski nfp_prog_offset_to_index(struct nfp_prog *nfp_prog, unsigned int offset) 105d9ae7f2bSJakub Kicinski { 106d9ae7f2bSJakub Kicinski return offset - nfp_prog->start_off; 107d9ae7f2bSJakub Kicinski } 108d9ae7f2bSJakub Kicinski 109d9ae7f2bSJakub Kicinski /* --- Emitters --- */ 110d9ae7f2bSJakub Kicinski static void 111d9ae7f2bSJakub Kicinski __emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, 1125468a8b9SJakub Kicinski u8 mode, u8 xfer, u8 areg, u8 breg, u8 size, bool sync, bool indir) 113d9ae7f2bSJakub Kicinski { 114d9ae7f2bSJakub Kicinski enum cmd_ctx_swap ctx; 115d9ae7f2bSJakub Kicinski u64 insn; 116d9ae7f2bSJakub Kicinski 117d9ae7f2bSJakub Kicinski if (sync) 118d9ae7f2bSJakub Kicinski ctx = CMD_CTX_SWAP; 119d9ae7f2bSJakub Kicinski else 120d9ae7f2bSJakub Kicinski ctx = CMD_CTX_NO_SWAP; 121d9ae7f2bSJakub Kicinski 122d9ae7f2bSJakub Kicinski insn = FIELD_PREP(OP_CMD_A_SRC, areg) | 123d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_CTX, ctx) | 124d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_B_SRC, breg) | 125d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_TOKEN, cmd_tgt_act[op].token) | 126d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_XFER, xfer) | 127d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_CNT, size) | 128d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_SIG, sync) | 129d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_TGT_CMD, cmd_tgt_act[op].tgt_cmd) | 1305468a8b9SJakub Kicinski FIELD_PREP(OP_CMD_INDIR, indir) | 131d9ae7f2bSJakub Kicinski FIELD_PREP(OP_CMD_MODE, mode); 132d9ae7f2bSJakub Kicinski 133d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 134d9ae7f2bSJakub Kicinski } 135d9ae7f2bSJakub Kicinski 136d9ae7f2bSJakub Kicinski static void 1375468a8b9SJakub Kicinski emit_cmd_any(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer, 1385468a8b9SJakub Kicinski swreg lreg, swreg rreg, u8 size, bool sync, bool indir) 139d9ae7f2bSJakub Kicinski { 140d9ae7f2bSJakub Kicinski struct nfp_insn_re_regs reg; 141d9ae7f2bSJakub Kicinski int err; 142d9ae7f2bSJakub Kicinski 143d9ae7f2bSJakub Kicinski err = swreg_to_restricted(reg_none(), lreg, rreg, ®, false); 144d9ae7f2bSJakub Kicinski if (err) { 145d9ae7f2bSJakub Kicinski nfp_prog->error = err; 146d9ae7f2bSJakub Kicinski return; 147d9ae7f2bSJakub Kicinski } 148d9ae7f2bSJakub Kicinski if (reg.swap) { 149d9ae7f2bSJakub Kicinski pr_err("cmd can't swap arguments\n"); 150d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 151d9ae7f2bSJakub Kicinski return; 152d9ae7f2bSJakub Kicinski } 153995e101fSJakub Kicinski if (reg.dst_lmextn || reg.src_lmextn) { 154995e101fSJakub Kicinski pr_err("cmd can't use LMextn\n"); 155995e101fSJakub Kicinski nfp_prog->error = -EFAULT; 156995e101fSJakub Kicinski return; 157995e101fSJakub Kicinski } 158d9ae7f2bSJakub Kicinski 1595468a8b9SJakub Kicinski __emit_cmd(nfp_prog, op, mode, xfer, reg.areg, reg.breg, size, sync, 1605468a8b9SJakub Kicinski indir); 1615468a8b9SJakub Kicinski } 1625468a8b9SJakub Kicinski 1635468a8b9SJakub Kicinski static void 1645468a8b9SJakub Kicinski emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer, 1655468a8b9SJakub Kicinski swreg lreg, swreg rreg, u8 size, bool sync) 1665468a8b9SJakub Kicinski { 1675468a8b9SJakub Kicinski emit_cmd_any(nfp_prog, op, mode, xfer, lreg, rreg, size, sync, false); 168d9ae7f2bSJakub Kicinski } 169d9ae7f2bSJakub Kicinski 170d9ae7f2bSJakub Kicinski static void 1719879a381SJiong Wang emit_cmd_indir(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer, 1729879a381SJiong Wang swreg lreg, swreg rreg, u8 size, bool sync) 1739879a381SJiong Wang { 1749879a381SJiong Wang emit_cmd_any(nfp_prog, op, mode, xfer, lreg, rreg, size, sync, true); 1759879a381SJiong Wang } 1769879a381SJiong Wang 1779879a381SJiong Wang static void 178d9ae7f2bSJakub Kicinski __emit_br(struct nfp_prog *nfp_prog, enum br_mask mask, enum br_ev_pip ev_pip, 179d9ae7f2bSJakub Kicinski enum br_ctx_signal_state css, u16 addr, u8 defer) 180d9ae7f2bSJakub Kicinski { 181d9ae7f2bSJakub Kicinski u16 addr_lo, addr_hi; 182d9ae7f2bSJakub Kicinski u64 insn; 183d9ae7f2bSJakub Kicinski 184d9ae7f2bSJakub Kicinski addr_lo = addr & (OP_BR_ADDR_LO >> __bf_shf(OP_BR_ADDR_LO)); 185d9ae7f2bSJakub Kicinski addr_hi = addr != addr_lo; 186d9ae7f2bSJakub Kicinski 187d9ae7f2bSJakub Kicinski insn = OP_BR_BASE | 188d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_MASK, mask) | 189d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_EV_PIP, ev_pip) | 190d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_CSS, css) | 191d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_DEFBR, defer) | 192d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_ADDR_LO, addr_lo) | 193d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_ADDR_HI, addr_hi); 194d9ae7f2bSJakub Kicinski 195d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 196d9ae7f2bSJakub Kicinski } 197d9ae7f2bSJakub Kicinski 198d9ae7f2bSJakub Kicinski static void emit_br_def(struct nfp_prog *nfp_prog, u16 addr, u8 defer) 199d9ae7f2bSJakub Kicinski { 200d9ae7f2bSJakub Kicinski if (defer > 2) { 201d9ae7f2bSJakub Kicinski pr_err("BUG: branch defer out of bounds %d\n", defer); 202d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 203d9ae7f2bSJakub Kicinski return; 204d9ae7f2bSJakub Kicinski } 205d9ae7f2bSJakub Kicinski __emit_br(nfp_prog, BR_UNC, BR_EV_PIP_UNCOND, BR_CSS_NONE, addr, defer); 206d9ae7f2bSJakub Kicinski } 207d9ae7f2bSJakub Kicinski 208d9ae7f2bSJakub Kicinski static void 209d9ae7f2bSJakub Kicinski emit_br(struct nfp_prog *nfp_prog, enum br_mask mask, u16 addr, u8 defer) 210d9ae7f2bSJakub Kicinski { 211d9ae7f2bSJakub Kicinski __emit_br(nfp_prog, mask, 212d9ae7f2bSJakub Kicinski mask != BR_UNC ? BR_EV_PIP_COND : BR_EV_PIP_UNCOND, 213d9ae7f2bSJakub Kicinski BR_CSS_NONE, addr, defer); 214d9ae7f2bSJakub Kicinski } 215d9ae7f2bSJakub Kicinski 216d9ae7f2bSJakub Kicinski static void 217d9ae7f2bSJakub Kicinski __emit_immed(struct nfp_prog *nfp_prog, u16 areg, u16 breg, u16 imm_hi, 218d9ae7f2bSJakub Kicinski enum immed_width width, bool invert, 219995e101fSJakub Kicinski enum immed_shift shift, bool wr_both, 220995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 221d9ae7f2bSJakub Kicinski { 222d9ae7f2bSJakub Kicinski u64 insn; 223d9ae7f2bSJakub Kicinski 224d9ae7f2bSJakub Kicinski insn = OP_IMMED_BASE | 225d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_A_SRC, areg) | 226d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_B_SRC, breg) | 227d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_IMM, imm_hi) | 228d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_WIDTH, width) | 229d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_INV, invert) | 230d9ae7f2bSJakub Kicinski FIELD_PREP(OP_IMMED_SHIFT, shift) | 231995e101fSJakub Kicinski FIELD_PREP(OP_IMMED_WR_AB, wr_both) | 232995e101fSJakub Kicinski FIELD_PREP(OP_IMMED_SRC_LMEXTN, src_lmextn) | 233995e101fSJakub Kicinski FIELD_PREP(OP_IMMED_DST_LMEXTN, dst_lmextn); 234d9ae7f2bSJakub Kicinski 235d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 236d9ae7f2bSJakub Kicinski } 237d9ae7f2bSJakub Kicinski 238d9ae7f2bSJakub Kicinski static void 239b3f868dfSJakub Kicinski emit_immed(struct nfp_prog *nfp_prog, swreg dst, u16 imm, 240d9ae7f2bSJakub Kicinski enum immed_width width, bool invert, enum immed_shift shift) 241d9ae7f2bSJakub Kicinski { 242d9ae7f2bSJakub Kicinski struct nfp_insn_ur_regs reg; 243d9ae7f2bSJakub Kicinski int err; 244d9ae7f2bSJakub Kicinski 245b3f868dfSJakub Kicinski if (swreg_type(dst) == NN_REG_IMM) { 246d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 247d9ae7f2bSJakub Kicinski return; 248d9ae7f2bSJakub Kicinski } 249d9ae7f2bSJakub Kicinski 250d9ae7f2bSJakub Kicinski err = swreg_to_unrestricted(dst, dst, reg_imm(imm & 0xff), ®); 251d9ae7f2bSJakub Kicinski if (err) { 252d9ae7f2bSJakub Kicinski nfp_prog->error = err; 253d9ae7f2bSJakub Kicinski return; 254d9ae7f2bSJakub Kicinski } 255d9ae7f2bSJakub Kicinski 2563239e7bbSJiong Wang /* Use reg.dst when destination is No-Dest. */ 2573239e7bbSJiong Wang __emit_immed(nfp_prog, 2583239e7bbSJiong Wang swreg_type(dst) == NN_REG_NONE ? reg.dst : reg.areg, 2593239e7bbSJiong Wang reg.breg, imm >> 8, width, invert, shift, 2603239e7bbSJiong Wang reg.wr_both, reg.dst_lmextn, reg.src_lmextn); 261d9ae7f2bSJakub Kicinski } 262d9ae7f2bSJakub Kicinski 263d9ae7f2bSJakub Kicinski static void 264d9ae7f2bSJakub Kicinski __emit_shf(struct nfp_prog *nfp_prog, u16 dst, enum alu_dst_ab dst_ab, 265d9ae7f2bSJakub Kicinski enum shf_sc sc, u8 shift, 266995e101fSJakub Kicinski u16 areg, enum shf_op op, u16 breg, bool i8, bool sw, bool wr_both, 267995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 268d9ae7f2bSJakub Kicinski { 269d9ae7f2bSJakub Kicinski u64 insn; 270d9ae7f2bSJakub Kicinski 271d9ae7f2bSJakub Kicinski if (!FIELD_FIT(OP_SHF_SHIFT, shift)) { 272d9ae7f2bSJakub Kicinski nfp_prog->error = -EFAULT; 273d9ae7f2bSJakub Kicinski return; 274d9ae7f2bSJakub Kicinski } 275d9ae7f2bSJakub Kicinski 276d9ae7f2bSJakub Kicinski if (sc == SHF_SC_L_SHF) 277d9ae7f2bSJakub Kicinski shift = 32 - shift; 278d9ae7f2bSJakub Kicinski 279d9ae7f2bSJakub Kicinski insn = OP_SHF_BASE | 280d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_A_SRC, areg) | 281d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_SC, sc) | 282d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_B_SRC, breg) | 283d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_I8, i8) | 284d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_SW, sw) | 285d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_DST, dst) | 286d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_SHIFT, shift) | 287d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_OP, op) | 288d9ae7f2bSJakub Kicinski FIELD_PREP(OP_SHF_DST_AB, dst_ab) | 289995e101fSJakub Kicinski FIELD_PREP(OP_SHF_WR_AB, wr_both) | 290995e101fSJakub Kicinski FIELD_PREP(OP_SHF_SRC_LMEXTN, src_lmextn) | 291995e101fSJakub Kicinski FIELD_PREP(OP_SHF_DST_LMEXTN, dst_lmextn); 292d9ae7f2bSJakub Kicinski 293d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 294d9ae7f2bSJakub Kicinski } 295d9ae7f2bSJakub Kicinski 296d9ae7f2bSJakub Kicinski static void 297b3f868dfSJakub Kicinski emit_shf(struct nfp_prog *nfp_prog, swreg dst, 298b3f868dfSJakub Kicinski swreg lreg, enum shf_op op, swreg rreg, enum shf_sc sc, u8 shift) 299d9ae7f2bSJakub Kicinski { 300d9ae7f2bSJakub Kicinski struct nfp_insn_re_regs reg; 301d9ae7f2bSJakub Kicinski int err; 302d9ae7f2bSJakub Kicinski 303d9ae7f2bSJakub Kicinski err = swreg_to_restricted(dst, lreg, rreg, ®, true); 304d9ae7f2bSJakub Kicinski if (err) { 305d9ae7f2bSJakub Kicinski nfp_prog->error = err; 306d9ae7f2bSJakub Kicinski return; 307d9ae7f2bSJakub Kicinski } 308d9ae7f2bSJakub Kicinski 309d9ae7f2bSJakub Kicinski __emit_shf(nfp_prog, reg.dst, reg.dst_ab, sc, shift, 310995e101fSJakub Kicinski reg.areg, op, reg.breg, reg.i8, reg.swap, reg.wr_both, 311995e101fSJakub Kicinski reg.dst_lmextn, reg.src_lmextn); 312d9ae7f2bSJakub Kicinski } 313d9ae7f2bSJakub Kicinski 314d9ae7f2bSJakub Kicinski static void 315d9ae7f2bSJakub Kicinski __emit_alu(struct nfp_prog *nfp_prog, u16 dst, enum alu_dst_ab dst_ab, 316995e101fSJakub Kicinski u16 areg, enum alu_op op, u16 breg, bool swap, bool wr_both, 317995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 318d9ae7f2bSJakub Kicinski { 319d9ae7f2bSJakub Kicinski u64 insn; 320d9ae7f2bSJakub Kicinski 321d9ae7f2bSJakub Kicinski insn = OP_ALU_BASE | 322d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_A_SRC, areg) | 323d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_B_SRC, breg) | 324d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_DST, dst) | 325d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_SW, swap) | 326d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_OP, op) | 327d9ae7f2bSJakub Kicinski FIELD_PREP(OP_ALU_DST_AB, dst_ab) | 328995e101fSJakub Kicinski FIELD_PREP(OP_ALU_WR_AB, wr_both) | 329995e101fSJakub Kicinski FIELD_PREP(OP_ALU_SRC_LMEXTN, src_lmextn) | 330995e101fSJakub Kicinski FIELD_PREP(OP_ALU_DST_LMEXTN, dst_lmextn); 331d9ae7f2bSJakub Kicinski 332d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 333d9ae7f2bSJakub Kicinski } 334d9ae7f2bSJakub Kicinski 335d9ae7f2bSJakub Kicinski static void 336b3f868dfSJakub Kicinski emit_alu(struct nfp_prog *nfp_prog, swreg dst, 337b3f868dfSJakub Kicinski swreg lreg, enum alu_op op, swreg rreg) 338d9ae7f2bSJakub Kicinski { 339d9ae7f2bSJakub Kicinski struct nfp_insn_ur_regs reg; 340d9ae7f2bSJakub Kicinski int err; 341d9ae7f2bSJakub Kicinski 342d9ae7f2bSJakub Kicinski err = swreg_to_unrestricted(dst, lreg, rreg, ®); 343d9ae7f2bSJakub Kicinski if (err) { 344d9ae7f2bSJakub Kicinski nfp_prog->error = err; 345d9ae7f2bSJakub Kicinski return; 346d9ae7f2bSJakub Kicinski } 347d9ae7f2bSJakub Kicinski 348d9ae7f2bSJakub Kicinski __emit_alu(nfp_prog, reg.dst, reg.dst_ab, 349995e101fSJakub Kicinski reg.areg, op, reg.breg, reg.swap, reg.wr_both, 350995e101fSJakub Kicinski reg.dst_lmextn, reg.src_lmextn); 351d9ae7f2bSJakub Kicinski } 352d9ae7f2bSJakub Kicinski 353d9ae7f2bSJakub Kicinski static void 354d9ae7f2bSJakub Kicinski __emit_ld_field(struct nfp_prog *nfp_prog, enum shf_sc sc, 355d9ae7f2bSJakub Kicinski u8 areg, u8 bmask, u8 breg, u8 shift, bool imm8, 356995e101fSJakub Kicinski bool zero, bool swap, bool wr_both, 357995e101fSJakub Kicinski bool dst_lmextn, bool src_lmextn) 358d9ae7f2bSJakub Kicinski { 359d9ae7f2bSJakub Kicinski u64 insn; 360d9ae7f2bSJakub Kicinski 361d9ae7f2bSJakub Kicinski insn = OP_LDF_BASE | 362d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_A_SRC, areg) | 363d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_SC, sc) | 364d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_B_SRC, breg) | 365d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_I8, imm8) | 366d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_SW, swap) | 367d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_ZF, zero) | 368d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_BMASK, bmask) | 369d9ae7f2bSJakub Kicinski FIELD_PREP(OP_LDF_SHF, shift) | 370995e101fSJakub Kicinski FIELD_PREP(OP_LDF_WR_AB, wr_both) | 371995e101fSJakub Kicinski FIELD_PREP(OP_LDF_SRC_LMEXTN, src_lmextn) | 372995e101fSJakub Kicinski FIELD_PREP(OP_LDF_DST_LMEXTN, dst_lmextn); 373d9ae7f2bSJakub Kicinski 374d9ae7f2bSJakub Kicinski nfp_prog_push(nfp_prog, insn); 375d9ae7f2bSJakub Kicinski } 376d9ae7f2bSJakub Kicinski 377d9ae7f2bSJakub Kicinski static void 378bc8c80a8SJakub Kicinski emit_ld_field_any(struct nfp_prog *nfp_prog, swreg dst, u8 bmask, swreg src, 379bc8c80a8SJakub Kicinski enum shf_sc sc, u8 shift, bool zero) 380d9ae7f2bSJakub Kicinski { 381d9ae7f2bSJakub Kicinski struct nfp_insn_re_regs reg; 382d9ae7f2bSJakub Kicinski int err; 383d9ae7f2bSJakub Kicinski 3842de1be1dSJakub Kicinski /* Note: ld_field is special as it uses one of the src regs as dst */ 3852de1be1dSJakub Kicinski err = swreg_to_restricted(dst, dst, src, ®, true); 386d9ae7f2bSJakub Kicinski if (err) { 387d9ae7f2bSJakub Kicinski nfp_prog->error = err; 388d9ae7f2bSJakub Kicinski return; 389d9ae7f2bSJakub Kicinski } 390d9ae7f2bSJakub Kicinski 391d9ae7f2bSJakub Kicinski __emit_ld_field(nfp_prog, sc, reg.areg, bmask, reg.breg, shift, 392995e101fSJakub Kicinski reg.i8, zero, reg.swap, reg.wr_both, 393995e101fSJakub Kicinski reg.dst_lmextn, reg.src_lmextn); 394d9ae7f2bSJakub Kicinski } 395d9ae7f2bSJakub Kicinski 396d9ae7f2bSJakub Kicinski static void 397b3f868dfSJakub Kicinski emit_ld_field(struct nfp_prog *nfp_prog, swreg dst, u8 bmask, swreg src, 398d9ae7f2bSJakub Kicinski enum shf_sc sc, u8 shift) 399d9ae7f2bSJakub Kicinski { 400bc8c80a8SJakub Kicinski emit_ld_field_any(nfp_prog, dst, bmask, src, sc, shift, false); 401d9ae7f2bSJakub Kicinski } 402d9ae7f2bSJakub Kicinski 4032df03a50SJakub Kicinski static void 4042df03a50SJakub Kicinski __emit_lcsr(struct nfp_prog *nfp_prog, u16 areg, u16 breg, bool wr, u16 addr, 4052df03a50SJakub Kicinski bool dst_lmextn, bool src_lmextn) 4062df03a50SJakub Kicinski { 4072df03a50SJakub Kicinski u64 insn; 4082df03a50SJakub Kicinski 4092df03a50SJakub Kicinski insn = OP_LCSR_BASE | 4102df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_A_SRC, areg) | 4112df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_B_SRC, breg) | 4122df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_WRITE, wr) | 4132df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_ADDR, addr) | 4142df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_SRC_LMEXTN, src_lmextn) | 4152df03a50SJakub Kicinski FIELD_PREP(OP_LCSR_DST_LMEXTN, dst_lmextn); 4162df03a50SJakub Kicinski 4172df03a50SJakub Kicinski nfp_prog_push(nfp_prog, insn); 4182df03a50SJakub Kicinski } 4192df03a50SJakub Kicinski 4202df03a50SJakub Kicinski static void emit_csr_wr(struct nfp_prog *nfp_prog, swreg src, u16 addr) 4212df03a50SJakub Kicinski { 4222df03a50SJakub Kicinski struct nfp_insn_ur_regs reg; 4232df03a50SJakub Kicinski int err; 4242df03a50SJakub Kicinski 4252df03a50SJakub Kicinski /* This instruction takes immeds instead of reg_none() for the ignored 4262df03a50SJakub Kicinski * operand, but we can't encode 2 immeds in one instr with our normal 4272df03a50SJakub Kicinski * swreg infra so if param is an immed, we encode as reg_none() and 4282df03a50SJakub Kicinski * copy the immed to both operands. 4292df03a50SJakub Kicinski */ 4302df03a50SJakub Kicinski if (swreg_type(src) == NN_REG_IMM) { 4312df03a50SJakub Kicinski err = swreg_to_unrestricted(reg_none(), src, reg_none(), ®); 4322df03a50SJakub Kicinski reg.breg = reg.areg; 4332df03a50SJakub Kicinski } else { 4342df03a50SJakub Kicinski err = swreg_to_unrestricted(reg_none(), src, reg_imm(0), ®); 4352df03a50SJakub Kicinski } 4362df03a50SJakub Kicinski if (err) { 4372df03a50SJakub Kicinski nfp_prog->error = err; 4382df03a50SJakub Kicinski return; 4392df03a50SJakub Kicinski } 4402df03a50SJakub Kicinski 4412df03a50SJakub Kicinski __emit_lcsr(nfp_prog, reg.areg, reg.breg, true, addr / 4, 4422df03a50SJakub Kicinski false, reg.src_lmextn); 4432df03a50SJakub Kicinski } 4442df03a50SJakub Kicinski 4451c03e03fSJakub Kicinski static void emit_nop(struct nfp_prog *nfp_prog) 4461c03e03fSJakub Kicinski { 4471c03e03fSJakub Kicinski __emit_immed(nfp_prog, UR_REG_IMM, UR_REG_IMM, 0, 0, 0, 0, 0, 0, 0); 4481c03e03fSJakub Kicinski } 4491c03e03fSJakub Kicinski 450d9ae7f2bSJakub Kicinski /* --- Wrappers --- */ 451d9ae7f2bSJakub Kicinski static bool pack_immed(u32 imm, u16 *val, enum immed_shift *shift) 452d9ae7f2bSJakub Kicinski { 453d9ae7f2bSJakub Kicinski if (!(imm & 0xffff0000)) { 454d9ae7f2bSJakub Kicinski *val = imm; 455d9ae7f2bSJakub Kicinski *shift = IMMED_SHIFT_0B; 456d9ae7f2bSJakub Kicinski } else if (!(imm & 0xff0000ff)) { 457d9ae7f2bSJakub Kicinski *val = imm >> 8; 458d9ae7f2bSJakub Kicinski *shift = IMMED_SHIFT_1B; 459d9ae7f2bSJakub Kicinski } else if (!(imm & 0x0000ffff)) { 460d9ae7f2bSJakub Kicinski *val = imm >> 16; 461d9ae7f2bSJakub Kicinski *shift = IMMED_SHIFT_2B; 462d9ae7f2bSJakub Kicinski } else { 463d9ae7f2bSJakub Kicinski return false; 464d9ae7f2bSJakub Kicinski } 465d9ae7f2bSJakub Kicinski 466d9ae7f2bSJakub Kicinski return true; 467d9ae7f2bSJakub Kicinski } 468d9ae7f2bSJakub Kicinski 469b3f868dfSJakub Kicinski static void wrp_immed(struct nfp_prog *nfp_prog, swreg dst, u32 imm) 470d9ae7f2bSJakub Kicinski { 471d9ae7f2bSJakub Kicinski enum immed_shift shift; 472d9ae7f2bSJakub Kicinski u16 val; 473d9ae7f2bSJakub Kicinski 474d9ae7f2bSJakub Kicinski if (pack_immed(imm, &val, &shift)) { 475d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, val, IMMED_WIDTH_ALL, false, shift); 476d9ae7f2bSJakub Kicinski } else if (pack_immed(~imm, &val, &shift)) { 477d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, val, IMMED_WIDTH_ALL, true, shift); 478d9ae7f2bSJakub Kicinski } else { 479d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, imm & 0xffff, IMMED_WIDTH_ALL, 480d9ae7f2bSJakub Kicinski false, IMMED_SHIFT_0B); 481d9ae7f2bSJakub Kicinski emit_immed(nfp_prog, dst, imm >> 16, IMMED_WIDTH_WORD, 482d9ae7f2bSJakub Kicinski false, IMMED_SHIFT_2B); 483d9ae7f2bSJakub Kicinski } 484d9ae7f2bSJakub Kicinski } 485d9ae7f2bSJakub Kicinski 486d9ae7f2bSJakub Kicinski /* ur_load_imm_any() - encode immediate or use tmp register (unrestricted) 487d9ae7f2bSJakub Kicinski * If the @imm is small enough encode it directly in operand and return 488d9ae7f2bSJakub Kicinski * otherwise load @imm to a spare register and return its encoding. 489d9ae7f2bSJakub Kicinski */ 490b3f868dfSJakub Kicinski static swreg ur_load_imm_any(struct nfp_prog *nfp_prog, u32 imm, swreg tmp_reg) 491d9ae7f2bSJakub Kicinski { 492d9ae7f2bSJakub Kicinski if (FIELD_FIT(UR_REG_IMM_MAX, imm)) 493d9ae7f2bSJakub Kicinski return reg_imm(imm); 494d9ae7f2bSJakub Kicinski 495d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, tmp_reg, imm); 496d9ae7f2bSJakub Kicinski return tmp_reg; 497d9ae7f2bSJakub Kicinski } 498d9ae7f2bSJakub Kicinski 499d9ae7f2bSJakub Kicinski /* re_load_imm_any() - encode immediate or use tmp register (restricted) 500d9ae7f2bSJakub Kicinski * If the @imm is small enough encode it directly in operand and return 501d9ae7f2bSJakub Kicinski * otherwise load @imm to a spare register and return its encoding. 502d9ae7f2bSJakub Kicinski */ 503b3f868dfSJakub Kicinski static swreg re_load_imm_any(struct nfp_prog *nfp_prog, u32 imm, swreg tmp_reg) 504d9ae7f2bSJakub Kicinski { 505d9ae7f2bSJakub Kicinski if (FIELD_FIT(RE_REG_IMM_MAX, imm)) 506d9ae7f2bSJakub Kicinski return reg_imm(imm); 507d9ae7f2bSJakub Kicinski 508d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, tmp_reg, imm); 509d9ae7f2bSJakub Kicinski return tmp_reg; 510d9ae7f2bSJakub Kicinski } 511d9ae7f2bSJakub Kicinski 512ff42bb9fSJakub Kicinski static void wrp_nops(struct nfp_prog *nfp_prog, unsigned int count) 513ff42bb9fSJakub Kicinski { 514ff42bb9fSJakub Kicinski while (count--) 515ff42bb9fSJakub Kicinski emit_nop(nfp_prog); 516ff42bb9fSJakub Kicinski } 517ff42bb9fSJakub Kicinski 518d9ae7f2bSJakub Kicinski static void 519d9ae7f2bSJakub Kicinski wrp_br_special(struct nfp_prog *nfp_prog, enum br_mask mask, 520d9ae7f2bSJakub Kicinski enum br_special special) 521d9ae7f2bSJakub Kicinski { 522d9ae7f2bSJakub Kicinski emit_br(nfp_prog, mask, 0, 0); 523d9ae7f2bSJakub Kicinski 524d9ae7f2bSJakub Kicinski nfp_prog->prog[nfp_prog->prog_len - 1] |= 525d9ae7f2bSJakub Kicinski FIELD_PREP(OP_BR_SPECIAL, special); 526d9ae7f2bSJakub Kicinski } 527d9ae7f2bSJakub Kicinski 528c000dfb5SJakub Kicinski static void wrp_mov(struct nfp_prog *nfp_prog, swreg dst, swreg src) 529c000dfb5SJakub Kicinski { 530c000dfb5SJakub Kicinski emit_alu(nfp_prog, dst, reg_none(), ALU_OP_NONE, src); 531c000dfb5SJakub Kicinski } 532c000dfb5SJakub Kicinski 533d9ae7f2bSJakub Kicinski static void wrp_reg_mov(struct nfp_prog *nfp_prog, u16 dst, u16 src) 534d9ae7f2bSJakub Kicinski { 535c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst), reg_b(src)); 536d9ae7f2bSJakub Kicinski } 537d9ae7f2bSJakub Kicinski 5389879a381SJiong Wang /* wrp_reg_subpart() - load @field_len bytes from @offset of @src, write the 5399879a381SJiong Wang * result to @dst from low end. 5409879a381SJiong Wang */ 5419879a381SJiong Wang static void 5429879a381SJiong Wang wrp_reg_subpart(struct nfp_prog *nfp_prog, swreg dst, swreg src, u8 field_len, 5439879a381SJiong Wang u8 offset) 5449879a381SJiong Wang { 5459879a381SJiong Wang enum shf_sc sc = offset ? SHF_SC_R_SHF : SHF_SC_NONE; 5469879a381SJiong Wang u8 mask = (1 << field_len) - 1; 5479879a381SJiong Wang 5489879a381SJiong Wang emit_ld_field_any(nfp_prog, dst, mask, src, sc, offset * 8, true); 5499879a381SJiong Wang } 5509879a381SJiong Wang 5519879a381SJiong Wang /* NFP has Command Push Pull bus which supports bluk memory operations. */ 5529879a381SJiong Wang static int nfp_cpp_memcpy(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 5539879a381SJiong Wang { 5549879a381SJiong Wang bool descending_seq = meta->ldst_gather_len < 0; 5559879a381SJiong Wang s16 len = abs(meta->ldst_gather_len); 5569879a381SJiong Wang swreg src_base, off; 5579879a381SJiong Wang unsigned int i; 5589879a381SJiong Wang u8 xfer_num; 5599879a381SJiong Wang 5609879a381SJiong Wang off = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 5619879a381SJiong Wang src_base = reg_a(meta->insn.src_reg * 2); 5629879a381SJiong Wang xfer_num = round_up(len, 4) / 4; 5639879a381SJiong Wang 5648c900538SJiong Wang /* Setup PREV_ALU fields to override memory read length. */ 5658c900538SJiong Wang if (len > 32) 5668c900538SJiong Wang wrp_immed(nfp_prog, reg_none(), 5678c900538SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, xfer_num - 1)); 5688c900538SJiong Wang 5699879a381SJiong Wang /* Memory read from source addr into transfer-in registers. */ 5708c900538SJiong Wang emit_cmd_any(nfp_prog, CMD_TGT_READ32_SWAP, CMD_MODE_32b, 0, src_base, 5718c900538SJiong Wang off, xfer_num - 1, true, len > 32); 5729879a381SJiong Wang 5739879a381SJiong Wang /* Move from transfer-in to transfer-out. */ 5749879a381SJiong Wang for (i = 0; i < xfer_num; i++) 5759879a381SJiong Wang wrp_mov(nfp_prog, reg_xfer(i), reg_xfer(i)); 5769879a381SJiong Wang 5779879a381SJiong Wang off = re_load_imm_any(nfp_prog, meta->paired_st->off, imm_b(nfp_prog)); 5789879a381SJiong Wang 5799879a381SJiong Wang if (len <= 8) { 5809879a381SJiong Wang /* Use single direct_ref write8. */ 5819879a381SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 5829879a381SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, len - 1, 5839879a381SJiong Wang true); 5848c900538SJiong Wang } else if (len <= 32 && IS_ALIGNED(len, 4)) { 5859879a381SJiong Wang /* Use single direct_ref write32. */ 5869879a381SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 5879879a381SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, xfer_num - 1, 5889879a381SJiong Wang true); 5898c900538SJiong Wang } else if (len <= 32) { 5909879a381SJiong Wang /* Use single indirect_ref write8. */ 5919879a381SJiong Wang wrp_immed(nfp_prog, reg_none(), 5929879a381SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, len - 1)); 5939879a381SJiong Wang emit_cmd_indir(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 5949879a381SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 5959879a381SJiong Wang len - 1, true); 5968c900538SJiong Wang } else if (IS_ALIGNED(len, 4)) { 5978c900538SJiong Wang /* Use single indirect_ref write32. */ 5988c900538SJiong Wang wrp_immed(nfp_prog, reg_none(), 5998c900538SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, xfer_num - 1)); 6008c900538SJiong Wang emit_cmd_indir(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 6018c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 6028c900538SJiong Wang xfer_num - 1, true); 6038c900538SJiong Wang } else if (len <= 40) { 6048c900538SJiong Wang /* Use one direct_ref write32 to write the first 32-bytes, then 6058c900538SJiong Wang * another direct_ref write8 to write the remaining bytes. 6068c900538SJiong Wang */ 6078c900538SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 6088c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 7, 6098c900538SJiong Wang true); 6108c900538SJiong Wang 6118c900538SJiong Wang off = re_load_imm_any(nfp_prog, meta->paired_st->off + 32, 6128c900538SJiong Wang imm_b(nfp_prog)); 6138c900538SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 8, 6148c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, len - 33, 6158c900538SJiong Wang true); 6168c900538SJiong Wang } else { 6178c900538SJiong Wang /* Use one indirect_ref write32 to write 4-bytes aligned length, 6188c900538SJiong Wang * then another direct_ref write8 to write the remaining bytes. 6198c900538SJiong Wang */ 6208c900538SJiong Wang u8 new_off; 6218c900538SJiong Wang 6228c900538SJiong Wang wrp_immed(nfp_prog, reg_none(), 6238c900538SJiong Wang CMD_OVE_LEN | FIELD_PREP(CMD_OV_LEN, xfer_num - 2)); 6248c900538SJiong Wang emit_cmd_indir(nfp_prog, CMD_TGT_WRITE32_SWAP, CMD_MODE_32b, 0, 6258c900538SJiong Wang reg_a(meta->paired_st->dst_reg * 2), off, 6268c900538SJiong Wang xfer_num - 2, true); 6278c900538SJiong Wang new_off = meta->paired_st->off + (xfer_num - 1) * 4; 6288c900538SJiong Wang off = re_load_imm_any(nfp_prog, new_off, imm_b(nfp_prog)); 6298c900538SJiong Wang emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 6308c900538SJiong Wang xfer_num - 1, reg_a(meta->paired_st->dst_reg * 2), off, 6318c900538SJiong Wang (len & 0x3) - 1, true); 6329879a381SJiong Wang } 6339879a381SJiong Wang 6349879a381SJiong Wang /* TODO: The following extra load is to make sure data flow be identical 6359879a381SJiong Wang * before and after we do memory copy optimization. 6369879a381SJiong Wang * 6379879a381SJiong Wang * The load destination register is not guaranteed to be dead, so we 6389879a381SJiong Wang * need to make sure it is loaded with the value the same as before 6399879a381SJiong Wang * this transformation. 6409879a381SJiong Wang * 6419879a381SJiong Wang * These extra loads could be removed once we have accurate register 6429879a381SJiong Wang * usage information. 6439879a381SJiong Wang */ 6449879a381SJiong Wang if (descending_seq) 6459879a381SJiong Wang xfer_num = 0; 6469879a381SJiong Wang else if (BPF_SIZE(meta->insn.code) != BPF_DW) 6479879a381SJiong Wang xfer_num = xfer_num - 1; 6489879a381SJiong Wang else 6499879a381SJiong Wang xfer_num = xfer_num - 2; 6509879a381SJiong Wang 6519879a381SJiong Wang switch (BPF_SIZE(meta->insn.code)) { 6529879a381SJiong Wang case BPF_B: 6539879a381SJiong Wang wrp_reg_subpart(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6549879a381SJiong Wang reg_xfer(xfer_num), 1, 6559879a381SJiong Wang IS_ALIGNED(len, 4) ? 3 : (len & 3) - 1); 6569879a381SJiong Wang break; 6579879a381SJiong Wang case BPF_H: 6589879a381SJiong Wang wrp_reg_subpart(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6599879a381SJiong Wang reg_xfer(xfer_num), 2, (len & 3) ^ 2); 6609879a381SJiong Wang break; 6619879a381SJiong Wang case BPF_W: 6629879a381SJiong Wang wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6639879a381SJiong Wang reg_xfer(0)); 6649879a381SJiong Wang break; 6659879a381SJiong Wang case BPF_DW: 6669879a381SJiong Wang wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2), 6679879a381SJiong Wang reg_xfer(xfer_num)); 6689879a381SJiong Wang wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 6699879a381SJiong Wang reg_xfer(xfer_num + 1)); 6709879a381SJiong Wang break; 6719879a381SJiong Wang } 6729879a381SJiong Wang 6739879a381SJiong Wang if (BPF_SIZE(meta->insn.code) != BPF_DW) 6749879a381SJiong Wang wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 6759879a381SJiong Wang 6769879a381SJiong Wang return 0; 6779879a381SJiong Wang } 6789879a381SJiong Wang 679d9ae7f2bSJakub Kicinski static int 6800a793977SJakub Kicinski data_ld(struct nfp_prog *nfp_prog, swreg offset, u8 dst_gpr, int size) 681d9ae7f2bSJakub Kicinski { 682d9ae7f2bSJakub Kicinski unsigned int i; 683d9ae7f2bSJakub Kicinski u16 shift, sz; 684d9ae7f2bSJakub Kicinski 685d9ae7f2bSJakub Kicinski /* We load the value from the address indicated in @offset and then 686d9ae7f2bSJakub Kicinski * shift out the data we don't need. Note: this is big endian! 687d9ae7f2bSJakub Kicinski */ 6880a793977SJakub Kicinski sz = max(size, 4); 689d9ae7f2bSJakub Kicinski shift = size < 4 ? 4 - size : 0; 690d9ae7f2bSJakub Kicinski 6910a793977SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_READ8, CMD_MODE_32b, 0, 6920a793977SJakub Kicinski pptr_reg(nfp_prog), offset, sz - 1, true); 6930a793977SJakub Kicinski 6940a793977SJakub Kicinski i = 0; 6950a793977SJakub Kicinski if (shift) 6960a793977SJakub Kicinski emit_shf(nfp_prog, reg_both(dst_gpr), reg_none(), SHF_OP_NONE, 6970a793977SJakub Kicinski reg_xfer(0), SHF_SC_R_SHF, shift * 8); 6980a793977SJakub Kicinski else 6990a793977SJakub Kicinski for (; i * 4 < size; i++) 7000a793977SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst_gpr + i), reg_xfer(i)); 7010a793977SJakub Kicinski 7020a793977SJakub Kicinski if (i < 2) 7030a793977SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst_gpr + 1), 0); 7040a793977SJakub Kicinski 7050a793977SJakub Kicinski return 0; 7060a793977SJakub Kicinski } 7070a793977SJakub Kicinski 7080a793977SJakub Kicinski static int 7092ca71441SJakub Kicinski data_ld_host_order(struct nfp_prog *nfp_prog, u8 src_gpr, swreg offset, 7102ca71441SJakub Kicinski u8 dst_gpr, int size) 7112ca71441SJakub Kicinski { 7122ca71441SJakub Kicinski unsigned int i; 7132ca71441SJakub Kicinski u8 mask, sz; 7142ca71441SJakub Kicinski 7152ca71441SJakub Kicinski /* We load the value from the address indicated in @offset and then 7162ca71441SJakub Kicinski * mask out the data we don't need. Note: this is little endian! 7172ca71441SJakub Kicinski */ 7182ca71441SJakub Kicinski sz = max(size, 4); 7192ca71441SJakub Kicinski mask = size < 4 ? GENMASK(size - 1, 0) : 0; 7202ca71441SJakub Kicinski 7212ca71441SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_READ32_SWAP, CMD_MODE_32b, 0, 7222ca71441SJakub Kicinski reg_a(src_gpr), offset, sz / 4 - 1, true); 7232ca71441SJakub Kicinski 7242ca71441SJakub Kicinski i = 0; 7252ca71441SJakub Kicinski if (mask) 7262ca71441SJakub Kicinski emit_ld_field_any(nfp_prog, reg_both(dst_gpr), mask, 7272ca71441SJakub Kicinski reg_xfer(0), SHF_SC_NONE, 0, true); 7282ca71441SJakub Kicinski else 7292ca71441SJakub Kicinski for (; i * 4 < size; i++) 7302ca71441SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst_gpr + i), reg_xfer(i)); 7312ca71441SJakub Kicinski 7322ca71441SJakub Kicinski if (i < 2) 7332ca71441SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst_gpr + 1), 0); 7342ca71441SJakub Kicinski 7352ca71441SJakub Kicinski return 0; 7362ca71441SJakub Kicinski } 7372ca71441SJakub Kicinski 7382ca71441SJakub Kicinski static int 7390a793977SJakub Kicinski construct_data_ind_ld(struct nfp_prog *nfp_prog, u16 offset, u16 src, u8 size) 7400a793977SJakub Kicinski { 7410a793977SJakub Kicinski swreg tmp_reg; 7420a793977SJakub Kicinski 743d9ae7f2bSJakub Kicinski /* Calculate the true offset (src_reg + imm) */ 744d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, offset, imm_b(nfp_prog)); 7450a793977SJakub Kicinski emit_alu(nfp_prog, imm_both(nfp_prog), reg_a(src), ALU_OP_ADD, tmp_reg); 7460a793977SJakub Kicinski 747d9ae7f2bSJakub Kicinski /* Check packet length (size guaranteed to fit b/c it's u8) */ 748d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_a(nfp_prog), 749d9ae7f2bSJakub Kicinski imm_a(nfp_prog), ALU_OP_ADD, reg_imm(size)); 750d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 75118e53b6cSJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, imm_a(nfp_prog)); 752d9ae7f2bSJakub Kicinski wrp_br_special(nfp_prog, BR_BLO, OP_BR_GO_ABORT); 7530a793977SJakub Kicinski 754d9ae7f2bSJakub Kicinski /* Load data */ 7550a793977SJakub Kicinski return data_ld(nfp_prog, imm_b(nfp_prog), 0, size); 756d9ae7f2bSJakub Kicinski } 757d9ae7f2bSJakub Kicinski 758d9ae7f2bSJakub Kicinski static int construct_data_ld(struct nfp_prog *nfp_prog, u16 offset, u8 size) 759d9ae7f2bSJakub Kicinski { 7600a793977SJakub Kicinski swreg tmp_reg; 7610a793977SJakub Kicinski 7620a793977SJakub Kicinski /* Check packet length */ 7630a793977SJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, offset + size, imm_a(nfp_prog)); 7640a793977SJakub Kicinski emit_alu(nfp_prog, reg_none(), plen_reg(nfp_prog), ALU_OP_SUB, tmp_reg); 7650a793977SJakub Kicinski wrp_br_special(nfp_prog, BR_BLO, OP_BR_GO_ABORT); 7660a793977SJakub Kicinski 7670a793977SJakub Kicinski /* Load data */ 7680a793977SJakub Kicinski tmp_reg = re_load_imm_any(nfp_prog, offset, imm_b(nfp_prog)); 7690a793977SJakub Kicinski return data_ld(nfp_prog, tmp_reg, 0, size); 770d9ae7f2bSJakub Kicinski } 771d9ae7f2bSJakub Kicinski 772e663fe38SJakub Kicinski static int 773e663fe38SJakub Kicinski data_stx_host_order(struct nfp_prog *nfp_prog, u8 dst_gpr, swreg offset, 774e663fe38SJakub Kicinski u8 src_gpr, u8 size) 775e663fe38SJakub Kicinski { 776e663fe38SJakub Kicinski unsigned int i; 777e663fe38SJakub Kicinski 778e663fe38SJakub Kicinski for (i = 0; i * 4 < size; i++) 779e663fe38SJakub Kicinski wrp_mov(nfp_prog, reg_xfer(i), reg_a(src_gpr + i)); 780e663fe38SJakub Kicinski 781e663fe38SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 782e663fe38SJakub Kicinski reg_a(dst_gpr), offset, size - 1, true); 783e663fe38SJakub Kicinski 784e663fe38SJakub Kicinski return 0; 785e663fe38SJakub Kicinski } 786e663fe38SJakub Kicinski 787e663fe38SJakub Kicinski static int 788e663fe38SJakub Kicinski data_st_host_order(struct nfp_prog *nfp_prog, u8 dst_gpr, swreg offset, 789e663fe38SJakub Kicinski u64 imm, u8 size) 790e663fe38SJakub Kicinski { 791e663fe38SJakub Kicinski wrp_immed(nfp_prog, reg_xfer(0), imm); 792e663fe38SJakub Kicinski if (size == 8) 793e663fe38SJakub Kicinski wrp_immed(nfp_prog, reg_xfer(1), imm >> 32); 794e663fe38SJakub Kicinski 795e663fe38SJakub Kicinski emit_cmd(nfp_prog, CMD_TGT_WRITE8_SWAP, CMD_MODE_32b, 0, 796e663fe38SJakub Kicinski reg_a(dst_gpr), offset, size - 1, true); 797e663fe38SJakub Kicinski 798e663fe38SJakub Kicinski return 0; 799e663fe38SJakub Kicinski } 800e663fe38SJakub Kicinski 801ee9133a8SJakub Kicinski typedef int 802ee9133a8SJakub Kicinski (*lmem_step)(struct nfp_prog *nfp_prog, u8 gpr, u8 gpr_byte, s32 off, 8032df03a50SJakub Kicinski unsigned int size, bool first, bool new_gpr, bool last, bool lm3, 8042df03a50SJakub Kicinski bool needs_inc); 805a82b23fbSJakub Kicinski 806a82b23fbSJakub Kicinski static int 807a82b23fbSJakub Kicinski wrp_lmem_load(struct nfp_prog *nfp_prog, u8 dst, u8 dst_byte, s32 off, 8082df03a50SJakub Kicinski unsigned int size, bool first, bool new_gpr, bool last, bool lm3, 8092df03a50SJakub Kicinski bool needs_inc) 810a82b23fbSJakub Kicinski { 8112df03a50SJakub Kicinski bool should_inc = needs_inc && new_gpr && !last; 812a82b23fbSJakub Kicinski u32 idx, src_byte; 813a82b23fbSJakub Kicinski enum shf_sc sc; 814a82b23fbSJakub Kicinski swreg reg; 815a82b23fbSJakub Kicinski int shf; 816a82b23fbSJakub Kicinski u8 mask; 817a82b23fbSJakub Kicinski 818a82b23fbSJakub Kicinski if (WARN_ON_ONCE(dst_byte + size > 4 || off % 4 + size > 4)) 819a82b23fbSJakub Kicinski return -EOPNOTSUPP; 820a82b23fbSJakub Kicinski 821a82b23fbSJakub Kicinski idx = off / 4; 822a82b23fbSJakub Kicinski 823a82b23fbSJakub Kicinski /* Move the entire word */ 824a82b23fbSJakub Kicinski if (size == 4) { 8252df03a50SJakub Kicinski wrp_mov(nfp_prog, reg_both(dst), 8262df03a50SJakub Kicinski should_inc ? reg_lm_inc(3) : reg_lm(lm3 ? 3 : 0, idx)); 827a82b23fbSJakub Kicinski return 0; 828a82b23fbSJakub Kicinski } 829a82b23fbSJakub Kicinski 8302df03a50SJakub Kicinski if (WARN_ON_ONCE(lm3 && idx > RE_REG_LM_IDX_MAX)) 8312df03a50SJakub Kicinski return -EOPNOTSUPP; 8322df03a50SJakub Kicinski 833a82b23fbSJakub Kicinski src_byte = off % 4; 834a82b23fbSJakub Kicinski 835a82b23fbSJakub Kicinski mask = (1 << size) - 1; 836a82b23fbSJakub Kicinski mask <<= dst_byte; 837a82b23fbSJakub Kicinski 838a82b23fbSJakub Kicinski if (WARN_ON_ONCE(mask > 0xf)) 839a82b23fbSJakub Kicinski return -EOPNOTSUPP; 840a82b23fbSJakub Kicinski 841a82b23fbSJakub Kicinski shf = abs(src_byte - dst_byte) * 8; 842a82b23fbSJakub Kicinski if (src_byte == dst_byte) { 843a82b23fbSJakub Kicinski sc = SHF_SC_NONE; 844a82b23fbSJakub Kicinski } else if (src_byte < dst_byte) { 845a82b23fbSJakub Kicinski shf = 32 - shf; 846a82b23fbSJakub Kicinski sc = SHF_SC_L_SHF; 847a82b23fbSJakub Kicinski } else { 848a82b23fbSJakub Kicinski sc = SHF_SC_R_SHF; 849a82b23fbSJakub Kicinski } 850a82b23fbSJakub Kicinski 851a82b23fbSJakub Kicinski /* ld_field can address fewer indexes, if offset too large do RMW. 852a82b23fbSJakub Kicinski * Because we RMV twice we waste 2 cycles on unaligned 8 byte writes. 853a82b23fbSJakub Kicinski */ 854a82b23fbSJakub Kicinski if (idx <= RE_REG_LM_IDX_MAX) { 8552df03a50SJakub Kicinski reg = reg_lm(lm3 ? 3 : 0, idx); 856a82b23fbSJakub Kicinski } else { 857a82b23fbSJakub Kicinski reg = imm_a(nfp_prog); 8589a90c83cSJakub Kicinski /* If it's not the first part of the load and we start a new GPR 8599a90c83cSJakub Kicinski * that means we are loading a second part of the LMEM word into 8609a90c83cSJakub Kicinski * a new GPR. IOW we've already looked that LMEM word and 8619a90c83cSJakub Kicinski * therefore it has been loaded into imm_a(). 8629a90c83cSJakub Kicinski */ 8639a90c83cSJakub Kicinski if (first || !new_gpr) 864a82b23fbSJakub Kicinski wrp_mov(nfp_prog, reg, reg_lm(0, idx)); 865a82b23fbSJakub Kicinski } 866a82b23fbSJakub Kicinski 867a82b23fbSJakub Kicinski emit_ld_field_any(nfp_prog, reg_both(dst), mask, reg, sc, shf, new_gpr); 868a82b23fbSJakub Kicinski 8692df03a50SJakub Kicinski if (should_inc) 8702df03a50SJakub Kicinski wrp_mov(nfp_prog, reg_none(), reg_lm_inc(3)); 8712df03a50SJakub Kicinski 872a82b23fbSJakub Kicinski return 0; 873a82b23fbSJakub Kicinski } 874ee9133a8SJakub Kicinski 875ee9133a8SJakub Kicinski static int 876ee9133a8SJakub Kicinski wrp_lmem_store(struct nfp_prog *nfp_prog, u8 src, u8 src_byte, s32 off, 8772df03a50SJakub Kicinski unsigned int size, bool first, bool new_gpr, bool last, bool lm3, 8782df03a50SJakub Kicinski bool needs_inc) 879ee9133a8SJakub Kicinski { 8802df03a50SJakub Kicinski bool should_inc = needs_inc && new_gpr && !last; 881ee9133a8SJakub Kicinski u32 idx, dst_byte; 882ee9133a8SJakub Kicinski enum shf_sc sc; 883ee9133a8SJakub Kicinski swreg reg; 884ee9133a8SJakub Kicinski int shf; 885ee9133a8SJakub Kicinski u8 mask; 886ee9133a8SJakub Kicinski 887ee9133a8SJakub Kicinski if (WARN_ON_ONCE(src_byte + size > 4 || off % 4 + size > 4)) 888ee9133a8SJakub Kicinski return -EOPNOTSUPP; 889ee9133a8SJakub Kicinski 890ee9133a8SJakub Kicinski idx = off / 4; 891ee9133a8SJakub Kicinski 892ee9133a8SJakub Kicinski /* Move the entire word */ 893ee9133a8SJakub Kicinski if (size == 4) { 8942df03a50SJakub Kicinski wrp_mov(nfp_prog, 8952df03a50SJakub Kicinski should_inc ? reg_lm_inc(3) : reg_lm(lm3 ? 3 : 0, idx), 8962df03a50SJakub Kicinski reg_b(src)); 897ee9133a8SJakub Kicinski return 0; 898ee9133a8SJakub Kicinski } 899ee9133a8SJakub Kicinski 9002df03a50SJakub Kicinski if (WARN_ON_ONCE(lm3 && idx > RE_REG_LM_IDX_MAX)) 9012df03a50SJakub Kicinski return -EOPNOTSUPP; 9022df03a50SJakub Kicinski 903ee9133a8SJakub Kicinski dst_byte = off % 4; 904ee9133a8SJakub Kicinski 905ee9133a8SJakub Kicinski mask = (1 << size) - 1; 906ee9133a8SJakub Kicinski mask <<= dst_byte; 907ee9133a8SJakub Kicinski 908ee9133a8SJakub Kicinski if (WARN_ON_ONCE(mask > 0xf)) 909ee9133a8SJakub Kicinski return -EOPNOTSUPP; 910ee9133a8SJakub Kicinski 911ee9133a8SJakub Kicinski shf = abs(src_byte - dst_byte) * 8; 912ee9133a8SJakub Kicinski if (src_byte == dst_byte) { 913ee9133a8SJakub Kicinski sc = SHF_SC_NONE; 914ee9133a8SJakub Kicinski } else if (src_byte < dst_byte) { 915ee9133a8SJakub Kicinski shf = 32 - shf; 916ee9133a8SJakub Kicinski sc = SHF_SC_L_SHF; 917ee9133a8SJakub Kicinski } else { 918ee9133a8SJakub Kicinski sc = SHF_SC_R_SHF; 919ee9133a8SJakub Kicinski } 920ee9133a8SJakub Kicinski 921ee9133a8SJakub Kicinski /* ld_field can address fewer indexes, if offset too large do RMW. 922ee9133a8SJakub Kicinski * Because we RMV twice we waste 2 cycles on unaligned 8 byte writes. 923ee9133a8SJakub Kicinski */ 924ee9133a8SJakub Kicinski if (idx <= RE_REG_LM_IDX_MAX) { 9252df03a50SJakub Kicinski reg = reg_lm(lm3 ? 3 : 0, idx); 926ee9133a8SJakub Kicinski } else { 927ee9133a8SJakub Kicinski reg = imm_a(nfp_prog); 9289a90c83cSJakub Kicinski /* Only first and last LMEM locations are going to need RMW, 9299a90c83cSJakub Kicinski * the middle location will be overwritten fully. 9309a90c83cSJakub Kicinski */ 9319a90c83cSJakub Kicinski if (first || last) 932ee9133a8SJakub Kicinski wrp_mov(nfp_prog, reg, reg_lm(0, idx)); 933ee9133a8SJakub Kicinski } 934ee9133a8SJakub Kicinski 935ee9133a8SJakub Kicinski emit_ld_field(nfp_prog, reg, mask, reg_b(src), sc, shf); 936ee9133a8SJakub Kicinski 9379a90c83cSJakub Kicinski if (new_gpr || last) { 938ee9133a8SJakub Kicinski if (idx > RE_REG_LM_IDX_MAX) 939ee9133a8SJakub Kicinski wrp_mov(nfp_prog, reg_lm(0, idx), reg); 9402df03a50SJakub Kicinski if (should_inc) 9412df03a50SJakub Kicinski wrp_mov(nfp_prog, reg_none(), reg_lm_inc(3)); 9429a90c83cSJakub Kicinski } 943ee9133a8SJakub Kicinski 944ee9133a8SJakub Kicinski return 0; 945ee9133a8SJakub Kicinski } 946ee9133a8SJakub Kicinski 947ee9133a8SJakub Kicinski static int 948ee9133a8SJakub Kicinski mem_op_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 949b14157eeSJakub Kicinski unsigned int size, unsigned int ptr_off, u8 gpr, u8 ptr_gpr, 950b14157eeSJakub Kicinski bool clr_gpr, lmem_step step) 951ee9133a8SJakub Kicinski { 952d3488480SJakub Kicinski s32 off = nfp_prog->stack_depth + meta->insn.off + ptr_off; 9539a90c83cSJakub Kicinski bool first = true, last; 9542df03a50SJakub Kicinski bool needs_inc = false; 9552df03a50SJakub Kicinski swreg stack_off_reg; 956a82b23fbSJakub Kicinski u8 prev_gpr = 255; 957ee9133a8SJakub Kicinski u32 gpr_byte = 0; 9582df03a50SJakub Kicinski bool lm3 = true; 959ee9133a8SJakub Kicinski int ret; 960ee9133a8SJakub Kicinski 961b14157eeSJakub Kicinski if (meta->ptr_not_const) { 962b14157eeSJakub Kicinski /* Use of the last encountered ptr_off is OK, they all have 963b14157eeSJakub Kicinski * the same alignment. Depend on low bits of value being 964b14157eeSJakub Kicinski * discarded when written to LMaddr register. 965b14157eeSJakub Kicinski */ 966b14157eeSJakub Kicinski stack_off_reg = ur_load_imm_any(nfp_prog, meta->insn.off, 967b14157eeSJakub Kicinski stack_imm(nfp_prog)); 968b14157eeSJakub Kicinski 969b14157eeSJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 970b14157eeSJakub Kicinski reg_a(ptr_gpr), ALU_OP_ADD, stack_off_reg); 971b14157eeSJakub Kicinski 972b14157eeSJakub Kicinski needs_inc = true; 973b14157eeSJakub Kicinski } else if (off + size <= 64) { 9742df03a50SJakub Kicinski /* We can reach bottom 64B with LMaddr0 */ 9752df03a50SJakub Kicinski lm3 = false; 9762df03a50SJakub Kicinski } else if (round_down(off, 32) == round_down(off + size - 1, 32)) { 9772df03a50SJakub Kicinski /* We have to set up a new pointer. If we know the offset 9782df03a50SJakub Kicinski * and the entire access falls into a single 32 byte aligned 9792df03a50SJakub Kicinski * window we won't have to increment the LM pointer. 9802df03a50SJakub Kicinski * The 32 byte alignment is imporant because offset is ORed in 9812df03a50SJakub Kicinski * not added when doing *l$indexN[off]. 9822df03a50SJakub Kicinski */ 9832df03a50SJakub Kicinski stack_off_reg = ur_load_imm_any(nfp_prog, round_down(off, 32), 9842df03a50SJakub Kicinski stack_imm(nfp_prog)); 9852df03a50SJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 9862df03a50SJakub Kicinski stack_reg(nfp_prog), ALU_OP_ADD, stack_off_reg); 9872df03a50SJakub Kicinski 9882df03a50SJakub Kicinski off %= 32; 9892df03a50SJakub Kicinski } else { 9902df03a50SJakub Kicinski stack_off_reg = ur_load_imm_any(nfp_prog, round_down(off, 4), 9912df03a50SJakub Kicinski stack_imm(nfp_prog)); 9922df03a50SJakub Kicinski 9932df03a50SJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 9942df03a50SJakub Kicinski stack_reg(nfp_prog), ALU_OP_ADD, stack_off_reg); 9952df03a50SJakub Kicinski 9962df03a50SJakub Kicinski needs_inc = true; 9972df03a50SJakub Kicinski } 9982df03a50SJakub Kicinski if (lm3) { 9992df03a50SJakub Kicinski emit_csr_wr(nfp_prog, imm_b(nfp_prog), NFP_CSR_ACT_LM_ADDR3); 10002df03a50SJakub Kicinski /* For size < 4 one slot will be filled by zeroing of upper. */ 10012df03a50SJakub Kicinski wrp_nops(nfp_prog, clr_gpr && size < 8 ? 2 : 3); 10022df03a50SJakub Kicinski } 10032df03a50SJakub Kicinski 1004a82b23fbSJakub Kicinski if (clr_gpr && size < 8) 1005a82b23fbSJakub Kicinski wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 1006a82b23fbSJakub Kicinski 1007ee9133a8SJakub Kicinski while (size) { 1008ee9133a8SJakub Kicinski u32 slice_end; 1009ee9133a8SJakub Kicinski u8 slice_size; 1010ee9133a8SJakub Kicinski 1011ee9133a8SJakub Kicinski slice_size = min(size, 4 - gpr_byte); 1012ee9133a8SJakub Kicinski slice_end = min(off + slice_size, round_up(off + 1, 4)); 1013ee9133a8SJakub Kicinski slice_size = slice_end - off; 1014ee9133a8SJakub Kicinski 10159a90c83cSJakub Kicinski last = slice_size == size; 10169a90c83cSJakub Kicinski 10172df03a50SJakub Kicinski if (needs_inc) 10182df03a50SJakub Kicinski off %= 4; 10192df03a50SJakub Kicinski 1020a82b23fbSJakub Kicinski ret = step(nfp_prog, gpr, gpr_byte, off, slice_size, 10212df03a50SJakub Kicinski first, gpr != prev_gpr, last, lm3, needs_inc); 1022ee9133a8SJakub Kicinski if (ret) 1023ee9133a8SJakub Kicinski return ret; 1024ee9133a8SJakub Kicinski 1025a82b23fbSJakub Kicinski prev_gpr = gpr; 10269a90c83cSJakub Kicinski first = false; 10279a90c83cSJakub Kicinski 1028ee9133a8SJakub Kicinski gpr_byte += slice_size; 1029ee9133a8SJakub Kicinski if (gpr_byte >= 4) { 1030ee9133a8SJakub Kicinski gpr_byte -= 4; 1031ee9133a8SJakub Kicinski gpr++; 1032ee9133a8SJakub Kicinski } 1033ee9133a8SJakub Kicinski 1034ee9133a8SJakub Kicinski size -= slice_size; 1035ee9133a8SJakub Kicinski off += slice_size; 1036ee9133a8SJakub Kicinski } 1037ee9133a8SJakub Kicinski 1038ee9133a8SJakub Kicinski return 0; 1039ee9133a8SJakub Kicinski } 1040ee9133a8SJakub Kicinski 1041d9ae7f2bSJakub Kicinski static void 1042d9ae7f2bSJakub Kicinski wrp_alu_imm(struct nfp_prog *nfp_prog, u8 dst, enum alu_op alu_op, u32 imm) 1043d9ae7f2bSJakub Kicinski { 1044b3f868dfSJakub Kicinski swreg tmp_reg; 1045d9ae7f2bSJakub Kicinski 1046d9ae7f2bSJakub Kicinski if (alu_op == ALU_OP_AND) { 1047d9ae7f2bSJakub Kicinski if (!imm) 1048d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), 0); 1049d9ae7f2bSJakub Kicinski if (!imm || !~imm) 1050d9ae7f2bSJakub Kicinski return; 1051d9ae7f2bSJakub Kicinski } 1052d9ae7f2bSJakub Kicinski if (alu_op == ALU_OP_OR) { 1053d9ae7f2bSJakub Kicinski if (!~imm) 1054d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), ~0U); 1055d9ae7f2bSJakub Kicinski if (!imm || !~imm) 1056d9ae7f2bSJakub Kicinski return; 1057d9ae7f2bSJakub Kicinski } 1058d9ae7f2bSJakub Kicinski if (alu_op == ALU_OP_XOR) { 1059d9ae7f2bSJakub Kicinski if (!~imm) 1060d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_none(), 10615d42ced1SJiong Wang ALU_OP_NOT, reg_b(dst)); 1062d9ae7f2bSJakub Kicinski if (!imm || !~imm) 1063d9ae7f2bSJakub Kicinski return; 1064d9ae7f2bSJakub Kicinski } 1065d9ae7f2bSJakub Kicinski 1066d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm, imm_b(nfp_prog)); 1067d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, tmp_reg); 1068d9ae7f2bSJakub Kicinski } 1069d9ae7f2bSJakub Kicinski 1070d9ae7f2bSJakub Kicinski static int 1071d9ae7f2bSJakub Kicinski wrp_alu64_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1072d9ae7f2bSJakub Kicinski enum alu_op alu_op, bool skip) 1073d9ae7f2bSJakub Kicinski { 1074d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1075d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1076d9ae7f2bSJakub Kicinski 1077d9ae7f2bSJakub Kicinski if (skip) { 1078d9ae7f2bSJakub Kicinski meta->skip = true; 1079d9ae7f2bSJakub Kicinski return 0; 1080d9ae7f2bSJakub Kicinski } 1081d9ae7f2bSJakub Kicinski 1082d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, alu_op, imm & ~0U); 1083d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, alu_op, imm >> 32); 1084d9ae7f2bSJakub Kicinski 1085d9ae7f2bSJakub Kicinski return 0; 1086d9ae7f2bSJakub Kicinski } 1087d9ae7f2bSJakub Kicinski 1088d9ae7f2bSJakub Kicinski static int 1089d9ae7f2bSJakub Kicinski wrp_alu64_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1090d9ae7f2bSJakub Kicinski enum alu_op alu_op) 1091d9ae7f2bSJakub Kicinski { 1092d9ae7f2bSJakub Kicinski u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2; 1093d9ae7f2bSJakub Kicinski 1094d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, reg_b(src)); 1095d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst + 1), 1096d9ae7f2bSJakub Kicinski reg_a(dst + 1), alu_op, reg_b(src + 1)); 1097d9ae7f2bSJakub Kicinski 1098d9ae7f2bSJakub Kicinski return 0; 1099d9ae7f2bSJakub Kicinski } 1100d9ae7f2bSJakub Kicinski 1101d9ae7f2bSJakub Kicinski static int 1102d9ae7f2bSJakub Kicinski wrp_alu32_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1103d9ae7f2bSJakub Kicinski enum alu_op alu_op, bool skip) 1104d9ae7f2bSJakub Kicinski { 1105d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1106d9ae7f2bSJakub Kicinski 1107d9ae7f2bSJakub Kicinski if (skip) { 1108d9ae7f2bSJakub Kicinski meta->skip = true; 1109d9ae7f2bSJakub Kicinski return 0; 1110d9ae7f2bSJakub Kicinski } 1111d9ae7f2bSJakub Kicinski 1112d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, alu_op, insn->imm); 1113d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1114d9ae7f2bSJakub Kicinski 1115d9ae7f2bSJakub Kicinski return 0; 1116d9ae7f2bSJakub Kicinski } 1117d9ae7f2bSJakub Kicinski 1118d9ae7f2bSJakub Kicinski static int 1119d9ae7f2bSJakub Kicinski wrp_alu32_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1120d9ae7f2bSJakub Kicinski enum alu_op alu_op) 1121d9ae7f2bSJakub Kicinski { 1122d9ae7f2bSJakub Kicinski u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2; 1123d9ae7f2bSJakub Kicinski 1124d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), reg_a(dst), alu_op, reg_b(src)); 1125d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1126d9ae7f2bSJakub Kicinski 1127d9ae7f2bSJakub Kicinski return 0; 1128d9ae7f2bSJakub Kicinski } 1129d9ae7f2bSJakub Kicinski 1130d9ae7f2bSJakub Kicinski static void 1131d9ae7f2bSJakub Kicinski wrp_test_reg_one(struct nfp_prog *nfp_prog, u8 dst, enum alu_op alu_op, u8 src, 1132d9ae7f2bSJakub Kicinski enum br_mask br_mask, u16 off) 1133d9ae7f2bSJakub Kicinski { 1134d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(dst), alu_op, reg_b(src)); 1135d9ae7f2bSJakub Kicinski emit_br(nfp_prog, br_mask, off, 0); 1136d9ae7f2bSJakub Kicinski } 1137d9ae7f2bSJakub Kicinski 1138d9ae7f2bSJakub Kicinski static int 1139d9ae7f2bSJakub Kicinski wrp_test_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1140d9ae7f2bSJakub Kicinski enum alu_op alu_op, enum br_mask br_mask) 1141d9ae7f2bSJakub Kicinski { 1142d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1143d9ae7f2bSJakub Kicinski 1144d9ae7f2bSJakub Kicinski wrp_test_reg_one(nfp_prog, insn->dst_reg * 2, alu_op, 1145d9ae7f2bSJakub Kicinski insn->src_reg * 2, br_mask, insn->off); 1146d9ae7f2bSJakub Kicinski wrp_test_reg_one(nfp_prog, insn->dst_reg * 2 + 1, alu_op, 1147d9ae7f2bSJakub Kicinski insn->src_reg * 2 + 1, br_mask, insn->off); 1148d9ae7f2bSJakub Kicinski 1149d9ae7f2bSJakub Kicinski return 0; 1150d9ae7f2bSJakub Kicinski } 1151d9ae7f2bSJakub Kicinski 1152d9ae7f2bSJakub Kicinski static int 1153d9ae7f2bSJakub Kicinski wrp_cmp_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1154d9ae7f2bSJakub Kicinski enum br_mask br_mask, bool swap) 1155d9ae7f2bSJakub Kicinski { 1156d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1157d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1158d9ae7f2bSJakub Kicinski u8 reg = insn->dst_reg * 2; 1159b3f868dfSJakub Kicinski swreg tmp_reg; 1160d9ae7f2bSJakub Kicinski 1161d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1162d9ae7f2bSJakub Kicinski if (!swap) 1163d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(reg), ALU_OP_SUB, tmp_reg); 1164d9ae7f2bSJakub Kicinski else 1165d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), tmp_reg, ALU_OP_SUB, reg_a(reg)); 1166d9ae7f2bSJakub Kicinski 1167d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1168d9ae7f2bSJakub Kicinski if (!swap) 1169d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1170d9ae7f2bSJakub Kicinski reg_a(reg + 1), ALU_OP_SUB_C, tmp_reg); 1171d9ae7f2bSJakub Kicinski else 1172d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1173d9ae7f2bSJakub Kicinski tmp_reg, ALU_OP_SUB_C, reg_a(reg + 1)); 1174d9ae7f2bSJakub Kicinski 1175d9ae7f2bSJakub Kicinski emit_br(nfp_prog, br_mask, insn->off, 0); 1176d9ae7f2bSJakub Kicinski 1177d9ae7f2bSJakub Kicinski return 0; 1178d9ae7f2bSJakub Kicinski } 1179d9ae7f2bSJakub Kicinski 1180d9ae7f2bSJakub Kicinski static int 1181d9ae7f2bSJakub Kicinski wrp_cmp_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1182d9ae7f2bSJakub Kicinski enum br_mask br_mask, bool swap) 1183d9ae7f2bSJakub Kicinski { 1184d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 118526fa818dSJakub Kicinski u8 areg, breg; 118626fa818dSJakub Kicinski 118726fa818dSJakub Kicinski areg = insn->dst_reg * 2; 118826fa818dSJakub Kicinski breg = insn->src_reg * 2; 1189d9ae7f2bSJakub Kicinski 1190d9ae7f2bSJakub Kicinski if (swap) { 1191d9ae7f2bSJakub Kicinski areg ^= breg; 1192d9ae7f2bSJakub Kicinski breg ^= areg; 1193d9ae7f2bSJakub Kicinski areg ^= breg; 1194d9ae7f2bSJakub Kicinski } 1195d9ae7f2bSJakub Kicinski 1196d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(areg), ALU_OP_SUB, reg_b(breg)); 1197d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1198d9ae7f2bSJakub Kicinski reg_a(areg + 1), ALU_OP_SUB_C, reg_b(breg + 1)); 1199d9ae7f2bSJakub Kicinski emit_br(nfp_prog, br_mask, insn->off, 0); 1200d9ae7f2bSJakub Kicinski 1201d9ae7f2bSJakub Kicinski return 0; 1202d9ae7f2bSJakub Kicinski } 1203d9ae7f2bSJakub Kicinski 12043119d1fdSJakub Kicinski static void wrp_end32(struct nfp_prog *nfp_prog, swreg reg_in, u8 gpr_out) 12053119d1fdSJakub Kicinski { 12063119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr_out), 0xf, reg_in, 12073119d1fdSJakub Kicinski SHF_SC_R_ROT, 8); 12083119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr_out), 0x5, reg_a(gpr_out), 12093119d1fdSJakub Kicinski SHF_SC_R_ROT, 16); 12103119d1fdSJakub Kicinski } 12113119d1fdSJakub Kicinski 12120d49eaf4SJakub Kicinski static int adjust_head(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 12130d49eaf4SJakub Kicinski { 12140d49eaf4SJakub Kicinski swreg tmp = imm_a(nfp_prog), tmp_len = imm_b(nfp_prog); 12150d49eaf4SJakub Kicinski struct nfp_bpf_cap_adjust_head *adjust_head; 12160d49eaf4SJakub Kicinski u32 ret_einval, end; 12170d49eaf4SJakub Kicinski 12180d49eaf4SJakub Kicinski adjust_head = &nfp_prog->bpf->adjust_head; 12190d49eaf4SJakub Kicinski 12200d49eaf4SJakub Kicinski ret_einval = nfp_prog_current_offset(nfp_prog) + 14; 12210d49eaf4SJakub Kicinski end = ret_einval + 2; 12220d49eaf4SJakub Kicinski 12230d49eaf4SJakub Kicinski /* We need to use a temp because offset is just a part of the pkt ptr */ 12240d49eaf4SJakub Kicinski emit_alu(nfp_prog, tmp, 12250d49eaf4SJakub Kicinski reg_a(2 * 2), ALU_OP_ADD_2B, pptr_reg(nfp_prog)); 12260d49eaf4SJakub Kicinski 12270d49eaf4SJakub Kicinski /* Validate result will fit within FW datapath constraints */ 12280d49eaf4SJakub Kicinski emit_alu(nfp_prog, reg_none(), 12290d49eaf4SJakub Kicinski tmp, ALU_OP_SUB, reg_imm(adjust_head->off_min)); 12300d49eaf4SJakub Kicinski emit_br(nfp_prog, BR_BLO, ret_einval, 0); 12310d49eaf4SJakub Kicinski emit_alu(nfp_prog, reg_none(), 12320d49eaf4SJakub Kicinski reg_imm(adjust_head->off_max), ALU_OP_SUB, tmp); 12330d49eaf4SJakub Kicinski emit_br(nfp_prog, BR_BLO, ret_einval, 0); 12340d49eaf4SJakub Kicinski 12350d49eaf4SJakub Kicinski /* Validate the length is at least ETH_HLEN */ 12360d49eaf4SJakub Kicinski emit_alu(nfp_prog, tmp_len, 12370d49eaf4SJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12380d49eaf4SJakub Kicinski emit_alu(nfp_prog, reg_none(), 12390d49eaf4SJakub Kicinski tmp_len, ALU_OP_SUB, reg_imm(ETH_HLEN)); 12400d49eaf4SJakub Kicinski emit_br(nfp_prog, BR_BMI, ret_einval, 0); 12410d49eaf4SJakub Kicinski 12420d49eaf4SJakub Kicinski /* Load the ret code */ 12430d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(0), 0); 12440d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(1), 0); 12450d49eaf4SJakub Kicinski 12460d49eaf4SJakub Kicinski /* Modify the packet metadata */ 12470d49eaf4SJakub Kicinski emit_ld_field(nfp_prog, pptr_reg(nfp_prog), 0x3, tmp, SHF_SC_NONE, 0); 12480d49eaf4SJakub Kicinski 12490d49eaf4SJakub Kicinski /* Skip over the -EINVAL ret code (defer 2) */ 12500d49eaf4SJakub Kicinski emit_br_def(nfp_prog, end, 2); 12510d49eaf4SJakub Kicinski 12520d49eaf4SJakub Kicinski emit_alu(nfp_prog, plen_reg(nfp_prog), 12530d49eaf4SJakub Kicinski plen_reg(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12540d49eaf4SJakub Kicinski emit_alu(nfp_prog, pv_len(nfp_prog), 12550d49eaf4SJakub Kicinski pv_len(nfp_prog), ALU_OP_SUB, reg_a(2 * 2)); 12560d49eaf4SJakub Kicinski 12570d49eaf4SJakub Kicinski /* return -EINVAL target */ 12580d49eaf4SJakub Kicinski if (!nfp_prog_confirm_current_offset(nfp_prog, ret_einval)) 12590d49eaf4SJakub Kicinski return -EINVAL; 12600d49eaf4SJakub Kicinski 12610d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(0), -22); 12620d49eaf4SJakub Kicinski wrp_immed(nfp_prog, reg_both(1), ~0); 12630d49eaf4SJakub Kicinski 12640d49eaf4SJakub Kicinski if (!nfp_prog_confirm_current_offset(nfp_prog, end)) 12650d49eaf4SJakub Kicinski return -EINVAL; 12660d49eaf4SJakub Kicinski 12670d49eaf4SJakub Kicinski return 0; 12680d49eaf4SJakub Kicinski } 12690d49eaf4SJakub Kicinski 1270d9ae7f2bSJakub Kicinski /* --- Callbacks --- */ 1271d9ae7f2bSJakub Kicinski static int mov_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1272d9ae7f2bSJakub Kicinski { 1273d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1274b14157eeSJakub Kicinski u8 dst = insn->dst_reg * 2; 1275b14157eeSJakub Kicinski u8 src = insn->src_reg * 2; 1276d9ae7f2bSJakub Kicinski 1277b14157eeSJakub Kicinski if (insn->src_reg == BPF_REG_10) { 1278b14157eeSJakub Kicinski swreg stack_depth_reg; 1279b14157eeSJakub Kicinski 1280b14157eeSJakub Kicinski stack_depth_reg = ur_load_imm_any(nfp_prog, 1281b14157eeSJakub Kicinski nfp_prog->stack_depth, 1282b14157eeSJakub Kicinski stack_imm(nfp_prog)); 1283b14157eeSJakub Kicinski emit_alu(nfp_prog, reg_both(dst), 1284b14157eeSJakub Kicinski stack_reg(nfp_prog), ALU_OP_ADD, stack_depth_reg); 1285b14157eeSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), 0); 1286b14157eeSJakub Kicinski } else { 1287b14157eeSJakub Kicinski wrp_reg_mov(nfp_prog, dst, src); 1288b14157eeSJakub Kicinski wrp_reg_mov(nfp_prog, dst + 1, src + 1); 1289b14157eeSJakub Kicinski } 1290d9ae7f2bSJakub Kicinski 1291d9ae7f2bSJakub Kicinski return 0; 1292d9ae7f2bSJakub Kicinski } 1293d9ae7f2bSJakub Kicinski 1294d9ae7f2bSJakub Kicinski static int mov_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1295d9ae7f2bSJakub Kicinski { 1296d9ae7f2bSJakub Kicinski u64 imm = meta->insn.imm; /* sign extend */ 1297d9ae7f2bSJakub Kicinski 1298d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2), imm & ~0U); 1299d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), imm >> 32); 1300d9ae7f2bSJakub Kicinski 1301d9ae7f2bSJakub Kicinski return 0; 1302d9ae7f2bSJakub Kicinski } 1303d9ae7f2bSJakub Kicinski 1304d9ae7f2bSJakub Kicinski static int xor_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1305d9ae7f2bSJakub Kicinski { 1306d9ae7f2bSJakub Kicinski return wrp_alu64_reg(nfp_prog, meta, ALU_OP_XOR); 1307d9ae7f2bSJakub Kicinski } 1308d9ae7f2bSJakub Kicinski 1309d9ae7f2bSJakub Kicinski static int xor_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1310d9ae7f2bSJakub Kicinski { 1311d9ae7f2bSJakub Kicinski return wrp_alu64_imm(nfp_prog, meta, ALU_OP_XOR, !meta->insn.imm); 1312d9ae7f2bSJakub Kicinski } 1313d9ae7f2bSJakub Kicinski 1314d9ae7f2bSJakub Kicinski static int and_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1315d9ae7f2bSJakub Kicinski { 1316d9ae7f2bSJakub Kicinski return wrp_alu64_reg(nfp_prog, meta, ALU_OP_AND); 1317d9ae7f2bSJakub Kicinski } 1318d9ae7f2bSJakub Kicinski 1319d9ae7f2bSJakub Kicinski static int and_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1320d9ae7f2bSJakub Kicinski { 1321d9ae7f2bSJakub Kicinski return wrp_alu64_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm); 1322d9ae7f2bSJakub Kicinski } 1323d9ae7f2bSJakub Kicinski 1324d9ae7f2bSJakub Kicinski static int or_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1325d9ae7f2bSJakub Kicinski { 1326d9ae7f2bSJakub Kicinski return wrp_alu64_reg(nfp_prog, meta, ALU_OP_OR); 1327d9ae7f2bSJakub Kicinski } 1328d9ae7f2bSJakub Kicinski 1329d9ae7f2bSJakub Kicinski static int or_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1330d9ae7f2bSJakub Kicinski { 1331d9ae7f2bSJakub Kicinski return wrp_alu64_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm); 1332d9ae7f2bSJakub Kicinski } 1333d9ae7f2bSJakub Kicinski 1334d9ae7f2bSJakub Kicinski static int add_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1335d9ae7f2bSJakub Kicinski { 1336d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1337d9ae7f2bSJakub Kicinski 1338d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), 1339d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_ADD, 1340d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2)); 1341d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 1342d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_ADD_C, 1343d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2 + 1)); 1344d9ae7f2bSJakub Kicinski 1345d9ae7f2bSJakub Kicinski return 0; 1346d9ae7f2bSJakub Kicinski } 1347d9ae7f2bSJakub Kicinski 1348d9ae7f2bSJakub Kicinski static int add_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1349d9ae7f2bSJakub Kicinski { 1350d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1351d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1352d9ae7f2bSJakub Kicinski 1353d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, ALU_OP_ADD, imm & ~0U); 1354d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, ALU_OP_ADD_C, imm >> 32); 1355d9ae7f2bSJakub Kicinski 1356d9ae7f2bSJakub Kicinski return 0; 1357d9ae7f2bSJakub Kicinski } 1358d9ae7f2bSJakub Kicinski 1359d9ae7f2bSJakub Kicinski static int sub_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1360d9ae7f2bSJakub Kicinski { 1361d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1362d9ae7f2bSJakub Kicinski 1363d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), 1364d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_SUB, 1365d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2)); 1366d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 1367d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_SUB_C, 1368d9ae7f2bSJakub Kicinski reg_b(insn->src_reg * 2 + 1)); 1369d9ae7f2bSJakub Kicinski 1370d9ae7f2bSJakub Kicinski return 0; 1371d9ae7f2bSJakub Kicinski } 1372d9ae7f2bSJakub Kicinski 1373d9ae7f2bSJakub Kicinski static int sub_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1374d9ae7f2bSJakub Kicinski { 1375d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1376d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1377d9ae7f2bSJakub Kicinski 1378d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2, ALU_OP_SUB, imm & ~0U); 1379d9ae7f2bSJakub Kicinski wrp_alu_imm(nfp_prog, insn->dst_reg * 2 + 1, ALU_OP_SUB_C, imm >> 32); 1380d9ae7f2bSJakub Kicinski 1381d9ae7f2bSJakub Kicinski return 0; 1382d9ae7f2bSJakub Kicinski } 1383d9ae7f2bSJakub Kicinski 1384254ef4d7SJiong Wang static int neg_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1385254ef4d7SJiong Wang { 1386254ef4d7SJiong Wang const struct bpf_insn *insn = &meta->insn; 1387254ef4d7SJiong Wang 1388254ef4d7SJiong Wang emit_alu(nfp_prog, reg_both(insn->dst_reg * 2), reg_imm(0), 1389254ef4d7SJiong Wang ALU_OP_SUB, reg_b(insn->dst_reg * 2)); 1390254ef4d7SJiong Wang emit_alu(nfp_prog, reg_both(insn->dst_reg * 2 + 1), reg_imm(0), 1391254ef4d7SJiong Wang ALU_OP_SUB_C, reg_b(insn->dst_reg * 2 + 1)); 1392254ef4d7SJiong Wang 1393254ef4d7SJiong Wang return 0; 1394254ef4d7SJiong Wang } 1395254ef4d7SJiong Wang 1396d9ae7f2bSJakub Kicinski static int shl_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1397d9ae7f2bSJakub Kicinski { 1398d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 13993cae1319SJakub Kicinski u8 dst = insn->dst_reg * 2; 1400d9ae7f2bSJakub Kicinski 14013cae1319SJakub Kicinski if (insn->imm < 32) { 14023cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst + 1), 14033cae1319SJakub Kicinski reg_a(dst + 1), SHF_OP_NONE, reg_b(dst), 14043cae1319SJakub Kicinski SHF_SC_R_DSHF, 32 - insn->imm); 14053cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst), 14063cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst), 14073cae1319SJakub Kicinski SHF_SC_L_SHF, insn->imm); 14083cae1319SJakub Kicinski } else if (insn->imm == 32) { 14093cae1319SJakub Kicinski wrp_reg_mov(nfp_prog, dst + 1, dst); 14103cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), 0); 14113cae1319SJakub Kicinski } else if (insn->imm > 32) { 14123cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst + 1), 14133cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst), 14143cae1319SJakub Kicinski SHF_SC_L_SHF, insn->imm - 32); 14153cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), 0); 14163cae1319SJakub Kicinski } 1417d9ae7f2bSJakub Kicinski 1418d9ae7f2bSJakub Kicinski return 0; 1419d9ae7f2bSJakub Kicinski } 1420d9ae7f2bSJakub Kicinski 1421d9ae7f2bSJakub Kicinski static int shr_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1422d9ae7f2bSJakub Kicinski { 1423d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 14243cae1319SJakub Kicinski u8 dst = insn->dst_reg * 2; 1425d9ae7f2bSJakub Kicinski 14263cae1319SJakub Kicinski if (insn->imm < 32) { 14273cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst), 14283cae1319SJakub Kicinski reg_a(dst + 1), SHF_OP_NONE, reg_b(dst), 14293cae1319SJakub Kicinski SHF_SC_R_DSHF, insn->imm); 14303cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst + 1), 14313cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst + 1), 14323cae1319SJakub Kicinski SHF_SC_R_SHF, insn->imm); 14333cae1319SJakub Kicinski } else if (insn->imm == 32) { 14343cae1319SJakub Kicinski wrp_reg_mov(nfp_prog, dst, dst + 1); 14353cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), 0); 14363cae1319SJakub Kicinski } else if (insn->imm > 32) { 14373cae1319SJakub Kicinski emit_shf(nfp_prog, reg_both(dst), 14383cae1319SJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(dst + 1), 14393cae1319SJakub Kicinski SHF_SC_R_SHF, insn->imm - 32); 14403cae1319SJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), 0); 14413cae1319SJakub Kicinski } 1442d9ae7f2bSJakub Kicinski 1443d9ae7f2bSJakub Kicinski return 0; 1444d9ae7f2bSJakub Kicinski } 1445d9ae7f2bSJakub Kicinski 1446d9ae7f2bSJakub Kicinski static int mov_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1447d9ae7f2bSJakub Kicinski { 1448d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1449d9ae7f2bSJakub Kicinski 1450d9ae7f2bSJakub Kicinski wrp_reg_mov(nfp_prog, insn->dst_reg * 2, insn->src_reg * 2); 1451d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1452d9ae7f2bSJakub Kicinski 1453d9ae7f2bSJakub Kicinski return 0; 1454d9ae7f2bSJakub Kicinski } 1455d9ae7f2bSJakub Kicinski 1456d9ae7f2bSJakub Kicinski static int mov_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1457d9ae7f2bSJakub Kicinski { 1458d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1459d9ae7f2bSJakub Kicinski 1460d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2), insn->imm); 1461d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1462d9ae7f2bSJakub Kicinski 1463d9ae7f2bSJakub Kicinski return 0; 1464d9ae7f2bSJakub Kicinski } 1465d9ae7f2bSJakub Kicinski 1466d9ae7f2bSJakub Kicinski static int xor_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1467d9ae7f2bSJakub Kicinski { 1468d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_XOR); 1469d9ae7f2bSJakub Kicinski } 1470d9ae7f2bSJakub Kicinski 1471d9ae7f2bSJakub Kicinski static int xor_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1472d9ae7f2bSJakub Kicinski { 1473d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_XOR, !~meta->insn.imm); 1474d9ae7f2bSJakub Kicinski } 1475d9ae7f2bSJakub Kicinski 1476d9ae7f2bSJakub Kicinski static int and_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1477d9ae7f2bSJakub Kicinski { 1478d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_AND); 1479d9ae7f2bSJakub Kicinski } 1480d9ae7f2bSJakub Kicinski 1481d9ae7f2bSJakub Kicinski static int and_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1482d9ae7f2bSJakub Kicinski { 1483d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm); 1484d9ae7f2bSJakub Kicinski } 1485d9ae7f2bSJakub Kicinski 1486d9ae7f2bSJakub Kicinski static int or_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1487d9ae7f2bSJakub Kicinski { 1488d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_OR); 1489d9ae7f2bSJakub Kicinski } 1490d9ae7f2bSJakub Kicinski 1491d9ae7f2bSJakub Kicinski static int or_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1492d9ae7f2bSJakub Kicinski { 1493d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm); 1494d9ae7f2bSJakub Kicinski } 1495d9ae7f2bSJakub Kicinski 1496d9ae7f2bSJakub Kicinski static int add_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1497d9ae7f2bSJakub Kicinski { 1498d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_ADD); 1499d9ae7f2bSJakub Kicinski } 1500d9ae7f2bSJakub Kicinski 1501d9ae7f2bSJakub Kicinski static int add_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1502d9ae7f2bSJakub Kicinski { 1503d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_ADD, !meta->insn.imm); 1504d9ae7f2bSJakub Kicinski } 1505d9ae7f2bSJakub Kicinski 1506d9ae7f2bSJakub Kicinski static int sub_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1507d9ae7f2bSJakub Kicinski { 1508d9ae7f2bSJakub Kicinski return wrp_alu32_reg(nfp_prog, meta, ALU_OP_SUB); 1509d9ae7f2bSJakub Kicinski } 1510d9ae7f2bSJakub Kicinski 1511d9ae7f2bSJakub Kicinski static int sub_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1512d9ae7f2bSJakub Kicinski { 1513d9ae7f2bSJakub Kicinski return wrp_alu32_imm(nfp_prog, meta, ALU_OP_SUB, !meta->insn.imm); 1514d9ae7f2bSJakub Kicinski } 1515d9ae7f2bSJakub Kicinski 1516254ef4d7SJiong Wang static int neg_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1517254ef4d7SJiong Wang { 1518254ef4d7SJiong Wang u8 dst = meta->insn.dst_reg * 2; 1519254ef4d7SJiong Wang 1520254ef4d7SJiong Wang emit_alu(nfp_prog, reg_both(dst), reg_imm(0), ALU_OP_SUB, reg_b(dst)); 1521254ef4d7SJiong Wang wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1522254ef4d7SJiong Wang 1523254ef4d7SJiong Wang return 0; 1524254ef4d7SJiong Wang } 1525254ef4d7SJiong Wang 1526d9ae7f2bSJakub Kicinski static int shl_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1527d9ae7f2bSJakub Kicinski { 1528d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1529d9ae7f2bSJakub Kicinski 1530d9ae7f2bSJakub Kicinski if (!insn->imm) 1531d9ae7f2bSJakub Kicinski return 1; /* TODO: zero shift means indirect */ 1532d9ae7f2bSJakub Kicinski 1533d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_both(insn->dst_reg * 2), 1534d9ae7f2bSJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(insn->dst_reg * 2), 1535d9ae7f2bSJakub Kicinski SHF_SC_L_SHF, insn->imm); 1536d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_both(insn->dst_reg * 2 + 1), 0); 1537d9ae7f2bSJakub Kicinski 1538d9ae7f2bSJakub Kicinski return 0; 1539d9ae7f2bSJakub Kicinski } 1540d9ae7f2bSJakub Kicinski 15413119d1fdSJakub Kicinski static int end_reg32(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 15423119d1fdSJakub Kicinski { 15433119d1fdSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 15443119d1fdSJakub Kicinski u8 gpr = insn->dst_reg * 2; 15453119d1fdSJakub Kicinski 15463119d1fdSJakub Kicinski switch (insn->imm) { 15473119d1fdSJakub Kicinski case 16: 15483119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr), 0x9, reg_b(gpr), 15493119d1fdSJakub Kicinski SHF_SC_R_ROT, 8); 15503119d1fdSJakub Kicinski emit_ld_field(nfp_prog, reg_both(gpr), 0xe, reg_a(gpr), 15513119d1fdSJakub Kicinski SHF_SC_R_SHF, 16); 15523119d1fdSJakub Kicinski 15533119d1fdSJakub Kicinski wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 15543119d1fdSJakub Kicinski break; 15553119d1fdSJakub Kicinski case 32: 15563119d1fdSJakub Kicinski wrp_end32(nfp_prog, reg_a(gpr), gpr); 15573119d1fdSJakub Kicinski wrp_immed(nfp_prog, reg_both(gpr + 1), 0); 15583119d1fdSJakub Kicinski break; 15593119d1fdSJakub Kicinski case 64: 15603119d1fdSJakub Kicinski wrp_mov(nfp_prog, imm_a(nfp_prog), reg_b(gpr + 1)); 15613119d1fdSJakub Kicinski 15623119d1fdSJakub Kicinski wrp_end32(nfp_prog, reg_a(gpr), gpr + 1); 15633119d1fdSJakub Kicinski wrp_end32(nfp_prog, imm_a(nfp_prog), gpr); 15643119d1fdSJakub Kicinski break; 15653119d1fdSJakub Kicinski } 15663119d1fdSJakub Kicinski 15673119d1fdSJakub Kicinski return 0; 15683119d1fdSJakub Kicinski } 15693119d1fdSJakub Kicinski 1570d9ae7f2bSJakub Kicinski static int imm_ld8_part2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1571d9ae7f2bSJakub Kicinski { 15729f16c8abSJakub Kicinski struct nfp_insn_meta *prev = nfp_meta_prev(meta); 15739f16c8abSJakub Kicinski u32 imm_lo, imm_hi; 15749f16c8abSJakub Kicinski u8 dst; 15759f16c8abSJakub Kicinski 15769f16c8abSJakub Kicinski dst = prev->insn.dst_reg * 2; 15779f16c8abSJakub Kicinski imm_lo = prev->insn.imm; 15789f16c8abSJakub Kicinski imm_hi = meta->insn.imm; 15799f16c8abSJakub Kicinski 15809f16c8abSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst), imm_lo); 15819f16c8abSJakub Kicinski 15829f16c8abSJakub Kicinski /* mov is always 1 insn, load imm may be two, so try to use mov */ 15839f16c8abSJakub Kicinski if (imm_hi == imm_lo) 15849f16c8abSJakub Kicinski wrp_mov(nfp_prog, reg_both(dst + 1), reg_a(dst)); 15859f16c8abSJakub Kicinski else 15869f16c8abSJakub Kicinski wrp_immed(nfp_prog, reg_both(dst + 1), imm_hi); 1587d9ae7f2bSJakub Kicinski 1588d9ae7f2bSJakub Kicinski return 0; 1589d9ae7f2bSJakub Kicinski } 1590d9ae7f2bSJakub Kicinski 1591d9ae7f2bSJakub Kicinski static int imm_ld8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1592d9ae7f2bSJakub Kicinski { 1593d9ae7f2bSJakub Kicinski meta->double_cb = imm_ld8_part2; 1594d9ae7f2bSJakub Kicinski return 0; 1595d9ae7f2bSJakub Kicinski } 1596d9ae7f2bSJakub Kicinski 1597d9ae7f2bSJakub Kicinski static int data_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1598d9ae7f2bSJakub Kicinski { 1599d9ae7f2bSJakub Kicinski return construct_data_ld(nfp_prog, meta->insn.imm, 1); 1600d9ae7f2bSJakub Kicinski } 1601d9ae7f2bSJakub Kicinski 1602d9ae7f2bSJakub Kicinski static int data_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1603d9ae7f2bSJakub Kicinski { 1604d9ae7f2bSJakub Kicinski return construct_data_ld(nfp_prog, meta->insn.imm, 2); 1605d9ae7f2bSJakub Kicinski } 1606d9ae7f2bSJakub Kicinski 1607d9ae7f2bSJakub Kicinski static int data_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1608d9ae7f2bSJakub Kicinski { 1609d9ae7f2bSJakub Kicinski return construct_data_ld(nfp_prog, meta->insn.imm, 4); 1610d9ae7f2bSJakub Kicinski } 1611d9ae7f2bSJakub Kicinski 1612d9ae7f2bSJakub Kicinski static int data_ind_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1613d9ae7f2bSJakub Kicinski { 1614d9ae7f2bSJakub Kicinski return construct_data_ind_ld(nfp_prog, meta->insn.imm, 16150a793977SJakub Kicinski meta->insn.src_reg * 2, 1); 1616d9ae7f2bSJakub Kicinski } 1617d9ae7f2bSJakub Kicinski 1618d9ae7f2bSJakub Kicinski static int data_ind_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1619d9ae7f2bSJakub Kicinski { 1620d9ae7f2bSJakub Kicinski return construct_data_ind_ld(nfp_prog, meta->insn.imm, 16210a793977SJakub Kicinski meta->insn.src_reg * 2, 2); 1622d9ae7f2bSJakub Kicinski } 1623d9ae7f2bSJakub Kicinski 1624d9ae7f2bSJakub Kicinski static int data_ind_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1625d9ae7f2bSJakub Kicinski { 1626d9ae7f2bSJakub Kicinski return construct_data_ind_ld(nfp_prog, meta->insn.imm, 16270a793977SJakub Kicinski meta->insn.src_reg * 2, 4); 1628d9ae7f2bSJakub Kicinski } 1629d9ae7f2bSJakub Kicinski 1630a82b23fbSJakub Kicinski static int 1631a82b23fbSJakub Kicinski mem_ldx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1632d3488480SJakub Kicinski unsigned int size, unsigned int ptr_off) 1633a82b23fbSJakub Kicinski { 1634d3488480SJakub Kicinski return mem_op_stack(nfp_prog, meta, size, ptr_off, 1635b14157eeSJakub Kicinski meta->insn.dst_reg * 2, meta->insn.src_reg * 2, 1636b14157eeSJakub Kicinski true, wrp_lmem_load); 1637a82b23fbSJakub Kicinski } 1638a82b23fbSJakub Kicinski 1639943c57b9SJakub Kicinski static int mem_ldx_skb(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1640943c57b9SJakub Kicinski u8 size) 1641d9ae7f2bSJakub Kicinski { 1642bfddbc8aSJakub Kicinski swreg dst = reg_both(meta->insn.dst_reg * 2); 1643bfddbc8aSJakub Kicinski 1644943c57b9SJakub Kicinski switch (meta->insn.off) { 1645c6c580d7SJakub Kicinski case offsetof(struct __sk_buff, len): 1646c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct __sk_buff, len)) 1647943c57b9SJakub Kicinski return -EOPNOTSUPP; 1648bfddbc8aSJakub Kicinski wrp_mov(nfp_prog, dst, plen_reg(nfp_prog)); 1649bfddbc8aSJakub Kicinski break; 1650c6c580d7SJakub Kicinski case offsetof(struct __sk_buff, data): 1651c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct __sk_buff, data)) 1652bfddbc8aSJakub Kicinski return -EOPNOTSUPP; 1653bfddbc8aSJakub Kicinski wrp_mov(nfp_prog, dst, pptr_reg(nfp_prog)); 1654bfddbc8aSJakub Kicinski break; 1655c6c580d7SJakub Kicinski case offsetof(struct __sk_buff, data_end): 1656c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct __sk_buff, data_end)) 1657bfddbc8aSJakub Kicinski return -EOPNOTSUPP; 1658bfddbc8aSJakub Kicinski emit_alu(nfp_prog, dst, 1659bfddbc8aSJakub Kicinski plen_reg(nfp_prog), ALU_OP_ADD, pptr_reg(nfp_prog)); 1660943c57b9SJakub Kicinski break; 1661943c57b9SJakub Kicinski default: 1662d9ae7f2bSJakub Kicinski return -EOPNOTSUPP; 1663943c57b9SJakub Kicinski } 1664943c57b9SJakub Kicinski 1665943c57b9SJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1666d9ae7f2bSJakub Kicinski 1667d9ae7f2bSJakub Kicinski return 0; 1668d9ae7f2bSJakub Kicinski } 1669d9ae7f2bSJakub Kicinski 1670943c57b9SJakub Kicinski static int mem_ldx_xdp(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1671943c57b9SJakub Kicinski u8 size) 1672d9ae7f2bSJakub Kicinski { 1673b3f868dfSJakub Kicinski swreg dst = reg_both(meta->insn.dst_reg * 2); 1674d9ae7f2bSJakub Kicinski 1675943c57b9SJakub Kicinski switch (meta->insn.off) { 1676c6c580d7SJakub Kicinski case offsetof(struct xdp_md, data): 1677c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct xdp_md, data)) 1678c6c580d7SJakub Kicinski return -EOPNOTSUPP; 1679c000dfb5SJakub Kicinski wrp_mov(nfp_prog, dst, pptr_reg(nfp_prog)); 1680943c57b9SJakub Kicinski break; 1681c6c580d7SJakub Kicinski case offsetof(struct xdp_md, data_end): 1682c6c580d7SJakub Kicinski if (size != FIELD_SIZEOF(struct xdp_md, data_end)) 1683c6c580d7SJakub Kicinski return -EOPNOTSUPP; 1684943c57b9SJakub Kicinski emit_alu(nfp_prog, dst, 1685943c57b9SJakub Kicinski plen_reg(nfp_prog), ALU_OP_ADD, pptr_reg(nfp_prog)); 1686943c57b9SJakub Kicinski break; 1687943c57b9SJakub Kicinski default: 1688943c57b9SJakub Kicinski return -EOPNOTSUPP; 1689943c57b9SJakub Kicinski } 1690d9ae7f2bSJakub Kicinski 1691943c57b9SJakub Kicinski wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0); 1692d9ae7f2bSJakub Kicinski 1693d9ae7f2bSJakub Kicinski return 0; 1694d9ae7f2bSJakub Kicinski } 1695d9ae7f2bSJakub Kicinski 16962ca71441SJakub Kicinski static int 16972ca71441SJakub Kicinski mem_ldx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 16982ca71441SJakub Kicinski unsigned int size) 16992ca71441SJakub Kicinski { 17002ca71441SJakub Kicinski swreg tmp_reg; 17012ca71441SJakub Kicinski 17022ca71441SJakub Kicinski tmp_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 17032ca71441SJakub Kicinski 17042ca71441SJakub Kicinski return data_ld_host_order(nfp_prog, meta->insn.src_reg * 2, tmp_reg, 17052ca71441SJakub Kicinski meta->insn.dst_reg * 2, size); 17062ca71441SJakub Kicinski } 17072ca71441SJakub Kicinski 17082ca71441SJakub Kicinski static int 17092ca71441SJakub Kicinski mem_ldx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 17102ca71441SJakub Kicinski unsigned int size) 17112ca71441SJakub Kicinski { 17129879a381SJiong Wang if (meta->ldst_gather_len) 17139879a381SJiong Wang return nfp_cpp_memcpy(nfp_prog, meta); 17149879a381SJiong Wang 17152ca71441SJakub Kicinski if (meta->ptr.type == PTR_TO_CTX) { 1716012bb8a8SJakub Kicinski if (nfp_prog->type == BPF_PROG_TYPE_XDP) 17172ca71441SJakub Kicinski return mem_ldx_xdp(nfp_prog, meta, size); 17182ca71441SJakub Kicinski else 17192ca71441SJakub Kicinski return mem_ldx_skb(nfp_prog, meta, size); 17202ca71441SJakub Kicinski } 17212ca71441SJakub Kicinski 17222ca71441SJakub Kicinski if (meta->ptr.type == PTR_TO_PACKET) 17232ca71441SJakub Kicinski return mem_ldx_data(nfp_prog, meta, size); 17242ca71441SJakub Kicinski 1725a82b23fbSJakub Kicinski if (meta->ptr.type == PTR_TO_STACK) 1726d3488480SJakub Kicinski return mem_ldx_stack(nfp_prog, meta, size, 1727d3488480SJakub Kicinski meta->ptr.off + meta->ptr.var_off.value); 1728a82b23fbSJakub Kicinski 17292ca71441SJakub Kicinski return -EOPNOTSUPP; 17302ca71441SJakub Kicinski } 17312ca71441SJakub Kicinski 17322ca71441SJakub Kicinski static int mem_ldx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 17332ca71441SJakub Kicinski { 17342ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 1); 17352ca71441SJakub Kicinski } 17362ca71441SJakub Kicinski 17372ca71441SJakub Kicinski static int mem_ldx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 17382ca71441SJakub Kicinski { 17392ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 2); 17402ca71441SJakub Kicinski } 17412ca71441SJakub Kicinski 1742d9ae7f2bSJakub Kicinski static int mem_ldx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1743d9ae7f2bSJakub Kicinski { 17442ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 4); 17452ca71441SJakub Kicinski } 17462ca71441SJakub Kicinski 17472ca71441SJakub Kicinski static int mem_ldx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 17482ca71441SJakub Kicinski { 17492ca71441SJakub Kicinski return mem_ldx(nfp_prog, meta, 8); 1750d9ae7f2bSJakub Kicinski } 1751d9ae7f2bSJakub Kicinski 1752e663fe38SJakub Kicinski static int 1753e663fe38SJakub Kicinski mem_st_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1754e663fe38SJakub Kicinski unsigned int size) 1755d9ae7f2bSJakub Kicinski { 1756e663fe38SJakub Kicinski u64 imm = meta->insn.imm; /* sign extend */ 1757e663fe38SJakub Kicinski swreg off_reg; 1758e663fe38SJakub Kicinski 1759e663fe38SJakub Kicinski off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1760e663fe38SJakub Kicinski 1761e663fe38SJakub Kicinski return data_st_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg, 1762e663fe38SJakub Kicinski imm, size); 1763e663fe38SJakub Kicinski } 1764e663fe38SJakub Kicinski 1765e663fe38SJakub Kicinski static int mem_st(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1766e663fe38SJakub Kicinski unsigned int size) 1767e663fe38SJakub Kicinski { 1768e663fe38SJakub Kicinski if (meta->ptr.type == PTR_TO_PACKET) 1769e663fe38SJakub Kicinski return mem_st_data(nfp_prog, meta, size); 1770e663fe38SJakub Kicinski 1771d9ae7f2bSJakub Kicinski return -EOPNOTSUPP; 1772d9ae7f2bSJakub Kicinski } 1773d9ae7f2bSJakub Kicinski 1774e663fe38SJakub Kicinski static int mem_st1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1775d9ae7f2bSJakub Kicinski { 1776e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 1); 1777e663fe38SJakub Kicinski } 1778e663fe38SJakub Kicinski 1779e663fe38SJakub Kicinski static int mem_st2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1780e663fe38SJakub Kicinski { 1781e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 2); 1782e663fe38SJakub Kicinski } 1783e663fe38SJakub Kicinski 1784e663fe38SJakub Kicinski static int mem_st4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1785e663fe38SJakub Kicinski { 1786e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 4); 1787e663fe38SJakub Kicinski } 1788e663fe38SJakub Kicinski 1789e663fe38SJakub Kicinski static int mem_st8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1790e663fe38SJakub Kicinski { 1791e663fe38SJakub Kicinski return mem_st(nfp_prog, meta, 8); 1792e663fe38SJakub Kicinski } 1793e663fe38SJakub Kicinski 1794e663fe38SJakub Kicinski static int 1795e663fe38SJakub Kicinski mem_stx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1796e663fe38SJakub Kicinski unsigned int size) 1797e663fe38SJakub Kicinski { 1798e663fe38SJakub Kicinski swreg off_reg; 1799e663fe38SJakub Kicinski 1800e663fe38SJakub Kicinski off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog)); 1801e663fe38SJakub Kicinski 1802e663fe38SJakub Kicinski return data_stx_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg, 1803e663fe38SJakub Kicinski meta->insn.src_reg * 2, size); 1804e663fe38SJakub Kicinski } 1805e663fe38SJakub Kicinski 1806e663fe38SJakub Kicinski static int 1807ee9133a8SJakub Kicinski mem_stx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1808d3488480SJakub Kicinski unsigned int size, unsigned int ptr_off) 1809ee9133a8SJakub Kicinski { 1810d3488480SJakub Kicinski return mem_op_stack(nfp_prog, meta, size, ptr_off, 1811b14157eeSJakub Kicinski meta->insn.src_reg * 2, meta->insn.dst_reg * 2, 1812b14157eeSJakub Kicinski false, wrp_lmem_store); 1813ee9133a8SJakub Kicinski } 1814ee9133a8SJakub Kicinski 1815ee9133a8SJakub Kicinski static int 1816e663fe38SJakub Kicinski mem_stx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, 1817e663fe38SJakub Kicinski unsigned int size) 1818e663fe38SJakub Kicinski { 1819e663fe38SJakub Kicinski if (meta->ptr.type == PTR_TO_PACKET) 1820e663fe38SJakub Kicinski return mem_stx_data(nfp_prog, meta, size); 1821e663fe38SJakub Kicinski 1822ee9133a8SJakub Kicinski if (meta->ptr.type == PTR_TO_STACK) 1823d3488480SJakub Kicinski return mem_stx_stack(nfp_prog, meta, size, 1824d3488480SJakub Kicinski meta->ptr.off + meta->ptr.var_off.value); 1825ee9133a8SJakub Kicinski 1826d9ae7f2bSJakub Kicinski return -EOPNOTSUPP; 1827d9ae7f2bSJakub Kicinski } 1828d9ae7f2bSJakub Kicinski 1829e663fe38SJakub Kicinski static int mem_stx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1830e663fe38SJakub Kicinski { 1831e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 1); 1832e663fe38SJakub Kicinski } 1833e663fe38SJakub Kicinski 1834e663fe38SJakub Kicinski static int mem_stx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1835e663fe38SJakub Kicinski { 1836e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 2); 1837e663fe38SJakub Kicinski } 1838e663fe38SJakub Kicinski 1839d9ae7f2bSJakub Kicinski static int mem_stx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1840d9ae7f2bSJakub Kicinski { 1841e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 4); 1842e663fe38SJakub Kicinski } 18432ca71441SJakub Kicinski 1844e663fe38SJakub Kicinski static int mem_stx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1845e663fe38SJakub Kicinski { 1846e663fe38SJakub Kicinski return mem_stx(nfp_prog, meta, 8); 1847d9ae7f2bSJakub Kicinski } 1848d9ae7f2bSJakub Kicinski 1849d9ae7f2bSJakub Kicinski static int jump(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1850d9ae7f2bSJakub Kicinski { 1851d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_UNC, meta->insn.off, 0); 1852d9ae7f2bSJakub Kicinski 1853d9ae7f2bSJakub Kicinski return 0; 1854d9ae7f2bSJakub Kicinski } 1855d9ae7f2bSJakub Kicinski 1856d9ae7f2bSJakub Kicinski static int jeq_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1857d9ae7f2bSJakub Kicinski { 1858d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1859d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1860b3f868dfSJakub Kicinski swreg or1, or2, tmp_reg; 1861b3f868dfSJakub Kicinski 1862b3f868dfSJakub Kicinski or1 = reg_a(insn->dst_reg * 2); 1863b3f868dfSJakub Kicinski or2 = reg_b(insn->dst_reg * 2 + 1); 1864d9ae7f2bSJakub Kicinski 1865d9ae7f2bSJakub Kicinski if (imm & ~0U) { 1866d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1867d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_a(nfp_prog), 1868d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_XOR, tmp_reg); 1869d9ae7f2bSJakub Kicinski or1 = imm_a(nfp_prog); 1870d9ae7f2bSJakub Kicinski } 1871d9ae7f2bSJakub Kicinski 1872d9ae7f2bSJakub Kicinski if (imm >> 32) { 1873d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1874d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), 1875d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_XOR, tmp_reg); 1876d9ae7f2bSJakub Kicinski or2 = imm_b(nfp_prog); 1877d9ae7f2bSJakub Kicinski } 1878d9ae7f2bSJakub Kicinski 1879d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), or1, ALU_OP_OR, or2); 1880d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BEQ, insn->off, 0); 1881d9ae7f2bSJakub Kicinski 1882d9ae7f2bSJakub Kicinski return 0; 1883d9ae7f2bSJakub Kicinski } 1884d9ae7f2bSJakub Kicinski 1885d9ae7f2bSJakub Kicinski static int jgt_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1886d9ae7f2bSJakub Kicinski { 188726fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BLO, true); 1888d9ae7f2bSJakub Kicinski } 1889d9ae7f2bSJakub Kicinski 1890d9ae7f2bSJakub Kicinski static int jge_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1891d9ae7f2bSJakub Kicinski { 189226fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BHS, false); 1893d9ae7f2bSJakub Kicinski } 1894d9ae7f2bSJakub Kicinski 18955dd294d4SDaniel Borkmann static int jlt_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 18965dd294d4SDaniel Borkmann { 189726fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BLO, false); 18985dd294d4SDaniel Borkmann } 18995dd294d4SDaniel Borkmann 19005dd294d4SDaniel Borkmann static int jle_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 19015dd294d4SDaniel Borkmann { 190226fa818dSJakub Kicinski return wrp_cmp_imm(nfp_prog, meta, BR_BHS, true); 19035dd294d4SDaniel Borkmann } 19045dd294d4SDaniel Borkmann 1905d9ae7f2bSJakub Kicinski static int jset_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1906d9ae7f2bSJakub Kicinski { 1907d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1908d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1909b3f868dfSJakub Kicinski swreg tmp_reg; 1910d9ae7f2bSJakub Kicinski 1911d9ae7f2bSJakub Kicinski if (!imm) { 1912d9ae7f2bSJakub Kicinski meta->skip = true; 1913d9ae7f2bSJakub Kicinski return 0; 1914d9ae7f2bSJakub Kicinski } 1915d9ae7f2bSJakub Kicinski 1916d9ae7f2bSJakub Kicinski if (imm & ~0U) { 1917d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1918d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1919d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_AND, tmp_reg); 1920d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1921d9ae7f2bSJakub Kicinski } 1922d9ae7f2bSJakub Kicinski 1923d9ae7f2bSJakub Kicinski if (imm >> 32) { 1924d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1925d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1926d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_AND, tmp_reg); 1927d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1928d9ae7f2bSJakub Kicinski } 1929d9ae7f2bSJakub Kicinski 1930d9ae7f2bSJakub Kicinski return 0; 1931d9ae7f2bSJakub Kicinski } 1932d9ae7f2bSJakub Kicinski 1933d9ae7f2bSJakub Kicinski static int jne_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1934d9ae7f2bSJakub Kicinski { 1935d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1936d9ae7f2bSJakub Kicinski u64 imm = insn->imm; /* sign extend */ 1937b3f868dfSJakub Kicinski swreg tmp_reg; 1938d9ae7f2bSJakub Kicinski 1939d9ae7f2bSJakub Kicinski if (!imm) { 1940d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(insn->dst_reg * 2), 1941d9ae7f2bSJakub Kicinski ALU_OP_OR, reg_b(insn->dst_reg * 2 + 1)); 1942d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 194382837370SJakub Kicinski return 0; 1944d9ae7f2bSJakub Kicinski } 1945d9ae7f2bSJakub Kicinski 1946d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm & ~0U, imm_b(nfp_prog)); 1947d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1948d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2), ALU_OP_XOR, tmp_reg); 1949d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1950d9ae7f2bSJakub Kicinski 1951d9ae7f2bSJakub Kicinski tmp_reg = ur_load_imm_any(nfp_prog, imm >> 32, imm_b(nfp_prog)); 1952d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1953d9ae7f2bSJakub Kicinski reg_a(insn->dst_reg * 2 + 1), ALU_OP_XOR, tmp_reg); 1954d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BNE, insn->off, 0); 1955d9ae7f2bSJakub Kicinski 1956d9ae7f2bSJakub Kicinski return 0; 1957d9ae7f2bSJakub Kicinski } 1958d9ae7f2bSJakub Kicinski 1959d9ae7f2bSJakub Kicinski static int jeq_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1960d9ae7f2bSJakub Kicinski { 1961d9ae7f2bSJakub Kicinski const struct bpf_insn *insn = &meta->insn; 1962d9ae7f2bSJakub Kicinski 1963d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_a(nfp_prog), reg_a(insn->dst_reg * 2), 1964d9ae7f2bSJakub Kicinski ALU_OP_XOR, reg_b(insn->src_reg * 2)); 1965d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, imm_b(nfp_prog), reg_a(insn->dst_reg * 2 + 1), 1966d9ae7f2bSJakub Kicinski ALU_OP_XOR, reg_b(insn->src_reg * 2 + 1)); 1967d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), 1968d9ae7f2bSJakub Kicinski imm_a(nfp_prog), ALU_OP_OR, imm_b(nfp_prog)); 1969d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BEQ, insn->off, 0); 1970d9ae7f2bSJakub Kicinski 1971d9ae7f2bSJakub Kicinski return 0; 1972d9ae7f2bSJakub Kicinski } 1973d9ae7f2bSJakub Kicinski 1974d9ae7f2bSJakub Kicinski static int jgt_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1975d9ae7f2bSJakub Kicinski { 197626fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BLO, true); 1977d9ae7f2bSJakub Kicinski } 1978d9ae7f2bSJakub Kicinski 1979d9ae7f2bSJakub Kicinski static int jge_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1980d9ae7f2bSJakub Kicinski { 198126fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BHS, false); 1982d9ae7f2bSJakub Kicinski } 1983d9ae7f2bSJakub Kicinski 19845dd294d4SDaniel Borkmann static int jlt_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 19855dd294d4SDaniel Borkmann { 198626fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BLO, false); 19875dd294d4SDaniel Borkmann } 19885dd294d4SDaniel Borkmann 19895dd294d4SDaniel Borkmann static int jle_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 19905dd294d4SDaniel Borkmann { 199126fa818dSJakub Kicinski return wrp_cmp_reg(nfp_prog, meta, BR_BHS, true); 19925dd294d4SDaniel Borkmann } 19935dd294d4SDaniel Borkmann 1994d9ae7f2bSJakub Kicinski static int jset_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 1995d9ae7f2bSJakub Kicinski { 1996d9ae7f2bSJakub Kicinski return wrp_test_reg(nfp_prog, meta, ALU_OP_AND, BR_BNE); 1997d9ae7f2bSJakub Kicinski } 1998d9ae7f2bSJakub Kicinski 1999d9ae7f2bSJakub Kicinski static int jne_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 2000d9ae7f2bSJakub Kicinski { 2001d9ae7f2bSJakub Kicinski return wrp_test_reg(nfp_prog, meta, ALU_OP_XOR, BR_BNE); 2002d9ae7f2bSJakub Kicinski } 2003d9ae7f2bSJakub Kicinski 20042cb230bdSJakub Kicinski static int call(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 20052cb230bdSJakub Kicinski { 20062cb230bdSJakub Kicinski switch (meta->insn.imm) { 20070d49eaf4SJakub Kicinski case BPF_FUNC_xdp_adjust_head: 20080d49eaf4SJakub Kicinski return adjust_head(nfp_prog, meta); 20092cb230bdSJakub Kicinski default: 20102cb230bdSJakub Kicinski WARN_ONCE(1, "verifier allowed unsupported function\n"); 20112cb230bdSJakub Kicinski return -EOPNOTSUPP; 20122cb230bdSJakub Kicinski } 20132cb230bdSJakub Kicinski } 20142cb230bdSJakub Kicinski 2015d9ae7f2bSJakub Kicinski static int goto_out(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta) 2016d9ae7f2bSJakub Kicinski { 2017d9ae7f2bSJakub Kicinski wrp_br_special(nfp_prog, BR_UNC, OP_BR_GO_OUT); 2018d9ae7f2bSJakub Kicinski 2019d9ae7f2bSJakub Kicinski return 0; 2020d9ae7f2bSJakub Kicinski } 2021d9ae7f2bSJakub Kicinski 2022d9ae7f2bSJakub Kicinski static const instr_cb_t instr_cb[256] = { 2023d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_MOV | BPF_X] = mov_reg64, 2024d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_MOV | BPF_K] = mov_imm64, 2025d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_XOR | BPF_X] = xor_reg64, 2026d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_XOR | BPF_K] = xor_imm64, 2027d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_AND | BPF_X] = and_reg64, 2028d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_AND | BPF_K] = and_imm64, 2029d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_OR | BPF_X] = or_reg64, 2030d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_OR | BPF_K] = or_imm64, 2031d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_ADD | BPF_X] = add_reg64, 2032d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_ADD | BPF_K] = add_imm64, 2033d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_SUB | BPF_X] = sub_reg64, 2034d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_SUB | BPF_K] = sub_imm64, 2035254ef4d7SJiong Wang [BPF_ALU64 | BPF_NEG] = neg_reg64, 2036d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_LSH | BPF_K] = shl_imm64, 2037d9ae7f2bSJakub Kicinski [BPF_ALU64 | BPF_RSH | BPF_K] = shr_imm64, 2038d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_MOV | BPF_X] = mov_reg, 2039d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_MOV | BPF_K] = mov_imm, 2040d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_XOR | BPF_X] = xor_reg, 2041d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_XOR | BPF_K] = xor_imm, 2042d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_AND | BPF_X] = and_reg, 2043d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_AND | BPF_K] = and_imm, 2044d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_OR | BPF_X] = or_reg, 2045d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_OR | BPF_K] = or_imm, 2046d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_ADD | BPF_X] = add_reg, 2047d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_ADD | BPF_K] = add_imm, 2048d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_SUB | BPF_X] = sub_reg, 2049d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_SUB | BPF_K] = sub_imm, 2050254ef4d7SJiong Wang [BPF_ALU | BPF_NEG] = neg_reg, 2051d9ae7f2bSJakub Kicinski [BPF_ALU | BPF_LSH | BPF_K] = shl_imm, 20523119d1fdSJakub Kicinski [BPF_ALU | BPF_END | BPF_X] = end_reg32, 2053d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IMM | BPF_DW] = imm_ld8, 2054d9ae7f2bSJakub Kicinski [BPF_LD | BPF_ABS | BPF_B] = data_ld1, 2055d9ae7f2bSJakub Kicinski [BPF_LD | BPF_ABS | BPF_H] = data_ld2, 2056d9ae7f2bSJakub Kicinski [BPF_LD | BPF_ABS | BPF_W] = data_ld4, 2057d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IND | BPF_B] = data_ind_ld1, 2058d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IND | BPF_H] = data_ind_ld2, 2059d9ae7f2bSJakub Kicinski [BPF_LD | BPF_IND | BPF_W] = data_ind_ld4, 20602ca71441SJakub Kicinski [BPF_LDX | BPF_MEM | BPF_B] = mem_ldx1, 20612ca71441SJakub Kicinski [BPF_LDX | BPF_MEM | BPF_H] = mem_ldx2, 2062d9ae7f2bSJakub Kicinski [BPF_LDX | BPF_MEM | BPF_W] = mem_ldx4, 20632ca71441SJakub Kicinski [BPF_LDX | BPF_MEM | BPF_DW] = mem_ldx8, 2064e663fe38SJakub Kicinski [BPF_STX | BPF_MEM | BPF_B] = mem_stx1, 2065e663fe38SJakub Kicinski [BPF_STX | BPF_MEM | BPF_H] = mem_stx2, 2066d9ae7f2bSJakub Kicinski [BPF_STX | BPF_MEM | BPF_W] = mem_stx4, 2067e663fe38SJakub Kicinski [BPF_STX | BPF_MEM | BPF_DW] = mem_stx8, 2068e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_B] = mem_st1, 2069e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_H] = mem_st2, 2070e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_W] = mem_st4, 2071e663fe38SJakub Kicinski [BPF_ST | BPF_MEM | BPF_DW] = mem_st8, 2072d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JA | BPF_K] = jump, 2073d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JEQ | BPF_K] = jeq_imm, 2074d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGT | BPF_K] = jgt_imm, 2075d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGE | BPF_K] = jge_imm, 20765dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLT | BPF_K] = jlt_imm, 20775dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLE | BPF_K] = jle_imm, 2078d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JSET | BPF_K] = jset_imm, 2079d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JNE | BPF_K] = jne_imm, 2080d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JEQ | BPF_X] = jeq_reg, 2081d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGT | BPF_X] = jgt_reg, 2082d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JGE | BPF_X] = jge_reg, 20835dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLT | BPF_X] = jlt_reg, 20845dd294d4SDaniel Borkmann [BPF_JMP | BPF_JLE | BPF_X] = jle_reg, 2085d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JSET | BPF_X] = jset_reg, 2086d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_JNE | BPF_X] = jne_reg, 20872cb230bdSJakub Kicinski [BPF_JMP | BPF_CALL] = call, 2088d9ae7f2bSJakub Kicinski [BPF_JMP | BPF_EXIT] = goto_out, 2089d9ae7f2bSJakub Kicinski }; 2090d9ae7f2bSJakub Kicinski 2091d9ae7f2bSJakub Kicinski /* --- Misc code --- */ 2092d9ae7f2bSJakub Kicinski static void br_set_offset(u64 *instr, u16 offset) 2093d9ae7f2bSJakub Kicinski { 2094d9ae7f2bSJakub Kicinski u16 addr_lo, addr_hi; 2095d9ae7f2bSJakub Kicinski 2096d9ae7f2bSJakub Kicinski addr_lo = offset & (OP_BR_ADDR_LO >> __bf_shf(OP_BR_ADDR_LO)); 2097d9ae7f2bSJakub Kicinski addr_hi = offset != addr_lo; 2098d9ae7f2bSJakub Kicinski *instr &= ~(OP_BR_ADDR_HI | OP_BR_ADDR_LO); 2099d9ae7f2bSJakub Kicinski *instr |= FIELD_PREP(OP_BR_ADDR_HI, addr_hi); 2100d9ae7f2bSJakub Kicinski *instr |= FIELD_PREP(OP_BR_ADDR_LO, addr_lo); 2101d9ae7f2bSJakub Kicinski } 2102d9ae7f2bSJakub Kicinski 2103d9ae7f2bSJakub Kicinski /* --- Assembler logic --- */ 2104d9ae7f2bSJakub Kicinski static int nfp_fixup_branches(struct nfp_prog *nfp_prog) 2105d9ae7f2bSJakub Kicinski { 21065b674140SJiong Wang struct nfp_insn_meta *meta, *jmp_dst; 2107854dc87dSJiong Wang u32 idx, br_idx; 2108d9ae7f2bSJakub Kicinski 2109854dc87dSJiong Wang list_for_each_entry(meta, &nfp_prog->insns, l) { 2110d9ae7f2bSJakub Kicinski if (meta->skip) 2111d9ae7f2bSJakub Kicinski continue; 21122cb230bdSJakub Kicinski if (meta->insn.code == (BPF_JMP | BPF_CALL)) 21132cb230bdSJakub Kicinski continue; 2114d9ae7f2bSJakub Kicinski if (BPF_CLASS(meta->insn.code) != BPF_JMP) 2115d9ae7f2bSJakub Kicinski continue; 2116d9ae7f2bSJakub Kicinski 21175b674140SJiong Wang if (list_is_last(&meta->l, &nfp_prog->insns)) 2118854dc87dSJiong Wang idx = nfp_prog->last_bpf_off; 21195b674140SJiong Wang else 21205b674140SJiong Wang idx = list_next_entry(meta, l)->off - 1; 2121854dc87dSJiong Wang 2122854dc87dSJiong Wang br_idx = nfp_prog_offset_to_index(nfp_prog, idx); 2123854dc87dSJiong Wang 2124d9ae7f2bSJakub Kicinski if (!nfp_is_br(nfp_prog->prog[br_idx])) { 2125d9ae7f2bSJakub Kicinski pr_err("Fixup found block not ending in branch %d %02x %016llx!!\n", 2126d9ae7f2bSJakub Kicinski br_idx, meta->insn.code, nfp_prog->prog[br_idx]); 2127d9ae7f2bSJakub Kicinski return -ELOOP; 2128d9ae7f2bSJakub Kicinski } 2129d9ae7f2bSJakub Kicinski /* Leave special branches for later */ 2130d9ae7f2bSJakub Kicinski if (FIELD_GET(OP_BR_SPECIAL, nfp_prog->prog[br_idx])) 2131d9ae7f2bSJakub Kicinski continue; 2132d9ae7f2bSJakub Kicinski 21335b674140SJiong Wang if (!meta->jmp_dst) { 21345b674140SJiong Wang pr_err("Non-exit jump doesn't have destination info recorded!!\n"); 2135d9ae7f2bSJakub Kicinski return -ELOOP; 2136d9ae7f2bSJakub Kicinski } 2137d9ae7f2bSJakub Kicinski 21385b674140SJiong Wang jmp_dst = meta->jmp_dst; 2139854dc87dSJiong Wang 21405b674140SJiong Wang if (jmp_dst->skip) { 2141d9ae7f2bSJakub Kicinski pr_err("Branch landing on removed instruction!!\n"); 2142d9ae7f2bSJakub Kicinski return -ELOOP; 2143d9ae7f2bSJakub Kicinski } 2144d9ae7f2bSJakub Kicinski 2145d9ae7f2bSJakub Kicinski for (idx = nfp_prog_offset_to_index(nfp_prog, meta->off); 2146d9ae7f2bSJakub Kicinski idx <= br_idx; idx++) { 2147d9ae7f2bSJakub Kicinski if (!nfp_is_br(nfp_prog->prog[idx])) 2148d9ae7f2bSJakub Kicinski continue; 21495b674140SJiong Wang br_set_offset(&nfp_prog->prog[idx], jmp_dst->off); 2150d9ae7f2bSJakub Kicinski } 2151d9ae7f2bSJakub Kicinski } 2152d9ae7f2bSJakub Kicinski 2153d9ae7f2bSJakub Kicinski /* Fixup 'goto out's separately, they can be scattered around */ 2154d9ae7f2bSJakub Kicinski for (br_idx = 0; br_idx < nfp_prog->prog_len; br_idx++) { 2155d9ae7f2bSJakub Kicinski enum br_special special; 2156d9ae7f2bSJakub Kicinski 2157d9ae7f2bSJakub Kicinski if ((nfp_prog->prog[br_idx] & OP_BR_BASE_MASK) != OP_BR_BASE) 2158d9ae7f2bSJakub Kicinski continue; 2159d9ae7f2bSJakub Kicinski 2160d9ae7f2bSJakub Kicinski special = FIELD_GET(OP_BR_SPECIAL, nfp_prog->prog[br_idx]); 2161d9ae7f2bSJakub Kicinski switch (special) { 2162d9ae7f2bSJakub Kicinski case OP_BR_NORMAL: 2163d9ae7f2bSJakub Kicinski break; 2164d9ae7f2bSJakub Kicinski case OP_BR_GO_OUT: 2165d9ae7f2bSJakub Kicinski br_set_offset(&nfp_prog->prog[br_idx], 2166d9ae7f2bSJakub Kicinski nfp_prog->tgt_out); 2167d9ae7f2bSJakub Kicinski break; 2168d9ae7f2bSJakub Kicinski case OP_BR_GO_ABORT: 2169d9ae7f2bSJakub Kicinski br_set_offset(&nfp_prog->prog[br_idx], 2170d9ae7f2bSJakub Kicinski nfp_prog->tgt_abort); 2171d9ae7f2bSJakub Kicinski break; 2172d9ae7f2bSJakub Kicinski } 2173d9ae7f2bSJakub Kicinski 2174d9ae7f2bSJakub Kicinski nfp_prog->prog[br_idx] &= ~OP_BR_SPECIAL; 2175d9ae7f2bSJakub Kicinski } 2176d9ae7f2bSJakub Kicinski 2177d9ae7f2bSJakub Kicinski return 0; 2178d9ae7f2bSJakub Kicinski } 2179d9ae7f2bSJakub Kicinski 2180d9ae7f2bSJakub Kicinski static void nfp_intro(struct nfp_prog *nfp_prog) 2181d9ae7f2bSJakub Kicinski { 218218e53b6cSJakub Kicinski wrp_immed(nfp_prog, plen_reg(nfp_prog), GENMASK(13, 0)); 218318e53b6cSJakub Kicinski emit_alu(nfp_prog, plen_reg(nfp_prog), 218418e53b6cSJakub Kicinski plen_reg(nfp_prog), ALU_OP_AND, pv_len(nfp_prog)); 2185d9ae7f2bSJakub Kicinski } 2186d9ae7f2bSJakub Kicinski 2187d9ae7f2bSJakub Kicinski static void nfp_outro_tc_da(struct nfp_prog *nfp_prog) 2188d9ae7f2bSJakub Kicinski { 2189d9ae7f2bSJakub Kicinski /* TC direct-action mode: 2190d9ae7f2bSJakub Kicinski * 0,1 ok NOT SUPPORTED[1] 2191d9ae7f2bSJakub Kicinski * 2 drop 0x22 -> drop, count as stat1 2192d9ae7f2bSJakub Kicinski * 4,5 nuke 0x02 -> drop 2193d9ae7f2bSJakub Kicinski * 7 redir 0x44 -> redir, count as stat2 2194d9ae7f2bSJakub Kicinski * * unspec 0x11 -> pass, count as stat0 2195d9ae7f2bSJakub Kicinski * 2196d9ae7f2bSJakub Kicinski * [1] We can't support OK and RECLASSIFY because we can't tell TC 2197d9ae7f2bSJakub Kicinski * the exact decision made. We are forced to support UNSPEC 2198d9ae7f2bSJakub Kicinski * to handle aborts so that's the only one we handle for passing 2199d9ae7f2bSJakub Kicinski * packets up the stack. 2200d9ae7f2bSJakub Kicinski */ 2201d9ae7f2bSJakub Kicinski /* Target for aborts */ 2202d9ae7f2bSJakub Kicinski nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 2203d9ae7f2bSJakub Kicinski 2204d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2205d9ae7f2bSJakub Kicinski 2206c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2207d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x11), SHF_SC_L_SHF, 16); 2208d9ae7f2bSJakub Kicinski 2209d9ae7f2bSJakub Kicinski /* Target for normal exits */ 2210d9ae7f2bSJakub Kicinski nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 2211d9ae7f2bSJakub Kicinski 2212d9ae7f2bSJakub Kicinski /* if R0 > 7 jump to abort */ 2213d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_imm(7), ALU_OP_SUB, reg_b(0)); 2214d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BLO, nfp_prog->tgt_abort, 0); 2215c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2216d9ae7f2bSJakub Kicinski 2217d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_b(2), 0x41221211); 2218d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_b(3), 0x41001211); 2219d9ae7f2bSJakub Kicinski 2220d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_a(1), 2221d9ae7f2bSJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(0), SHF_SC_L_SHF, 2); 2222d9ae7f2bSJakub Kicinski 2223d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 2224d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_a(2), 2225d9ae7f2bSJakub Kicinski reg_imm(0xf), SHF_OP_AND, reg_b(2), SHF_SC_R_SHF, 0); 2226d9ae7f2bSJakub Kicinski 2227d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 2228d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_b(2), 2229d9ae7f2bSJakub Kicinski reg_imm(0xf), SHF_OP_AND, reg_b(3), SHF_SC_R_SHF, 0); 2230d9ae7f2bSJakub Kicinski 2231d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2232d9ae7f2bSJakub Kicinski 2233d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_b(2), 2234d9ae7f2bSJakub Kicinski reg_a(2), SHF_OP_OR, reg_b(2), SHF_SC_L_SHF, 4); 2235d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_b(2), SHF_SC_L_SHF, 16); 2236d9ae7f2bSJakub Kicinski } 2237d9ae7f2bSJakub Kicinski 2238d9ae7f2bSJakub Kicinski static void nfp_outro_xdp(struct nfp_prog *nfp_prog) 2239d9ae7f2bSJakub Kicinski { 2240d9ae7f2bSJakub Kicinski /* XDP return codes: 2241d9ae7f2bSJakub Kicinski * 0 aborted 0x82 -> drop, count as stat3 2242d9ae7f2bSJakub Kicinski * 1 drop 0x22 -> drop, count as stat1 2243d9ae7f2bSJakub Kicinski * 2 pass 0x11 -> pass, count as stat0 2244d9ae7f2bSJakub Kicinski * 3 tx 0x44 -> redir, count as stat2 2245d9ae7f2bSJakub Kicinski * * unknown 0x82 -> drop, count as stat3 2246d9ae7f2bSJakub Kicinski */ 2247d9ae7f2bSJakub Kicinski /* Target for aborts */ 2248d9ae7f2bSJakub Kicinski nfp_prog->tgt_abort = nfp_prog_current_offset(nfp_prog); 2249d9ae7f2bSJakub Kicinski 2250d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2251d9ae7f2bSJakub Kicinski 2252c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2253d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_imm(0x82), SHF_SC_L_SHF, 16); 2254d9ae7f2bSJakub Kicinski 2255d9ae7f2bSJakub Kicinski /* Target for normal exits */ 2256d9ae7f2bSJakub Kicinski nfp_prog->tgt_out = nfp_prog_current_offset(nfp_prog); 2257d9ae7f2bSJakub Kicinski 2258d9ae7f2bSJakub Kicinski /* if R0 > 3 jump to abort */ 2259d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_imm(3), ALU_OP_SUB, reg_b(0)); 2260d9ae7f2bSJakub Kicinski emit_br(nfp_prog, BR_BLO, nfp_prog->tgt_abort, 0); 2261d9ae7f2bSJakub Kicinski 2262d9ae7f2bSJakub Kicinski wrp_immed(nfp_prog, reg_b(2), 0x44112282); 2263d9ae7f2bSJakub Kicinski 2264d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_a(1), 2265d9ae7f2bSJakub Kicinski reg_none(), SHF_OP_NONE, reg_b(0), SHF_SC_L_SHF, 3); 2266d9ae7f2bSJakub Kicinski 2267d9ae7f2bSJakub Kicinski emit_alu(nfp_prog, reg_none(), reg_a(1), ALU_OP_OR, reg_imm(0)); 2268d9ae7f2bSJakub Kicinski emit_shf(nfp_prog, reg_b(2), 2269d9ae7f2bSJakub Kicinski reg_imm(0xff), SHF_OP_AND, reg_b(2), SHF_SC_R_SHF, 0); 2270d9ae7f2bSJakub Kicinski 2271d9ae7f2bSJakub Kicinski emit_br_def(nfp_prog, nfp_prog->tgt_done, 2); 2272d9ae7f2bSJakub Kicinski 2273c000dfb5SJakub Kicinski wrp_mov(nfp_prog, reg_a(0), NFP_BPF_ABI_FLAGS); 2274d9ae7f2bSJakub Kicinski emit_ld_field(nfp_prog, reg_a(0), 0xc, reg_b(2), SHF_SC_L_SHF, 16); 2275d9ae7f2bSJakub Kicinski } 2276d9ae7f2bSJakub Kicinski 2277d9ae7f2bSJakub Kicinski static void nfp_outro(struct nfp_prog *nfp_prog) 2278d9ae7f2bSJakub Kicinski { 2279012bb8a8SJakub Kicinski switch (nfp_prog->type) { 2280012bb8a8SJakub Kicinski case BPF_PROG_TYPE_SCHED_CLS: 2281d9ae7f2bSJakub Kicinski nfp_outro_tc_da(nfp_prog); 2282d9ae7f2bSJakub Kicinski break; 2283012bb8a8SJakub Kicinski case BPF_PROG_TYPE_XDP: 2284d9ae7f2bSJakub Kicinski nfp_outro_xdp(nfp_prog); 2285d9ae7f2bSJakub Kicinski break; 2286012bb8a8SJakub Kicinski default: 2287012bb8a8SJakub Kicinski WARN_ON(1); 2288d9ae7f2bSJakub Kicinski } 2289d9ae7f2bSJakub Kicinski } 2290d9ae7f2bSJakub Kicinski 2291d9ae7f2bSJakub Kicinski static int nfp_translate(struct nfp_prog *nfp_prog) 2292d9ae7f2bSJakub Kicinski { 2293d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta; 2294ff42bb9fSJakub Kicinski int err; 2295d9ae7f2bSJakub Kicinski 2296d9ae7f2bSJakub Kicinski nfp_intro(nfp_prog); 2297d9ae7f2bSJakub Kicinski if (nfp_prog->error) 2298d9ae7f2bSJakub Kicinski return nfp_prog->error; 2299d9ae7f2bSJakub Kicinski 2300d9ae7f2bSJakub Kicinski list_for_each_entry(meta, &nfp_prog->insns, l) { 2301d9ae7f2bSJakub Kicinski instr_cb_t cb = instr_cb[meta->insn.code]; 2302d9ae7f2bSJakub Kicinski 2303d9ae7f2bSJakub Kicinski meta->off = nfp_prog_current_offset(nfp_prog); 2304d9ae7f2bSJakub Kicinski 2305d9ae7f2bSJakub Kicinski if (meta->skip) { 2306d9ae7f2bSJakub Kicinski nfp_prog->n_translated++; 2307d9ae7f2bSJakub Kicinski continue; 2308d9ae7f2bSJakub Kicinski } 2309d9ae7f2bSJakub Kicinski 2310d9ae7f2bSJakub Kicinski if (nfp_meta_has_prev(nfp_prog, meta) && 2311d9ae7f2bSJakub Kicinski nfp_meta_prev(meta)->double_cb) 2312d9ae7f2bSJakub Kicinski cb = nfp_meta_prev(meta)->double_cb; 2313d9ae7f2bSJakub Kicinski if (!cb) 2314d9ae7f2bSJakub Kicinski return -ENOENT; 2315d9ae7f2bSJakub Kicinski err = cb(nfp_prog, meta); 2316d9ae7f2bSJakub Kicinski if (err) 2317d9ae7f2bSJakub Kicinski return err; 2318d9ae7f2bSJakub Kicinski 2319d9ae7f2bSJakub Kicinski nfp_prog->n_translated++; 2320d9ae7f2bSJakub Kicinski } 2321d9ae7f2bSJakub Kicinski 2322854dc87dSJiong Wang nfp_prog->last_bpf_off = nfp_prog_current_offset(nfp_prog) - 1; 2323854dc87dSJiong Wang 2324d9ae7f2bSJakub Kicinski nfp_outro(nfp_prog); 2325d9ae7f2bSJakub Kicinski if (nfp_prog->error) 2326d9ae7f2bSJakub Kicinski return nfp_prog->error; 2327d9ae7f2bSJakub Kicinski 2328ff42bb9fSJakub Kicinski wrp_nops(nfp_prog, NFP_USTORE_PREFETCH_WINDOW); 23291c03e03fSJakub Kicinski if (nfp_prog->error) 23301c03e03fSJakub Kicinski return nfp_prog->error; 23311c03e03fSJakub Kicinski 2332d9ae7f2bSJakub Kicinski return nfp_fixup_branches(nfp_prog); 2333d9ae7f2bSJakub Kicinski } 2334d9ae7f2bSJakub Kicinski 2335d9ae7f2bSJakub Kicinski /* --- Optimizations --- */ 2336d9ae7f2bSJakub Kicinski static void nfp_bpf_opt_reg_init(struct nfp_prog *nfp_prog) 2337d9ae7f2bSJakub Kicinski { 2338d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta; 2339d9ae7f2bSJakub Kicinski 2340d9ae7f2bSJakub Kicinski list_for_each_entry(meta, &nfp_prog->insns, l) { 2341d9ae7f2bSJakub Kicinski struct bpf_insn insn = meta->insn; 2342d9ae7f2bSJakub Kicinski 2343d9ae7f2bSJakub Kicinski /* Programs converted from cBPF start with register xoring */ 2344d9ae7f2bSJakub Kicinski if (insn.code == (BPF_ALU64 | BPF_XOR | BPF_X) && 2345d9ae7f2bSJakub Kicinski insn.src_reg == insn.dst_reg) 2346d9ae7f2bSJakub Kicinski continue; 2347d9ae7f2bSJakub Kicinski 2348d9ae7f2bSJakub Kicinski /* Programs start with R6 = R1 but we ignore the skb pointer */ 2349d9ae7f2bSJakub Kicinski if (insn.code == (BPF_ALU64 | BPF_MOV | BPF_X) && 2350d9ae7f2bSJakub Kicinski insn.src_reg == 1 && insn.dst_reg == 6) 2351d9ae7f2bSJakub Kicinski meta->skip = true; 2352d9ae7f2bSJakub Kicinski 2353d9ae7f2bSJakub Kicinski /* Return as soon as something doesn't match */ 2354d9ae7f2bSJakub Kicinski if (!meta->skip) 2355d9ae7f2bSJakub Kicinski return; 2356d9ae7f2bSJakub Kicinski } 2357d9ae7f2bSJakub Kicinski } 2358d9ae7f2bSJakub Kicinski 2359d9ae7f2bSJakub Kicinski /* Remove masking after load since our load guarantees this is not needed */ 2360d9ae7f2bSJakub Kicinski static void nfp_bpf_opt_ld_mask(struct nfp_prog *nfp_prog) 2361d9ae7f2bSJakub Kicinski { 2362d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta1, *meta2; 2363d9ae7f2bSJakub Kicinski const s32 exp_mask[] = { 2364d9ae7f2bSJakub Kicinski [BPF_B] = 0x000000ffU, 2365d9ae7f2bSJakub Kicinski [BPF_H] = 0x0000ffffU, 2366d9ae7f2bSJakub Kicinski [BPF_W] = 0xffffffffU, 2367d9ae7f2bSJakub Kicinski }; 2368d9ae7f2bSJakub Kicinski 2369d9ae7f2bSJakub Kicinski nfp_for_each_insn_walk2(nfp_prog, meta1, meta2) { 2370d9ae7f2bSJakub Kicinski struct bpf_insn insn, next; 2371d9ae7f2bSJakub Kicinski 2372d9ae7f2bSJakub Kicinski insn = meta1->insn; 2373d9ae7f2bSJakub Kicinski next = meta2->insn; 2374d9ae7f2bSJakub Kicinski 2375d9ae7f2bSJakub Kicinski if (BPF_CLASS(insn.code) != BPF_LD) 2376d9ae7f2bSJakub Kicinski continue; 2377d9ae7f2bSJakub Kicinski if (BPF_MODE(insn.code) != BPF_ABS && 2378d9ae7f2bSJakub Kicinski BPF_MODE(insn.code) != BPF_IND) 2379d9ae7f2bSJakub Kicinski continue; 2380d9ae7f2bSJakub Kicinski 2381d9ae7f2bSJakub Kicinski if (next.code != (BPF_ALU64 | BPF_AND | BPF_K)) 2382d9ae7f2bSJakub Kicinski continue; 2383d9ae7f2bSJakub Kicinski 2384d9ae7f2bSJakub Kicinski if (!exp_mask[BPF_SIZE(insn.code)]) 2385d9ae7f2bSJakub Kicinski continue; 2386d9ae7f2bSJakub Kicinski if (exp_mask[BPF_SIZE(insn.code)] != next.imm) 2387d9ae7f2bSJakub Kicinski continue; 2388d9ae7f2bSJakub Kicinski 2389d9ae7f2bSJakub Kicinski if (next.src_reg || next.dst_reg) 2390d9ae7f2bSJakub Kicinski continue; 2391d9ae7f2bSJakub Kicinski 23921266f5d6SJiong Wang if (meta2->flags & FLAG_INSN_IS_JUMP_DST) 23931266f5d6SJiong Wang continue; 23941266f5d6SJiong Wang 2395d9ae7f2bSJakub Kicinski meta2->skip = true; 2396d9ae7f2bSJakub Kicinski } 2397d9ae7f2bSJakub Kicinski } 2398d9ae7f2bSJakub Kicinski 2399d9ae7f2bSJakub Kicinski static void nfp_bpf_opt_ld_shift(struct nfp_prog *nfp_prog) 2400d9ae7f2bSJakub Kicinski { 2401d9ae7f2bSJakub Kicinski struct nfp_insn_meta *meta1, *meta2, *meta3; 2402d9ae7f2bSJakub Kicinski 2403d9ae7f2bSJakub Kicinski nfp_for_each_insn_walk3(nfp_prog, meta1, meta2, meta3) { 2404d9ae7f2bSJakub Kicinski struct bpf_insn insn, next1, next2; 2405d9ae7f2bSJakub Kicinski 2406d9ae7f2bSJakub Kicinski insn = meta1->insn; 2407d9ae7f2bSJakub Kicinski next1 = meta2->insn; 2408d9ae7f2bSJakub Kicinski next2 = meta3->insn; 2409d9ae7f2bSJakub Kicinski 2410d9ae7f2bSJakub Kicinski if (BPF_CLASS(insn.code) != BPF_LD) 2411d9ae7f2bSJakub Kicinski continue; 2412d9ae7f2bSJakub Kicinski if (BPF_MODE(insn.code) != BPF_ABS && 2413d9ae7f2bSJakub Kicinski BPF_MODE(insn.code) != BPF_IND) 2414d9ae7f2bSJakub Kicinski continue; 2415d9ae7f2bSJakub Kicinski if (BPF_SIZE(insn.code) != BPF_W) 2416d9ae7f2bSJakub Kicinski continue; 2417d9ae7f2bSJakub Kicinski 2418d9ae7f2bSJakub Kicinski if (!(next1.code == (BPF_LSH | BPF_K | BPF_ALU64) && 2419d9ae7f2bSJakub Kicinski next2.code == (BPF_RSH | BPF_K | BPF_ALU64)) && 2420d9ae7f2bSJakub Kicinski !(next1.code == (BPF_RSH | BPF_K | BPF_ALU64) && 2421d9ae7f2bSJakub Kicinski next2.code == (BPF_LSH | BPF_K | BPF_ALU64))) 2422d9ae7f2bSJakub Kicinski continue; 2423d9ae7f2bSJakub Kicinski 2424d9ae7f2bSJakub Kicinski if (next1.src_reg || next1.dst_reg || 2425d9ae7f2bSJakub Kicinski next2.src_reg || next2.dst_reg) 2426d9ae7f2bSJakub Kicinski continue; 2427d9ae7f2bSJakub Kicinski 2428d9ae7f2bSJakub Kicinski if (next1.imm != 0x20 || next2.imm != 0x20) 2429d9ae7f2bSJakub Kicinski continue; 2430d9ae7f2bSJakub Kicinski 243129fe46efSJiong Wang if (meta2->flags & FLAG_INSN_IS_JUMP_DST || 243229fe46efSJiong Wang meta3->flags & FLAG_INSN_IS_JUMP_DST) 243329fe46efSJiong Wang continue; 243429fe46efSJiong Wang 2435d9ae7f2bSJakub Kicinski meta2->skip = true; 2436d9ae7f2bSJakub Kicinski meta3->skip = true; 2437d9ae7f2bSJakub Kicinski } 2438d9ae7f2bSJakub Kicinski } 2439d9ae7f2bSJakub Kicinski 24406bc7103cSJiong Wang /* load/store pair that forms memory copy sould look like the following: 24416bc7103cSJiong Wang * 24426bc7103cSJiong Wang * ld_width R, [addr_src + offset_src] 24436bc7103cSJiong Wang * st_width [addr_dest + offset_dest], R 24446bc7103cSJiong Wang * 24456bc7103cSJiong Wang * The destination register of load and source register of store should 24466bc7103cSJiong Wang * be the same, load and store should also perform at the same width. 24476bc7103cSJiong Wang * If either of addr_src or addr_dest is stack pointer, we don't do the 24486bc7103cSJiong Wang * CPP optimization as stack is modelled by registers on NFP. 24496bc7103cSJiong Wang */ 24506bc7103cSJiong Wang static bool 24516bc7103cSJiong Wang curr_pair_is_memcpy(struct nfp_insn_meta *ld_meta, 24526bc7103cSJiong Wang struct nfp_insn_meta *st_meta) 24536bc7103cSJiong Wang { 24546bc7103cSJiong Wang struct bpf_insn *ld = &ld_meta->insn; 24556bc7103cSJiong Wang struct bpf_insn *st = &st_meta->insn; 24566bc7103cSJiong Wang 24576bc7103cSJiong Wang if (!is_mbpf_load(ld_meta) || !is_mbpf_store(st_meta)) 24586bc7103cSJiong Wang return false; 24596bc7103cSJiong Wang 24606bc7103cSJiong Wang if (ld_meta->ptr.type != PTR_TO_PACKET) 24616bc7103cSJiong Wang return false; 24626bc7103cSJiong Wang 24636bc7103cSJiong Wang if (st_meta->ptr.type != PTR_TO_PACKET) 24646bc7103cSJiong Wang return false; 24656bc7103cSJiong Wang 24666bc7103cSJiong Wang if (BPF_SIZE(ld->code) != BPF_SIZE(st->code)) 24676bc7103cSJiong Wang return false; 24686bc7103cSJiong Wang 24696bc7103cSJiong Wang if (ld->dst_reg != st->src_reg) 24706bc7103cSJiong Wang return false; 24716bc7103cSJiong Wang 24726bc7103cSJiong Wang /* There is jump to the store insn in this pair. */ 24736bc7103cSJiong Wang if (st_meta->flags & FLAG_INSN_IS_JUMP_DST) 24746bc7103cSJiong Wang return false; 24756bc7103cSJiong Wang 24766bc7103cSJiong Wang return true; 24776bc7103cSJiong Wang } 24786bc7103cSJiong Wang 24796bc7103cSJiong Wang /* Currently, we only support chaining load/store pairs if: 24806bc7103cSJiong Wang * 24816bc7103cSJiong Wang * - Their address base registers are the same. 24826bc7103cSJiong Wang * - Their address offsets are in the same order. 24836bc7103cSJiong Wang * - They operate at the same memory width. 24846bc7103cSJiong Wang * - There is no jump into the middle of them. 24856bc7103cSJiong Wang */ 24866bc7103cSJiong Wang static bool 24876bc7103cSJiong Wang curr_pair_chain_with_previous(struct nfp_insn_meta *ld_meta, 24886bc7103cSJiong Wang struct nfp_insn_meta *st_meta, 24896bc7103cSJiong Wang struct bpf_insn *prev_ld, 24906bc7103cSJiong Wang struct bpf_insn *prev_st) 24916bc7103cSJiong Wang { 24926bc7103cSJiong Wang u8 prev_size, curr_size, prev_ld_base, prev_st_base, prev_ld_dst; 24936bc7103cSJiong Wang struct bpf_insn *ld = &ld_meta->insn; 24946bc7103cSJiong Wang struct bpf_insn *st = &st_meta->insn; 24956bc7103cSJiong Wang s16 prev_ld_off, prev_st_off; 24966bc7103cSJiong Wang 24976bc7103cSJiong Wang /* This pair is the start pair. */ 24986bc7103cSJiong Wang if (!prev_ld) 24996bc7103cSJiong Wang return true; 25006bc7103cSJiong Wang 25016bc7103cSJiong Wang prev_size = BPF_LDST_BYTES(prev_ld); 25026bc7103cSJiong Wang curr_size = BPF_LDST_BYTES(ld); 25036bc7103cSJiong Wang prev_ld_base = prev_ld->src_reg; 25046bc7103cSJiong Wang prev_st_base = prev_st->dst_reg; 25056bc7103cSJiong Wang prev_ld_dst = prev_ld->dst_reg; 25066bc7103cSJiong Wang prev_ld_off = prev_ld->off; 25076bc7103cSJiong Wang prev_st_off = prev_st->off; 25086bc7103cSJiong Wang 25096bc7103cSJiong Wang if (ld->dst_reg != prev_ld_dst) 25106bc7103cSJiong Wang return false; 25116bc7103cSJiong Wang 25126bc7103cSJiong Wang if (ld->src_reg != prev_ld_base || st->dst_reg != prev_st_base) 25136bc7103cSJiong Wang return false; 25146bc7103cSJiong Wang 25156bc7103cSJiong Wang if (curr_size != prev_size) 25166bc7103cSJiong Wang return false; 25176bc7103cSJiong Wang 25186bc7103cSJiong Wang /* There is jump to the head of this pair. */ 25196bc7103cSJiong Wang if (ld_meta->flags & FLAG_INSN_IS_JUMP_DST) 25206bc7103cSJiong Wang return false; 25216bc7103cSJiong Wang 25226bc7103cSJiong Wang /* Both in ascending order. */ 25236bc7103cSJiong Wang if (prev_ld_off + prev_size == ld->off && 25246bc7103cSJiong Wang prev_st_off + prev_size == st->off) 25256bc7103cSJiong Wang return true; 25266bc7103cSJiong Wang 25276bc7103cSJiong Wang /* Both in descending order. */ 25286bc7103cSJiong Wang if (ld->off + curr_size == prev_ld_off && 25296bc7103cSJiong Wang st->off + curr_size == prev_st_off) 25306bc7103cSJiong Wang return true; 25316bc7103cSJiong Wang 25326bc7103cSJiong Wang return false; 25336bc7103cSJiong Wang } 25346bc7103cSJiong Wang 25356bc7103cSJiong Wang /* Return TRUE if cross memory access happens. Cross memory access means 25366bc7103cSJiong Wang * store area is overlapping with load area that a later load might load 25376bc7103cSJiong Wang * the value from previous store, for this case we can't treat the sequence 25386bc7103cSJiong Wang * as an memory copy. 25396bc7103cSJiong Wang */ 25406bc7103cSJiong Wang static bool 25416bc7103cSJiong Wang cross_mem_access(struct bpf_insn *ld, struct nfp_insn_meta *head_ld_meta, 25426bc7103cSJiong Wang struct nfp_insn_meta *head_st_meta) 25436bc7103cSJiong Wang { 25446bc7103cSJiong Wang s16 head_ld_off, head_st_off, ld_off; 25456bc7103cSJiong Wang 25466bc7103cSJiong Wang /* Different pointer types does not overlap. */ 25476bc7103cSJiong Wang if (head_ld_meta->ptr.type != head_st_meta->ptr.type) 25486bc7103cSJiong Wang return false; 25496bc7103cSJiong Wang 25506bc7103cSJiong Wang /* load and store are both PTR_TO_PACKET, check ID info. */ 25516bc7103cSJiong Wang if (head_ld_meta->ptr.id != head_st_meta->ptr.id) 25526bc7103cSJiong Wang return true; 25536bc7103cSJiong Wang 25546bc7103cSJiong Wang /* Canonicalize the offsets. Turn all of them against the original 25556bc7103cSJiong Wang * base register. 25566bc7103cSJiong Wang */ 25576bc7103cSJiong Wang head_ld_off = head_ld_meta->insn.off + head_ld_meta->ptr.off; 25586bc7103cSJiong Wang head_st_off = head_st_meta->insn.off + head_st_meta->ptr.off; 25596bc7103cSJiong Wang ld_off = ld->off + head_ld_meta->ptr.off; 25606bc7103cSJiong Wang 25616bc7103cSJiong Wang /* Ascending order cross. */ 25626bc7103cSJiong Wang if (ld_off > head_ld_off && 25636bc7103cSJiong Wang head_ld_off < head_st_off && ld_off >= head_st_off) 25646bc7103cSJiong Wang return true; 25656bc7103cSJiong Wang 25666bc7103cSJiong Wang /* Descending order cross. */ 25676bc7103cSJiong Wang if (ld_off < head_ld_off && 25686bc7103cSJiong Wang head_ld_off > head_st_off && ld_off <= head_st_off) 25696bc7103cSJiong Wang return true; 25706bc7103cSJiong Wang 25716bc7103cSJiong Wang return false; 25726bc7103cSJiong Wang } 25736bc7103cSJiong Wang 25746bc7103cSJiong Wang /* This pass try to identify the following instructoin sequences. 25756bc7103cSJiong Wang * 25766bc7103cSJiong Wang * load R, [regA + offA] 25776bc7103cSJiong Wang * store [regB + offB], R 25786bc7103cSJiong Wang * load R, [regA + offA + const_imm_A] 25796bc7103cSJiong Wang * store [regB + offB + const_imm_A], R 25806bc7103cSJiong Wang * load R, [regA + offA + 2 * const_imm_A] 25816bc7103cSJiong Wang * store [regB + offB + 2 * const_imm_A], R 25826bc7103cSJiong Wang * ... 25836bc7103cSJiong Wang * 25846bc7103cSJiong Wang * Above sequence is typically generated by compiler when lowering 25856bc7103cSJiong Wang * memcpy. NFP prefer using CPP instructions to accelerate it. 25866bc7103cSJiong Wang */ 25876bc7103cSJiong Wang static void nfp_bpf_opt_ldst_gather(struct nfp_prog *nfp_prog) 25886bc7103cSJiong Wang { 25896bc7103cSJiong Wang struct nfp_insn_meta *head_ld_meta = NULL; 25906bc7103cSJiong Wang struct nfp_insn_meta *head_st_meta = NULL; 25916bc7103cSJiong Wang struct nfp_insn_meta *meta1, *meta2; 25926bc7103cSJiong Wang struct bpf_insn *prev_ld = NULL; 25936bc7103cSJiong Wang struct bpf_insn *prev_st = NULL; 25946bc7103cSJiong Wang u8 count = 0; 25956bc7103cSJiong Wang 25966bc7103cSJiong Wang nfp_for_each_insn_walk2(nfp_prog, meta1, meta2) { 25976bc7103cSJiong Wang struct bpf_insn *ld = &meta1->insn; 25986bc7103cSJiong Wang struct bpf_insn *st = &meta2->insn; 25996bc7103cSJiong Wang 26006bc7103cSJiong Wang /* Reset record status if any of the following if true: 26016bc7103cSJiong Wang * - The current insn pair is not load/store. 26026bc7103cSJiong Wang * - The load/store pair doesn't chain with previous one. 26036bc7103cSJiong Wang * - The chained load/store pair crossed with previous pair. 26046bc7103cSJiong Wang * - The chained load/store pair has a total size of memory 26056bc7103cSJiong Wang * copy beyond 128 bytes which is the maximum length a 26066bc7103cSJiong Wang * single NFP CPP command can transfer. 26076bc7103cSJiong Wang */ 26086bc7103cSJiong Wang if (!curr_pair_is_memcpy(meta1, meta2) || 26096bc7103cSJiong Wang !curr_pair_chain_with_previous(meta1, meta2, prev_ld, 26106bc7103cSJiong Wang prev_st) || 26116bc7103cSJiong Wang (head_ld_meta && (cross_mem_access(ld, head_ld_meta, 26126bc7103cSJiong Wang head_st_meta) || 26136bc7103cSJiong Wang head_ld_meta->ldst_gather_len >= 128))) { 26146bc7103cSJiong Wang if (!count) 26156bc7103cSJiong Wang continue; 26166bc7103cSJiong Wang 26176bc7103cSJiong Wang if (count > 1) { 26186bc7103cSJiong Wang s16 prev_ld_off = prev_ld->off; 26196bc7103cSJiong Wang s16 prev_st_off = prev_st->off; 26206bc7103cSJiong Wang s16 head_ld_off = head_ld_meta->insn.off; 26216bc7103cSJiong Wang 26226bc7103cSJiong Wang if (prev_ld_off < head_ld_off) { 26236bc7103cSJiong Wang head_ld_meta->insn.off = prev_ld_off; 26246bc7103cSJiong Wang head_st_meta->insn.off = prev_st_off; 26256bc7103cSJiong Wang head_ld_meta->ldst_gather_len = 26266bc7103cSJiong Wang -head_ld_meta->ldst_gather_len; 26276bc7103cSJiong Wang } 26286bc7103cSJiong Wang 26296bc7103cSJiong Wang head_ld_meta->paired_st = &head_st_meta->insn; 26306bc7103cSJiong Wang head_st_meta->skip = true; 26316bc7103cSJiong Wang } else { 26326bc7103cSJiong Wang head_ld_meta->ldst_gather_len = 0; 26336bc7103cSJiong Wang } 26346bc7103cSJiong Wang 26356bc7103cSJiong Wang /* If the chain is ended by an load/store pair then this 26366bc7103cSJiong Wang * could serve as the new head of the the next chain. 26376bc7103cSJiong Wang */ 26386bc7103cSJiong Wang if (curr_pair_is_memcpy(meta1, meta2)) { 26396bc7103cSJiong Wang head_ld_meta = meta1; 26406bc7103cSJiong Wang head_st_meta = meta2; 26416bc7103cSJiong Wang head_ld_meta->ldst_gather_len = 26426bc7103cSJiong Wang BPF_LDST_BYTES(ld); 26436bc7103cSJiong Wang meta1 = nfp_meta_next(meta1); 26446bc7103cSJiong Wang meta2 = nfp_meta_next(meta2); 26456bc7103cSJiong Wang prev_ld = ld; 26466bc7103cSJiong Wang prev_st = st; 26476bc7103cSJiong Wang count = 1; 26486bc7103cSJiong Wang } else { 26496bc7103cSJiong Wang head_ld_meta = NULL; 26506bc7103cSJiong Wang head_st_meta = NULL; 26516bc7103cSJiong Wang prev_ld = NULL; 26526bc7103cSJiong Wang prev_st = NULL; 26536bc7103cSJiong Wang count = 0; 26546bc7103cSJiong Wang } 26556bc7103cSJiong Wang 26566bc7103cSJiong Wang continue; 26576bc7103cSJiong Wang } 26586bc7103cSJiong Wang 26596bc7103cSJiong Wang if (!head_ld_meta) { 26606bc7103cSJiong Wang head_ld_meta = meta1; 26616bc7103cSJiong Wang head_st_meta = meta2; 26626bc7103cSJiong Wang } else { 26636bc7103cSJiong Wang meta1->skip = true; 26646bc7103cSJiong Wang meta2->skip = true; 26656bc7103cSJiong Wang } 26666bc7103cSJiong Wang 26676bc7103cSJiong Wang head_ld_meta->ldst_gather_len += BPF_LDST_BYTES(ld); 26686bc7103cSJiong Wang meta1 = nfp_meta_next(meta1); 26696bc7103cSJiong Wang meta2 = nfp_meta_next(meta2); 26706bc7103cSJiong Wang prev_ld = ld; 26716bc7103cSJiong Wang prev_st = st; 26726bc7103cSJiong Wang count++; 26736bc7103cSJiong Wang } 26746bc7103cSJiong Wang } 26756bc7103cSJiong Wang 2676d9ae7f2bSJakub Kicinski static int nfp_bpf_optimize(struct nfp_prog *nfp_prog) 2677d9ae7f2bSJakub Kicinski { 2678d9ae7f2bSJakub Kicinski nfp_bpf_opt_reg_init(nfp_prog); 2679d9ae7f2bSJakub Kicinski 2680d9ae7f2bSJakub Kicinski nfp_bpf_opt_ld_mask(nfp_prog); 2681d9ae7f2bSJakub Kicinski nfp_bpf_opt_ld_shift(nfp_prog); 26826bc7103cSJiong Wang nfp_bpf_opt_ldst_gather(nfp_prog); 2683d9ae7f2bSJakub Kicinski 2684d9ae7f2bSJakub Kicinski return 0; 2685d9ae7f2bSJakub Kicinski } 2686d9ae7f2bSJakub Kicinski 26872e85d388SJakub Kicinski static int nfp_bpf_ustore_calc(struct nfp_prog *nfp_prog, __le64 *ustore) 2688fd068ddcSJakub Kicinski { 2689fd068ddcSJakub Kicinski int i; 2690fd068ddcSJakub Kicinski 2691fd068ddcSJakub Kicinski for (i = 0; i < nfp_prog->prog_len; i++) { 2692fd068ddcSJakub Kicinski int err; 2693fd068ddcSJakub Kicinski 2694fd068ddcSJakub Kicinski err = nfp_ustore_check_valid_no_ecc(nfp_prog->prog[i]); 2695fd068ddcSJakub Kicinski if (err) 2696fd068ddcSJakub Kicinski return err; 2697fd068ddcSJakub Kicinski 2698fd068ddcSJakub Kicinski nfp_prog->prog[i] = nfp_ustore_calc_ecc_insn(nfp_prog->prog[i]); 26992e85d388SJakub Kicinski 27002e85d388SJakub Kicinski ustore[i] = cpu_to_le64(nfp_prog->prog[i]); 2701fd068ddcSJakub Kicinski } 2702fd068ddcSJakub Kicinski 2703fd068ddcSJakub Kicinski return 0; 2704fd068ddcSJakub Kicinski } 2705fd068ddcSJakub Kicinski 2706c6c580d7SJakub Kicinski int nfp_bpf_jit(struct nfp_prog *nfp_prog) 2707d9ae7f2bSJakub Kicinski { 2708d9ae7f2bSJakub Kicinski int ret; 2709d9ae7f2bSJakub Kicinski 2710d9ae7f2bSJakub Kicinski ret = nfp_bpf_optimize(nfp_prog); 2711d9ae7f2bSJakub Kicinski if (ret) 27129314c442SJakub Kicinski return ret; 2713d9ae7f2bSJakub Kicinski 2714d9ae7f2bSJakub Kicinski ret = nfp_translate(nfp_prog); 2715d9ae7f2bSJakub Kicinski if (ret) { 2716d9ae7f2bSJakub Kicinski pr_err("Translation failed with error %d (translated: %u)\n", 2717d9ae7f2bSJakub Kicinski ret, nfp_prog->n_translated); 27189314c442SJakub Kicinski return -EINVAL; 2719d9ae7f2bSJakub Kicinski } 2720d9ae7f2bSJakub Kicinski 27219314c442SJakub Kicinski return nfp_bpf_ustore_calc(nfp_prog, (__force __le64 *)nfp_prog->prog); 2722d9ae7f2bSJakub Kicinski } 2723