// SPDX-License-Identifier: GPL-2.0+ /* Microchip Sparx5 Switch driver VCAP implementation * * Copyright (c) 2022 Microchip Technology Inc. and its subsidiaries. * * The Sparx5 Chip Register Model can be browsed at this location: * https://github.com/microchip-ung/sparx-5_reginfo */ #include #include #include "vcap_api.h" #include "vcap_api_client.h" #include "sparx5_main_regs.h" #include "sparx5_main.h" #include "sparx5_vcap_impl.h" #include "sparx5_vcap_ag_api.h" #define SUPER_VCAP_BLK_SIZE 3072 /* addresses per Super VCAP block */ #define STREAMSIZE (64 * 4) /* bytes in the VCAP cache area */ #define SPARX5_IS2_LOOKUPS 4 /* IS2 port keyset selection control */ /* IS2 non-ethernet traffic type keyset generation */ enum vcap_is2_port_sel_noneth { VCAP_IS2_PS_NONETH_MAC_ETYPE, VCAP_IS2_PS_NONETH_CUSTOM_1, VCAP_IS2_PS_NONETH_CUSTOM_2, VCAP_IS2_PS_NONETH_NO_LOOKUP }; /* IS2 IPv4 unicast traffic type keyset generation */ enum vcap_is2_port_sel_ipv4_uc { VCAP_IS2_PS_IPV4_UC_MAC_ETYPE, VCAP_IS2_PS_IPV4_UC_IP4_TCP_UDP_OTHER, VCAP_IS2_PS_IPV4_UC_IP_7TUPLE, }; /* IS2 IPv4 multicast traffic type keyset generation */ enum vcap_is2_port_sel_ipv4_mc { VCAP_IS2_PS_IPV4_MC_MAC_ETYPE, VCAP_IS2_PS_IPV4_MC_IP4_TCP_UDP_OTHER, VCAP_IS2_PS_IPV4_MC_IP_7TUPLE, VCAP_IS2_PS_IPV4_MC_IP4_VID, }; /* IS2 IPv6 unicast traffic type keyset generation */ enum vcap_is2_port_sel_ipv6_uc { VCAP_IS2_PS_IPV6_UC_MAC_ETYPE, VCAP_IS2_PS_IPV6_UC_IP_7TUPLE, VCAP_IS2_PS_IPV6_UC_IP6_STD, VCAP_IS2_PS_IPV6_UC_IP4_TCP_UDP_OTHER, }; /* IS2 IPv6 multicast traffic type keyset generation */ enum vcap_is2_port_sel_ipv6_mc { VCAP_IS2_PS_IPV6_MC_MAC_ETYPE, VCAP_IS2_PS_IPV6_MC_IP_7TUPLE, VCAP_IS2_PS_IPV6_MC_IP6_VID, VCAP_IS2_PS_IPV6_MC_IP6_STD, VCAP_IS2_PS_IPV6_MC_IP4_TCP_UDP_OTHER, }; /* IS2 ARP traffic type keyset generation */ enum vcap_is2_port_sel_arp { VCAP_IS2_PS_ARP_MAC_ETYPE, VCAP_IS2_PS_ARP_ARP, }; static struct sparx5_vcap_inst { enum vcap_type vtype; /* type of vcap */ int vinst; /* instance number within the same type */ int lookups; /* number of lookups in this vcap type */ int lookups_per_instance; /* number of lookups in this instance */ int first_cid; /* first chain id in this vcap */ int last_cid; /* last chain id in this vcap */ int count; /* number of available addresses, not in super vcap */ int map_id; /* id in the super vcap block mapping (if applicable) */ int blockno; /* starting block in super vcap (if applicable) */ int blocks; /* number of blocks in super vcap (if applicable) */ } sparx5_vcap_inst_cfg[] = { { .vtype = VCAP_TYPE_IS2, /* IS2-0 */ .vinst = 0, .map_id = 4, .lookups = SPARX5_IS2_LOOKUPS, .lookups_per_instance = SPARX5_IS2_LOOKUPS / 2, .first_cid = SPARX5_VCAP_CID_IS2_L0, .last_cid = SPARX5_VCAP_CID_IS2_L2 - 1, .blockno = 0, /* Maps block 0-1 */ .blocks = 2, }, { .vtype = VCAP_TYPE_IS2, /* IS2-1 */ .vinst = 1, .map_id = 5, .lookups = SPARX5_IS2_LOOKUPS, .lookups_per_instance = SPARX5_IS2_LOOKUPS / 2, .first_cid = SPARX5_VCAP_CID_IS2_L2, .last_cid = SPARX5_VCAP_CID_IS2_MAX, .blockno = 2, /* Maps block 2-3 */ .blocks = 2, }, }; /* Await the super VCAP completion of the current operation */ static void sparx5_vcap_wait_super_update(struct sparx5 *sparx5) { u32 value; read_poll_timeout(spx5_rd, value, !VCAP_SUPER_CTRL_UPDATE_SHOT_GET(value), 500, 10000, false, sparx5, VCAP_SUPER_CTRL); } /* Initializing a VCAP address range: only IS2 for now */ static void _sparx5_vcap_range_init(struct sparx5 *sparx5, struct vcap_admin *admin, u32 addr, u32 count) { u32 size = count - 1; spx5_wr(VCAP_SUPER_CFG_MV_NUM_POS_SET(0) | VCAP_SUPER_CFG_MV_SIZE_SET(size), sparx5, VCAP_SUPER_CFG); spx5_wr(VCAP_SUPER_CTRL_UPDATE_CMD_SET(VCAP_CMD_INITIALIZE) | VCAP_SUPER_CTRL_UPDATE_ENTRY_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_ACTION_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_CNT_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_ADDR_SET(addr) | VCAP_SUPER_CTRL_CLEAR_CACHE_SET(true) | VCAP_SUPER_CTRL_UPDATE_SHOT_SET(true), sparx5, VCAP_SUPER_CTRL); sparx5_vcap_wait_super_update(sparx5); } /* Initializing VCAP rule data area */ static void sparx5_vcap_block_init(struct sparx5 *sparx5, struct vcap_admin *admin) { _sparx5_vcap_range_init(sparx5, admin, admin->first_valid_addr, admin->last_valid_addr - admin->first_valid_addr); } /* Get the keyset name from the sparx5 VCAP model */ static const char *sparx5_vcap_keyset_name(struct net_device *ndev, enum vcap_keyfield_set keyset) { struct sparx5_port *port = netdev_priv(ndev); return port->sparx5->vcap_ctrl->stats->keyfield_set_names[keyset]; } /* Check if this is the first lookup of IS2 */ static bool sparx5_vcap_is2_is_first_chain(struct vcap_rule *rule) { return (rule->vcap_chain_id >= SPARX5_VCAP_CID_IS2_L0 && rule->vcap_chain_id < SPARX5_VCAP_CID_IS2_L1) || ((rule->vcap_chain_id >= SPARX5_VCAP_CID_IS2_L2 && rule->vcap_chain_id < SPARX5_VCAP_CID_IS2_L3)); } /* Set the narrow range ingress port mask on a rule */ static void sparx5_vcap_add_range_port_mask(struct vcap_rule *rule, struct net_device *ndev) { struct sparx5_port *port = netdev_priv(ndev); u32 port_mask; u32 range; range = port->portno / BITS_PER_TYPE(u32); /* Port bit set to match-any */ port_mask = ~BIT(port->portno % BITS_PER_TYPE(u32)); vcap_rule_add_key_u32(rule, VCAP_KF_IF_IGR_PORT_MASK_SEL, 0, 0xf); vcap_rule_add_key_u32(rule, VCAP_KF_IF_IGR_PORT_MASK_RNG, range, 0xf); vcap_rule_add_key_u32(rule, VCAP_KF_IF_IGR_PORT_MASK, 0, port_mask); } /* Set the wide range ingress port mask on a rule */ static void sparx5_vcap_add_wide_port_mask(struct vcap_rule *rule, struct net_device *ndev) { struct sparx5_port *port = netdev_priv(ndev); struct vcap_u72_key port_mask; u32 range; /* Port bit set to match-any */ memset(port_mask.value, 0, sizeof(port_mask.value)); memset(port_mask.mask, 0xff, sizeof(port_mask.mask)); range = port->portno / BITS_PER_BYTE; port_mask.mask[range] = ~BIT(port->portno % BITS_PER_BYTE); vcap_rule_add_key_u72(rule, VCAP_KF_IF_IGR_PORT_MASK, &port_mask); } /* API callback used for validating a field keyset (check the port keysets) */ static enum vcap_keyfield_set sparx5_vcap_validate_keyset(struct net_device *ndev, struct vcap_admin *admin, struct vcap_rule *rule, struct vcap_keyset_list *kslist, u16 l3_proto) { if (!kslist || kslist->cnt == 0) return VCAP_KFS_NO_VALUE; /* for now just return whatever the API suggests */ return kslist->keysets[0]; } /* API callback used for adding default fields to a rule */ static void sparx5_vcap_add_default_fields(struct net_device *ndev, struct vcap_admin *admin, struct vcap_rule *rule) { const struct vcap_field *field; field = vcap_lookup_keyfield(rule, VCAP_KF_IF_IGR_PORT_MASK); if (field && field->width == SPX5_PORTS) sparx5_vcap_add_wide_port_mask(rule, ndev); else if (field && field->width == BITS_PER_TYPE(u32)) sparx5_vcap_add_range_port_mask(rule, ndev); else pr_err("%s:%d: %s: could not add an ingress port mask for: %s\n", __func__, __LINE__, netdev_name(ndev), sparx5_vcap_keyset_name(ndev, rule->keyset)); /* add the lookup bit */ if (sparx5_vcap_is2_is_first_chain(rule)) vcap_rule_add_key_bit(rule, VCAP_KF_LOOKUP_FIRST_IS, VCAP_BIT_1); else vcap_rule_add_key_bit(rule, VCAP_KF_LOOKUP_FIRST_IS, VCAP_BIT_0); } /* API callback used for erasing the vcap cache area (not the register area) */ static void sparx5_vcap_cache_erase(struct vcap_admin *admin) { memset(admin->cache.keystream, 0, STREAMSIZE); memset(admin->cache.maskstream, 0, STREAMSIZE); memset(admin->cache.actionstream, 0, STREAMSIZE); memset(&admin->cache.counter, 0, sizeof(admin->cache.counter)); } /* API callback used for writing to the VCAP cache */ static void sparx5_vcap_cache_write(struct net_device *ndev, struct vcap_admin *admin, enum vcap_selection sel, u32 start, u32 count) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; u32 *keystr, *mskstr, *actstr; int idx; keystr = &admin->cache.keystream[start]; mskstr = &admin->cache.maskstream[start]; actstr = &admin->cache.actionstream[start]; switch (sel) { case VCAP_SEL_ENTRY: for (idx = 0; idx < count; ++idx) { /* Avoid 'match-off' by setting value & mask */ spx5_wr(keystr[idx] & mskstr[idx], sparx5, VCAP_SUPER_VCAP_ENTRY_DAT(idx)); spx5_wr(~mskstr[idx], sparx5, VCAP_SUPER_VCAP_MASK_DAT(idx)); } break; case VCAP_SEL_ACTION: for (idx = 0; idx < count; ++idx) spx5_wr(actstr[idx], sparx5, VCAP_SUPER_VCAP_ACTION_DAT(idx)); break; case VCAP_SEL_ALL: pr_err("%s:%d: cannot write all streams at once\n", __func__, __LINE__); break; default: break; } } /* API callback used for reading from the VCAP into the VCAP cache */ static void sparx5_vcap_cache_read(struct net_device *ndev, struct vcap_admin *admin, enum vcap_selection sel, u32 start, u32 count) { /* this will be added later */ } /* API callback used for initializing a VCAP address range */ static void sparx5_vcap_range_init(struct net_device *ndev, struct vcap_admin *admin, u32 addr, u32 count) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; _sparx5_vcap_range_init(sparx5, admin, addr, count); } /* API callback used for updating the VCAP cache */ static void sparx5_vcap_update(struct net_device *ndev, struct vcap_admin *admin, enum vcap_command cmd, enum vcap_selection sel, u32 addr) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; bool clear; clear = (cmd == VCAP_CMD_INITIALIZE); spx5_wr(VCAP_SUPER_CFG_MV_NUM_POS_SET(0) | VCAP_SUPER_CFG_MV_SIZE_SET(0), sparx5, VCAP_SUPER_CFG); spx5_wr(VCAP_SUPER_CTRL_UPDATE_CMD_SET(cmd) | VCAP_SUPER_CTRL_UPDATE_ENTRY_DIS_SET((VCAP_SEL_ENTRY & sel) == 0) | VCAP_SUPER_CTRL_UPDATE_ACTION_DIS_SET((VCAP_SEL_ACTION & sel) == 0) | VCAP_SUPER_CTRL_UPDATE_CNT_DIS_SET((VCAP_SEL_COUNTER & sel) == 0) | VCAP_SUPER_CTRL_UPDATE_ADDR_SET(addr) | VCAP_SUPER_CTRL_CLEAR_CACHE_SET(clear) | VCAP_SUPER_CTRL_UPDATE_SHOT_SET(true), sparx5, VCAP_SUPER_CTRL); sparx5_vcap_wait_super_update(sparx5); } /* API callback used for moving a block of rules in the VCAP */ static void sparx5_vcap_move(struct net_device *ndev, struct vcap_admin *admin, u32 addr, int offset, int count) { /* this will be added later */ } /* Provide port information via a callback interface */ static int sparx5_port_info(struct net_device *ndev, enum vcap_type vtype, int (*pf)(void *out, int arg, const char *fmt, ...), void *out, int arg) { /* this will be added later */ return 0; } /* API callback operations: only IS2 is supported for now */ static struct vcap_operations sparx5_vcap_ops = { .validate_keyset = sparx5_vcap_validate_keyset, .add_default_fields = sparx5_vcap_add_default_fields, .cache_erase = sparx5_vcap_cache_erase, .cache_write = sparx5_vcap_cache_write, .cache_read = sparx5_vcap_cache_read, .init = sparx5_vcap_range_init, .update = sparx5_vcap_update, .move = sparx5_vcap_move, .port_info = sparx5_port_info, }; /* Enable lookups per port and set the keyset generation: only IS2 for now */ static void sparx5_vcap_port_key_selection(struct sparx5 *sparx5, struct vcap_admin *admin) { int portno, lookup; u32 keysel; /* enable all 4 lookups on all ports */ for (portno = 0; portno < SPX5_PORTS; ++portno) spx5_wr(ANA_ACL_VCAP_S2_CFG_SEC_ENA_SET(0xf), sparx5, ANA_ACL_VCAP_S2_CFG(portno)); /* all traffic types generate the MAC_ETYPE keyset for now in all * lookups on all ports */ keysel = ANA_ACL_VCAP_S2_KEY_SEL_KEY_SEL_ENA_SET(true) | ANA_ACL_VCAP_S2_KEY_SEL_NON_ETH_KEY_SEL_SET(VCAP_IS2_PS_NONETH_MAC_ETYPE) | ANA_ACL_VCAP_S2_KEY_SEL_IP4_MC_KEY_SEL_SET(VCAP_IS2_PS_IPV4_MC_MAC_ETYPE) | ANA_ACL_VCAP_S2_KEY_SEL_IP4_UC_KEY_SEL_SET(VCAP_IS2_PS_IPV4_UC_MAC_ETYPE) | ANA_ACL_VCAP_S2_KEY_SEL_IP6_MC_KEY_SEL_SET(VCAP_IS2_PS_IPV6_MC_MAC_ETYPE) | ANA_ACL_VCAP_S2_KEY_SEL_IP6_UC_KEY_SEL_SET(VCAP_IS2_PS_IPV6_UC_MAC_ETYPE) | ANA_ACL_VCAP_S2_KEY_SEL_ARP_KEY_SEL_SET(VCAP_IS2_PS_ARP_MAC_ETYPE); for (lookup = 0; lookup < admin->lookups; ++lookup) { for (portno = 0; portno < SPX5_PORTS; ++portno) { spx5_wr(keysel, sparx5, ANA_ACL_VCAP_S2_KEY_SEL(portno, lookup)); } } } /* Disable lookups per port and set the keyset generation: only IS2 for now */ static void sparx5_vcap_port_key_deselection(struct sparx5 *sparx5, struct vcap_admin *admin) { int portno; for (portno = 0; portno < SPX5_PORTS; ++portno) spx5_rmw(ANA_ACL_VCAP_S2_CFG_SEC_ENA_SET(0), ANA_ACL_VCAP_S2_CFG_SEC_ENA, sparx5, ANA_ACL_VCAP_S2_CFG(portno)); } static void sparx5_vcap_admin_free(struct vcap_admin *admin) { if (!admin) return; kfree(admin->cache.keystream); kfree(admin->cache.maskstream); kfree(admin->cache.actionstream); kfree(admin); } /* Allocate a vcap instance with a rule list and a cache area */ static struct vcap_admin * sparx5_vcap_admin_alloc(struct sparx5 *sparx5, struct vcap_control *ctrl, const struct sparx5_vcap_inst *cfg) { struct vcap_admin *admin; admin = kzalloc(sizeof(*admin), GFP_KERNEL); if (!admin) return ERR_PTR(-ENOMEM); INIT_LIST_HEAD(&admin->list); INIT_LIST_HEAD(&admin->rules); admin->vtype = cfg->vtype; admin->vinst = cfg->vinst; admin->lookups = cfg->lookups; admin->lookups_per_instance = cfg->lookups_per_instance; admin->first_cid = cfg->first_cid; admin->last_cid = cfg->last_cid; admin->cache.keystream = kzalloc(STREAMSIZE, GFP_KERNEL); admin->cache.maskstream = kzalloc(STREAMSIZE, GFP_KERNEL); admin->cache.actionstream = kzalloc(STREAMSIZE, GFP_KERNEL); if (!admin->cache.keystream || !admin->cache.maskstream || !admin->cache.actionstream) { sparx5_vcap_admin_free(admin); return ERR_PTR(-ENOMEM); } return admin; } /* Do block allocations and provide addresses for VCAP instances */ static void sparx5_vcap_block_alloc(struct sparx5 *sparx5, struct vcap_admin *admin, const struct sparx5_vcap_inst *cfg) { int idx; /* Super VCAP block mapping and address configuration. Block 0 * is assigned addresses 0 through 3071, block 1 is assigned * addresses 3072 though 6143, and so on. */ for (idx = cfg->blockno; idx < cfg->blockno + cfg->blocks; ++idx) { spx5_wr(VCAP_SUPER_IDX_CORE_IDX_SET(idx), sparx5, VCAP_SUPER_IDX); spx5_wr(VCAP_SUPER_MAP_CORE_MAP_SET(cfg->map_id), sparx5, VCAP_SUPER_MAP); } admin->first_valid_addr = cfg->blockno * SUPER_VCAP_BLK_SIZE; admin->last_used_addr = admin->first_valid_addr + cfg->blocks * SUPER_VCAP_BLK_SIZE; admin->last_valid_addr = admin->last_used_addr - 1; } /* Allocate a vcap control and vcap instances and configure the system */ int sparx5_vcap_init(struct sparx5 *sparx5) { const struct sparx5_vcap_inst *cfg; struct vcap_control *ctrl; struct vcap_admin *admin; int err = 0, idx; /* Create a VCAP control instance that owns the platform specific VCAP * model with VCAP instances and information about keysets, keys, * actionsets and actions * - Create administrative state for each available VCAP * - Lists of rules * - Address information * - Initialize VCAP blocks * - Configure port keysets */ ctrl = kzalloc(sizeof(*ctrl), GFP_KERNEL); if (!ctrl) return -ENOMEM; sparx5->vcap_ctrl = ctrl; /* select the sparx5 VCAP model */ ctrl->vcaps = sparx5_vcaps; ctrl->stats = &sparx5_vcap_stats; /* Setup callbacks to allow the API to use the VCAP HW */ ctrl->ops = &sparx5_vcap_ops; INIT_LIST_HEAD(&ctrl->list); for (idx = 0; idx < ARRAY_SIZE(sparx5_vcap_inst_cfg); ++idx) { cfg = &sparx5_vcap_inst_cfg[idx]; admin = sparx5_vcap_admin_alloc(sparx5, ctrl, cfg); if (IS_ERR(admin)) { err = PTR_ERR(admin); pr_err("%s:%d: vcap allocation failed: %d\n", __func__, __LINE__, err); return err; } sparx5_vcap_block_alloc(sparx5, admin, cfg); sparx5_vcap_block_init(sparx5, admin); if (cfg->vinst == 0) sparx5_vcap_port_key_selection(sparx5, admin); list_add_tail(&admin->list, &ctrl->list); } return err; } void sparx5_vcap_destroy(struct sparx5 *sparx5) { struct vcap_control *ctrl = sparx5->vcap_ctrl; struct vcap_admin *admin, *admin_next; if (!ctrl) return; list_for_each_entry_safe(admin, admin_next, &ctrl->list, list) { sparx5_vcap_port_key_deselection(sparx5, admin); vcap_del_rules(ctrl, admin); list_del(&admin->list); sparx5_vcap_admin_free(admin); } kfree(ctrl); }