// SPDX-License-Identifier: GPL-2.0+ /* Microchip Sparx5 Switch driver VCAP implementation * * Copyright (c) 2022 Microchip Technology Inc. and its subsidiaries. * * The Sparx5 Chip Register Model can be browsed at this location: * https://github.com/microchip-ung/sparx-5_reginfo */ #include "vcap_api_debugfs.h" #include "sparx5_main_regs.h" #include "sparx5_main.h" #include "sparx5_vcap_impl.h" #include "sparx5_vcap_ag_api.h" #include "sparx5_vcap_debugfs.h" #define SUPER_VCAP_BLK_SIZE 3072 /* addresses per Super VCAP block */ #define STREAMSIZE (64 * 4) /* bytes in the VCAP cache area */ #define SPARX5_IS2_LOOKUPS 4 #define VCAP_IS2_KEYSEL(_ena, _noneth, _v4_mc, _v4_uc, _v6_mc, _v6_uc, _arp) \ (ANA_ACL_VCAP_S2_KEY_SEL_KEY_SEL_ENA_SET(_ena) | \ ANA_ACL_VCAP_S2_KEY_SEL_NON_ETH_KEY_SEL_SET(_noneth) | \ ANA_ACL_VCAP_S2_KEY_SEL_IP4_MC_KEY_SEL_SET(_v4_mc) | \ ANA_ACL_VCAP_S2_KEY_SEL_IP4_UC_KEY_SEL_SET(_v4_uc) | \ ANA_ACL_VCAP_S2_KEY_SEL_IP6_MC_KEY_SEL_SET(_v6_mc) | \ ANA_ACL_VCAP_S2_KEY_SEL_IP6_UC_KEY_SEL_SET(_v6_uc) | \ ANA_ACL_VCAP_S2_KEY_SEL_ARP_KEY_SEL_SET(_arp)) static struct sparx5_vcap_inst { enum vcap_type vtype; /* type of vcap */ int vinst; /* instance number within the same type */ int lookups; /* number of lookups in this vcap type */ int lookups_per_instance; /* number of lookups in this instance */ int first_cid; /* first chain id in this vcap */ int last_cid; /* last chain id in this vcap */ int count; /* number of available addresses, not in super vcap */ int map_id; /* id in the super vcap block mapping (if applicable) */ int blockno; /* starting block in super vcap (if applicable) */ int blocks; /* number of blocks in super vcap (if applicable) */ } sparx5_vcap_inst_cfg[] = { { .vtype = VCAP_TYPE_IS2, /* IS2-0 */ .vinst = 0, .map_id = 4, .lookups = SPARX5_IS2_LOOKUPS, .lookups_per_instance = SPARX5_IS2_LOOKUPS / 2, .first_cid = SPARX5_VCAP_CID_IS2_L0, .last_cid = SPARX5_VCAP_CID_IS2_L2 - 1, .blockno = 0, /* Maps block 0-1 */ .blocks = 2, }, { .vtype = VCAP_TYPE_IS2, /* IS2-1 */ .vinst = 1, .map_id = 5, .lookups = SPARX5_IS2_LOOKUPS, .lookups_per_instance = SPARX5_IS2_LOOKUPS / 2, .first_cid = SPARX5_VCAP_CID_IS2_L2, .last_cid = SPARX5_VCAP_CID_IS2_MAX, .blockno = 2, /* Maps block 2-3 */ .blocks = 2, }, }; /* Await the super VCAP completion of the current operation */ static void sparx5_vcap_wait_super_update(struct sparx5 *sparx5) { u32 value; read_poll_timeout(spx5_rd, value, !VCAP_SUPER_CTRL_UPDATE_SHOT_GET(value), 500, 10000, false, sparx5, VCAP_SUPER_CTRL); } /* Initializing a VCAP address range: only IS2 for now */ static void _sparx5_vcap_range_init(struct sparx5 *sparx5, struct vcap_admin *admin, u32 addr, u32 count) { u32 size = count - 1; spx5_wr(VCAP_SUPER_CFG_MV_NUM_POS_SET(0) | VCAP_SUPER_CFG_MV_SIZE_SET(size), sparx5, VCAP_SUPER_CFG); spx5_wr(VCAP_SUPER_CTRL_UPDATE_CMD_SET(VCAP_CMD_INITIALIZE) | VCAP_SUPER_CTRL_UPDATE_ENTRY_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_ACTION_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_CNT_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_ADDR_SET(addr) | VCAP_SUPER_CTRL_CLEAR_CACHE_SET(true) | VCAP_SUPER_CTRL_UPDATE_SHOT_SET(true), sparx5, VCAP_SUPER_CTRL); sparx5_vcap_wait_super_update(sparx5); } /* Initializing VCAP rule data area */ static void sparx5_vcap_block_init(struct sparx5 *sparx5, struct vcap_admin *admin) { _sparx5_vcap_range_init(sparx5, admin, admin->first_valid_addr, admin->last_valid_addr - admin->first_valid_addr); } /* Get the keyset name from the sparx5 VCAP model */ static const char *sparx5_vcap_keyset_name(struct net_device *ndev, enum vcap_keyfield_set keyset) { struct sparx5_port *port = netdev_priv(ndev); return vcap_keyset_name(port->sparx5->vcap_ctrl, keyset); } /* Check if this is the first lookup of IS2 */ static bool sparx5_vcap_is2_is_first_chain(struct vcap_rule *rule) { return (rule->vcap_chain_id >= SPARX5_VCAP_CID_IS2_L0 && rule->vcap_chain_id < SPARX5_VCAP_CID_IS2_L1) || ((rule->vcap_chain_id >= SPARX5_VCAP_CID_IS2_L2 && rule->vcap_chain_id < SPARX5_VCAP_CID_IS2_L3)); } /* Set the narrow range ingress port mask on a rule */ static void sparx5_vcap_add_range_port_mask(struct vcap_rule *rule, struct net_device *ndev) { struct sparx5_port *port = netdev_priv(ndev); u32 port_mask; u32 range; range = port->portno / BITS_PER_TYPE(u32); /* Port bit set to match-any */ port_mask = ~BIT(port->portno % BITS_PER_TYPE(u32)); vcap_rule_add_key_u32(rule, VCAP_KF_IF_IGR_PORT_MASK_SEL, 0, 0xf); vcap_rule_add_key_u32(rule, VCAP_KF_IF_IGR_PORT_MASK_RNG, range, 0xf); vcap_rule_add_key_u32(rule, VCAP_KF_IF_IGR_PORT_MASK, 0, port_mask); } /* Set the wide range ingress port mask on a rule */ static void sparx5_vcap_add_wide_port_mask(struct vcap_rule *rule, struct net_device *ndev) { struct sparx5_port *port = netdev_priv(ndev); struct vcap_u72_key port_mask; u32 range; /* Port bit set to match-any */ memset(port_mask.value, 0, sizeof(port_mask.value)); memset(port_mask.mask, 0xff, sizeof(port_mask.mask)); range = port->portno / BITS_PER_BYTE; port_mask.mask[range] = ~BIT(port->portno % BITS_PER_BYTE); vcap_rule_add_key_u72(rule, VCAP_KF_IF_IGR_PORT_MASK, &port_mask); } /* Convert chain id to vcap lookup id */ static int sparx5_vcap_cid_to_lookup(int cid) { int lookup = 0; /* For now only handle IS2 */ if (cid >= SPARX5_VCAP_CID_IS2_L1 && cid < SPARX5_VCAP_CID_IS2_L2) lookup = 1; else if (cid >= SPARX5_VCAP_CID_IS2_L2 && cid < SPARX5_VCAP_CID_IS2_L3) lookup = 2; else if (cid >= SPARX5_VCAP_CID_IS2_L3 && cid < SPARX5_VCAP_CID_IS2_MAX) lookup = 3; return lookup; } /* Return the list of keysets for the vcap port configuration */ static int sparx5_vcap_is2_get_port_keysets(struct net_device *ndev, int lookup, struct vcap_keyset_list *keysetlist, u16 l3_proto) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; int portno = port->portno; u32 value; /* Check if the port keyset selection is enabled */ value = spx5_rd(sparx5, ANA_ACL_VCAP_S2_KEY_SEL(portno, lookup)); if (!ANA_ACL_VCAP_S2_KEY_SEL_KEY_SEL_ENA_GET(value)) return -ENOENT; /* Collect all keysets for the port in a list */ if (l3_proto == ETH_P_ALL || l3_proto == ETH_P_ARP) { switch (ANA_ACL_VCAP_S2_KEY_SEL_ARP_KEY_SEL_GET(value)) { case VCAP_IS2_PS_ARP_MAC_ETYPE: vcap_keyset_list_add(keysetlist, VCAP_KFS_MAC_ETYPE); break; case VCAP_IS2_PS_ARP_ARP: vcap_keyset_list_add(keysetlist, VCAP_KFS_ARP); break; } } if (l3_proto == ETH_P_ALL || l3_proto == ETH_P_IP) { switch (ANA_ACL_VCAP_S2_KEY_SEL_IP4_UC_KEY_SEL_GET(value)) { case VCAP_IS2_PS_IPV4_UC_MAC_ETYPE: vcap_keyset_list_add(keysetlist, VCAP_KFS_MAC_ETYPE); break; case VCAP_IS2_PS_IPV4_UC_IP4_TCP_UDP_OTHER: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_TCP_UDP); vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_OTHER); break; case VCAP_IS2_PS_IPV4_UC_IP_7TUPLE: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP_7TUPLE); break; } switch (ANA_ACL_VCAP_S2_KEY_SEL_IP4_MC_KEY_SEL_GET(value)) { case VCAP_IS2_PS_IPV4_MC_MAC_ETYPE: vcap_keyset_list_add(keysetlist, VCAP_KFS_MAC_ETYPE); break; case VCAP_IS2_PS_IPV4_MC_IP4_TCP_UDP_OTHER: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_TCP_UDP); vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_OTHER); break; case VCAP_IS2_PS_IPV4_MC_IP_7TUPLE: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP_7TUPLE); break; } } if (l3_proto == ETH_P_ALL || l3_proto == ETH_P_IPV6) { switch (ANA_ACL_VCAP_S2_KEY_SEL_IP6_UC_KEY_SEL_GET(value)) { case VCAP_IS2_PS_IPV6_UC_MAC_ETYPE: vcap_keyset_list_add(keysetlist, VCAP_KFS_MAC_ETYPE); break; case VCAP_IS2_PS_IPV6_UC_IP_7TUPLE: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP_7TUPLE); break; case VCAP_IS2_PS_IPV6_UC_IP6_STD: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP6_STD); break; case VCAP_IS2_PS_IPV6_UC_IP4_TCP_UDP_OTHER: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_TCP_UDP); vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_OTHER); break; } switch (ANA_ACL_VCAP_S2_KEY_SEL_IP6_MC_KEY_SEL_GET(value)) { case VCAP_IS2_PS_IPV6_MC_MAC_ETYPE: vcap_keyset_list_add(keysetlist, VCAP_KFS_MAC_ETYPE); break; case VCAP_IS2_PS_IPV6_MC_IP_7TUPLE: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP_7TUPLE); break; case VCAP_IS2_PS_IPV6_MC_IP6_STD: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP6_STD); break; case VCAP_IS2_PS_IPV6_MC_IP4_TCP_UDP_OTHER: vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_TCP_UDP); vcap_keyset_list_add(keysetlist, VCAP_KFS_IP4_OTHER); break; case VCAP_IS2_PS_IPV6_MC_IP6_VID: /* Not used */ break; } } if (l3_proto != ETH_P_ARP && l3_proto != ETH_P_IP && l3_proto != ETH_P_IPV6) { switch (ANA_ACL_VCAP_S2_KEY_SEL_NON_ETH_KEY_SEL_GET(value)) { case VCAP_IS2_PS_NONETH_MAC_ETYPE: /* IS2 non-classified frames generate MAC_ETYPE */ vcap_keyset_list_add(keysetlist, VCAP_KFS_MAC_ETYPE); break; } } return 0; } /* Get the port keyset for the vcap lookup */ int sparx5_vcap_get_port_keyset(struct net_device *ndev, struct vcap_admin *admin, int cid, u16 l3_proto, struct vcap_keyset_list *kslist) { int lookup; lookup = sparx5_vcap_cid_to_lookup(cid); return sparx5_vcap_is2_get_port_keysets(ndev, lookup, kslist, l3_proto); } /* API callback used for validating a field keyset (check the port keysets) */ static enum vcap_keyfield_set sparx5_vcap_validate_keyset(struct net_device *ndev, struct vcap_admin *admin, struct vcap_rule *rule, struct vcap_keyset_list *kslist, u16 l3_proto) { struct vcap_keyset_list keysetlist = {}; enum vcap_keyfield_set keysets[10] = {}; int idx, jdx, lookup; if (!kslist || kslist->cnt == 0) return VCAP_KFS_NO_VALUE; /* Get a list of currently configured keysets in the lookups */ lookup = sparx5_vcap_cid_to_lookup(rule->vcap_chain_id); keysetlist.max = ARRAY_SIZE(keysets); keysetlist.keysets = keysets; sparx5_vcap_is2_get_port_keysets(ndev, lookup, &keysetlist, l3_proto); /* Check if there is a match and return the match */ for (idx = 0; idx < kslist->cnt; ++idx) for (jdx = 0; jdx < keysetlist.cnt; ++jdx) if (kslist->keysets[idx] == keysets[jdx]) return kslist->keysets[idx]; pr_err("%s:%d: %s not supported in port key selection\n", __func__, __LINE__, sparx5_vcap_keyset_name(ndev, kslist->keysets[0])); return -ENOENT; } /* API callback used for adding default fields to a rule */ static void sparx5_vcap_add_default_fields(struct net_device *ndev, struct vcap_admin *admin, struct vcap_rule *rule) { const struct vcap_field *field; field = vcap_lookup_keyfield(rule, VCAP_KF_IF_IGR_PORT_MASK); if (field && field->width == SPX5_PORTS) sparx5_vcap_add_wide_port_mask(rule, ndev); else if (field && field->width == BITS_PER_TYPE(u32)) sparx5_vcap_add_range_port_mask(rule, ndev); else pr_err("%s:%d: %s: could not add an ingress port mask for: %s\n", __func__, __LINE__, netdev_name(ndev), sparx5_vcap_keyset_name(ndev, rule->keyset)); /* add the lookup bit */ if (sparx5_vcap_is2_is_first_chain(rule)) vcap_rule_add_key_bit(rule, VCAP_KF_LOOKUP_FIRST_IS, VCAP_BIT_1); else vcap_rule_add_key_bit(rule, VCAP_KF_LOOKUP_FIRST_IS, VCAP_BIT_0); } /* API callback used for erasing the vcap cache area (not the register area) */ static void sparx5_vcap_cache_erase(struct vcap_admin *admin) { memset(admin->cache.keystream, 0, STREAMSIZE); memset(admin->cache.maskstream, 0, STREAMSIZE); memset(admin->cache.actionstream, 0, STREAMSIZE); memset(&admin->cache.counter, 0, sizeof(admin->cache.counter)); } /* API callback used for writing to the VCAP cache */ static void sparx5_vcap_cache_write(struct net_device *ndev, struct vcap_admin *admin, enum vcap_selection sel, u32 start, u32 count) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; u32 *keystr, *mskstr, *actstr; int idx; keystr = &admin->cache.keystream[start]; mskstr = &admin->cache.maskstream[start]; actstr = &admin->cache.actionstream[start]; switch (sel) { case VCAP_SEL_ENTRY: for (idx = 0; idx < count; ++idx) { /* Avoid 'match-off' by setting value & mask */ spx5_wr(keystr[idx] & mskstr[idx], sparx5, VCAP_SUPER_VCAP_ENTRY_DAT(idx)); spx5_wr(~mskstr[idx], sparx5, VCAP_SUPER_VCAP_MASK_DAT(idx)); } break; case VCAP_SEL_ACTION: for (idx = 0; idx < count; ++idx) spx5_wr(actstr[idx], sparx5, VCAP_SUPER_VCAP_ACTION_DAT(idx)); break; case VCAP_SEL_ALL: pr_err("%s:%d: cannot write all streams at once\n", __func__, __LINE__); break; default: break; } if (sel & VCAP_SEL_COUNTER) { start = start & 0xfff; /* counter limit */ if (admin->vinst == 0) spx5_wr(admin->cache.counter, sparx5, ANA_ACL_CNT_A(start)); else spx5_wr(admin->cache.counter, sparx5, ANA_ACL_CNT_B(start)); spx5_wr(admin->cache.sticky, sparx5, VCAP_SUPER_VCAP_CNT_DAT(0)); } } /* API callback used for reading from the VCAP into the VCAP cache */ static void sparx5_vcap_cache_read(struct net_device *ndev, struct vcap_admin *admin, enum vcap_selection sel, u32 start, u32 count) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; u32 *keystr, *mskstr, *actstr; int idx; keystr = &admin->cache.keystream[start]; mskstr = &admin->cache.maskstream[start]; actstr = &admin->cache.actionstream[start]; if (sel & VCAP_SEL_ENTRY) { for (idx = 0; idx < count; ++idx) { keystr[idx] = spx5_rd(sparx5, VCAP_SUPER_VCAP_ENTRY_DAT(idx)); mskstr[idx] = ~spx5_rd(sparx5, VCAP_SUPER_VCAP_MASK_DAT(idx)); } } if (sel & VCAP_SEL_ACTION) { for (idx = 0; idx < count; ++idx) actstr[idx] = spx5_rd(sparx5, VCAP_SUPER_VCAP_ACTION_DAT(idx)); } if (sel & VCAP_SEL_COUNTER) { start = start & 0xfff; /* counter limit */ if (admin->vinst == 0) admin->cache.counter = spx5_rd(sparx5, ANA_ACL_CNT_A(start)); else admin->cache.counter = spx5_rd(sparx5, ANA_ACL_CNT_B(start)); admin->cache.sticky = spx5_rd(sparx5, VCAP_SUPER_VCAP_CNT_DAT(0)); } } /* API callback used for initializing a VCAP address range */ static void sparx5_vcap_range_init(struct net_device *ndev, struct vcap_admin *admin, u32 addr, u32 count) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; _sparx5_vcap_range_init(sparx5, admin, addr, count); } /* API callback used for updating the VCAP cache */ static void sparx5_vcap_update(struct net_device *ndev, struct vcap_admin *admin, enum vcap_command cmd, enum vcap_selection sel, u32 addr) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; bool clear; clear = (cmd == VCAP_CMD_INITIALIZE); spx5_wr(VCAP_SUPER_CFG_MV_NUM_POS_SET(0) | VCAP_SUPER_CFG_MV_SIZE_SET(0), sparx5, VCAP_SUPER_CFG); spx5_wr(VCAP_SUPER_CTRL_UPDATE_CMD_SET(cmd) | VCAP_SUPER_CTRL_UPDATE_ENTRY_DIS_SET((VCAP_SEL_ENTRY & sel) == 0) | VCAP_SUPER_CTRL_UPDATE_ACTION_DIS_SET((VCAP_SEL_ACTION & sel) == 0) | VCAP_SUPER_CTRL_UPDATE_CNT_DIS_SET((VCAP_SEL_COUNTER & sel) == 0) | VCAP_SUPER_CTRL_UPDATE_ADDR_SET(addr) | VCAP_SUPER_CTRL_CLEAR_CACHE_SET(clear) | VCAP_SUPER_CTRL_UPDATE_SHOT_SET(true), sparx5, VCAP_SUPER_CTRL); sparx5_vcap_wait_super_update(sparx5); } /* API callback used for moving a block of rules in the VCAP */ static void sparx5_vcap_move(struct net_device *ndev, struct vcap_admin *admin, u32 addr, int offset, int count) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5 = port->sparx5; enum vcap_command cmd; u16 mv_num_pos; u16 mv_size; mv_size = count - 1; if (offset > 0) { mv_num_pos = offset - 1; cmd = VCAP_CMD_MOVE_DOWN; } else { mv_num_pos = -offset - 1; cmd = VCAP_CMD_MOVE_UP; } spx5_wr(VCAP_SUPER_CFG_MV_NUM_POS_SET(mv_num_pos) | VCAP_SUPER_CFG_MV_SIZE_SET(mv_size), sparx5, VCAP_SUPER_CFG); spx5_wr(VCAP_SUPER_CTRL_UPDATE_CMD_SET(cmd) | VCAP_SUPER_CTRL_UPDATE_ENTRY_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_ACTION_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_CNT_DIS_SET(0) | VCAP_SUPER_CTRL_UPDATE_ADDR_SET(addr) | VCAP_SUPER_CTRL_CLEAR_CACHE_SET(false) | VCAP_SUPER_CTRL_UPDATE_SHOT_SET(true), sparx5, VCAP_SUPER_CTRL); sparx5_vcap_wait_super_update(sparx5); } /* Enable all lookups in the VCAP instance */ static int sparx5_vcap_enable(struct net_device *ndev, struct vcap_admin *admin, bool enable) { struct sparx5_port *port = netdev_priv(ndev); struct sparx5 *sparx5; int portno; sparx5 = port->sparx5; portno = port->portno; /* For now we only consider IS2 */ if (enable) spx5_wr(ANA_ACL_VCAP_S2_CFG_SEC_ENA_SET(0xf), sparx5, ANA_ACL_VCAP_S2_CFG(portno)); else spx5_wr(ANA_ACL_VCAP_S2_CFG_SEC_ENA_SET(0), sparx5, ANA_ACL_VCAP_S2_CFG(portno)); return 0; } /* API callback operations: only IS2 is supported for now */ static struct vcap_operations sparx5_vcap_ops = { .validate_keyset = sparx5_vcap_validate_keyset, .add_default_fields = sparx5_vcap_add_default_fields, .cache_erase = sparx5_vcap_cache_erase, .cache_write = sparx5_vcap_cache_write, .cache_read = sparx5_vcap_cache_read, .init = sparx5_vcap_range_init, .update = sparx5_vcap_update, .move = sparx5_vcap_move, .port_info = sparx5_port_info, .enable = sparx5_vcap_enable, }; /* Enable lookups per port and set the keyset generation: only IS2 for now */ static void sparx5_vcap_port_key_selection(struct sparx5 *sparx5, struct vcap_admin *admin) { int portno, lookup; u32 keysel; /* all traffic types generate the MAC_ETYPE keyset for now in all * lookups on all ports */ keysel = VCAP_IS2_KEYSEL(true, VCAP_IS2_PS_NONETH_MAC_ETYPE, VCAP_IS2_PS_IPV4_MC_IP4_TCP_UDP_OTHER, VCAP_IS2_PS_IPV4_UC_IP4_TCP_UDP_OTHER, VCAP_IS2_PS_IPV6_MC_IP_7TUPLE, VCAP_IS2_PS_IPV6_UC_IP_7TUPLE, VCAP_IS2_PS_ARP_ARP); for (lookup = 0; lookup < admin->lookups; ++lookup) { for (portno = 0; portno < SPX5_PORTS; ++portno) { spx5_wr(keysel, sparx5, ANA_ACL_VCAP_S2_KEY_SEL(portno, lookup)); } } } /* Disable lookups per port and set the keyset generation: only IS2 for now */ static void sparx5_vcap_port_key_deselection(struct sparx5 *sparx5, struct vcap_admin *admin) { int portno; for (portno = 0; portno < SPX5_PORTS; ++portno) spx5_rmw(ANA_ACL_VCAP_S2_CFG_SEC_ENA_SET(0), ANA_ACL_VCAP_S2_CFG_SEC_ENA, sparx5, ANA_ACL_VCAP_S2_CFG(portno)); } static void sparx5_vcap_admin_free(struct vcap_admin *admin) { if (!admin) return; mutex_destroy(&admin->lock); kfree(admin->cache.keystream); kfree(admin->cache.maskstream); kfree(admin->cache.actionstream); kfree(admin); } /* Allocate a vcap instance with a rule list and a cache area */ static struct vcap_admin * sparx5_vcap_admin_alloc(struct sparx5 *sparx5, struct vcap_control *ctrl, const struct sparx5_vcap_inst *cfg) { struct vcap_admin *admin; admin = kzalloc(sizeof(*admin), GFP_KERNEL); if (!admin) return ERR_PTR(-ENOMEM); INIT_LIST_HEAD(&admin->list); INIT_LIST_HEAD(&admin->rules); INIT_LIST_HEAD(&admin->enabled); mutex_init(&admin->lock); admin->vtype = cfg->vtype; admin->vinst = cfg->vinst; admin->lookups = cfg->lookups; admin->lookups_per_instance = cfg->lookups_per_instance; admin->first_cid = cfg->first_cid; admin->last_cid = cfg->last_cid; admin->cache.keystream = kzalloc(STREAMSIZE, GFP_KERNEL); admin->cache.maskstream = kzalloc(STREAMSIZE, GFP_KERNEL); admin->cache.actionstream = kzalloc(STREAMSIZE, GFP_KERNEL); if (!admin->cache.keystream || !admin->cache.maskstream || !admin->cache.actionstream) { sparx5_vcap_admin_free(admin); return ERR_PTR(-ENOMEM); } return admin; } /* Do block allocations and provide addresses for VCAP instances */ static void sparx5_vcap_block_alloc(struct sparx5 *sparx5, struct vcap_admin *admin, const struct sparx5_vcap_inst *cfg) { int idx; /* Super VCAP block mapping and address configuration. Block 0 * is assigned addresses 0 through 3071, block 1 is assigned * addresses 3072 though 6143, and so on. */ for (idx = cfg->blockno; idx < cfg->blockno + cfg->blocks; ++idx) { spx5_wr(VCAP_SUPER_IDX_CORE_IDX_SET(idx), sparx5, VCAP_SUPER_IDX); spx5_wr(VCAP_SUPER_MAP_CORE_MAP_SET(cfg->map_id), sparx5, VCAP_SUPER_MAP); } admin->first_valid_addr = cfg->blockno * SUPER_VCAP_BLK_SIZE; admin->last_used_addr = admin->first_valid_addr + cfg->blocks * SUPER_VCAP_BLK_SIZE; admin->last_valid_addr = admin->last_used_addr - 1; } /* Allocate a vcap control and vcap instances and configure the system */ int sparx5_vcap_init(struct sparx5 *sparx5) { const struct sparx5_vcap_inst *cfg; struct vcap_control *ctrl; struct vcap_admin *admin; struct dentry *dir; int err = 0, idx; /* Create a VCAP control instance that owns the platform specific VCAP * model with VCAP instances and information about keysets, keys, * actionsets and actions * - Create administrative state for each available VCAP * - Lists of rules * - Address information * - Initialize VCAP blocks * - Configure port keysets */ ctrl = kzalloc(sizeof(*ctrl), GFP_KERNEL); if (!ctrl) return -ENOMEM; sparx5->vcap_ctrl = ctrl; /* select the sparx5 VCAP model */ ctrl->vcaps = sparx5_vcaps; ctrl->stats = &sparx5_vcap_stats; /* Setup callbacks to allow the API to use the VCAP HW */ ctrl->ops = &sparx5_vcap_ops; INIT_LIST_HEAD(&ctrl->list); for (idx = 0; idx < ARRAY_SIZE(sparx5_vcap_inst_cfg); ++idx) { cfg = &sparx5_vcap_inst_cfg[idx]; admin = sparx5_vcap_admin_alloc(sparx5, ctrl, cfg); if (IS_ERR(admin)) { err = PTR_ERR(admin); pr_err("%s:%d: vcap allocation failed: %d\n", __func__, __LINE__, err); return err; } sparx5_vcap_block_alloc(sparx5, admin, cfg); sparx5_vcap_block_init(sparx5, admin); if (cfg->vinst == 0) sparx5_vcap_port_key_selection(sparx5, admin); list_add_tail(&admin->list, &ctrl->list); } dir = vcap_debugfs(sparx5->dev, sparx5->debugfs_root, ctrl); for (idx = 0; idx < SPX5_PORTS; ++idx) if (sparx5->ports[idx]) vcap_port_debugfs(sparx5->dev, dir, ctrl, sparx5->ports[idx]->ndev); return err; } void sparx5_vcap_destroy(struct sparx5 *sparx5) { struct vcap_control *ctrl = sparx5->vcap_ctrl; struct vcap_admin *admin, *admin_next; if (!ctrl) return; list_for_each_entry_safe(admin, admin_next, &ctrl->list, list) { sparx5_vcap_port_key_deselection(sparx5, admin); vcap_del_rules(ctrl, admin); list_del(&admin->list); sparx5_vcap_admin_free(admin); } kfree(ctrl); }