13ea9bd5dSMichal Swiatkowski // SPDX-License-Identifier: GPL-2.0 23ea9bd5dSMichal Swiatkowski /* Copyright (C) 2019-2021, Intel Corporation. */ 33ea9bd5dSMichal Swiatkowski 43ea9bd5dSMichal Swiatkowski #include "ice.h" 51a1c40dfSGrzegorz Nitka #include "ice_lib.h" 63ea9bd5dSMichal Swiatkowski #include "ice_eswitch.h" 71a1c40dfSGrzegorz Nitka #include "ice_fltr.h" 81a1c40dfSGrzegorz Nitka #include "ice_repr.h" 93ea9bd5dSMichal Swiatkowski #include "ice_devlink.h" 107fde6d8bSMichal Swiatkowski #include "ice_tc_lib.h" 113ea9bd5dSMichal Swiatkowski 123ea9bd5dSMichal Swiatkowski /** 13c1e5da5dSWojciech Drewek * ice_eswitch_add_vf_mac_rule - add adv rule with VF's MAC 14c1e5da5dSWojciech Drewek * @pf: pointer to PF struct 15c1e5da5dSWojciech Drewek * @vf: pointer to VF struct 16c1e5da5dSWojciech Drewek * @mac: VF's MAC address 17c1e5da5dSWojciech Drewek * 18c1e5da5dSWojciech Drewek * This function adds advanced rule that forwards packets with 19c1e5da5dSWojciech Drewek * VF's MAC address (src MAC) to the corresponding switchdev ctrl VSI queue. 20c1e5da5dSWojciech Drewek */ 21c1e5da5dSWojciech Drewek int 22c1e5da5dSWojciech Drewek ice_eswitch_add_vf_mac_rule(struct ice_pf *pf, struct ice_vf *vf, const u8 *mac) 23c1e5da5dSWojciech Drewek { 24c1e5da5dSWojciech Drewek struct ice_vsi *ctrl_vsi = pf->switchdev.control_vsi; 25c1e5da5dSWojciech Drewek struct ice_adv_rule_info rule_info = { 0 }; 26c1e5da5dSWojciech Drewek struct ice_adv_lkup_elem *list; 27c1e5da5dSWojciech Drewek struct ice_hw *hw = &pf->hw; 28c1e5da5dSWojciech Drewek const u16 lkups_cnt = 1; 29c1e5da5dSWojciech Drewek int err; 30c1e5da5dSWojciech Drewek 31c1e5da5dSWojciech Drewek list = kcalloc(lkups_cnt, sizeof(*list), GFP_ATOMIC); 32c1e5da5dSWojciech Drewek if (!list) 33c1e5da5dSWojciech Drewek return -ENOMEM; 34c1e5da5dSWojciech Drewek 35c1e5da5dSWojciech Drewek list[0].type = ICE_MAC_OFOS; 36c1e5da5dSWojciech Drewek ether_addr_copy(list[0].h_u.eth_hdr.src_addr, mac); 37c1e5da5dSWojciech Drewek eth_broadcast_addr(list[0].m_u.eth_hdr.src_addr); 38c1e5da5dSWojciech Drewek 39c1e5da5dSWojciech Drewek rule_info.sw_act.flag |= ICE_FLTR_TX; 40c1e5da5dSWojciech Drewek rule_info.sw_act.vsi_handle = ctrl_vsi->idx; 41c1e5da5dSWojciech Drewek rule_info.sw_act.fltr_act = ICE_FWD_TO_Q; 42c1e5da5dSWojciech Drewek rule_info.rx = false; 43c1e5da5dSWojciech Drewek rule_info.sw_act.fwd_id.q_id = hw->func_caps.common_cap.rxq_first_id + 44c1e5da5dSWojciech Drewek ctrl_vsi->rxq_map[vf->vf_id]; 45c1e5da5dSWojciech Drewek rule_info.flags_info.act |= ICE_SINGLE_ACT_LB_ENABLE; 46c1e5da5dSWojciech Drewek rule_info.flags_info.act_valid = true; 47c1e5da5dSWojciech Drewek 48c1e5da5dSWojciech Drewek err = ice_add_adv_rule(hw, list, lkups_cnt, &rule_info, 49c1e5da5dSWojciech Drewek vf->repr->mac_rule); 50c1e5da5dSWojciech Drewek if (err) 51c1e5da5dSWojciech Drewek dev_err(ice_pf_to_dev(pf), "Unable to add VF mac rule in switchdev mode for VF %d", 52c1e5da5dSWojciech Drewek vf->vf_id); 53c1e5da5dSWojciech Drewek else 54c1e5da5dSWojciech Drewek vf->repr->rule_added = true; 55c1e5da5dSWojciech Drewek 56c1e5da5dSWojciech Drewek kfree(list); 57c1e5da5dSWojciech Drewek return err; 58c1e5da5dSWojciech Drewek } 59c1e5da5dSWojciech Drewek 60c1e5da5dSWojciech Drewek /** 61c1e5da5dSWojciech Drewek * ice_eswitch_replay_vf_mac_rule - replay adv rule with VF's MAC 62c1e5da5dSWojciech Drewek * @vf: pointer to vF struct 63c1e5da5dSWojciech Drewek * 64c1e5da5dSWojciech Drewek * This function replays VF's MAC rule after reset. 65c1e5da5dSWojciech Drewek */ 66c1e5da5dSWojciech Drewek void ice_eswitch_replay_vf_mac_rule(struct ice_vf *vf) 67c1e5da5dSWojciech Drewek { 68c1e5da5dSWojciech Drewek int err; 69c1e5da5dSWojciech Drewek 70c1e5da5dSWojciech Drewek if (!ice_is_switchdev_running(vf->pf)) 71c1e5da5dSWojciech Drewek return; 72c1e5da5dSWojciech Drewek 73c1e5da5dSWojciech Drewek if (is_valid_ether_addr(vf->hw_lan_addr.addr)) { 74c1e5da5dSWojciech Drewek err = ice_eswitch_add_vf_mac_rule(vf->pf, vf, 75c1e5da5dSWojciech Drewek vf->hw_lan_addr.addr); 76c1e5da5dSWojciech Drewek if (err) { 77c1e5da5dSWojciech Drewek dev_err(ice_pf_to_dev(vf->pf), "Failed to add MAC %pM for VF %d\n, error %d\n", 78c1e5da5dSWojciech Drewek vf->hw_lan_addr.addr, vf->vf_id, err); 79c1e5da5dSWojciech Drewek return; 80c1e5da5dSWojciech Drewek } 81c1e5da5dSWojciech Drewek vf->num_mac++; 82c1e5da5dSWojciech Drewek 83c1e5da5dSWojciech Drewek ether_addr_copy(vf->dev_lan_addr.addr, vf->hw_lan_addr.addr); 84c1e5da5dSWojciech Drewek } 85c1e5da5dSWojciech Drewek } 86c1e5da5dSWojciech Drewek 87c1e5da5dSWojciech Drewek /** 88c1e5da5dSWojciech Drewek * ice_eswitch_del_vf_mac_rule - delete adv rule with VF's MAC 89c1e5da5dSWojciech Drewek * @vf: pointer to the VF struct 90c1e5da5dSWojciech Drewek * 91c1e5da5dSWojciech Drewek * Delete the advanced rule that was used to forward packets with the VF's MAC 92c1e5da5dSWojciech Drewek * address (src MAC) to the corresponding switchdev ctrl VSI queue. 93c1e5da5dSWojciech Drewek */ 94c1e5da5dSWojciech Drewek void ice_eswitch_del_vf_mac_rule(struct ice_vf *vf) 95c1e5da5dSWojciech Drewek { 96c1e5da5dSWojciech Drewek if (!ice_is_switchdev_running(vf->pf)) 97c1e5da5dSWojciech Drewek return; 98c1e5da5dSWojciech Drewek 99c1e5da5dSWojciech Drewek if (!vf->repr->rule_added) 100c1e5da5dSWojciech Drewek return; 101c1e5da5dSWojciech Drewek 102c1e5da5dSWojciech Drewek ice_rem_adv_rule_by_id(&vf->pf->hw, vf->repr->mac_rule); 103c1e5da5dSWojciech Drewek vf->repr->rule_added = false; 104c1e5da5dSWojciech Drewek } 105c1e5da5dSWojciech Drewek 106c1e5da5dSWojciech Drewek /** 1071a1c40dfSGrzegorz Nitka * ice_eswitch_setup_env - configure switchdev HW filters 1081a1c40dfSGrzegorz Nitka * @pf: pointer to PF struct 1091a1c40dfSGrzegorz Nitka * 1101a1c40dfSGrzegorz Nitka * This function adds HW filters configuration specific for switchdev 1111a1c40dfSGrzegorz Nitka * mode. 1121a1c40dfSGrzegorz Nitka */ 1131a1c40dfSGrzegorz Nitka static int ice_eswitch_setup_env(struct ice_pf *pf) 1141a1c40dfSGrzegorz Nitka { 1151a1c40dfSGrzegorz Nitka struct ice_vsi *uplink_vsi = pf->switchdev.uplink_vsi; 116c79bb28eSMarcin Szycik struct net_device *uplink_netdev = uplink_vsi->netdev; 1171a1c40dfSGrzegorz Nitka struct ice_vsi *ctrl_vsi = pf->switchdev.control_vsi; 118*c31af68aSBrett Creeley struct ice_vsi_vlan_ops *vlan_ops; 1191a1c40dfSGrzegorz Nitka bool rule_added = false; 1201a1c40dfSGrzegorz Nitka 121*c31af68aSBrett Creeley vlan_ops = ice_get_compat_vsi_vlan_ops(ctrl_vsi); 122*c31af68aSBrett Creeley if (vlan_ops->dis_stripping(ctrl_vsi)) 123*c31af68aSBrett Creeley return -ENODEV; 1241a1c40dfSGrzegorz Nitka 1251a1c40dfSGrzegorz Nitka ice_remove_vsi_fltr(&pf->hw, uplink_vsi->idx); 1261a1c40dfSGrzegorz Nitka 127c79bb28eSMarcin Szycik netif_addr_lock_bh(uplink_netdev); 128c79bb28eSMarcin Szycik __dev_uc_unsync(uplink_netdev, NULL); 129c79bb28eSMarcin Szycik __dev_mc_unsync(uplink_netdev, NULL); 130c79bb28eSMarcin Szycik netif_addr_unlock_bh(uplink_netdev); 131c79bb28eSMarcin Szycik 1323e0b5971SBrett Creeley if (ice_vsi_add_vlan_zero(uplink_vsi)) 1331a1c40dfSGrzegorz Nitka goto err_def_rx; 1341a1c40dfSGrzegorz Nitka 1351a1c40dfSGrzegorz Nitka if (!ice_is_dflt_vsi_in_use(uplink_vsi->vsw)) { 1361a1c40dfSGrzegorz Nitka if (ice_set_dflt_vsi(uplink_vsi->vsw, uplink_vsi)) 1371a1c40dfSGrzegorz Nitka goto err_def_rx; 1381a1c40dfSGrzegorz Nitka rule_added = true; 1391a1c40dfSGrzegorz Nitka } 1401a1c40dfSGrzegorz Nitka 1411a1c40dfSGrzegorz Nitka if (ice_vsi_update_security(uplink_vsi, ice_vsi_ctx_set_allow_override)) 1421a1c40dfSGrzegorz Nitka goto err_override_uplink; 1431a1c40dfSGrzegorz Nitka 1441a1c40dfSGrzegorz Nitka if (ice_vsi_update_security(ctrl_vsi, ice_vsi_ctx_set_allow_override)) 1451a1c40dfSGrzegorz Nitka goto err_override_control; 1461a1c40dfSGrzegorz Nitka 1471a1c40dfSGrzegorz Nitka return 0; 1481a1c40dfSGrzegorz Nitka 1491a1c40dfSGrzegorz Nitka err_override_control: 1501a1c40dfSGrzegorz Nitka ice_vsi_update_security(uplink_vsi, ice_vsi_ctx_clear_allow_override); 1511a1c40dfSGrzegorz Nitka err_override_uplink: 1521a1c40dfSGrzegorz Nitka if (rule_added) 1531a1c40dfSGrzegorz Nitka ice_clear_dflt_vsi(uplink_vsi->vsw); 1541a1c40dfSGrzegorz Nitka err_def_rx: 1551a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(uplink_vsi, 1561a1c40dfSGrzegorz Nitka uplink_vsi->port_info->mac.perm_addr, 1571a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 1581a1c40dfSGrzegorz Nitka return -ENODEV; 1591a1c40dfSGrzegorz Nitka } 1601a1c40dfSGrzegorz Nitka 1611a1c40dfSGrzegorz Nitka /** 1621a1c40dfSGrzegorz Nitka * ice_eswitch_remap_rings_to_vectors - reconfigure rings of switchdev ctrl VSI 1631a1c40dfSGrzegorz Nitka * @pf: pointer to PF struct 1641a1c40dfSGrzegorz Nitka * 1651a1c40dfSGrzegorz Nitka * In switchdev number of allocated Tx/Rx rings is equal. 1661a1c40dfSGrzegorz Nitka * 1671a1c40dfSGrzegorz Nitka * This function fills q_vectors structures associated with representor and 1681a1c40dfSGrzegorz Nitka * move each ring pairs to port representor netdevs. Each port representor 1691a1c40dfSGrzegorz Nitka * will have dedicated 1 Tx/Rx ring pair, so number of rings pair is equal to 1701a1c40dfSGrzegorz Nitka * number of VFs. 1711a1c40dfSGrzegorz Nitka */ 1721a1c40dfSGrzegorz Nitka static void ice_eswitch_remap_rings_to_vectors(struct ice_pf *pf) 1731a1c40dfSGrzegorz Nitka { 1741a1c40dfSGrzegorz Nitka struct ice_vsi *vsi = pf->switchdev.control_vsi; 1751a1c40dfSGrzegorz Nitka int q_id; 1761a1c40dfSGrzegorz Nitka 1771a1c40dfSGrzegorz Nitka ice_for_each_txq(vsi, q_id) { 1781a1c40dfSGrzegorz Nitka struct ice_repr *repr = pf->vf[q_id].repr; 1791a1c40dfSGrzegorz Nitka struct ice_q_vector *q_vector = repr->q_vector; 180e72bba21SMaciej Fijalkowski struct ice_tx_ring *tx_ring = vsi->tx_rings[q_id]; 181e72bba21SMaciej Fijalkowski struct ice_rx_ring *rx_ring = vsi->rx_rings[q_id]; 1821a1c40dfSGrzegorz Nitka 1831a1c40dfSGrzegorz Nitka q_vector->vsi = vsi; 1841a1c40dfSGrzegorz Nitka q_vector->reg_idx = vsi->q_vectors[0]->reg_idx; 1851a1c40dfSGrzegorz Nitka 1861a1c40dfSGrzegorz Nitka q_vector->num_ring_tx = 1; 187e72bba21SMaciej Fijalkowski q_vector->tx.tx_ring = tx_ring; 188e72bba21SMaciej Fijalkowski tx_ring->q_vector = q_vector; 189e72bba21SMaciej Fijalkowski tx_ring->next = NULL; 190e72bba21SMaciej Fijalkowski tx_ring->netdev = repr->netdev; 1911a1c40dfSGrzegorz Nitka /* In switchdev mode, from OS stack perspective, there is only 1921a1c40dfSGrzegorz Nitka * one queue for given netdev, so it needs to be indexed as 0. 1931a1c40dfSGrzegorz Nitka */ 1941a1c40dfSGrzegorz Nitka tx_ring->q_index = 0; 1951a1c40dfSGrzegorz Nitka 1961a1c40dfSGrzegorz Nitka q_vector->num_ring_rx = 1; 197e72bba21SMaciej Fijalkowski q_vector->rx.rx_ring = rx_ring; 198e72bba21SMaciej Fijalkowski rx_ring->q_vector = q_vector; 199e72bba21SMaciej Fijalkowski rx_ring->next = NULL; 200e72bba21SMaciej Fijalkowski rx_ring->netdev = repr->netdev; 2011a1c40dfSGrzegorz Nitka } 2021a1c40dfSGrzegorz Nitka } 2031a1c40dfSGrzegorz Nitka 2041a1c40dfSGrzegorz Nitka /** 2051a1c40dfSGrzegorz Nitka * ice_eswitch_setup_reprs - configure port reprs to run in switchdev mode 2061a1c40dfSGrzegorz Nitka * @pf: pointer to PF struct 2071a1c40dfSGrzegorz Nitka */ 2081a1c40dfSGrzegorz Nitka static int ice_eswitch_setup_reprs(struct ice_pf *pf) 2091a1c40dfSGrzegorz Nitka { 2101a1c40dfSGrzegorz Nitka struct ice_vsi *ctrl_vsi = pf->switchdev.control_vsi; 2111a1c40dfSGrzegorz Nitka int max_vsi_num = 0; 2121a1c40dfSGrzegorz Nitka int i; 2131a1c40dfSGrzegorz Nitka 2141a1c40dfSGrzegorz Nitka ice_for_each_vf(pf, i) { 2151a1c40dfSGrzegorz Nitka struct ice_vsi *vsi = pf->vf[i].repr->src_vsi; 2161a1c40dfSGrzegorz Nitka struct ice_vf *vf = &pf->vf[i]; 2171a1c40dfSGrzegorz Nitka 2181a1c40dfSGrzegorz Nitka ice_remove_vsi_fltr(&pf->hw, vsi->idx); 2191a1c40dfSGrzegorz Nitka vf->repr->dst = metadata_dst_alloc(0, METADATA_HW_PORT_MUX, 2201a1c40dfSGrzegorz Nitka GFP_KERNEL); 2211a1c40dfSGrzegorz Nitka if (!vf->repr->dst) { 2221a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(vsi, 2231a1c40dfSGrzegorz Nitka vf->hw_lan_addr.addr, 2241a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 2251a1c40dfSGrzegorz Nitka goto err; 2261a1c40dfSGrzegorz Nitka } 2271a1c40dfSGrzegorz Nitka 2281a1c40dfSGrzegorz Nitka if (ice_vsi_update_security(vsi, ice_vsi_ctx_clear_antispoof)) { 2291a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(vsi, 2301a1c40dfSGrzegorz Nitka vf->hw_lan_addr.addr, 2311a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 2321a1c40dfSGrzegorz Nitka metadata_dst_free(vf->repr->dst); 2331a1c40dfSGrzegorz Nitka goto err; 2341a1c40dfSGrzegorz Nitka } 2351a1c40dfSGrzegorz Nitka 2363e0b5971SBrett Creeley if (ice_vsi_add_vlan_zero(vsi)) { 2371a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(vsi, 2381a1c40dfSGrzegorz Nitka vf->hw_lan_addr.addr, 2391a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 2401a1c40dfSGrzegorz Nitka metadata_dst_free(vf->repr->dst); 2411a1c40dfSGrzegorz Nitka ice_vsi_update_security(vsi, ice_vsi_ctx_set_antispoof); 2421a1c40dfSGrzegorz Nitka goto err; 2431a1c40dfSGrzegorz Nitka } 2441a1c40dfSGrzegorz Nitka 2451a1c40dfSGrzegorz Nitka if (max_vsi_num < vsi->vsi_num) 2461a1c40dfSGrzegorz Nitka max_vsi_num = vsi->vsi_num; 2471a1c40dfSGrzegorz Nitka 2481a1c40dfSGrzegorz Nitka netif_napi_add(vf->repr->netdev, &vf->repr->q_vector->napi, ice_napi_poll, 2491a1c40dfSGrzegorz Nitka NAPI_POLL_WEIGHT); 2501a1c40dfSGrzegorz Nitka 2511a1c40dfSGrzegorz Nitka netif_keep_dst(vf->repr->netdev); 2521a1c40dfSGrzegorz Nitka } 2531a1c40dfSGrzegorz Nitka 2541a1c40dfSGrzegorz Nitka ice_for_each_vf(pf, i) { 2551a1c40dfSGrzegorz Nitka struct ice_repr *repr = pf->vf[i].repr; 2561a1c40dfSGrzegorz Nitka struct ice_vsi *vsi = repr->src_vsi; 2571a1c40dfSGrzegorz Nitka struct metadata_dst *dst; 2581a1c40dfSGrzegorz Nitka 2591a1c40dfSGrzegorz Nitka dst = repr->dst; 2601a1c40dfSGrzegorz Nitka dst->u.port_info.port_id = vsi->vsi_num; 2611a1c40dfSGrzegorz Nitka dst->u.port_info.lower_dev = repr->netdev; 2621a1c40dfSGrzegorz Nitka ice_repr_set_traffic_vsi(repr, ctrl_vsi); 2631a1c40dfSGrzegorz Nitka } 2641a1c40dfSGrzegorz Nitka 2651a1c40dfSGrzegorz Nitka return 0; 2661a1c40dfSGrzegorz Nitka 2671a1c40dfSGrzegorz Nitka err: 2681a1c40dfSGrzegorz Nitka for (i = i - 1; i >= 0; i--) { 2691a1c40dfSGrzegorz Nitka struct ice_vsi *vsi = pf->vf[i].repr->src_vsi; 2701a1c40dfSGrzegorz Nitka struct ice_vf *vf = &pf->vf[i]; 2711a1c40dfSGrzegorz Nitka 2721a1c40dfSGrzegorz Nitka ice_vsi_update_security(vsi, ice_vsi_ctx_set_antispoof); 2731a1c40dfSGrzegorz Nitka metadata_dst_free(vf->repr->dst); 2741a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(vsi, vf->hw_lan_addr.addr, 2751a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 2761a1c40dfSGrzegorz Nitka } 2771a1c40dfSGrzegorz Nitka 2781a1c40dfSGrzegorz Nitka return -ENODEV; 2791a1c40dfSGrzegorz Nitka } 2801a1c40dfSGrzegorz Nitka 2811a1c40dfSGrzegorz Nitka /** 2821a1c40dfSGrzegorz Nitka * ice_eswitch_release_reprs - clear PR VSIs configuration 2831a1c40dfSGrzegorz Nitka * @pf: poiner to PF struct 2841a1c40dfSGrzegorz Nitka * @ctrl_vsi: pointer to switchdev control VSI 2851a1c40dfSGrzegorz Nitka */ 2861a1c40dfSGrzegorz Nitka static void 2871a1c40dfSGrzegorz Nitka ice_eswitch_release_reprs(struct ice_pf *pf, struct ice_vsi *ctrl_vsi) 2881a1c40dfSGrzegorz Nitka { 2891a1c40dfSGrzegorz Nitka int i; 2901a1c40dfSGrzegorz Nitka 2911a1c40dfSGrzegorz Nitka ice_for_each_vf(pf, i) { 2921a1c40dfSGrzegorz Nitka struct ice_vsi *vsi = pf->vf[i].repr->src_vsi; 2931a1c40dfSGrzegorz Nitka struct ice_vf *vf = &pf->vf[i]; 2941a1c40dfSGrzegorz Nitka 2951a1c40dfSGrzegorz Nitka ice_vsi_update_security(vsi, ice_vsi_ctx_set_antispoof); 2961a1c40dfSGrzegorz Nitka metadata_dst_free(vf->repr->dst); 2971a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(vsi, vf->hw_lan_addr.addr, 2981a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 2991a1c40dfSGrzegorz Nitka 3001a1c40dfSGrzegorz Nitka netif_napi_del(&vf->repr->q_vector->napi); 3011a1c40dfSGrzegorz Nitka } 3021a1c40dfSGrzegorz Nitka } 3031a1c40dfSGrzegorz Nitka 3041a1c40dfSGrzegorz Nitka /** 3051c54c839SGrzegorz Nitka * ice_eswitch_update_repr - reconfigure VF port representor 3061c54c839SGrzegorz Nitka * @vsi: VF VSI for which port representor is configured 3071c54c839SGrzegorz Nitka */ 3081c54c839SGrzegorz Nitka void ice_eswitch_update_repr(struct ice_vsi *vsi) 3091c54c839SGrzegorz Nitka { 3101c54c839SGrzegorz Nitka struct ice_pf *pf = vsi->back; 3111c54c839SGrzegorz Nitka struct ice_repr *repr; 3121c54c839SGrzegorz Nitka struct ice_vf *vf; 3131c54c839SGrzegorz Nitka int ret; 3141c54c839SGrzegorz Nitka 3151c54c839SGrzegorz Nitka if (!ice_is_switchdev_running(pf)) 3161c54c839SGrzegorz Nitka return; 3171c54c839SGrzegorz Nitka 3181c54c839SGrzegorz Nitka vf = &pf->vf[vsi->vf_id]; 3191c54c839SGrzegorz Nitka repr = vf->repr; 3201c54c839SGrzegorz Nitka repr->src_vsi = vsi; 3211c54c839SGrzegorz Nitka repr->dst->u.port_info.port_id = vsi->vsi_num; 3221c54c839SGrzegorz Nitka 3231c54c839SGrzegorz Nitka ret = ice_vsi_update_security(vsi, ice_vsi_ctx_clear_antispoof); 3241c54c839SGrzegorz Nitka if (ret) { 3251c54c839SGrzegorz Nitka ice_fltr_add_mac_and_broadcast(vsi, vf->hw_lan_addr.addr, ICE_FWD_TO_VSI); 3261c54c839SGrzegorz Nitka dev_err(ice_pf_to_dev(pf), "Failed to update VF %d port representor", vsi->vf_id); 3271c54c839SGrzegorz Nitka } 3281c54c839SGrzegorz Nitka } 3291c54c839SGrzegorz Nitka 3301c54c839SGrzegorz Nitka /** 331f5396b8aSGrzegorz Nitka * ice_eswitch_port_start_xmit - callback for packets transmit 332f5396b8aSGrzegorz Nitka * @skb: send buffer 333f5396b8aSGrzegorz Nitka * @netdev: network interface device structure 334f5396b8aSGrzegorz Nitka * 335f5396b8aSGrzegorz Nitka * Returns NETDEV_TX_OK if sent, else an error code 336f5396b8aSGrzegorz Nitka */ 337f5396b8aSGrzegorz Nitka netdev_tx_t 338f5396b8aSGrzegorz Nitka ice_eswitch_port_start_xmit(struct sk_buff *skb, struct net_device *netdev) 339f5396b8aSGrzegorz Nitka { 340f5396b8aSGrzegorz Nitka struct ice_netdev_priv *np; 341f5396b8aSGrzegorz Nitka struct ice_repr *repr; 342f5396b8aSGrzegorz Nitka struct ice_vsi *vsi; 343f5396b8aSGrzegorz Nitka 344f5396b8aSGrzegorz Nitka np = netdev_priv(netdev); 345f5396b8aSGrzegorz Nitka vsi = np->vsi; 346f5396b8aSGrzegorz Nitka 347f5396b8aSGrzegorz Nitka if (ice_is_reset_in_progress(vsi->back->state)) 348f5396b8aSGrzegorz Nitka return NETDEV_TX_BUSY; 349f5396b8aSGrzegorz Nitka 350f5396b8aSGrzegorz Nitka repr = ice_netdev_to_repr(netdev); 351f5396b8aSGrzegorz Nitka skb_dst_drop(skb); 352f5396b8aSGrzegorz Nitka dst_hold((struct dst_entry *)repr->dst); 353f5396b8aSGrzegorz Nitka skb_dst_set(skb, (struct dst_entry *)repr->dst); 354f5396b8aSGrzegorz Nitka skb->queue_mapping = repr->vf->vf_id; 355f5396b8aSGrzegorz Nitka 356f5396b8aSGrzegorz Nitka return ice_start_xmit(skb, netdev); 357f5396b8aSGrzegorz Nitka } 358f5396b8aSGrzegorz Nitka 359f5396b8aSGrzegorz Nitka /** 360f5396b8aSGrzegorz Nitka * ice_eswitch_set_target_vsi - set switchdev context in Tx context descriptor 361f5396b8aSGrzegorz Nitka * @skb: pointer to send buffer 362f5396b8aSGrzegorz Nitka * @off: pointer to offload struct 363f5396b8aSGrzegorz Nitka */ 364f5396b8aSGrzegorz Nitka void 365f5396b8aSGrzegorz Nitka ice_eswitch_set_target_vsi(struct sk_buff *skb, 366f5396b8aSGrzegorz Nitka struct ice_tx_offload_params *off) 367f5396b8aSGrzegorz Nitka { 368f5396b8aSGrzegorz Nitka struct metadata_dst *dst = skb_metadata_dst(skb); 369f5396b8aSGrzegorz Nitka u64 cd_cmd, dst_vsi; 370f5396b8aSGrzegorz Nitka 371f5396b8aSGrzegorz Nitka if (!dst) { 372f5396b8aSGrzegorz Nitka cd_cmd = ICE_TX_CTX_DESC_SWTCH_UPLINK << ICE_TXD_CTX_QW1_CMD_S; 373f5396b8aSGrzegorz Nitka off->cd_qw1 |= (cd_cmd | ICE_TX_DESC_DTYPE_CTX); 374f5396b8aSGrzegorz Nitka } else { 375f5396b8aSGrzegorz Nitka cd_cmd = ICE_TX_CTX_DESC_SWTCH_VSI << ICE_TXD_CTX_QW1_CMD_S; 376f5396b8aSGrzegorz Nitka dst_vsi = ((u64)dst->u.port_info.port_id << 377f5396b8aSGrzegorz Nitka ICE_TXD_CTX_QW1_VSI_S) & ICE_TXD_CTX_QW1_VSI_M; 378f5396b8aSGrzegorz Nitka off->cd_qw1 = cd_cmd | dst_vsi | ICE_TX_DESC_DTYPE_CTX; 379f5396b8aSGrzegorz Nitka } 380f5396b8aSGrzegorz Nitka } 381f5396b8aSGrzegorz Nitka 382f5396b8aSGrzegorz Nitka /** 3831a1c40dfSGrzegorz Nitka * ice_eswitch_release_env - clear switchdev HW filters 3841a1c40dfSGrzegorz Nitka * @pf: pointer to PF struct 3851a1c40dfSGrzegorz Nitka * 3861a1c40dfSGrzegorz Nitka * This function removes HW filters configuration specific for switchdev 3871a1c40dfSGrzegorz Nitka * mode and restores default legacy mode settings. 3881a1c40dfSGrzegorz Nitka */ 3891a1c40dfSGrzegorz Nitka static void ice_eswitch_release_env(struct ice_pf *pf) 3901a1c40dfSGrzegorz Nitka { 3911a1c40dfSGrzegorz Nitka struct ice_vsi *uplink_vsi = pf->switchdev.uplink_vsi; 3921a1c40dfSGrzegorz Nitka struct ice_vsi *ctrl_vsi = pf->switchdev.control_vsi; 3931a1c40dfSGrzegorz Nitka 3941a1c40dfSGrzegorz Nitka ice_vsi_update_security(ctrl_vsi, ice_vsi_ctx_clear_allow_override); 3951a1c40dfSGrzegorz Nitka ice_vsi_update_security(uplink_vsi, ice_vsi_ctx_clear_allow_override); 3961a1c40dfSGrzegorz Nitka ice_clear_dflt_vsi(uplink_vsi->vsw); 3971a1c40dfSGrzegorz Nitka ice_fltr_add_mac_and_broadcast(uplink_vsi, 3981a1c40dfSGrzegorz Nitka uplink_vsi->port_info->mac.perm_addr, 3991a1c40dfSGrzegorz Nitka ICE_FWD_TO_VSI); 4001a1c40dfSGrzegorz Nitka } 4011a1c40dfSGrzegorz Nitka 4021a1c40dfSGrzegorz Nitka /** 4031a1c40dfSGrzegorz Nitka * ice_eswitch_vsi_setup - configure switchdev control VSI 4041a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 4051a1c40dfSGrzegorz Nitka * @pi: pointer to port_info structure 4061a1c40dfSGrzegorz Nitka */ 4071a1c40dfSGrzegorz Nitka static struct ice_vsi * 4081a1c40dfSGrzegorz Nitka ice_eswitch_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi) 4091a1c40dfSGrzegorz Nitka { 4100754d65bSKiran Patil return ice_vsi_setup(pf, pi, ICE_VSI_SWITCHDEV_CTRL, ICE_INVAL_VFID, NULL); 4111a1c40dfSGrzegorz Nitka } 4121a1c40dfSGrzegorz Nitka 4131a1c40dfSGrzegorz Nitka /** 414b3be918dSGrzegorz Nitka * ice_eswitch_napi_del - remove NAPI handle for all port representors 415b3be918dSGrzegorz Nitka * @pf: pointer to PF structure 416b3be918dSGrzegorz Nitka */ 417b3be918dSGrzegorz Nitka static void ice_eswitch_napi_del(struct ice_pf *pf) 418b3be918dSGrzegorz Nitka { 419b3be918dSGrzegorz Nitka int i; 420b3be918dSGrzegorz Nitka 421b3be918dSGrzegorz Nitka ice_for_each_vf(pf, i) 422b3be918dSGrzegorz Nitka netif_napi_del(&pf->vf[i].repr->q_vector->napi); 423b3be918dSGrzegorz Nitka } 424b3be918dSGrzegorz Nitka 425b3be918dSGrzegorz Nitka /** 4261a1c40dfSGrzegorz Nitka * ice_eswitch_napi_enable - enable NAPI for all port representors 4271a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 4281a1c40dfSGrzegorz Nitka */ 4291a1c40dfSGrzegorz Nitka static void ice_eswitch_napi_enable(struct ice_pf *pf) 4301a1c40dfSGrzegorz Nitka { 4311a1c40dfSGrzegorz Nitka int i; 4321a1c40dfSGrzegorz Nitka 4331a1c40dfSGrzegorz Nitka ice_for_each_vf(pf, i) 4341a1c40dfSGrzegorz Nitka napi_enable(&pf->vf[i].repr->q_vector->napi); 4351a1c40dfSGrzegorz Nitka } 4361a1c40dfSGrzegorz Nitka 4371a1c40dfSGrzegorz Nitka /** 4381a1c40dfSGrzegorz Nitka * ice_eswitch_napi_disable - disable NAPI for all port representors 4391a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 4401a1c40dfSGrzegorz Nitka */ 4411a1c40dfSGrzegorz Nitka static void ice_eswitch_napi_disable(struct ice_pf *pf) 4421a1c40dfSGrzegorz Nitka { 4431a1c40dfSGrzegorz Nitka int i; 4441a1c40dfSGrzegorz Nitka 4451a1c40dfSGrzegorz Nitka ice_for_each_vf(pf, i) 4461a1c40dfSGrzegorz Nitka napi_disable(&pf->vf[i].repr->q_vector->napi); 4471a1c40dfSGrzegorz Nitka } 4481a1c40dfSGrzegorz Nitka 4491a1c40dfSGrzegorz Nitka /** 4501a1c40dfSGrzegorz Nitka * ice_eswitch_enable_switchdev - configure eswitch in switchdev mode 4511a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 4521a1c40dfSGrzegorz Nitka */ 4531a1c40dfSGrzegorz Nitka static int ice_eswitch_enable_switchdev(struct ice_pf *pf) 4541a1c40dfSGrzegorz Nitka { 4551a1c40dfSGrzegorz Nitka struct ice_vsi *ctrl_vsi; 4561a1c40dfSGrzegorz Nitka 4571a1c40dfSGrzegorz Nitka pf->switchdev.control_vsi = ice_eswitch_vsi_setup(pf, pf->hw.port_info); 4581a1c40dfSGrzegorz Nitka if (!pf->switchdev.control_vsi) 4591a1c40dfSGrzegorz Nitka return -ENODEV; 4601a1c40dfSGrzegorz Nitka 4611a1c40dfSGrzegorz Nitka ctrl_vsi = pf->switchdev.control_vsi; 4621a1c40dfSGrzegorz Nitka pf->switchdev.uplink_vsi = ice_get_main_vsi(pf); 4631a1c40dfSGrzegorz Nitka if (!pf->switchdev.uplink_vsi) 4641a1c40dfSGrzegorz Nitka goto err_vsi; 4651a1c40dfSGrzegorz Nitka 4661a1c40dfSGrzegorz Nitka if (ice_eswitch_setup_env(pf)) 4671a1c40dfSGrzegorz Nitka goto err_vsi; 4681a1c40dfSGrzegorz Nitka 4691a1c40dfSGrzegorz Nitka if (ice_repr_add_for_all_vfs(pf)) 4701a1c40dfSGrzegorz Nitka goto err_repr_add; 4711a1c40dfSGrzegorz Nitka 4721a1c40dfSGrzegorz Nitka if (ice_eswitch_setup_reprs(pf)) 4731a1c40dfSGrzegorz Nitka goto err_setup_reprs; 4741a1c40dfSGrzegorz Nitka 4751a1c40dfSGrzegorz Nitka ice_eswitch_remap_rings_to_vectors(pf); 4761a1c40dfSGrzegorz Nitka 4771a1c40dfSGrzegorz Nitka if (ice_vsi_open(ctrl_vsi)) 4781a1c40dfSGrzegorz Nitka goto err_setup_reprs; 4791a1c40dfSGrzegorz Nitka 4801a1c40dfSGrzegorz Nitka ice_eswitch_napi_enable(pf); 4811a1c40dfSGrzegorz Nitka 4821a1c40dfSGrzegorz Nitka return 0; 4831a1c40dfSGrzegorz Nitka 4841a1c40dfSGrzegorz Nitka err_setup_reprs: 4851a1c40dfSGrzegorz Nitka ice_repr_rem_from_all_vfs(pf); 4861a1c40dfSGrzegorz Nitka err_repr_add: 4871a1c40dfSGrzegorz Nitka ice_eswitch_release_env(pf); 4881a1c40dfSGrzegorz Nitka err_vsi: 4891a1c40dfSGrzegorz Nitka ice_vsi_release(ctrl_vsi); 4901a1c40dfSGrzegorz Nitka return -ENODEV; 4911a1c40dfSGrzegorz Nitka } 4921a1c40dfSGrzegorz Nitka 4931a1c40dfSGrzegorz Nitka /** 4941a1c40dfSGrzegorz Nitka * ice_eswitch_disable_switchdev - disable switchdev resources 4951a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 4961a1c40dfSGrzegorz Nitka */ 4971a1c40dfSGrzegorz Nitka static void ice_eswitch_disable_switchdev(struct ice_pf *pf) 4981a1c40dfSGrzegorz Nitka { 4991a1c40dfSGrzegorz Nitka struct ice_vsi *ctrl_vsi = pf->switchdev.control_vsi; 5001a1c40dfSGrzegorz Nitka 5011a1c40dfSGrzegorz Nitka ice_eswitch_napi_disable(pf); 5021a1c40dfSGrzegorz Nitka ice_eswitch_release_env(pf); 503c1e5da5dSWojciech Drewek ice_rem_adv_rule_for_vsi(&pf->hw, ctrl_vsi->idx); 5041a1c40dfSGrzegorz Nitka ice_eswitch_release_reprs(pf, ctrl_vsi); 5051a1c40dfSGrzegorz Nitka ice_vsi_release(ctrl_vsi); 5061a1c40dfSGrzegorz Nitka ice_repr_rem_from_all_vfs(pf); 5071a1c40dfSGrzegorz Nitka } 5081a1c40dfSGrzegorz Nitka 5091a1c40dfSGrzegorz Nitka /** 5103ea9bd5dSMichal Swiatkowski * ice_eswitch_mode_set - set new eswitch mode 5113ea9bd5dSMichal Swiatkowski * @devlink: pointer to devlink structure 5123ea9bd5dSMichal Swiatkowski * @mode: eswitch mode to switch to 5133ea9bd5dSMichal Swiatkowski * @extack: pointer to extack structure 5143ea9bd5dSMichal Swiatkowski */ 5153ea9bd5dSMichal Swiatkowski int 5163ea9bd5dSMichal Swiatkowski ice_eswitch_mode_set(struct devlink *devlink, u16 mode, 5173ea9bd5dSMichal Swiatkowski struct netlink_ext_ack *extack) 5183ea9bd5dSMichal Swiatkowski { 5193ea9bd5dSMichal Swiatkowski struct ice_pf *pf = devlink_priv(devlink); 5203ea9bd5dSMichal Swiatkowski 5213ea9bd5dSMichal Swiatkowski if (pf->eswitch_mode == mode) 5223ea9bd5dSMichal Swiatkowski return 0; 5233ea9bd5dSMichal Swiatkowski 5243ea9bd5dSMichal Swiatkowski if (pf->num_alloc_vfs) { 5253ea9bd5dSMichal Swiatkowski dev_info(ice_pf_to_dev(pf), "Changing eswitch mode is allowed only if there is no VFs created"); 5263ea9bd5dSMichal Swiatkowski NL_SET_ERR_MSG_MOD(extack, "Changing eswitch mode is allowed only if there is no VFs created"); 5273ea9bd5dSMichal Swiatkowski return -EOPNOTSUPP; 5283ea9bd5dSMichal Swiatkowski } 5293ea9bd5dSMichal Swiatkowski 5303ea9bd5dSMichal Swiatkowski switch (mode) { 5313ea9bd5dSMichal Swiatkowski case DEVLINK_ESWITCH_MODE_LEGACY: 5323ea9bd5dSMichal Swiatkowski dev_info(ice_pf_to_dev(pf), "PF %d changed eswitch mode to legacy", 5333ea9bd5dSMichal Swiatkowski pf->hw.pf_id); 5343ea9bd5dSMichal Swiatkowski NL_SET_ERR_MSG_MOD(extack, "Changed eswitch mode to legacy"); 5353ea9bd5dSMichal Swiatkowski break; 5363ea9bd5dSMichal Swiatkowski case DEVLINK_ESWITCH_MODE_SWITCHDEV: 5373ea9bd5dSMichal Swiatkowski { 5383ea9bd5dSMichal Swiatkowski dev_info(ice_pf_to_dev(pf), "PF %d changed eswitch mode to switchdev", 5393ea9bd5dSMichal Swiatkowski pf->hw.pf_id); 5403ea9bd5dSMichal Swiatkowski NL_SET_ERR_MSG_MOD(extack, "Changed eswitch mode to switchdev"); 5413ea9bd5dSMichal Swiatkowski break; 5423ea9bd5dSMichal Swiatkowski } 5433ea9bd5dSMichal Swiatkowski default: 5443ea9bd5dSMichal Swiatkowski NL_SET_ERR_MSG_MOD(extack, "Unknown eswitch mode"); 5453ea9bd5dSMichal Swiatkowski return -EINVAL; 5463ea9bd5dSMichal Swiatkowski } 5473ea9bd5dSMichal Swiatkowski 5483ea9bd5dSMichal Swiatkowski pf->eswitch_mode = mode; 5493ea9bd5dSMichal Swiatkowski return 0; 5503ea9bd5dSMichal Swiatkowski } 5513ea9bd5dSMichal Swiatkowski 5523ea9bd5dSMichal Swiatkowski /** 5533ea9bd5dSMichal Swiatkowski * ice_eswitch_mode_get - get current eswitch mode 5543ea9bd5dSMichal Swiatkowski * @devlink: pointer to devlink structure 5553ea9bd5dSMichal Swiatkowski * @mode: output parameter for current eswitch mode 5563ea9bd5dSMichal Swiatkowski */ 5573ea9bd5dSMichal Swiatkowski int ice_eswitch_mode_get(struct devlink *devlink, u16 *mode) 5583ea9bd5dSMichal Swiatkowski { 5593ea9bd5dSMichal Swiatkowski struct ice_pf *pf = devlink_priv(devlink); 5603ea9bd5dSMichal Swiatkowski 5613ea9bd5dSMichal Swiatkowski *mode = pf->eswitch_mode; 5623ea9bd5dSMichal Swiatkowski return 0; 5633ea9bd5dSMichal Swiatkowski } 5641a1c40dfSGrzegorz Nitka 5651a1c40dfSGrzegorz Nitka /** 5661c54c839SGrzegorz Nitka * ice_is_eswitch_mode_switchdev - check if eswitch mode is set to switchdev 5671c54c839SGrzegorz Nitka * @pf: pointer to PF structure 5681c54c839SGrzegorz Nitka * 5691c54c839SGrzegorz Nitka * Returns true if eswitch mode is set to DEVLINK_ESWITCH_MODE_SWITCHDEV, 5701c54c839SGrzegorz Nitka * false otherwise. 5711c54c839SGrzegorz Nitka */ 5721c54c839SGrzegorz Nitka bool ice_is_eswitch_mode_switchdev(struct ice_pf *pf) 5731c54c839SGrzegorz Nitka { 5741c54c839SGrzegorz Nitka return pf->eswitch_mode == DEVLINK_ESWITCH_MODE_SWITCHDEV; 5751c54c839SGrzegorz Nitka } 5761c54c839SGrzegorz Nitka 5771c54c839SGrzegorz Nitka /** 5781a1c40dfSGrzegorz Nitka * ice_eswitch_release - cleanup eswitch 5791a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 5801a1c40dfSGrzegorz Nitka */ 5811a1c40dfSGrzegorz Nitka void ice_eswitch_release(struct ice_pf *pf) 5821a1c40dfSGrzegorz Nitka { 5831a1c40dfSGrzegorz Nitka if (pf->eswitch_mode == DEVLINK_ESWITCH_MODE_LEGACY) 5841a1c40dfSGrzegorz Nitka return; 5851a1c40dfSGrzegorz Nitka 5861a1c40dfSGrzegorz Nitka ice_eswitch_disable_switchdev(pf); 5871a1c40dfSGrzegorz Nitka pf->switchdev.is_running = false; 5881a1c40dfSGrzegorz Nitka } 5891a1c40dfSGrzegorz Nitka 5901a1c40dfSGrzegorz Nitka /** 5911a1c40dfSGrzegorz Nitka * ice_eswitch_configure - configure eswitch 5921a1c40dfSGrzegorz Nitka * @pf: pointer to PF structure 5931a1c40dfSGrzegorz Nitka */ 5941a1c40dfSGrzegorz Nitka int ice_eswitch_configure(struct ice_pf *pf) 5951a1c40dfSGrzegorz Nitka { 5961a1c40dfSGrzegorz Nitka int status; 5971a1c40dfSGrzegorz Nitka 5981a1c40dfSGrzegorz Nitka if (pf->eswitch_mode == DEVLINK_ESWITCH_MODE_LEGACY || pf->switchdev.is_running) 5991a1c40dfSGrzegorz Nitka return 0; 6001a1c40dfSGrzegorz Nitka 6011a1c40dfSGrzegorz Nitka status = ice_eswitch_enable_switchdev(pf); 6021a1c40dfSGrzegorz Nitka if (status) 6031a1c40dfSGrzegorz Nitka return status; 6041a1c40dfSGrzegorz Nitka 6051a1c40dfSGrzegorz Nitka pf->switchdev.is_running = true; 6061a1c40dfSGrzegorz Nitka return 0; 6071a1c40dfSGrzegorz Nitka } 608b3be918dSGrzegorz Nitka 609b3be918dSGrzegorz Nitka /** 610b3be918dSGrzegorz Nitka * ice_eswitch_start_all_tx_queues - start Tx queues of all port representors 611b3be918dSGrzegorz Nitka * @pf: pointer to PF structure 612b3be918dSGrzegorz Nitka */ 613b3be918dSGrzegorz Nitka static void ice_eswitch_start_all_tx_queues(struct ice_pf *pf) 614b3be918dSGrzegorz Nitka { 615b3be918dSGrzegorz Nitka struct ice_repr *repr; 616b3be918dSGrzegorz Nitka int i; 617b3be918dSGrzegorz Nitka 618b3be918dSGrzegorz Nitka if (test_bit(ICE_DOWN, pf->state)) 619b3be918dSGrzegorz Nitka return; 620b3be918dSGrzegorz Nitka 621b3be918dSGrzegorz Nitka ice_for_each_vf(pf, i) { 622b3be918dSGrzegorz Nitka repr = pf->vf[i].repr; 623b3be918dSGrzegorz Nitka if (repr) 624b3be918dSGrzegorz Nitka ice_repr_start_tx_queues(repr); 625b3be918dSGrzegorz Nitka } 626b3be918dSGrzegorz Nitka } 627b3be918dSGrzegorz Nitka 628b3be918dSGrzegorz Nitka /** 629b3be918dSGrzegorz Nitka * ice_eswitch_stop_all_tx_queues - stop Tx queues of all port representors 630b3be918dSGrzegorz Nitka * @pf: pointer to PF structure 631b3be918dSGrzegorz Nitka */ 632b3be918dSGrzegorz Nitka void ice_eswitch_stop_all_tx_queues(struct ice_pf *pf) 633b3be918dSGrzegorz Nitka { 634b3be918dSGrzegorz Nitka struct ice_repr *repr; 635b3be918dSGrzegorz Nitka int i; 636b3be918dSGrzegorz Nitka 637b3be918dSGrzegorz Nitka if (test_bit(ICE_DOWN, pf->state)) 638b3be918dSGrzegorz Nitka return; 639b3be918dSGrzegorz Nitka 640b3be918dSGrzegorz Nitka ice_for_each_vf(pf, i) { 641b3be918dSGrzegorz Nitka repr = pf->vf[i].repr; 642b3be918dSGrzegorz Nitka if (repr) 643b3be918dSGrzegorz Nitka ice_repr_stop_tx_queues(repr); 644b3be918dSGrzegorz Nitka } 645b3be918dSGrzegorz Nitka } 646b3be918dSGrzegorz Nitka 647b3be918dSGrzegorz Nitka /** 648b3be918dSGrzegorz Nitka * ice_eswitch_rebuild - rebuild eswitch 649b3be918dSGrzegorz Nitka * @pf: pointer to PF structure 650b3be918dSGrzegorz Nitka */ 651b3be918dSGrzegorz Nitka int ice_eswitch_rebuild(struct ice_pf *pf) 652b3be918dSGrzegorz Nitka { 653b3be918dSGrzegorz Nitka struct ice_vsi *ctrl_vsi = pf->switchdev.control_vsi; 654b3be918dSGrzegorz Nitka int status; 655b3be918dSGrzegorz Nitka 656b3be918dSGrzegorz Nitka ice_eswitch_napi_disable(pf); 657b3be918dSGrzegorz Nitka ice_eswitch_napi_del(pf); 658b3be918dSGrzegorz Nitka 659b3be918dSGrzegorz Nitka status = ice_eswitch_setup_env(pf); 660b3be918dSGrzegorz Nitka if (status) 661b3be918dSGrzegorz Nitka return status; 662b3be918dSGrzegorz Nitka 663b3be918dSGrzegorz Nitka status = ice_eswitch_setup_reprs(pf); 664b3be918dSGrzegorz Nitka if (status) 665b3be918dSGrzegorz Nitka return status; 666b3be918dSGrzegorz Nitka 667b3be918dSGrzegorz Nitka ice_eswitch_remap_rings_to_vectors(pf); 668b3be918dSGrzegorz Nitka 6697fde6d8bSMichal Swiatkowski ice_replay_tc_fltrs(pf); 6707fde6d8bSMichal Swiatkowski 671b3be918dSGrzegorz Nitka status = ice_vsi_open(ctrl_vsi); 672b3be918dSGrzegorz Nitka if (status) 673b3be918dSGrzegorz Nitka return status; 674b3be918dSGrzegorz Nitka 675b3be918dSGrzegorz Nitka ice_eswitch_napi_enable(pf); 676b3be918dSGrzegorz Nitka ice_eswitch_start_all_tx_queues(pf); 677b3be918dSGrzegorz Nitka 678b3be918dSGrzegorz Nitka return 0; 679b3be918dSGrzegorz Nitka } 680