1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright (C) 2022, Intel Corporation. */ 3 4 #include "ice_vf_lib_private.h" 5 #include "ice.h" 6 #include "ice_lib.h" 7 #include "ice_fltr.h" 8 #include "ice_virtchnl_allowlist.h" 9 10 /* Public functions which may be accessed by all driver files */ 11 12 /** 13 * ice_get_vf_by_id - Get pointer to VF by ID 14 * @pf: the PF private structure 15 * @vf_id: the VF ID to locate 16 * 17 * Locate and return a pointer to the VF structure associated with a given ID. 18 * Returns NULL if the ID does not have a valid VF structure associated with 19 * it. 20 * 21 * This function takes a reference to the VF, which must be released by 22 * calling ice_put_vf() once the caller is finished accessing the VF structure 23 * returned. 24 */ 25 struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id) 26 { 27 struct ice_vf *vf; 28 29 rcu_read_lock(); 30 hash_for_each_possible_rcu(pf->vfs.table, vf, entry, vf_id) { 31 if (vf->vf_id == vf_id) { 32 struct ice_vf *found; 33 34 if (kref_get_unless_zero(&vf->refcnt)) 35 found = vf; 36 else 37 found = NULL; 38 39 rcu_read_unlock(); 40 return found; 41 } 42 } 43 rcu_read_unlock(); 44 45 return NULL; 46 } 47 48 /** 49 * ice_release_vf - Release VF associated with a refcount 50 * @ref: the kref decremented to zero 51 * 52 * Callback function for kref_put to release a VF once its reference count has 53 * hit zero. 54 */ 55 static void ice_release_vf(struct kref *ref) 56 { 57 struct ice_vf *vf = container_of(ref, struct ice_vf, refcnt); 58 59 vf->vf_ops->free(vf); 60 } 61 62 /** 63 * ice_put_vf - Release a reference to a VF 64 * @vf: the VF structure to decrease reference count on 65 * 66 * Decrease the reference count for a VF, and free the entry if it is no 67 * longer in use. 68 * 69 * This must be called after ice_get_vf_by_id() once the reference to the VF 70 * structure is no longer used. Otherwise, the VF structure will never be 71 * freed. 72 */ 73 void ice_put_vf(struct ice_vf *vf) 74 { 75 kref_put(&vf->refcnt, ice_release_vf); 76 } 77 78 /** 79 * ice_has_vfs - Return true if the PF has any associated VFs 80 * @pf: the PF private structure 81 * 82 * Return whether or not the PF has any allocated VFs. 83 * 84 * Note that this function only guarantees that there are no VFs at the point 85 * of calling it. It does not guarantee that no more VFs will be added. 86 */ 87 bool ice_has_vfs(struct ice_pf *pf) 88 { 89 /* A simple check that the hash table is not empty does not require 90 * the mutex or rcu_read_lock. 91 */ 92 return !hash_empty(pf->vfs.table); 93 } 94 95 /** 96 * ice_get_num_vfs - Get number of allocated VFs 97 * @pf: the PF private structure 98 * 99 * Return the total number of allocated VFs. NOTE: VF IDs are not guaranteed 100 * to be contiguous. Do not assume that a VF ID is guaranteed to be less than 101 * the output of this function. 102 */ 103 u16 ice_get_num_vfs(struct ice_pf *pf) 104 { 105 struct ice_vf *vf; 106 unsigned int bkt; 107 u16 num_vfs = 0; 108 109 rcu_read_lock(); 110 ice_for_each_vf_rcu(pf, bkt, vf) 111 num_vfs++; 112 rcu_read_unlock(); 113 114 return num_vfs; 115 } 116 117 /** 118 * ice_get_vf_vsi - get VF's VSI based on the stored index 119 * @vf: VF used to get VSI 120 */ 121 struct ice_vsi *ice_get_vf_vsi(struct ice_vf *vf) 122 { 123 if (vf->lan_vsi_idx == ICE_NO_VSI) 124 return NULL; 125 126 return vf->pf->vsi[vf->lan_vsi_idx]; 127 } 128 129 /** 130 * ice_is_vf_disabled 131 * @vf: pointer to the VF info 132 * 133 * If the PF has been disabled, there is no need resetting VF until PF is 134 * active again. Similarly, if the VF has been disabled, this means something 135 * else is resetting the VF, so we shouldn't continue. 136 * 137 * Returns true if the caller should consider the VF as disabled whether 138 * because that single VF is explicitly disabled or because the PF is 139 * currently disabled. 140 */ 141 bool ice_is_vf_disabled(struct ice_vf *vf) 142 { 143 struct ice_pf *pf = vf->pf; 144 145 return (test_bit(ICE_VF_DIS, pf->state) || 146 test_bit(ICE_VF_STATE_DIS, vf->vf_states)); 147 } 148 149 /** 150 * ice_wait_on_vf_reset - poll to make sure a given VF is ready after reset 151 * @vf: The VF being resseting 152 * 153 * The max poll time is about ~800ms, which is about the maximum time it takes 154 * for a VF to be reset and/or a VF driver to be removed. 155 */ 156 static void ice_wait_on_vf_reset(struct ice_vf *vf) 157 { 158 int i; 159 160 for (i = 0; i < ICE_MAX_VF_RESET_TRIES; i++) { 161 if (test_bit(ICE_VF_STATE_INIT, vf->vf_states)) 162 break; 163 msleep(ICE_MAX_VF_RESET_SLEEP_MS); 164 } 165 } 166 167 /** 168 * ice_check_vf_ready_for_cfg - check if VF is ready to be configured/queried 169 * @vf: VF to check if it's ready to be configured/queried 170 * 171 * The purpose of this function is to make sure the VF is not in reset, not 172 * disabled, and initialized so it can be configured and/or queried by a host 173 * administrator. 174 */ 175 int ice_check_vf_ready_for_cfg(struct ice_vf *vf) 176 { 177 ice_wait_on_vf_reset(vf); 178 179 if (ice_is_vf_disabled(vf)) 180 return -EINVAL; 181 182 if (ice_check_vf_init(vf)) 183 return -EBUSY; 184 185 return 0; 186 } 187 188 /** 189 * ice_check_vf_ready_for_reset - check if VF is ready to be reset 190 * @vf: VF to check if it's ready to be reset 191 * 192 * The purpose of this function is to ensure that the VF is not in reset, 193 * disabled, and is both initialized and active, thus enabling us to safely 194 * initialize another reset. 195 */ 196 int ice_check_vf_ready_for_reset(struct ice_vf *vf) 197 { 198 int ret; 199 200 ret = ice_check_vf_ready_for_cfg(vf); 201 if (!ret && !test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states)) 202 ret = -EAGAIN; 203 204 return ret; 205 } 206 207 /** 208 * ice_trigger_vf_reset - Reset a VF on HW 209 * @vf: pointer to the VF structure 210 * @is_vflr: true if VFLR was issued, false if not 211 * @is_pfr: true if the reset was triggered due to a previous PFR 212 * 213 * Trigger hardware to start a reset for a particular VF. Expects the caller 214 * to wait the proper amount of time to allow hardware to reset the VF before 215 * it cleans up and restores VF functionality. 216 */ 217 static void ice_trigger_vf_reset(struct ice_vf *vf, bool is_vflr, bool is_pfr) 218 { 219 /* Inform VF that it is no longer active, as a warning */ 220 clear_bit(ICE_VF_STATE_ACTIVE, vf->vf_states); 221 222 /* Disable VF's configuration API during reset. The flag is re-enabled 223 * when it's safe again to access VF's VSI. 224 */ 225 clear_bit(ICE_VF_STATE_INIT, vf->vf_states); 226 227 /* VF_MBX_ARQLEN and VF_MBX_ATQLEN are cleared by PFR, so the driver 228 * needs to clear them in the case of VFR/VFLR. If this is done for 229 * PFR, it can mess up VF resets because the VF driver may already 230 * have started cleanup by the time we get here. 231 */ 232 if (!is_pfr) 233 vf->vf_ops->clear_mbx_register(vf); 234 235 vf->vf_ops->trigger_reset_register(vf, is_vflr); 236 } 237 238 static void ice_vf_clear_counters(struct ice_vf *vf) 239 { 240 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 241 242 if (vsi) 243 vsi->num_vlan = 0; 244 245 vf->num_mac = 0; 246 memset(&vf->mdd_tx_events, 0, sizeof(vf->mdd_tx_events)); 247 memset(&vf->mdd_rx_events, 0, sizeof(vf->mdd_rx_events)); 248 } 249 250 /** 251 * ice_vf_pre_vsi_rebuild - tasks to be done prior to VSI rebuild 252 * @vf: VF to perform pre VSI rebuild tasks 253 * 254 * These tasks are items that don't need to be amortized since they are most 255 * likely called in a for loop with all VF(s) in the reset_all_vfs() case. 256 */ 257 static void ice_vf_pre_vsi_rebuild(struct ice_vf *vf) 258 { 259 /* Close any IRQ mapping now */ 260 if (vf->vf_ops->irq_close) 261 vf->vf_ops->irq_close(vf); 262 263 ice_vf_clear_counters(vf); 264 vf->vf_ops->clear_reset_trigger(vf); 265 } 266 267 /** 268 * ice_vf_recreate_vsi - Release and re-create the VF's VSI 269 * @vf: VF to recreate the VSI for 270 * 271 * This is only called when a single VF is being reset (i.e. VVF, VFLR, host 272 * VF configuration change, etc) 273 * 274 * It releases and then re-creates a new VSI. 275 */ 276 static int ice_vf_recreate_vsi(struct ice_vf *vf) 277 { 278 struct ice_pf *pf = vf->pf; 279 int err; 280 281 ice_vf_vsi_release(vf); 282 283 err = vf->vf_ops->create_vsi(vf); 284 if (err) { 285 dev_err(ice_pf_to_dev(pf), 286 "Failed to recreate the VF%u's VSI, error %d\n", 287 vf->vf_id, err); 288 return err; 289 } 290 291 return 0; 292 } 293 294 /** 295 * ice_vf_rebuild_vsi - rebuild the VF's VSI 296 * @vf: VF to rebuild the VSI for 297 * 298 * This is only called when all VF(s) are being reset (i.e. PCIe Reset on the 299 * host, PFR, CORER, etc.). 300 * 301 * It reprograms the VSI configuration back into hardware. 302 */ 303 static int ice_vf_rebuild_vsi(struct ice_vf *vf) 304 { 305 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 306 struct ice_pf *pf = vf->pf; 307 308 if (WARN_ON(!vsi)) 309 return -EINVAL; 310 311 if (ice_vsi_rebuild(vsi, ICE_VSI_FLAG_INIT)) { 312 dev_err(ice_pf_to_dev(pf), "failed to rebuild VF %d VSI\n", 313 vf->vf_id); 314 return -EIO; 315 } 316 /* vsi->idx will remain the same in this case so don't update 317 * vf->lan_vsi_idx 318 */ 319 vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx); 320 vf->lan_vsi_num = vsi->vsi_num; 321 322 return 0; 323 } 324 325 /** 326 * ice_vf_post_vsi_rebuild - Reset tasks that occur after VSI rebuild 327 * @vf: the VF being reset 328 * 329 * Perform reset tasks which must occur after the VSI has been re-created or 330 * rebuilt during a VF reset. 331 */ 332 static void ice_vf_post_vsi_rebuild(struct ice_vf *vf) 333 { 334 ice_vf_rebuild_host_cfg(vf); 335 ice_vf_set_initialized(vf); 336 337 vf->vf_ops->post_vsi_rebuild(vf); 338 } 339 340 /** 341 * ice_is_any_vf_in_unicast_promisc - check if any VF(s) 342 * are in unicast promiscuous mode 343 * @pf: PF structure for accessing VF(s) 344 * 345 * Return false if no VF(s) are in unicast promiscuous mode, 346 * else return true 347 */ 348 bool ice_is_any_vf_in_unicast_promisc(struct ice_pf *pf) 349 { 350 bool is_vf_promisc = false; 351 struct ice_vf *vf; 352 unsigned int bkt; 353 354 rcu_read_lock(); 355 ice_for_each_vf_rcu(pf, bkt, vf) { 356 /* found a VF that has promiscuous mode configured */ 357 if (test_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states)) { 358 is_vf_promisc = true; 359 break; 360 } 361 } 362 rcu_read_unlock(); 363 364 return is_vf_promisc; 365 } 366 367 /** 368 * ice_vf_get_promisc_masks - Calculate masks for promiscuous modes 369 * @vf: the VF pointer 370 * @vsi: the VSI to configure 371 * @ucast_m: promiscuous mask to apply to unicast 372 * @mcast_m: promiscuous mask to apply to multicast 373 * 374 * Decide which mask should be used for unicast and multicast filter, 375 * based on presence of VLANs 376 */ 377 void 378 ice_vf_get_promisc_masks(struct ice_vf *vf, struct ice_vsi *vsi, 379 u8 *ucast_m, u8 *mcast_m) 380 { 381 if (ice_vf_is_port_vlan_ena(vf) || 382 ice_vsi_has_non_zero_vlans(vsi)) { 383 *mcast_m = ICE_MCAST_VLAN_PROMISC_BITS; 384 *ucast_m = ICE_UCAST_VLAN_PROMISC_BITS; 385 } else { 386 *mcast_m = ICE_MCAST_PROMISC_BITS; 387 *ucast_m = ICE_UCAST_PROMISC_BITS; 388 } 389 } 390 391 /** 392 * ice_vf_clear_all_promisc_modes - Clear promisc/allmulticast on VF VSI 393 * @vf: the VF pointer 394 * @vsi: the VSI to configure 395 * 396 * Clear all promiscuous/allmulticast filters for a VF 397 */ 398 static int 399 ice_vf_clear_all_promisc_modes(struct ice_vf *vf, struct ice_vsi *vsi) 400 { 401 struct ice_pf *pf = vf->pf; 402 u8 ucast_m, mcast_m; 403 int ret = 0; 404 405 ice_vf_get_promisc_masks(vf, vsi, &ucast_m, &mcast_m); 406 if (test_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states)) { 407 if (!test_bit(ICE_FLAG_VF_TRUE_PROMISC_ENA, pf->flags)) { 408 if (ice_is_dflt_vsi_in_use(vsi->port_info)) 409 ret = ice_clear_dflt_vsi(vsi); 410 } else { 411 ret = ice_vf_clear_vsi_promisc(vf, vsi, ucast_m); 412 } 413 414 if (ret) { 415 dev_err(ice_pf_to_dev(vf->pf), "Disabling promiscuous mode failed\n"); 416 } else { 417 clear_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states); 418 dev_info(ice_pf_to_dev(vf->pf), "Disabling promiscuous mode succeeded\n"); 419 } 420 } 421 422 if (test_bit(ICE_VF_STATE_MC_PROMISC, vf->vf_states)) { 423 ret = ice_vf_clear_vsi_promisc(vf, vsi, mcast_m); 424 if (ret) { 425 dev_err(ice_pf_to_dev(vf->pf), "Disabling allmulticast mode failed\n"); 426 } else { 427 clear_bit(ICE_VF_STATE_MC_PROMISC, vf->vf_states); 428 dev_info(ice_pf_to_dev(vf->pf), "Disabling allmulticast mode succeeded\n"); 429 } 430 } 431 return ret; 432 } 433 434 /** 435 * ice_vf_set_vsi_promisc - Enable promiscuous mode for a VF VSI 436 * @vf: the VF to configure 437 * @vsi: the VF's VSI 438 * @promisc_m: the promiscuous mode to enable 439 */ 440 int 441 ice_vf_set_vsi_promisc(struct ice_vf *vf, struct ice_vsi *vsi, u8 promisc_m) 442 { 443 struct ice_hw *hw = &vsi->back->hw; 444 int status; 445 446 if (ice_vf_is_port_vlan_ena(vf)) 447 status = ice_fltr_set_vsi_promisc(hw, vsi->idx, promisc_m, 448 ice_vf_get_port_vlan_id(vf)); 449 else if (ice_vsi_has_non_zero_vlans(vsi)) 450 status = ice_fltr_set_vlan_vsi_promisc(hw, vsi, promisc_m); 451 else 452 status = ice_fltr_set_vsi_promisc(hw, vsi->idx, promisc_m, 0); 453 454 if (status && status != -EEXIST) { 455 dev_err(ice_pf_to_dev(vsi->back), "enable Tx/Rx filter promiscuous mode on VF-%u failed, error: %d\n", 456 vf->vf_id, status); 457 return status; 458 } 459 460 return 0; 461 } 462 463 /** 464 * ice_vf_clear_vsi_promisc - Disable promiscuous mode for a VF VSI 465 * @vf: the VF to configure 466 * @vsi: the VF's VSI 467 * @promisc_m: the promiscuous mode to disable 468 */ 469 int 470 ice_vf_clear_vsi_promisc(struct ice_vf *vf, struct ice_vsi *vsi, u8 promisc_m) 471 { 472 struct ice_hw *hw = &vsi->back->hw; 473 int status; 474 475 if (ice_vf_is_port_vlan_ena(vf)) 476 status = ice_fltr_clear_vsi_promisc(hw, vsi->idx, promisc_m, 477 ice_vf_get_port_vlan_id(vf)); 478 else if (ice_vsi_has_non_zero_vlans(vsi)) 479 status = ice_fltr_clear_vlan_vsi_promisc(hw, vsi, promisc_m); 480 else 481 status = ice_fltr_clear_vsi_promisc(hw, vsi->idx, promisc_m, 0); 482 483 if (status && status != -ENOENT) { 484 dev_err(ice_pf_to_dev(vsi->back), "disable Tx/Rx filter promiscuous mode on VF-%u failed, error: %d\n", 485 vf->vf_id, status); 486 return status; 487 } 488 489 return 0; 490 } 491 492 /** 493 * ice_reset_all_vfs - reset all allocated VFs in one go 494 * @pf: pointer to the PF structure 495 * 496 * Reset all VFs at once, in response to a PF or other device reset. 497 * 498 * First, tell the hardware to reset each VF, then do all the waiting in one 499 * chunk, and finally finish restoring each VF after the wait. This is useful 500 * during PF routines which need to reset all VFs, as otherwise it must perform 501 * these resets in a serialized fashion. 502 */ 503 void ice_reset_all_vfs(struct ice_pf *pf) 504 { 505 struct device *dev = ice_pf_to_dev(pf); 506 struct ice_hw *hw = &pf->hw; 507 struct ice_vf *vf; 508 unsigned int bkt; 509 510 /* If we don't have any VFs, then there is nothing to reset */ 511 if (!ice_has_vfs(pf)) 512 return; 513 514 mutex_lock(&pf->vfs.table_lock); 515 516 /* clear all malicious info if the VFs are getting reset */ 517 ice_for_each_vf(pf, bkt, vf) 518 ice_mbx_clear_malvf(&vf->mbx_info); 519 520 /* If VFs have been disabled, there is no need to reset */ 521 if (test_and_set_bit(ICE_VF_DIS, pf->state)) { 522 mutex_unlock(&pf->vfs.table_lock); 523 return; 524 } 525 526 /* Begin reset on all VFs at once */ 527 ice_for_each_vf(pf, bkt, vf) 528 ice_trigger_vf_reset(vf, true, true); 529 530 /* HW requires some time to make sure it can flush the FIFO for a VF 531 * when it resets it. Now that we've triggered all of the VFs, iterate 532 * the table again and wait for each VF to complete. 533 */ 534 ice_for_each_vf(pf, bkt, vf) { 535 if (!vf->vf_ops->poll_reset_status(vf)) { 536 /* Display a warning if at least one VF didn't manage 537 * to reset in time, but continue on with the 538 * operation. 539 */ 540 dev_warn(dev, "VF %u reset check timeout\n", vf->vf_id); 541 break; 542 } 543 } 544 545 /* free VF resources to begin resetting the VSI state */ 546 ice_for_each_vf(pf, bkt, vf) { 547 mutex_lock(&vf->cfg_lock); 548 549 vf->driver_caps = 0; 550 ice_vc_set_default_allowlist(vf); 551 552 ice_vf_fdir_exit(vf); 553 ice_vf_fdir_init(vf); 554 /* clean VF control VSI when resetting VFs since it should be 555 * setup only when VF creates its first FDIR rule. 556 */ 557 if (vf->ctrl_vsi_idx != ICE_NO_VSI) 558 ice_vf_ctrl_invalidate_vsi(vf); 559 560 ice_vf_pre_vsi_rebuild(vf); 561 ice_vf_rebuild_vsi(vf); 562 ice_vf_post_vsi_rebuild(vf); 563 564 mutex_unlock(&vf->cfg_lock); 565 } 566 567 if (ice_is_eswitch_mode_switchdev(pf)) 568 if (ice_eswitch_rebuild(pf)) 569 dev_warn(dev, "eswitch rebuild failed\n"); 570 571 ice_flush(hw); 572 clear_bit(ICE_VF_DIS, pf->state); 573 574 mutex_unlock(&pf->vfs.table_lock); 575 } 576 577 /** 578 * ice_notify_vf_reset - Notify VF of a reset event 579 * @vf: pointer to the VF structure 580 */ 581 static void ice_notify_vf_reset(struct ice_vf *vf) 582 { 583 struct ice_hw *hw = &vf->pf->hw; 584 struct virtchnl_pf_event pfe; 585 586 /* Bail out if VF is in disabled state, neither initialized, nor active 587 * state - otherwise proceed with notifications 588 */ 589 if ((!test_bit(ICE_VF_STATE_INIT, vf->vf_states) && 590 !test_bit(ICE_VF_STATE_ACTIVE, vf->vf_states)) || 591 test_bit(ICE_VF_STATE_DIS, vf->vf_states)) 592 return; 593 594 pfe.event = VIRTCHNL_EVENT_RESET_IMPENDING; 595 pfe.severity = PF_EVENT_SEVERITY_CERTAIN_DOOM; 596 ice_aq_send_msg_to_vf(hw, vf->vf_id, VIRTCHNL_OP_EVENT, 597 VIRTCHNL_STATUS_SUCCESS, (u8 *)&pfe, sizeof(pfe), 598 NULL); 599 } 600 601 /** 602 * ice_reset_vf - Reset a particular VF 603 * @vf: pointer to the VF structure 604 * @flags: flags controlling behavior of the reset 605 * 606 * Flags: 607 * ICE_VF_RESET_VFLR - Indicates a reset is due to VFLR event 608 * ICE_VF_RESET_NOTIFY - Send VF a notification prior to reset 609 * ICE_VF_RESET_LOCK - Acquire VF cfg_lock before resetting 610 * 611 * Returns 0 if the VF is currently in reset, if resets are disabled, or if 612 * the VF resets successfully. Returns an error code if the VF fails to 613 * rebuild. 614 */ 615 int ice_reset_vf(struct ice_vf *vf, u32 flags) 616 { 617 struct ice_pf *pf = vf->pf; 618 struct ice_vsi *vsi; 619 struct device *dev; 620 int err = 0; 621 bool rsd; 622 623 dev = ice_pf_to_dev(pf); 624 625 if (flags & ICE_VF_RESET_NOTIFY) 626 ice_notify_vf_reset(vf); 627 628 if (test_bit(ICE_VF_RESETS_DISABLED, pf->state)) { 629 dev_dbg(dev, "Trying to reset VF %d, but all VF resets are disabled\n", 630 vf->vf_id); 631 return 0; 632 } 633 634 if (ice_is_vf_disabled(vf)) { 635 vsi = ice_get_vf_vsi(vf); 636 if (!vsi) { 637 dev_dbg(dev, "VF is already removed\n"); 638 return -EINVAL; 639 } 640 ice_vsi_stop_lan_tx_rings(vsi, ICE_NO_RESET, vf->vf_id); 641 642 if (ice_vsi_is_rx_queue_active(vsi)) 643 ice_vsi_stop_all_rx_rings(vsi); 644 645 dev_dbg(dev, "VF is already disabled, there is no need for resetting it, telling VM, all is fine %d\n", 646 vf->vf_id); 647 return 0; 648 } 649 650 if (flags & ICE_VF_RESET_LOCK) 651 mutex_lock(&vf->cfg_lock); 652 else 653 lockdep_assert_held(&vf->cfg_lock); 654 655 /* Set VF disable bit state here, before triggering reset */ 656 set_bit(ICE_VF_STATE_DIS, vf->vf_states); 657 ice_trigger_vf_reset(vf, flags & ICE_VF_RESET_VFLR, false); 658 659 vsi = ice_get_vf_vsi(vf); 660 if (WARN_ON(!vsi)) { 661 err = -EIO; 662 goto out_unlock; 663 } 664 665 ice_dis_vf_qs(vf); 666 667 /* Call Disable LAN Tx queue AQ whether or not queues are 668 * enabled. This is needed for successful completion of VFR. 669 */ 670 ice_dis_vsi_txq(vsi->port_info, vsi->idx, 0, 0, NULL, NULL, 671 NULL, vf->vf_ops->reset_type, vf->vf_id, NULL); 672 673 /* poll VPGEN_VFRSTAT reg to make sure 674 * that reset is complete 675 */ 676 rsd = vf->vf_ops->poll_reset_status(vf); 677 678 /* Display a warning if VF didn't manage to reset in time, but need to 679 * continue on with the operation. 680 */ 681 if (!rsd) 682 dev_warn(dev, "VF reset check timeout on VF %d\n", vf->vf_id); 683 684 vf->driver_caps = 0; 685 ice_vc_set_default_allowlist(vf); 686 687 /* disable promiscuous modes in case they were enabled 688 * ignore any error if disabling process failed 689 */ 690 ice_vf_clear_all_promisc_modes(vf, vsi); 691 692 ice_eswitch_del_vf_mac_rule(vf); 693 694 ice_vf_fdir_exit(vf); 695 ice_vf_fdir_init(vf); 696 /* clean VF control VSI when resetting VF since it should be setup 697 * only when VF creates its first FDIR rule. 698 */ 699 if (vf->ctrl_vsi_idx != ICE_NO_VSI) 700 ice_vf_ctrl_vsi_release(vf); 701 702 ice_vf_pre_vsi_rebuild(vf); 703 704 if (ice_vf_recreate_vsi(vf)) { 705 dev_err(dev, "Failed to release and setup the VF%u's VSI\n", 706 vf->vf_id); 707 err = -EFAULT; 708 goto out_unlock; 709 } 710 711 ice_vf_post_vsi_rebuild(vf); 712 vsi = ice_get_vf_vsi(vf); 713 if (WARN_ON(!vsi)) { 714 err = -EINVAL; 715 goto out_unlock; 716 } 717 718 ice_eswitch_update_repr(vsi); 719 ice_eswitch_replay_vf_mac_rule(vf); 720 721 /* if the VF has been reset allow it to come up again */ 722 ice_mbx_clear_malvf(&vf->mbx_info); 723 724 out_unlock: 725 if (flags & ICE_VF_RESET_LOCK) 726 mutex_unlock(&vf->cfg_lock); 727 728 return err; 729 } 730 731 /** 732 * ice_set_vf_state_qs_dis - Set VF queues state to disabled 733 * @vf: pointer to the VF structure 734 */ 735 static void ice_set_vf_state_qs_dis(struct ice_vf *vf) 736 { 737 /* Clear Rx/Tx enabled queues flag */ 738 bitmap_zero(vf->txq_ena, ICE_MAX_RSS_QS_PER_VF); 739 bitmap_zero(vf->rxq_ena, ICE_MAX_RSS_QS_PER_VF); 740 clear_bit(ICE_VF_STATE_QS_ENA, vf->vf_states); 741 } 742 743 /** 744 * ice_set_vf_state_dis - Set VF state to disabled 745 * @vf: pointer to the VF structure 746 */ 747 void ice_set_vf_state_dis(struct ice_vf *vf) 748 { 749 ice_set_vf_state_qs_dis(vf); 750 vf->vf_ops->clear_reset_state(vf); 751 } 752 753 /* Private functions only accessed from other virtualization files */ 754 755 /** 756 * ice_initialize_vf_entry - Initialize a VF entry 757 * @vf: pointer to the VF structure 758 */ 759 void ice_initialize_vf_entry(struct ice_vf *vf) 760 { 761 struct ice_pf *pf = vf->pf; 762 struct ice_vfs *vfs; 763 764 vfs = &pf->vfs; 765 766 /* assign default capabilities */ 767 vf->spoofchk = true; 768 vf->num_vf_qs = vfs->num_qps_per; 769 ice_vc_set_default_allowlist(vf); 770 ice_virtchnl_set_dflt_ops(vf); 771 772 /* ctrl_vsi_idx will be set to a valid value only when iAVF 773 * creates its first fdir rule. 774 */ 775 ice_vf_ctrl_invalidate_vsi(vf); 776 ice_vf_fdir_init(vf); 777 778 /* Initialize mailbox info for this VF */ 779 ice_mbx_init_vf_info(&pf->hw, &vf->mbx_info); 780 781 mutex_init(&vf->cfg_lock); 782 } 783 784 /** 785 * ice_dis_vf_qs - Disable the VF queues 786 * @vf: pointer to the VF structure 787 */ 788 void ice_dis_vf_qs(struct ice_vf *vf) 789 { 790 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 791 792 if (WARN_ON(!vsi)) 793 return; 794 795 ice_vsi_stop_lan_tx_rings(vsi, ICE_NO_RESET, vf->vf_id); 796 ice_vsi_stop_all_rx_rings(vsi); 797 ice_set_vf_state_qs_dis(vf); 798 } 799 800 /** 801 * ice_err_to_virt_err - translate errors for VF return code 802 * @err: error return code 803 */ 804 enum virtchnl_status_code ice_err_to_virt_err(int err) 805 { 806 switch (err) { 807 case 0: 808 return VIRTCHNL_STATUS_SUCCESS; 809 case -EINVAL: 810 case -ENODEV: 811 return VIRTCHNL_STATUS_ERR_PARAM; 812 case -ENOMEM: 813 return VIRTCHNL_STATUS_ERR_NO_MEMORY; 814 case -EALREADY: 815 case -EBUSY: 816 case -EIO: 817 case -ENOSPC: 818 return VIRTCHNL_STATUS_ERR_ADMIN_QUEUE_ERROR; 819 default: 820 return VIRTCHNL_STATUS_ERR_NOT_SUPPORTED; 821 } 822 } 823 824 /** 825 * ice_check_vf_init - helper to check if VF init complete 826 * @vf: the pointer to the VF to check 827 */ 828 int ice_check_vf_init(struct ice_vf *vf) 829 { 830 struct ice_pf *pf = vf->pf; 831 832 if (!test_bit(ICE_VF_STATE_INIT, vf->vf_states)) { 833 dev_err(ice_pf_to_dev(pf), "VF ID: %u in reset. Try again.\n", 834 vf->vf_id); 835 return -EBUSY; 836 } 837 return 0; 838 } 839 840 /** 841 * ice_vf_get_port_info - Get the VF's port info structure 842 * @vf: VF used to get the port info structure for 843 */ 844 struct ice_port_info *ice_vf_get_port_info(struct ice_vf *vf) 845 { 846 return vf->pf->hw.port_info; 847 } 848 849 /** 850 * ice_cfg_mac_antispoof - Configure MAC antispoof checking behavior 851 * @vsi: the VSI to configure 852 * @enable: whether to enable or disable the spoof checking 853 * 854 * Configure a VSI to enable (or disable) spoof checking behavior. 855 */ 856 static int ice_cfg_mac_antispoof(struct ice_vsi *vsi, bool enable) 857 { 858 struct ice_vsi_ctx *ctx; 859 int err; 860 861 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); 862 if (!ctx) 863 return -ENOMEM; 864 865 ctx->info.sec_flags = vsi->info.sec_flags; 866 ctx->info.valid_sections = cpu_to_le16(ICE_AQ_VSI_PROP_SECURITY_VALID); 867 868 if (enable) 869 ctx->info.sec_flags |= ICE_AQ_VSI_SEC_FLAG_ENA_MAC_ANTI_SPOOF; 870 else 871 ctx->info.sec_flags &= ~ICE_AQ_VSI_SEC_FLAG_ENA_MAC_ANTI_SPOOF; 872 873 err = ice_update_vsi(&vsi->back->hw, vsi->idx, ctx, NULL); 874 if (err) 875 dev_err(ice_pf_to_dev(vsi->back), "Failed to configure Tx MAC anti-spoof %s for VSI %d, error %d\n", 876 enable ? "ON" : "OFF", vsi->vsi_num, err); 877 else 878 vsi->info.sec_flags = ctx->info.sec_flags; 879 880 kfree(ctx); 881 882 return err; 883 } 884 885 /** 886 * ice_vsi_ena_spoofchk - enable Tx spoof checking for this VSI 887 * @vsi: VSI to enable Tx spoof checking for 888 */ 889 static int ice_vsi_ena_spoofchk(struct ice_vsi *vsi) 890 { 891 struct ice_vsi_vlan_ops *vlan_ops; 892 int err = 0; 893 894 vlan_ops = ice_get_compat_vsi_vlan_ops(vsi); 895 896 /* Allow VF with VLAN 0 only to send all tagged traffic */ 897 if (vsi->type != ICE_VSI_VF || ice_vsi_has_non_zero_vlans(vsi)) { 898 err = vlan_ops->ena_tx_filtering(vsi); 899 if (err) 900 return err; 901 } 902 903 return ice_cfg_mac_antispoof(vsi, true); 904 } 905 906 /** 907 * ice_vsi_dis_spoofchk - disable Tx spoof checking for this VSI 908 * @vsi: VSI to disable Tx spoof checking for 909 */ 910 static int ice_vsi_dis_spoofchk(struct ice_vsi *vsi) 911 { 912 struct ice_vsi_vlan_ops *vlan_ops; 913 int err; 914 915 vlan_ops = ice_get_compat_vsi_vlan_ops(vsi); 916 917 err = vlan_ops->dis_tx_filtering(vsi); 918 if (err) 919 return err; 920 921 return ice_cfg_mac_antispoof(vsi, false); 922 } 923 924 /** 925 * ice_vsi_apply_spoofchk - Apply Tx spoof checking setting to a VSI 926 * @vsi: VSI associated to the VF 927 * @enable: whether to enable or disable the spoof checking 928 */ 929 int ice_vsi_apply_spoofchk(struct ice_vsi *vsi, bool enable) 930 { 931 int err; 932 933 if (enable) 934 err = ice_vsi_ena_spoofchk(vsi); 935 else 936 err = ice_vsi_dis_spoofchk(vsi); 937 938 return err; 939 } 940 941 /** 942 * ice_is_vf_trusted 943 * @vf: pointer to the VF info 944 */ 945 bool ice_is_vf_trusted(struct ice_vf *vf) 946 { 947 return test_bit(ICE_VIRTCHNL_VF_CAP_PRIVILEGE, &vf->vf_caps); 948 } 949 950 /** 951 * ice_vf_has_no_qs_ena - check if the VF has any Rx or Tx queues enabled 952 * @vf: the VF to check 953 * 954 * Returns true if the VF has no Rx and no Tx queues enabled and returns false 955 * otherwise 956 */ 957 bool ice_vf_has_no_qs_ena(struct ice_vf *vf) 958 { 959 return (!bitmap_weight(vf->rxq_ena, ICE_MAX_RSS_QS_PER_VF) && 960 !bitmap_weight(vf->txq_ena, ICE_MAX_RSS_QS_PER_VF)); 961 } 962 963 /** 964 * ice_is_vf_link_up - check if the VF's link is up 965 * @vf: VF to check if link is up 966 */ 967 bool ice_is_vf_link_up(struct ice_vf *vf) 968 { 969 struct ice_port_info *pi = ice_vf_get_port_info(vf); 970 971 if (ice_check_vf_init(vf)) 972 return false; 973 974 if (ice_vf_has_no_qs_ena(vf)) 975 return false; 976 else if (vf->link_forced) 977 return vf->link_up; 978 else 979 return pi->phy.link_info.link_info & 980 ICE_AQ_LINK_UP; 981 } 982 983 /** 984 * ice_vf_set_host_trust_cfg - set trust setting based on pre-reset value 985 * @vf: VF to configure trust setting for 986 */ 987 static void ice_vf_set_host_trust_cfg(struct ice_vf *vf) 988 { 989 if (vf->trusted) 990 set_bit(ICE_VIRTCHNL_VF_CAP_PRIVILEGE, &vf->vf_caps); 991 else 992 clear_bit(ICE_VIRTCHNL_VF_CAP_PRIVILEGE, &vf->vf_caps); 993 } 994 995 /** 996 * ice_vf_rebuild_host_mac_cfg - add broadcast and the VF's perm_addr/LAA 997 * @vf: VF to add MAC filters for 998 * 999 * Called after a VF VSI has been re-added/rebuilt during reset. The PF driver 1000 * always re-adds a broadcast filter and the VF's perm_addr/LAA after reset. 1001 */ 1002 static int ice_vf_rebuild_host_mac_cfg(struct ice_vf *vf) 1003 { 1004 struct device *dev = ice_pf_to_dev(vf->pf); 1005 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 1006 u8 broadcast[ETH_ALEN]; 1007 int status; 1008 1009 if (WARN_ON(!vsi)) 1010 return -EINVAL; 1011 1012 if (ice_is_eswitch_mode_switchdev(vf->pf)) 1013 return 0; 1014 1015 eth_broadcast_addr(broadcast); 1016 status = ice_fltr_add_mac(vsi, broadcast, ICE_FWD_TO_VSI); 1017 if (status) { 1018 dev_err(dev, "failed to add broadcast MAC filter for VF %u, error %d\n", 1019 vf->vf_id, status); 1020 return status; 1021 } 1022 1023 vf->num_mac++; 1024 1025 if (is_valid_ether_addr(vf->hw_lan_addr)) { 1026 status = ice_fltr_add_mac(vsi, vf->hw_lan_addr, 1027 ICE_FWD_TO_VSI); 1028 if (status) { 1029 dev_err(dev, "failed to add default unicast MAC filter %pM for VF %u, error %d\n", 1030 &vf->hw_lan_addr[0], vf->vf_id, 1031 status); 1032 return status; 1033 } 1034 vf->num_mac++; 1035 1036 ether_addr_copy(vf->dev_lan_addr, vf->hw_lan_addr); 1037 } 1038 1039 return 0; 1040 } 1041 1042 /** 1043 * ice_vf_rebuild_host_vlan_cfg - add VLAN 0 filter or rebuild the Port VLAN 1044 * @vf: VF to add MAC filters for 1045 * @vsi: Pointer to VSI 1046 * 1047 * Called after a VF VSI has been re-added/rebuilt during reset. The PF driver 1048 * always re-adds either a VLAN 0 or port VLAN based filter after reset. 1049 */ 1050 static int ice_vf_rebuild_host_vlan_cfg(struct ice_vf *vf, struct ice_vsi *vsi) 1051 { 1052 struct ice_vsi_vlan_ops *vlan_ops = ice_get_compat_vsi_vlan_ops(vsi); 1053 struct device *dev = ice_pf_to_dev(vf->pf); 1054 int err; 1055 1056 if (ice_vf_is_port_vlan_ena(vf)) { 1057 err = vlan_ops->set_port_vlan(vsi, &vf->port_vlan_info); 1058 if (err) { 1059 dev_err(dev, "failed to configure port VLAN via VSI parameters for VF %u, error %d\n", 1060 vf->vf_id, err); 1061 return err; 1062 } 1063 1064 err = vlan_ops->add_vlan(vsi, &vf->port_vlan_info); 1065 } else { 1066 err = ice_vsi_add_vlan_zero(vsi); 1067 } 1068 1069 if (err) { 1070 dev_err(dev, "failed to add VLAN %u filter for VF %u during VF rebuild, error %d\n", 1071 ice_vf_is_port_vlan_ena(vf) ? 1072 ice_vf_get_port_vlan_id(vf) : 0, vf->vf_id, err); 1073 return err; 1074 } 1075 1076 err = vlan_ops->ena_rx_filtering(vsi); 1077 if (err) 1078 dev_warn(dev, "failed to enable Rx VLAN filtering for VF %d VSI %d during VF rebuild, error %d\n", 1079 vf->vf_id, vsi->idx, err); 1080 1081 return 0; 1082 } 1083 1084 /** 1085 * ice_vf_rebuild_host_tx_rate_cfg - re-apply the Tx rate limiting configuration 1086 * @vf: VF to re-apply the configuration for 1087 * 1088 * Called after a VF VSI has been re-added/rebuild during reset. The PF driver 1089 * needs to re-apply the host configured Tx rate limiting configuration. 1090 */ 1091 static int ice_vf_rebuild_host_tx_rate_cfg(struct ice_vf *vf) 1092 { 1093 struct device *dev = ice_pf_to_dev(vf->pf); 1094 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 1095 int err; 1096 1097 if (WARN_ON(!vsi)) 1098 return -EINVAL; 1099 1100 if (vf->min_tx_rate) { 1101 err = ice_set_min_bw_limit(vsi, (u64)vf->min_tx_rate * 1000); 1102 if (err) { 1103 dev_err(dev, "failed to set min Tx rate to %d Mbps for VF %u, error %d\n", 1104 vf->min_tx_rate, vf->vf_id, err); 1105 return err; 1106 } 1107 } 1108 1109 if (vf->max_tx_rate) { 1110 err = ice_set_max_bw_limit(vsi, (u64)vf->max_tx_rate * 1000); 1111 if (err) { 1112 dev_err(dev, "failed to set max Tx rate to %d Mbps for VF %u, error %d\n", 1113 vf->max_tx_rate, vf->vf_id, err); 1114 return err; 1115 } 1116 } 1117 1118 return 0; 1119 } 1120 1121 /** 1122 * ice_vf_rebuild_aggregator_node_cfg - rebuild aggregator node config 1123 * @vsi: Pointer to VSI 1124 * 1125 * This function moves VSI into corresponding scheduler aggregator node 1126 * based on cached value of "aggregator node info" per VSI 1127 */ 1128 static void ice_vf_rebuild_aggregator_node_cfg(struct ice_vsi *vsi) 1129 { 1130 struct ice_pf *pf = vsi->back; 1131 struct device *dev; 1132 int status; 1133 1134 if (!vsi->agg_node) 1135 return; 1136 1137 dev = ice_pf_to_dev(pf); 1138 if (vsi->agg_node->num_vsis == ICE_MAX_VSIS_IN_AGG_NODE) { 1139 dev_dbg(dev, 1140 "agg_id %u already has reached max_num_vsis %u\n", 1141 vsi->agg_node->agg_id, vsi->agg_node->num_vsis); 1142 return; 1143 } 1144 1145 status = ice_move_vsi_to_agg(pf->hw.port_info, vsi->agg_node->agg_id, 1146 vsi->idx, vsi->tc_cfg.ena_tc); 1147 if (status) 1148 dev_dbg(dev, "unable to move VSI idx %u into aggregator %u node", 1149 vsi->idx, vsi->agg_node->agg_id); 1150 else 1151 vsi->agg_node->num_vsis++; 1152 } 1153 1154 /** 1155 * ice_vf_rebuild_host_cfg - host admin configuration is persistent across reset 1156 * @vf: VF to rebuild host configuration on 1157 */ 1158 void ice_vf_rebuild_host_cfg(struct ice_vf *vf) 1159 { 1160 struct device *dev = ice_pf_to_dev(vf->pf); 1161 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 1162 1163 if (WARN_ON(!vsi)) 1164 return; 1165 1166 ice_vf_set_host_trust_cfg(vf); 1167 1168 if (ice_vf_rebuild_host_mac_cfg(vf)) 1169 dev_err(dev, "failed to rebuild default MAC configuration for VF %d\n", 1170 vf->vf_id); 1171 1172 if (ice_vf_rebuild_host_vlan_cfg(vf, vsi)) 1173 dev_err(dev, "failed to rebuild VLAN configuration for VF %u\n", 1174 vf->vf_id); 1175 1176 if (ice_vf_rebuild_host_tx_rate_cfg(vf)) 1177 dev_err(dev, "failed to rebuild Tx rate limiting configuration for VF %u\n", 1178 vf->vf_id); 1179 1180 if (ice_vsi_apply_spoofchk(vsi, vf->spoofchk)) 1181 dev_err(dev, "failed to rebuild spoofchk configuration for VF %d\n", 1182 vf->vf_id); 1183 1184 /* rebuild aggregator node config for main VF VSI */ 1185 ice_vf_rebuild_aggregator_node_cfg(vsi); 1186 } 1187 1188 /** 1189 * ice_vf_ctrl_invalidate_vsi - invalidate ctrl_vsi_idx to remove VSI access 1190 * @vf: VF that control VSI is being invalidated on 1191 */ 1192 void ice_vf_ctrl_invalidate_vsi(struct ice_vf *vf) 1193 { 1194 vf->ctrl_vsi_idx = ICE_NO_VSI; 1195 } 1196 1197 /** 1198 * ice_vf_ctrl_vsi_release - invalidate the VF's control VSI after freeing it 1199 * @vf: VF that control VSI is being released on 1200 */ 1201 void ice_vf_ctrl_vsi_release(struct ice_vf *vf) 1202 { 1203 ice_vsi_release(vf->pf->vsi[vf->ctrl_vsi_idx]); 1204 ice_vf_ctrl_invalidate_vsi(vf); 1205 } 1206 1207 /** 1208 * ice_vf_ctrl_vsi_setup - Set up a VF control VSI 1209 * @vf: VF to setup control VSI for 1210 * 1211 * Returns pointer to the successfully allocated VSI struct on success, 1212 * otherwise returns NULL on failure. 1213 */ 1214 struct ice_vsi *ice_vf_ctrl_vsi_setup(struct ice_vf *vf) 1215 { 1216 struct ice_vsi_cfg_params params = {}; 1217 struct ice_pf *pf = vf->pf; 1218 struct ice_vsi *vsi; 1219 1220 params.type = ICE_VSI_CTRL; 1221 params.pi = ice_vf_get_port_info(vf); 1222 params.vf = vf; 1223 params.flags = ICE_VSI_FLAG_INIT; 1224 1225 vsi = ice_vsi_setup(pf, ¶ms); 1226 if (!vsi) { 1227 dev_err(ice_pf_to_dev(pf), "Failed to create VF control VSI\n"); 1228 ice_vf_ctrl_invalidate_vsi(vf); 1229 } 1230 1231 return vsi; 1232 } 1233 1234 /** 1235 * ice_vf_init_host_cfg - Initialize host admin configuration 1236 * @vf: VF to initialize 1237 * @vsi: the VSI created at initialization 1238 * 1239 * Initialize the VF host configuration. Called during VF creation to setup 1240 * VLAN 0, add the VF VSI broadcast filter, and setup spoof checking. It 1241 * should only be called during VF creation. 1242 */ 1243 int ice_vf_init_host_cfg(struct ice_vf *vf, struct ice_vsi *vsi) 1244 { 1245 struct ice_vsi_vlan_ops *vlan_ops; 1246 struct ice_pf *pf = vf->pf; 1247 u8 broadcast[ETH_ALEN]; 1248 struct device *dev; 1249 int err; 1250 1251 dev = ice_pf_to_dev(pf); 1252 1253 err = ice_vsi_add_vlan_zero(vsi); 1254 if (err) { 1255 dev_warn(dev, "Failed to add VLAN 0 filter for VF %d\n", 1256 vf->vf_id); 1257 return err; 1258 } 1259 1260 vlan_ops = ice_get_compat_vsi_vlan_ops(vsi); 1261 err = vlan_ops->ena_rx_filtering(vsi); 1262 if (err) { 1263 dev_warn(dev, "Failed to enable Rx VLAN filtering for VF %d\n", 1264 vf->vf_id); 1265 return err; 1266 } 1267 1268 eth_broadcast_addr(broadcast); 1269 err = ice_fltr_add_mac(vsi, broadcast, ICE_FWD_TO_VSI); 1270 if (err) { 1271 dev_err(dev, "Failed to add broadcast MAC filter for VF %d, status %d\n", 1272 vf->vf_id, err); 1273 return err; 1274 } 1275 1276 vf->num_mac = 1; 1277 1278 err = ice_vsi_apply_spoofchk(vsi, vf->spoofchk); 1279 if (err) { 1280 dev_warn(dev, "Failed to initialize spoofchk setting for VF %d\n", 1281 vf->vf_id); 1282 return err; 1283 } 1284 1285 return 0; 1286 } 1287 1288 /** 1289 * ice_vf_invalidate_vsi - invalidate vsi_idx/vsi_num to remove VSI access 1290 * @vf: VF to remove access to VSI for 1291 */ 1292 void ice_vf_invalidate_vsi(struct ice_vf *vf) 1293 { 1294 vf->lan_vsi_idx = ICE_NO_VSI; 1295 vf->lan_vsi_num = ICE_NO_VSI; 1296 } 1297 1298 /** 1299 * ice_vf_vsi_release - Release the VF VSI and invalidate indexes 1300 * @vf: pointer to the VF structure 1301 * 1302 * Release the VF associated with this VSI and then invalidate the VSI 1303 * indexes. 1304 */ 1305 void ice_vf_vsi_release(struct ice_vf *vf) 1306 { 1307 struct ice_vsi *vsi = ice_get_vf_vsi(vf); 1308 1309 if (WARN_ON(!vsi)) 1310 return; 1311 1312 ice_vsi_release(vsi); 1313 ice_vf_invalidate_vsi(vf); 1314 } 1315 1316 /** 1317 * ice_vf_set_initialized - VF is ready for VIRTCHNL communication 1318 * @vf: VF to set in initialized state 1319 * 1320 * After this function the VF will be ready to receive/handle the 1321 * VIRTCHNL_OP_GET_VF_RESOURCES message 1322 */ 1323 void ice_vf_set_initialized(struct ice_vf *vf) 1324 { 1325 ice_set_vf_state_qs_dis(vf); 1326 clear_bit(ICE_VF_STATE_MC_PROMISC, vf->vf_states); 1327 clear_bit(ICE_VF_STATE_UC_PROMISC, vf->vf_states); 1328 clear_bit(ICE_VF_STATE_DIS, vf->vf_states); 1329 set_bit(ICE_VF_STATE_INIT, vf->vf_states); 1330 memset(&vf->vlan_v2_caps, 0, sizeof(vf->vlan_v2_caps)); 1331 } 1332