Lines Matching full:pf
62 struct ice_pf *pf = container_of(hw, struct ice_pf, hw); in ice_hw_to_dev() local
64 return &pf->pdev->dev; in ice_hw_to_dev()
72 static void ice_rebuild(struct ice_pf *pf, enum ice_reset_req reset_type);
74 static void ice_vsi_release_all(struct ice_pf *pf);
76 static int ice_rebuild_channels(struct ice_pf *pf);
110 * @pf: pointer to PF struct
112 static void ice_check_for_hang_subtask(struct ice_pf *pf) in ice_check_for_hang_subtask() argument
120 ice_for_each_vsi(pf, v) in ice_check_for_hang_subtask()
121 if (pf->vsi[v] && pf->vsi[v]->type == ICE_VSI_PF) { in ice_check_for_hang_subtask()
122 vsi = pf->vsi[v]; in ice_check_for_hang_subtask()
174 * @pf: board private structure
176 * Set initial set of MAC filters for PF VSI; configure filters for permanent
180 static int ice_init_mac_fltr(struct ice_pf *pf) in ice_init_mac_fltr() argument
185 vsi = ice_get_main_vsi(pf); in ice_init_mac_fltr()
258 * ice_set_promisc - Enable promiscuous mode for a given PF
287 * ice_clear_promisc - Disable promiscuous mode for a given PF
325 struct ice_pf *pf = vsi->back; in ice_vsi_sync_fltr() local
326 struct ice_hw *hw = &pf->hw; in ice_vsi_sync_fltr()
480 * @pf: board private structure
482 static void ice_sync_fltr_subtask(struct ice_pf *pf) in ice_sync_fltr_subtask() argument
486 if (!pf || !(test_bit(ICE_FLAG_FLTR_SYNC, pf->flags))) in ice_sync_fltr_subtask()
489 clear_bit(ICE_FLAG_FLTR_SYNC, pf->flags); in ice_sync_fltr_subtask()
491 ice_for_each_vsi(pf, v) in ice_sync_fltr_subtask()
492 if (pf->vsi[v] && ice_vsi_fltr_changed(pf->vsi[v]) && in ice_sync_fltr_subtask()
493 ice_vsi_sync_fltr(pf->vsi[v])) { in ice_sync_fltr_subtask()
495 set_bit(ICE_FLAG_FLTR_SYNC, pf->flags); in ice_sync_fltr_subtask()
501 * ice_pf_dis_all_vsi - Pause all VSIs on a PF
502 * @pf: the PF
505 static void ice_pf_dis_all_vsi(struct ice_pf *pf, bool locked) in ice_pf_dis_all_vsi() argument
510 ice_for_each_vsi(pf, v) in ice_pf_dis_all_vsi()
511 if (pf->vsi[v]) in ice_pf_dis_all_vsi()
512 ice_dis_vsi(pf->vsi[v], locked); in ice_pf_dis_all_vsi()
515 pf->pf_agg_node[node].num_vsis = 0; in ice_pf_dis_all_vsi()
518 pf->vf_agg_node[node].num_vsis = 0; in ice_pf_dis_all_vsi()
523 * @pf: board private structure
529 ice_prepare_for_reset(struct ice_pf *pf, enum ice_reset_req reset_type) in ice_prepare_for_reset() argument
531 struct ice_hw *hw = &pf->hw; in ice_prepare_for_reset()
536 dev_dbg(ice_pf_to_dev(pf), "reset_type=%d\n", reset_type); in ice_prepare_for_reset()
539 if (test_bit(ICE_PREPARED_FOR_RESET, pf->state)) in ice_prepare_for_reset()
542 synchronize_irq(pf->oicr_irq.virq); in ice_prepare_for_reset()
544 ice_unplug_aux_dev(pf); in ice_prepare_for_reset()
548 ice_vc_notify_reset(pf); in ice_prepare_for_reset()
551 mutex_lock(&pf->vfs.table_lock); in ice_prepare_for_reset()
552 ice_for_each_vf(pf, bkt, vf) in ice_prepare_for_reset()
554 mutex_unlock(&pf->vfs.table_lock); in ice_prepare_for_reset()
556 if (ice_is_eswitch_mode_switchdev(pf)) { in ice_prepare_for_reset()
558 ice_eswitch_br_fdb_flush(pf->eswitch.br_offloads->bridge); in ice_prepare_for_reset()
563 vsi = ice_get_main_vsi(pf); in ice_prepare_for_reset()
572 if (test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) { in ice_prepare_for_reset()
588 clear_bit(ICE_FLAG_TC_MQPRIO, pf->flags); in ice_prepare_for_reset()
600 set_bit(ICE_VSI_REBUILD_PENDING, ice_get_main_vsi(pf)->state); in ice_prepare_for_reset()
601 ice_pf_dis_all_vsi(pf, false); in ice_prepare_for_reset()
603 if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags)) in ice_prepare_for_reset()
604 ice_ptp_prepare_for_reset(pf); in ice_prepare_for_reset()
606 if (ice_is_feature_supported(pf, ICE_F_GNSS)) in ice_prepare_for_reset()
607 ice_gnss_exit(pf); in ice_prepare_for_reset()
614 set_bit(ICE_PREPARED_FOR_RESET, pf->state); in ice_prepare_for_reset()
619 * @pf: board private structure
622 static void ice_do_reset(struct ice_pf *pf, enum ice_reset_req reset_type) in ice_do_reset() argument
624 struct device *dev = ice_pf_to_dev(pf); in ice_do_reset()
625 struct ice_hw *hw = &pf->hw; in ice_do_reset()
629 if (pf->lag && pf->lag->bonded && reset_type == ICE_RESET_PFR) { in ice_do_reset()
634 ice_prepare_for_reset(pf, reset_type); in ice_do_reset()
639 set_bit(ICE_RESET_FAILED, pf->state); in ice_do_reset()
640 clear_bit(ICE_RESET_OICR_RECV, pf->state); in ice_do_reset()
641 clear_bit(ICE_PREPARED_FOR_RESET, pf->state); in ice_do_reset()
642 clear_bit(ICE_PFR_REQ, pf->state); in ice_do_reset()
643 clear_bit(ICE_CORER_REQ, pf->state); in ice_do_reset()
644 clear_bit(ICE_GLOBR_REQ, pf->state); in ice_do_reset()
645 wake_up(&pf->reset_wait_queue); in ice_do_reset()
654 pf->pfr_count++; in ice_do_reset()
655 ice_rebuild(pf, reset_type); in ice_do_reset()
656 clear_bit(ICE_PREPARED_FOR_RESET, pf->state); in ice_do_reset()
657 clear_bit(ICE_PFR_REQ, pf->state); in ice_do_reset()
658 wake_up(&pf->reset_wait_queue); in ice_do_reset()
659 ice_reset_all_vfs(pf); in ice_do_reset()
665 * @pf: board private structure
667 static void ice_reset_subtask(struct ice_pf *pf) in ice_reset_subtask() argument
673 * of reset is pending and sets bits in pf->state indicating the reset in ice_reset_subtask()
675 * prepare for pending reset if not already (for PF software-initiated in ice_reset_subtask()
681 if (test_bit(ICE_RESET_OICR_RECV, pf->state)) { in ice_reset_subtask()
683 if (test_and_clear_bit(ICE_CORER_RECV, pf->state)) in ice_reset_subtask()
685 if (test_and_clear_bit(ICE_GLOBR_RECV, pf->state)) in ice_reset_subtask()
687 if (test_and_clear_bit(ICE_EMPR_RECV, pf->state)) in ice_reset_subtask()
692 ice_prepare_for_reset(pf, reset_type); in ice_reset_subtask()
695 if (ice_check_reset(&pf->hw)) { in ice_reset_subtask()
696 set_bit(ICE_RESET_FAILED, pf->state); in ice_reset_subtask()
699 pf->hw.reset_ongoing = false; in ice_reset_subtask()
700 ice_rebuild(pf, reset_type); in ice_reset_subtask()
704 clear_bit(ICE_RESET_OICR_RECV, pf->state); in ice_reset_subtask()
705 clear_bit(ICE_PREPARED_FOR_RESET, pf->state); in ice_reset_subtask()
706 clear_bit(ICE_PFR_REQ, pf->state); in ice_reset_subtask()
707 clear_bit(ICE_CORER_REQ, pf->state); in ice_reset_subtask()
708 clear_bit(ICE_GLOBR_REQ, pf->state); in ice_reset_subtask()
709 wake_up(&pf->reset_wait_queue); in ice_reset_subtask()
710 ice_reset_all_vfs(pf); in ice_reset_subtask()
717 if (test_bit(ICE_PFR_REQ, pf->state)) { in ice_reset_subtask()
719 if (pf->lag && pf->lag->bonded) { in ice_reset_subtask()
720 dev_dbg(ice_pf_to_dev(pf), "PFR on a bonded interface, promoting to CORER\n"); in ice_reset_subtask()
724 if (test_bit(ICE_CORER_REQ, pf->state)) in ice_reset_subtask()
726 if (test_bit(ICE_GLOBR_REQ, pf->state)) in ice_reset_subtask()
733 if (!test_bit(ICE_DOWN, pf->state) && in ice_reset_subtask()
734 !test_bit(ICE_CFG_BUSY, pf->state)) { in ice_reset_subtask()
735 ice_do_reset(pf, reset_type); in ice_reset_subtask()
928 * @pf: private PF struct
937 static void ice_set_dflt_mib(struct ice_pf *pf) in ice_set_dflt_mib() argument
939 struct device *dev = ice_pf_to_dev(pf); in ice_set_dflt_mib()
943 struct ice_hw *hw = &pf->hw; in ice_set_dflt_mib()
1016 * @pf: pointer to PF struct
1021 static void ice_check_phy_fw_load(struct ice_pf *pf, u8 link_cfg_err) in ice_check_phy_fw_load() argument
1024 clear_bit(ICE_FLAG_PHY_FW_LOAD_FAILED, pf->flags); in ice_check_phy_fw_load()
1028 if (test_bit(ICE_FLAG_PHY_FW_LOAD_FAILED, pf->flags)) in ice_check_phy_fw_load()
1032 …dev_err(ice_pf_to_dev(pf), "Device failed to load the FW for the external PHY. Please download and… in ice_check_phy_fw_load()
1033 set_bit(ICE_FLAG_PHY_FW_LOAD_FAILED, pf->flags); in ice_check_phy_fw_load()
1039 * @pf: pointer to PF struct
1045 static void ice_check_module_power(struct ice_pf *pf, u8 link_cfg_err) in ice_check_module_power() argument
1050 clear_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags); in ice_check_module_power()
1057 if (test_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags)) in ice_check_module_power()
1061 …dev_err(ice_pf_to_dev(pf), "The installed module is incompatible with the device's NVM image. Cann… in ice_check_module_power()
1062 set_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags); in ice_check_module_power()
1064 …dev_err(ice_pf_to_dev(pf), "The module's power requirements exceed the device's power supply. Cann… in ice_check_module_power()
1065 set_bit(ICE_FLAG_MOD_POWER_UNSUPPORTED, pf->flags); in ice_check_module_power()
1071 * @pf: pointer to the PF struct
1077 static void ice_check_link_cfg_err(struct ice_pf *pf, u8 link_cfg_err) in ice_check_link_cfg_err() argument
1079 ice_check_module_power(pf, link_cfg_err); in ice_check_link_cfg_err()
1080 ice_check_phy_fw_load(pf, link_cfg_err); in ice_check_link_cfg_err()
1085 * @pf: PF that the link event is associated with
1093 ice_link_event(struct ice_pf *pf, struct ice_port_info *pi, bool link_up, in ice_link_event() argument
1096 struct device *dev = ice_pf_to_dev(pf); in ice_link_event()
1118 ice_check_link_cfg_err(pf, pi->phy.link_info.link_cfg_err); in ice_link_event()
1126 vsi = ice_get_main_vsi(pf); in ice_link_event()
1131 if (!test_bit(ICE_FLAG_NO_MEDIA, pf->flags) && in ice_link_event()
1133 set_bit(ICE_FLAG_NO_MEDIA, pf->flags); in ice_link_event()
1141 ice_ptp_link_change(pf, pf->hw.pf_id, link_up); in ice_link_event()
1143 if (ice_is_dcb_active(pf)) { in ice_link_event()
1144 if (test_bit(ICE_FLAG_DCB_ENA, pf->flags)) in ice_link_event()
1145 ice_dcb_rebuild(pf); in ice_link_event()
1148 ice_set_dflt_mib(pf); in ice_link_event()
1153 ice_vc_notify_link_state(pf); in ice_link_event()
1160 * @pf: board private structure
1162 static void ice_watchdog_subtask(struct ice_pf *pf) in ice_watchdog_subtask() argument
1167 if (test_bit(ICE_DOWN, pf->state) || in ice_watchdog_subtask()
1168 test_bit(ICE_CFG_BUSY, pf->state)) in ice_watchdog_subtask()
1173 pf->serv_tmr_prev + pf->serv_tmr_period)) in ice_watchdog_subtask()
1176 pf->serv_tmr_prev = jiffies; in ice_watchdog_subtask()
1181 ice_update_pf_stats(pf); in ice_watchdog_subtask()
1182 ice_for_each_vsi(pf, i) in ice_watchdog_subtask()
1183 if (pf->vsi[i] && pf->vsi[i]->netdev) in ice_watchdog_subtask()
1184 ice_update_vsi_stats(pf->vsi[i]); in ice_watchdog_subtask()
1218 * @pf: PF that the link event is associated with
1222 ice_handle_link_event(struct ice_pf *pf, struct ice_rq_event_info *event) in ice_handle_link_event() argument
1229 port_info = pf->hw.port_info; in ice_handle_link_event()
1233 status = ice_link_event(pf, port_info, in ice_handle_link_event()
1237 dev_dbg(ice_pf_to_dev(pf), "Could not process link event, error %d\n", in ice_handle_link_event()
1245 * @pf: pointer to the PF private structure
1250 * a given PF. Actual wait would be done by a call to ice_aq_wait_for_event().
1259 void ice_aq_prep_for_event(struct ice_pf *pf, struct ice_aq_task *task, in ice_aq_prep_for_event() argument
1266 spin_lock_bh(&pf->aq_wait_lock); in ice_aq_prep_for_event()
1267 hlist_add_head(&task->entry, &pf->aq_wait_list); in ice_aq_prep_for_event()
1268 spin_unlock_bh(&pf->aq_wait_lock); in ice_aq_prep_for_event()
1273 * @pf: pointer to the PF private structure
1277 * Waits for a specific AdminQ completion event on the ARQ for a given PF. The
1283 int ice_aq_wait_for_event(struct ice_pf *pf, struct ice_aq_task *task, in ice_aq_wait_for_event() argument
1287 struct device *dev = ice_pf_to_dev(pf); in ice_aq_wait_for_event()
1292 ret = wait_event_interruptible_timeout(pf->aq_wait_queue, in ice_aq_wait_for_event()
1320 spin_lock_bh(&pf->aq_wait_lock); in ice_aq_wait_for_event()
1322 spin_unlock_bh(&pf->aq_wait_lock); in ice_aq_wait_for_event()
1329 * @pf: pointer to the PF private structure
1345 static void ice_aq_check_events(struct ice_pf *pf, u16 opcode, in ice_aq_check_events() argument
1352 spin_lock_bh(&pf->aq_wait_lock); in ice_aq_check_events()
1353 hlist_for_each_entry(task, &pf->aq_wait_list, entry) { in ice_aq_check_events()
1373 spin_unlock_bh(&pf->aq_wait_lock); in ice_aq_check_events()
1376 wake_up(&pf->aq_wait_queue); in ice_aq_check_events()
1381 * @pf: the PF private structure
1386 static void ice_aq_cancel_waiting_tasks(struct ice_pf *pf) in ice_aq_cancel_waiting_tasks() argument
1390 spin_lock_bh(&pf->aq_wait_lock); in ice_aq_cancel_waiting_tasks()
1391 hlist_for_each_entry(task, &pf->aq_wait_list, entry) in ice_aq_cancel_waiting_tasks()
1393 spin_unlock_bh(&pf->aq_wait_lock); in ice_aq_cancel_waiting_tasks()
1395 wake_up(&pf->aq_wait_queue); in ice_aq_cancel_waiting_tasks()
1402 * @pf: ptr to struct ice_pf
1405 static int __ice_clean_ctrlq(struct ice_pf *pf, enum ice_ctl_q q_type) in __ice_clean_ctrlq() argument
1407 struct device *dev = ice_pf_to_dev(pf); in __ice_clean_ctrlq()
1409 struct ice_hw *hw = &pf->hw; in __ice_clean_ctrlq()
1415 /* Do not clean control queue if/when PF reset fails */ in __ice_clean_ctrlq()
1416 if (test_bit(ICE_RESET_FAILED, pf->state)) in __ice_clean_ctrlq()
1506 ice_aq_check_events(pf, opcode, &event); in __ice_clean_ctrlq()
1510 if (ice_handle_link_event(pf, &event)) in __ice_clean_ctrlq()
1514 ice_vf_lan_overflow_event(pf, &event); in __ice_clean_ctrlq()
1517 if (ice_is_feature_supported(pf, ICE_F_MBX_LIMIT)) { in __ice_clean_ctrlq()
1518 ice_vc_process_vf_msg(pf, &event, NULL); in __ice_clean_ctrlq()
1529 ice_vc_process_vf_msg(pf, &event, &data); in __ice_clean_ctrlq()
1536 ice_dcb_process_lldp_set_mib_change(pf, &event); in __ice_clean_ctrlq()
1567 * @pf: board private structure
1569 static void ice_clean_adminq_subtask(struct ice_pf *pf) in ice_clean_adminq_subtask() argument
1571 struct ice_hw *hw = &pf->hw; in ice_clean_adminq_subtask()
1573 if (!test_bit(ICE_ADMINQ_EVENT_PENDING, pf->state)) in ice_clean_adminq_subtask()
1576 if (__ice_clean_ctrlq(pf, ICE_CTL_Q_ADMIN)) in ice_clean_adminq_subtask()
1579 clear_bit(ICE_ADMINQ_EVENT_PENDING, pf->state); in ice_clean_adminq_subtask()
1587 __ice_clean_ctrlq(pf, ICE_CTL_Q_ADMIN); in ice_clean_adminq_subtask()
1594 * @pf: board private structure
1596 static void ice_clean_mailboxq_subtask(struct ice_pf *pf) in ice_clean_mailboxq_subtask() argument
1598 struct ice_hw *hw = &pf->hw; in ice_clean_mailboxq_subtask()
1600 if (!test_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state)) in ice_clean_mailboxq_subtask()
1603 if (__ice_clean_ctrlq(pf, ICE_CTL_Q_MAILBOX)) in ice_clean_mailboxq_subtask()
1606 clear_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state); in ice_clean_mailboxq_subtask()
1609 __ice_clean_ctrlq(pf, ICE_CTL_Q_MAILBOX); in ice_clean_mailboxq_subtask()
1616 * @pf: board private structure
1618 static void ice_clean_sbq_subtask(struct ice_pf *pf) in ice_clean_sbq_subtask() argument
1620 struct ice_hw *hw = &pf->hw; in ice_clean_sbq_subtask()
1624 clear_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state); in ice_clean_sbq_subtask()
1628 if (!test_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state)) in ice_clean_sbq_subtask()
1631 if (__ice_clean_ctrlq(pf, ICE_CTL_Q_SB)) in ice_clean_sbq_subtask()
1634 clear_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state); in ice_clean_sbq_subtask()
1637 __ice_clean_ctrlq(pf, ICE_CTL_Q_SB); in ice_clean_sbq_subtask()
1644 * @pf: board private structure
1648 void ice_service_task_schedule(struct ice_pf *pf) in ice_service_task_schedule() argument
1650 if (!test_bit(ICE_SERVICE_DIS, pf->state) && in ice_service_task_schedule()
1651 !test_and_set_bit(ICE_SERVICE_SCHED, pf->state) && in ice_service_task_schedule()
1652 !test_bit(ICE_NEEDS_RESTART, pf->state)) in ice_service_task_schedule()
1653 queue_work(ice_wq, &pf->serv_task); in ice_service_task_schedule()
1658 * @pf: board private structure
1660 static void ice_service_task_complete(struct ice_pf *pf) in ice_service_task_complete() argument
1662 WARN_ON(!test_bit(ICE_SERVICE_SCHED, pf->state)); in ice_service_task_complete()
1664 /* force memory (pf->state) to sync before next service task */ in ice_service_task_complete()
1666 clear_bit(ICE_SERVICE_SCHED, pf->state); in ice_service_task_complete()
1671 * @pf: board private structure
1676 static int ice_service_task_stop(struct ice_pf *pf) in ice_service_task_stop() argument
1680 ret = test_and_set_bit(ICE_SERVICE_DIS, pf->state); in ice_service_task_stop()
1682 if (pf->serv_tmr.function) in ice_service_task_stop()
1683 del_timer_sync(&pf->serv_tmr); in ice_service_task_stop()
1684 if (pf->serv_task.func) in ice_service_task_stop()
1685 cancel_work_sync(&pf->serv_task); in ice_service_task_stop()
1687 clear_bit(ICE_SERVICE_SCHED, pf->state); in ice_service_task_stop()
1693 * @pf: board private structure
1697 static void ice_service_task_restart(struct ice_pf *pf) in ice_service_task_restart() argument
1699 clear_bit(ICE_SERVICE_DIS, pf->state); in ice_service_task_restart()
1700 ice_service_task_schedule(pf); in ice_service_task_restart()
1709 struct ice_pf *pf = from_timer(pf, t, serv_tmr); in ice_service_timer() local
1711 mod_timer(&pf->serv_tmr, round_jiffies(pf->serv_tmr_period + jiffies)); in ice_service_timer()
1712 ice_service_task_schedule(pf); in ice_service_timer()
1717 * @pf: pointer to the PF structure
1720 * VF MDD logging is guarded by net_ratelimit. Additional PF and VF log
1722 * disable the queue, the PF can be configured to reset the VF using ethtool
1725 static void ice_handle_mdd_event(struct ice_pf *pf) in ice_handle_mdd_event() argument
1727 struct device *dev = ice_pf_to_dev(pf); in ice_handle_mdd_event()
1728 struct ice_hw *hw = &pf->hw; in ice_handle_mdd_event()
1733 if (!test_and_clear_bit(ICE_MDD_EVENT_PENDING, pf->state)) { in ice_handle_mdd_event()
1737 ice_print_vfs_mdd_events(pf); in ice_handle_mdd_event()
1753 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1754 dev_info(dev, "Malicious Driver Detection event %d on TX queue %d PF# %d VF# %d\n", in ice_handle_mdd_event()
1770 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1771 dev_info(dev, "Malicious Driver Detection event %d on TX queue %d PF# %d VF# %d\n", in ice_handle_mdd_event()
1787 if (netif_msg_rx_err(pf)) in ice_handle_mdd_event()
1788 dev_info(dev, "Malicious Driver Detection event %d on RX queue %d PF# %d VF# %d\n", in ice_handle_mdd_event()
1793 /* check to see if this PF caused an MDD event */ in ice_handle_mdd_event()
1797 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1798 dev_info(dev, "Malicious Driver Detection event TX_PQM detected on PF\n"); in ice_handle_mdd_event()
1804 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1805 dev_info(dev, "Malicious Driver Detection event TX_TCLAN detected on PF\n"); in ice_handle_mdd_event()
1811 if (netif_msg_rx_err(pf)) in ice_handle_mdd_event()
1812 dev_info(dev, "Malicious Driver Detection event RX detected on PF\n"); in ice_handle_mdd_event()
1818 mutex_lock(&pf->vfs.table_lock); in ice_handle_mdd_event()
1819 ice_for_each_vf(pf, bkt, vf) { in ice_handle_mdd_event()
1824 set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state); in ice_handle_mdd_event()
1825 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1834 set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state); in ice_handle_mdd_event()
1835 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1844 set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state); in ice_handle_mdd_event()
1845 if (netif_msg_tx_err(pf)) in ice_handle_mdd_event()
1854 set_bit(ICE_MDD_VF_PRINT_PENDING, pf->state); in ice_handle_mdd_event()
1855 if (netif_msg_rx_err(pf)) in ice_handle_mdd_event()
1860 * PF can be configured to reset the VF through ethtool in ice_handle_mdd_event()
1863 if (test_bit(ICE_FLAG_MDD_AUTO_RESET_VF, pf->flags)) { in ice_handle_mdd_event()
1872 mutex_unlock(&pf->vfs.table_lock); in ice_handle_mdd_event()
1874 ice_print_vfs_mdd_events(pf); in ice_handle_mdd_event()
1962 struct ice_pf *pf = pi->hw->back; in ice_init_nvm_phy_type() local
1973 dev_err(ice_pf_to_dev(pf), "Get PHY capability failed.\n"); in ice_init_nvm_phy_type()
1977 pf->nvm_phy_type_hi = pcaps->phy_type_high; in ice_init_nvm_phy_type()
1978 pf->nvm_phy_type_lo = pcaps->phy_type_low; in ice_init_nvm_phy_type()
1994 struct ice_pf *pf = pi->hw->back; in ice_init_link_dflt_override() local
1996 ldo = &pf->link_dflt_override; in ice_init_link_dflt_override()
2006 set_bit(ICE_FLAG_TOTAL_PORT_SHUTDOWN_ENA, pf->flags); in ice_init_link_dflt_override()
2007 set_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags); in ice_init_link_dflt_override()
2032 struct ice_pf *pf = pi->hw->back; in ice_init_phy_cfg_dflt_override() local
2034 ldo = &pf->link_dflt_override; in ice_init_phy_cfg_dflt_override()
2042 cfg->phy_type_low = pf->nvm_phy_type_lo & in ice_init_phy_cfg_dflt_override()
2044 cfg->phy_type_high = pf->nvm_phy_type_hi & in ice_init_phy_cfg_dflt_override()
2050 set_bit(ICE_LINK_DEFAULT_OVERRIDE_PENDING, pf->state); in ice_init_phy_cfg_dflt_override()
2071 struct ice_pf *pf = pi->hw->back; in ice_init_phy_user_cfg() local
2088 dev_err(ice_pf_to_dev(pf), "Get PHY capability failed.\n"); in ice_init_phy_user_cfg()
2098 set_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags); in ice_init_phy_user_cfg()
2105 (pf->link_dflt_override.options & ICE_LINK_OVERRIDE_EN)) { in ice_init_phy_user_cfg()
2120 set_bit(ICE_PHY_INIT_COMPLETE, pf->state); in ice_init_phy_user_cfg()
2141 struct ice_pf *pf = vsi->back; in ice_configure_phy() local
2150 if (!test_bit(ICE_FLAG_LINK_LENIENT_MODE_ENA, pf->flags) && in ice_configure_phy()
2154 if (test_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags)) in ice_configure_phy()
2240 err = ice_aq_set_phy_cfg(&pf->hw, pi, cfg, NULL); in ice_configure_phy()
2253 * @pf: pointer to PF struct
2258 static void ice_check_media_subtask(struct ice_pf *pf) in ice_check_media_subtask() argument
2265 if (!test_bit(ICE_FLAG_NO_MEDIA, pf->flags)) in ice_check_media_subtask()
2268 vsi = ice_get_main_vsi(pf); in ice_check_media_subtask()
2278 ice_check_link_cfg_err(pf, pi->phy.link_info.link_cfg_err); in ice_check_media_subtask()
2281 if (!test_bit(ICE_PHY_INIT_COMPLETE, pf->state)) in ice_check_media_subtask()
2293 clear_bit(ICE_FLAG_NO_MEDIA, pf->flags); in ice_check_media_subtask()
2303 * @work: pointer to work_struct contained by the PF struct
2307 struct ice_pf *pf = container_of(work, struct ice_pf, serv_task); in ice_service_task() local
2313 ice_reset_subtask(pf); in ice_service_task()
2316 if (ice_is_reset_in_progress(pf->state) || in ice_service_task()
2317 test_bit(ICE_SUSPENDED, pf->state) || in ice_service_task()
2318 test_bit(ICE_NEEDS_RESTART, pf->state)) { in ice_service_task()
2319 ice_service_task_complete(pf); in ice_service_task()
2323 if (test_and_clear_bit(ICE_AUX_ERR_PENDING, pf->state)) { in ice_service_task()
2330 swap(event->reg, pf->oicr_err_reg); in ice_service_task()
2331 ice_send_event_to_aux(pf, event); in ice_service_task()
2339 if (test_and_clear_bit(ICE_FLAG_UNPLUG_AUX_DEV, pf->flags)) in ice_service_task()
2340 ice_unplug_aux_dev(pf); in ice_service_task()
2343 if (test_and_clear_bit(ICE_FLAG_PLUG_AUX_DEV, pf->flags)) in ice_service_task()
2344 ice_plug_aux_dev(pf); in ice_service_task()
2346 if (test_and_clear_bit(ICE_FLAG_MTU_CHANGED, pf->flags)) { in ice_service_task()
2352 ice_send_event_to_aux(pf, event); in ice_service_task()
2357 ice_clean_adminq_subtask(pf); in ice_service_task()
2358 ice_check_media_subtask(pf); in ice_service_task()
2359 ice_check_for_hang_subtask(pf); in ice_service_task()
2360 ice_sync_fltr_subtask(pf); in ice_service_task()
2361 ice_handle_mdd_event(pf); in ice_service_task()
2362 ice_watchdog_subtask(pf); in ice_service_task()
2364 if (ice_is_safe_mode(pf)) { in ice_service_task()
2365 ice_service_task_complete(pf); in ice_service_task()
2369 ice_process_vflr_event(pf); in ice_service_task()
2370 ice_clean_mailboxq_subtask(pf); in ice_service_task()
2371 ice_clean_sbq_subtask(pf); in ice_service_task()
2372 ice_sync_arfs_fltrs(pf); in ice_service_task()
2373 ice_flush_fdir_ctx(pf); in ice_service_task()
2376 ice_service_task_complete(pf); in ice_service_task()
2382 if (time_after(jiffies, (start_time + pf->serv_tmr_period)) || in ice_service_task()
2383 test_bit(ICE_MDD_EVENT_PENDING, pf->state) || in ice_service_task()
2384 test_bit(ICE_VFLR_EVENT_PENDING, pf->state) || in ice_service_task()
2385 test_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state) || in ice_service_task()
2386 test_bit(ICE_FD_VF_FLUSH_CTX, pf->state) || in ice_service_task()
2387 test_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state) || in ice_service_task()
2388 test_bit(ICE_ADMINQ_EVENT_PENDING, pf->state)) in ice_service_task()
2389 mod_timer(&pf->serv_tmr, jiffies); in ice_service_task()
2414 * @pf: board private structure
2417 int ice_schedule_reset(struct ice_pf *pf, enum ice_reset_req reset) in ice_schedule_reset() argument
2419 struct device *dev = ice_pf_to_dev(pf); in ice_schedule_reset()
2422 if (test_bit(ICE_RESET_FAILED, pf->state)) { in ice_schedule_reset()
2427 if (ice_is_reset_in_progress(pf->state)) { in ice_schedule_reset()
2434 set_bit(ICE_PFR_REQ, pf->state); in ice_schedule_reset()
2437 set_bit(ICE_CORER_REQ, pf->state); in ice_schedule_reset()
2440 set_bit(ICE_GLOBR_REQ, pf->state); in ice_schedule_reset()
2446 ice_service_task_schedule(pf); in ice_schedule_reset()
2502 struct ice_pf *pf = vsi->back; in ice_vsi_req_irq_msix() local
2509 dev = ice_pf_to_dev(pf); in ice_vsi_req_irq_msix()
2666 struct ice_pf *pf = vsi->back; in ice_prepare_xdp_rings() local
2668 .qs_mutex = &pf->avail_q_mutex, in ice_prepare_xdp_rings()
2669 .pf_map = pf->avail_txqs, in ice_prepare_xdp_rings()
2670 .pf_map_size = pf->max_pf_txqs, in ice_prepare_xdp_rings()
2681 dev = ice_pf_to_dev(pf); in ice_prepare_xdp_rings()
2776 mutex_lock(&pf->avail_q_mutex); in ice_prepare_xdp_rings()
2778 clear_bit(vsi->txq_map[i + vsi->alloc_txq], pf->avail_txqs); in ice_prepare_xdp_rings()
2781 mutex_unlock(&pf->avail_q_mutex); in ice_prepare_xdp_rings()
2792 * Detach XDP rings from irq vectors, clean up the PF bitmap and free
2798 struct ice_pf *pf = vsi->back; in ice_destroy_xdp_rings() local
2820 mutex_lock(&pf->avail_q_mutex); in ice_destroy_xdp_rings()
2822 clear_bit(vsi->txq_map[i + vsi->alloc_txq], pf->avail_txqs); in ice_destroy_xdp_rings()
2825 mutex_unlock(&pf->avail_q_mutex); in ice_destroy_xdp_rings()
2839 devm_kfree(ice_pf_to_dev(pf), vsi->xdp_rings); in ice_destroy_xdp_rings()
3016 NL_SET_ERR_MSG_MOD(xdp->extack, "XDP can be loaded only on PF VSI"); in ice_xdp()
3039 * @pf: board private structure
3041 static void ice_ena_misc_vector(struct ice_pf *pf) in ice_ena_misc_vector() argument
3043 struct ice_hw *hw = &pf->hw; in ice_ena_misc_vector()
3070 wr32(hw, GLINT_DYN_CTL(pf->oicr_irq.index), in ice_ena_misc_vector()
3081 struct ice_pf *pf = (struct ice_pf *)data; in ice_misc_intr() local
3082 struct ice_hw *hw = &pf->hw; in ice_misc_intr()
3086 dev = ice_pf_to_dev(pf); in ice_misc_intr()
3087 set_bit(ICE_ADMINQ_EVENT_PENDING, pf->state); in ice_misc_intr()
3088 set_bit(ICE_MAILBOXQ_EVENT_PENDING, pf->state); in ice_misc_intr()
3089 set_bit(ICE_SIDEBANDQ_EVENT_PENDING, pf->state); in ice_misc_intr()
3096 pf->sw_int_count++; in ice_misc_intr()
3101 set_bit(ICE_MDD_EVENT_PENDING, pf->state); in ice_misc_intr()
3105 if (test_bit(ICE_VF_RESETS_DISABLED, pf->state)) { in ice_misc_intr()
3112 set_bit(ICE_VFLR_EVENT_PENDING, pf->state); in ice_misc_intr()
3125 pf->corer_count++; in ice_misc_intr()
3127 pf->globr_count++; in ice_misc_intr()
3129 pf->empr_count++; in ice_misc_intr()
3134 * pf->state so that the service task can start a reset/rebuild. in ice_misc_intr()
3136 if (!test_and_set_bit(ICE_RESET_OICR_RECV, pf->state)) { in ice_misc_intr()
3138 set_bit(ICE_CORER_RECV, pf->state); in ice_misc_intr()
3140 set_bit(ICE_GLOBR_RECV, pf->state); in ice_misc_intr()
3142 set_bit(ICE_EMPR_RECV, pf->state); in ice_misc_intr()
3150 * ICE_RESET_OICR_RECV in pf->state indicates in ice_misc_intr()
3164 set_bit(ICE_MISC_THREAD_TX_TSTAMP, pf->misc_thread); in ice_misc_intr()
3175 pf->ptp.ext_ts_irq |= gltsyn_stat & in ice_misc_intr()
3180 set_bit(ICE_MISC_THREAD_EXTTS_EVENT, pf->misc_thread); in ice_misc_intr()
3186 pf->oicr_err_reg |= oicr; in ice_misc_intr()
3187 set_bit(ICE_AUX_ERR_PENDING, pf->state); in ice_misc_intr()
3200 set_bit(ICE_PFR_REQ, pf->state); in ice_misc_intr()
3214 struct ice_pf *pf = data; in ice_misc_intr_thread_fn() local
3217 hw = &pf->hw; in ice_misc_intr_thread_fn()
3219 if (ice_is_reset_in_progress(pf->state)) in ice_misc_intr_thread_fn()
3222 ice_service_task_schedule(pf); in ice_misc_intr_thread_fn()
3224 if (test_and_clear_bit(ICE_MISC_THREAD_EXTTS_EVENT, pf->misc_thread)) in ice_misc_intr_thread_fn()
3225 ice_ptp_extts_event(pf); in ice_misc_intr_thread_fn()
3227 if (test_and_clear_bit(ICE_MISC_THREAD_TX_TSTAMP, pf->misc_thread)) { in ice_misc_intr_thread_fn()
3231 if (ice_ptp_process_ts(pf) == ICE_TX_TSTAMP_WORK_PENDING) { in ice_misc_intr_thread_fn()
3268 * @pf: board private structure
3270 static void ice_free_irq_msix_misc(struct ice_pf *pf) in ice_free_irq_msix_misc() argument
3272 int misc_irq_num = pf->oicr_irq.virq; in ice_free_irq_msix_misc()
3273 struct ice_hw *hw = &pf->hw; in ice_free_irq_msix_misc()
3282 devm_free_irq(ice_pf_to_dev(pf), misc_irq_num, pf); in ice_free_irq_msix_misc()
3284 ice_free_irq(pf, pf->oicr_irq); in ice_free_irq_msix_misc()
3320 * @pf: board private structure
3326 static int ice_req_irq_msix_misc(struct ice_pf *pf) in ice_req_irq_msix_misc() argument
3328 struct device *dev = ice_pf_to_dev(pf); in ice_req_irq_msix_misc()
3329 struct ice_hw *hw = &pf->hw; in ice_req_irq_msix_misc()
3333 if (!pf->int_name[0]) in ice_req_irq_msix_misc()
3334 snprintf(pf->int_name, sizeof(pf->int_name) - 1, "%s-%s:misc", in ice_req_irq_msix_misc()
3341 if (ice_is_reset_in_progress(pf->state)) in ice_req_irq_msix_misc()
3345 oicr_irq = ice_alloc_irq(pf, false); in ice_req_irq_msix_misc()
3349 pf->oicr_irq = oicr_irq; in ice_req_irq_msix_misc()
3350 err = devm_request_threaded_irq(dev, pf->oicr_irq.virq, ice_misc_intr, in ice_req_irq_msix_misc()
3352 pf->int_name, pf); in ice_req_irq_msix_misc()
3355 pf->int_name, err); in ice_req_irq_msix_misc()
3356 ice_free_irq(pf, pf->oicr_irq); in ice_req_irq_msix_misc()
3361 ice_ena_misc_vector(pf); in ice_req_irq_msix_misc()
3363 ice_ena_ctrlq_interrupts(hw, pf->oicr_irq.index); in ice_req_irq_msix_misc()
3364 wr32(hw, GLINT_ITR(ICE_RX_ITR, pf->oicr_irq.index), in ice_req_irq_msix_misc()
3400 struct ice_pf *pf = ice_netdev_to_pf(netdev); in ice_set_ops() local
3402 if (ice_is_safe_mode(pf)) { in ice_set_ops()
3409 netdev->udp_tunnel_nic_info = &pf->hw.udp_tunnel_nic; in ice_set_ops()
3427 struct ice_pf *pf = ice_netdev_to_pf(netdev); in ice_set_netdev_features() local
3428 bool is_dvm_ena = ice_is_dvm_ena(&pf->hw); in ice_set_netdev_features()
3434 if (ice_is_safe_mode(pf)) { in ice_set_netdev_features()
3526 * ice_pf_vsi_setup - Set up a PF VSI
3527 * @pf: board private structure
3534 ice_pf_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi) in ice_pf_vsi_setup() argument
3542 return ice_vsi_setup(pf, ¶ms); in ice_pf_vsi_setup()
3546 ice_chnl_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi, in ice_chnl_vsi_setup() argument
3556 return ice_vsi_setup(pf, ¶ms); in ice_chnl_vsi_setup()
3561 * @pf: board private structure
3568 ice_ctrl_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi) in ice_ctrl_vsi_setup() argument
3576 return ice_vsi_setup(pf, ¶ms); in ice_ctrl_vsi_setup()
3581 * @pf: board private structure
3588 ice_lb_vsi_setup(struct ice_pf *pf, struct ice_port_info *pi) in ice_lb_vsi_setup() argument
3596 return ice_vsi_setup(pf, ¶ms); in ice_lb_vsi_setup()
3796 * @pf: pointer to an ice_pf instance
3798 u16 ice_get_avail_txq_count(struct ice_pf *pf) in ice_get_avail_txq_count() argument
3800 return ice_get_avail_q_count(pf->avail_txqs, &pf->avail_q_mutex, in ice_get_avail_txq_count()
3801 pf->max_pf_txqs); in ice_get_avail_txq_count()
3806 * @pf: pointer to an ice_pf instance
3808 u16 ice_get_avail_rxq_count(struct ice_pf *pf) in ice_get_avail_rxq_count() argument
3810 return ice_get_avail_q_count(pf->avail_rxqs, &pf->avail_q_mutex, in ice_get_avail_rxq_count()
3811 pf->max_pf_rxqs); in ice_get_avail_rxq_count()
3816 * @pf: board private structure to initialize
3818 static void ice_deinit_pf(struct ice_pf *pf) in ice_deinit_pf() argument
3820 ice_service_task_stop(pf); in ice_deinit_pf()
3821 mutex_destroy(&pf->lag_mutex); in ice_deinit_pf()
3822 mutex_destroy(&pf->adev_mutex); in ice_deinit_pf()
3823 mutex_destroy(&pf->sw_mutex); in ice_deinit_pf()
3824 mutex_destroy(&pf->tc_mutex); in ice_deinit_pf()
3825 mutex_destroy(&pf->avail_q_mutex); in ice_deinit_pf()
3826 mutex_destroy(&pf->vfs.table_lock); in ice_deinit_pf()
3828 if (pf->avail_txqs) { in ice_deinit_pf()
3829 bitmap_free(pf->avail_txqs); in ice_deinit_pf()
3830 pf->avail_txqs = NULL; in ice_deinit_pf()
3833 if (pf->avail_rxqs) { in ice_deinit_pf()
3834 bitmap_free(pf->avail_rxqs); in ice_deinit_pf()
3835 pf->avail_rxqs = NULL; in ice_deinit_pf()
3838 if (pf->ptp.clock) in ice_deinit_pf()
3839 ptp_clock_unregister(pf->ptp.clock); in ice_deinit_pf()
3844 * @pf: pointer to the PF instance
3846 static void ice_set_pf_caps(struct ice_pf *pf) in ice_set_pf_caps() argument
3848 struct ice_hw_func_caps *func_caps = &pf->hw.func_caps; in ice_set_pf_caps()
3850 clear_bit(ICE_FLAG_RDMA_ENA, pf->flags); in ice_set_pf_caps()
3852 set_bit(ICE_FLAG_RDMA_ENA, pf->flags); in ice_set_pf_caps()
3853 clear_bit(ICE_FLAG_DCB_CAPABLE, pf->flags); in ice_set_pf_caps()
3855 set_bit(ICE_FLAG_DCB_CAPABLE, pf->flags); in ice_set_pf_caps()
3856 clear_bit(ICE_FLAG_SRIOV_CAPABLE, pf->flags); in ice_set_pf_caps()
3858 set_bit(ICE_FLAG_SRIOV_CAPABLE, pf->flags); in ice_set_pf_caps()
3859 pf->vfs.num_supported = min_t(int, func_caps->num_allocd_vfs, in ice_set_pf_caps()
3862 clear_bit(ICE_FLAG_RSS_ENA, pf->flags); in ice_set_pf_caps()
3864 set_bit(ICE_FLAG_RSS_ENA, pf->flags); in ice_set_pf_caps()
3866 clear_bit(ICE_FLAG_FD_ENA, pf->flags); in ice_set_pf_caps()
3873 pf->ctrl_vsi_idx = ICE_NO_VSI; in ice_set_pf_caps()
3874 set_bit(ICE_FLAG_FD_ENA, pf->flags); in ice_set_pf_caps()
3875 /* force guaranteed filter pool for PF */ in ice_set_pf_caps()
3876 ice_alloc_fd_guar_item(&pf->hw, &unused, in ice_set_pf_caps()
3878 /* force shared filter pool for PF */ in ice_set_pf_caps()
3879 ice_alloc_fd_shrd_item(&pf->hw, &unused, in ice_set_pf_caps()
3883 clear_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags); in ice_set_pf_caps()
3885 !(pf->hw.mac_type == ICE_MAC_E830)) in ice_set_pf_caps()
3886 set_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags); in ice_set_pf_caps()
3888 pf->max_pf_txqs = func_caps->common_cap.num_txq; in ice_set_pf_caps()
3889 pf->max_pf_rxqs = func_caps->common_cap.num_rxq; in ice_set_pf_caps()
3894 * @pf: board private structure to initialize
3896 static int ice_init_pf(struct ice_pf *pf) in ice_init_pf() argument
3898 ice_set_pf_caps(pf); in ice_init_pf()
3900 mutex_init(&pf->sw_mutex); in ice_init_pf()
3901 mutex_init(&pf->tc_mutex); in ice_init_pf()
3902 mutex_init(&pf->adev_mutex); in ice_init_pf()
3903 mutex_init(&pf->lag_mutex); in ice_init_pf()
3905 INIT_HLIST_HEAD(&pf->aq_wait_list); in ice_init_pf()
3906 spin_lock_init(&pf->aq_wait_lock); in ice_init_pf()
3907 init_waitqueue_head(&pf->aq_wait_queue); in ice_init_pf()
3909 init_waitqueue_head(&pf->reset_wait_queue); in ice_init_pf()
3912 timer_setup(&pf->serv_tmr, ice_service_timer, 0); in ice_init_pf()
3913 pf->serv_tmr_period = HZ; in ice_init_pf()
3914 INIT_WORK(&pf->serv_task, ice_service_task); in ice_init_pf()
3915 clear_bit(ICE_SERVICE_SCHED, pf->state); in ice_init_pf()
3917 mutex_init(&pf->avail_q_mutex); in ice_init_pf()
3918 pf->avail_txqs = bitmap_zalloc(pf->max_pf_txqs, GFP_KERNEL); in ice_init_pf()
3919 if (!pf->avail_txqs) in ice_init_pf()
3922 pf->avail_rxqs = bitmap_zalloc(pf->max_pf_rxqs, GFP_KERNEL); in ice_init_pf()
3923 if (!pf->avail_rxqs) { in ice_init_pf()
3924 bitmap_free(pf->avail_txqs); in ice_init_pf()
3925 pf->avail_txqs = NULL; in ice_init_pf()
3929 mutex_init(&pf->vfs.table_lock); in ice_init_pf()
3930 hash_init(pf->vfs.table); in ice_init_pf()
3931 if (ice_is_feature_supported(pf, ICE_F_MBX_LIMIT)) in ice_init_pf()
3932 wr32(&pf->hw, E830_MBX_PF_IN_FLIGHT_VF_MSGS_THRESH, in ice_init_pf()
3935 ice_mbx_init_snapshot(&pf->hw); in ice_init_pf()
3952 * word) indicates WoL is not supported on the corresponding PF ID. in ice_is_wol_supported()
3973 struct ice_pf *pf = vsi->back; in ice_vsi_recfg_qs() local
3979 while (test_and_set_bit(ICE_CFG_BUSY, pf->state)) { in ice_vsi_recfg_qs()
3996 dev_dbg(ice_pf_to_dev(pf), "Link is down, queue count change happens when link is brought up\n"); in ice_vsi_recfg_qs()
4012 ice_pf_dcb_recfg(pf, locked); in ice_vsi_recfg_qs()
4017 dev_err(ice_pf_to_dev(pf), "Error during VSI rebuild: %d. Unload and reload the driver.\n", in ice_vsi_recfg_qs()
4020 clear_bit(ICE_CFG_BUSY, pf->state); in ice_vsi_recfg_qs()
4025 * ice_set_safe_mode_vlan_cfg - configure PF VSI to allow all VLANs in safe mode
4026 * @pf: PF to configure
4028 * No VLAN offloads/filtering are advertised in safe mode so make sure the PF
4031 static void ice_set_safe_mode_vlan_cfg(struct ice_pf *pf) in ice_set_safe_mode_vlan_cfg() argument
4033 struct ice_vsi *vsi = ice_get_main_vsi(pf); in ice_set_safe_mode_vlan_cfg()
4045 hw = &pf->hw; in ice_set_safe_mode_vlan_cfg()
4084 struct ice_pf *pf = hw->back; in ice_log_pkg_init() local
4087 dev = ice_pf_to_dev(pf); in ice_log_pkg_init()
4161 * @pf: pointer to the PF instance
4167 ice_load_pkg(const struct firmware *firmware, struct ice_pf *pf) in ice_load_pkg() argument
4170 struct device *dev = ice_pf_to_dev(pf); in ice_load_pkg()
4171 struct ice_hw *hw = &pf->hw; in ice_load_pkg()
4188 clear_bit(ICE_FLAG_ADV_FEATURES, pf->flags); in ice_load_pkg()
4195 set_bit(ICE_FLAG_ADV_FEATURES, pf->flags); in ice_load_pkg()
4200 * @pf: pointer to the PF structure
4206 static void ice_verify_cacheline_size(struct ice_pf *pf) in ice_verify_cacheline_size() argument
4208 if (rd32(&pf->hw, GLPCI_CNF2) & GLPCI_CNF2_CACHELINE_SIZE_M) in ice_verify_cacheline_size()
4209 …dev_warn(ice_pf_to_dev(pf), "%d Byte cache line assumption is invalid, driver may have Tx timeouts… in ice_verify_cacheline_size()
4215 * @pf: PF struct
4219 static int ice_send_version(struct ice_pf *pf) in ice_send_version() argument
4229 return ice_aq_send_driver_ver(&pf->hw, &dv, NULL); in ice_send_version()
4234 * @pf: pointer to the PF instance
4238 static int ice_init_fdir(struct ice_pf *pf) in ice_init_fdir() argument
4240 struct device *dev = ice_pf_to_dev(pf); in ice_init_fdir()
4245 * Allocate it and store it in the PF. in ice_init_fdir()
4247 ctrl_vsi = ice_ctrl_vsi_setup(pf, pf->hw.port_info); in ice_init_fdir()
4259 mutex_init(&pf->hw.fdir_fltr_lock); in ice_init_fdir()
4261 err = ice_fdir_create_dflt_rules(pf); in ice_init_fdir()
4268 ice_fdir_release_flows(&pf->hw); in ice_init_fdir()
4272 if (pf->ctrl_vsi_idx != ICE_NO_VSI) { in ice_init_fdir()
4273 pf->vsi[pf->ctrl_vsi_idx] = NULL; in ice_init_fdir()
4274 pf->ctrl_vsi_idx = ICE_NO_VSI; in ice_init_fdir()
4279 static void ice_deinit_fdir(struct ice_pf *pf) in ice_deinit_fdir() argument
4281 struct ice_vsi *vsi = ice_get_ctrl_vsi(pf); in ice_deinit_fdir()
4288 if (pf->ctrl_vsi_idx != ICE_NO_VSI) { in ice_deinit_fdir()
4289 pf->vsi[pf->ctrl_vsi_idx] = NULL; in ice_deinit_fdir()
4290 pf->ctrl_vsi_idx = ICE_NO_VSI; in ice_deinit_fdir()
4293 mutex_destroy(&(&pf->hw)->fdir_fltr_lock); in ice_deinit_fdir()
4298 * @pf: pointer to the PF instance
4300 static char *ice_get_opt_fw_name(struct ice_pf *pf) in ice_get_opt_fw_name() argument
4305 struct pci_dev *pdev = pf->pdev; in ice_get_opt_fw_name()
4328 * @pf: pointer to the PF instance
4330 static void ice_request_fw(struct ice_pf *pf) in ice_request_fw() argument
4332 char *opt_fw_filename = ice_get_opt_fw_name(pf); in ice_request_fw()
4334 struct device *dev = ice_pf_to_dev(pf); in ice_request_fw()
4349 ice_load_pkg(firmware, pf); in ice_request_fw()
4363 ice_load_pkg(firmware, pf); in ice_request_fw()
4369 * @pf: pointer to the PF struct
4371 static void ice_print_wake_reason(struct ice_pf *pf) in ice_print_wake_reason() argument
4373 u32 wus = pf->wakeup_reason; in ice_print_wake_reason()
4391 dev_info(ice_pf_to_dev(pf), "Wake reason: %s", wake_str); in ice_print_wake_reason()
4494 static int ice_init_eth(struct ice_pf *pf) in ice_init_eth() argument
4496 struct ice_vsi *vsi = ice_get_main_vsi(pf); in ice_init_eth()
4511 err = ice_init_mac_fltr(pf); in ice_init_eth()
4515 err = ice_devlink_create_pf_port(pf); in ice_init_eth()
4519 SET_NETDEV_DEVLINK_PORT(vsi->netdev, &pf->devlink_port); in ice_init_eth()
4536 ice_devlink_destroy_pf_port(pf); in ice_init_eth()
4543 static void ice_deinit_eth(struct ice_pf *pf) in ice_deinit_eth() argument
4545 struct ice_vsi *vsi = ice_get_main_vsi(pf); in ice_deinit_eth()
4552 ice_devlink_destroy_pf_port(pf); in ice_deinit_eth()
4582 static int ice_init_dev(struct ice_pf *pf) in ice_init_dev() argument
4584 struct device *dev = ice_pf_to_dev(pf); in ice_init_dev()
4585 struct ice_hw *hw = &pf->hw; in ice_init_dev()
4606 ice_init_feature_support(pf); in ice_init_dev()
4608 ice_request_fw(pf); in ice_init_dev()
4611 * set in pf->state, which will cause ice_is_safe_mode to return in ice_init_dev()
4614 if (ice_is_safe_mode(pf)) { in ice_init_dev()
4623 err = ice_init_pf(pf); in ice_init_dev()
4629 pf->hw.udp_tunnel_nic.set_port = ice_udp_tunnel_set_port; in ice_init_dev()
4630 pf->hw.udp_tunnel_nic.unset_port = ice_udp_tunnel_unset_port; in ice_init_dev()
4631 pf->hw.udp_tunnel_nic.flags = UDP_TUNNEL_NIC_INFO_MAY_SLEEP; in ice_init_dev()
4632 pf->hw.udp_tunnel_nic.shared = &pf->hw.udp_tunnel_shared; in ice_init_dev()
4633 if (pf->hw.tnl.valid_count[TNL_VXLAN]) { in ice_init_dev()
4634 pf->hw.udp_tunnel_nic.tables[0].n_entries = in ice_init_dev()
4635 pf->hw.tnl.valid_count[TNL_VXLAN]; in ice_init_dev()
4636 pf->hw.udp_tunnel_nic.tables[0].tunnel_types = in ice_init_dev()
4639 if (pf->hw.tnl.valid_count[TNL_GENEVE]) { in ice_init_dev()
4640 pf->hw.udp_tunnel_nic.tables[1].n_entries = in ice_init_dev()
4641 pf->hw.tnl.valid_count[TNL_GENEVE]; in ice_init_dev()
4642 pf->hw.udp_tunnel_nic.tables[1].tunnel_types = in ice_init_dev()
4646 err = ice_init_interrupt_scheme(pf); in ice_init_dev()
4658 err = ice_req_irq_msix_misc(pf); in ice_init_dev()
4667 ice_clear_interrupt_scheme(pf); in ice_init_dev()
4669 ice_deinit_pf(pf); in ice_init_dev()
4675 static void ice_deinit_dev(struct ice_pf *pf) in ice_deinit_dev() argument
4677 ice_free_irq_msix_misc(pf); in ice_deinit_dev()
4678 ice_deinit_pf(pf); in ice_deinit_dev()
4679 ice_deinit_hw(&pf->hw); in ice_deinit_dev()
4682 ice_reset(&pf->hw, ICE_RESET_PFR); in ice_deinit_dev()
4683 pci_wait_for_pending_transaction(pf->pdev); in ice_deinit_dev()
4684 ice_clear_interrupt_scheme(pf); in ice_deinit_dev()
4687 static void ice_init_features(struct ice_pf *pf) in ice_init_features() argument
4689 struct device *dev = ice_pf_to_dev(pf); in ice_init_features()
4691 if (ice_is_safe_mode(pf)) in ice_init_features()
4695 if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags)) in ice_init_features()
4696 ice_ptp_init(pf); in ice_init_features()
4698 if (ice_is_feature_supported(pf, ICE_F_GNSS)) in ice_init_features()
4699 ice_gnss_init(pf); in ice_init_features()
4702 if (ice_init_fdir(pf)) in ice_init_features()
4706 if (ice_init_pf_dcb(pf, false)) { in ice_init_features()
4707 clear_bit(ICE_FLAG_DCB_CAPABLE, pf->flags); in ice_init_features()
4708 clear_bit(ICE_FLAG_DCB_ENA, pf->flags); in ice_init_features()
4710 ice_cfg_lldp_mib_change(&pf->hw, true); in ice_init_features()
4713 if (ice_init_lag(pf)) in ice_init_features()
4717 static void ice_deinit_features(struct ice_pf *pf) in ice_deinit_features() argument
4719 if (ice_is_safe_mode(pf)) in ice_deinit_features()
4722 ice_deinit_lag(pf); in ice_deinit_features()
4723 if (test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags)) in ice_deinit_features()
4724 ice_cfg_lldp_mib_change(&pf->hw, false); in ice_deinit_features()
4725 ice_deinit_fdir(pf); in ice_deinit_features()
4726 if (ice_is_feature_supported(pf, ICE_F_GNSS)) in ice_deinit_features()
4727 ice_gnss_exit(pf); in ice_deinit_features()
4728 if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags)) in ice_deinit_features()
4729 ice_ptp_release(pf); in ice_deinit_features()
4732 static void ice_init_wakeup(struct ice_pf *pf) in ice_init_wakeup() argument
4735 pf->wakeup_reason = rd32(&pf->hw, PFPM_WUS); in ice_init_wakeup()
4738 ice_print_wake_reason(pf); in ice_init_wakeup()
4741 wr32(&pf->hw, PFPM_WUS, U32_MAX); in ice_init_wakeup()
4744 device_set_wakeup_enable(ice_pf_to_dev(pf), false); in ice_init_wakeup()
4747 static int ice_init_link(struct ice_pf *pf) in ice_init_link() argument
4749 struct device *dev = ice_pf_to_dev(pf); in ice_init_link()
4752 err = ice_init_link_events(pf->hw.port_info); in ice_init_link()
4759 err = ice_init_nvm_phy_type(pf->hw.port_info); in ice_init_link()
4764 err = ice_update_link_info(pf->hw.port_info); in ice_init_link()
4768 ice_init_link_dflt_override(pf->hw.port_info); in ice_init_link()
4770 ice_check_link_cfg_err(pf, in ice_init_link()
4771 pf->hw.port_info->phy.link_info.link_cfg_err); in ice_init_link()
4774 if (pf->hw.port_info->phy.link_info.link_info & in ice_init_link()
4777 err = ice_init_phy_user_cfg(pf->hw.port_info); in ice_init_link()
4781 if (!test_bit(ICE_FLAG_LINK_DOWN_ON_CLOSE_ENA, pf->flags)) { in ice_init_link()
4782 struct ice_vsi *vsi = ice_get_main_vsi(pf); in ice_init_link()
4788 set_bit(ICE_FLAG_NO_MEDIA, pf->flags); in ice_init_link()
4794 static int ice_init_pf_sw(struct ice_pf *pf) in ice_init_pf_sw() argument
4796 bool dvm = ice_is_dvm_ena(&pf->hw); in ice_init_pf_sw()
4801 pf->first_sw = kzalloc(sizeof(*pf->first_sw), GFP_KERNEL); in ice_init_pf_sw()
4802 if (!pf->first_sw) in ice_init_pf_sw()
4805 if (pf->hw.evb_veb) in ice_init_pf_sw()
4806 pf->first_sw->bridge_mode = BRIDGE_MODE_VEB; in ice_init_pf_sw()
4808 pf->first_sw->bridge_mode = BRIDGE_MODE_VEPA; in ice_init_pf_sw()
4810 pf->first_sw->pf = pf; in ice_init_pf_sw()
4813 pf->first_sw->sw_id = pf->hw.port_info->sw_id; in ice_init_pf_sw()
4815 err = ice_aq_set_port_params(pf->hw.port_info, dvm, NULL); in ice_init_pf_sw()
4819 vsi = ice_pf_vsi_setup(pf, pf->hw.port_info); in ice_init_pf_sw()
4829 kfree(pf->first_sw); in ice_init_pf_sw()
4833 static void ice_deinit_pf_sw(struct ice_pf *pf) in ice_deinit_pf_sw() argument
4835 struct ice_vsi *vsi = ice_get_main_vsi(pf); in ice_deinit_pf_sw()
4841 kfree(pf->first_sw); in ice_deinit_pf_sw()
4844 static int ice_alloc_vsis(struct ice_pf *pf) in ice_alloc_vsis() argument
4846 struct device *dev = ice_pf_to_dev(pf); in ice_alloc_vsis()
4848 pf->num_alloc_vsi = pf->hw.func_caps.guar_num_vsi; in ice_alloc_vsis()
4849 if (!pf->num_alloc_vsi) in ice_alloc_vsis()
4852 if (pf->num_alloc_vsi > UDP_TUNNEL_NIC_MAX_SHARING_DEVICES) { in ice_alloc_vsis()
4855 pf->num_alloc_vsi, UDP_TUNNEL_NIC_MAX_SHARING_DEVICES); in ice_alloc_vsis()
4856 pf->num_alloc_vsi = UDP_TUNNEL_NIC_MAX_SHARING_DEVICES; in ice_alloc_vsis()
4859 pf->vsi = devm_kcalloc(dev, pf->num_alloc_vsi, sizeof(*pf->vsi), in ice_alloc_vsis()
4861 if (!pf->vsi) in ice_alloc_vsis()
4864 pf->vsi_stats = devm_kcalloc(dev, pf->num_alloc_vsi, in ice_alloc_vsis()
4865 sizeof(*pf->vsi_stats), GFP_KERNEL); in ice_alloc_vsis()
4866 if (!pf->vsi_stats) { in ice_alloc_vsis()
4867 devm_kfree(dev, pf->vsi); in ice_alloc_vsis()
4874 static void ice_dealloc_vsis(struct ice_pf *pf) in ice_dealloc_vsis() argument
4876 devm_kfree(ice_pf_to_dev(pf), pf->vsi_stats); in ice_dealloc_vsis()
4877 pf->vsi_stats = NULL; in ice_dealloc_vsis()
4879 pf->num_alloc_vsi = 0; in ice_dealloc_vsis()
4880 devm_kfree(ice_pf_to_dev(pf), pf->vsi); in ice_dealloc_vsis()
4881 pf->vsi = NULL; in ice_dealloc_vsis()
4884 static int ice_init_devlink(struct ice_pf *pf) in ice_init_devlink() argument
4888 err = ice_devlink_register_params(pf); in ice_init_devlink()
4892 ice_devlink_init_regions(pf); in ice_init_devlink()
4893 ice_devlink_register(pf); in ice_init_devlink()
4898 static void ice_deinit_devlink(struct ice_pf *pf) in ice_deinit_devlink() argument
4900 ice_devlink_unregister(pf); in ice_deinit_devlink()
4901 ice_devlink_destroy_regions(pf); in ice_deinit_devlink()
4902 ice_devlink_unregister_params(pf); in ice_deinit_devlink()
4905 static int ice_init(struct ice_pf *pf) in ice_init() argument
4909 err = ice_init_dev(pf); in ice_init()
4913 err = ice_alloc_vsis(pf); in ice_init()
4917 err = ice_init_pf_sw(pf); in ice_init()
4921 ice_init_wakeup(pf); in ice_init()
4923 err = ice_init_link(pf); in ice_init()
4927 err = ice_send_version(pf); in ice_init()
4931 ice_verify_cacheline_size(pf); in ice_init()
4933 if (ice_is_safe_mode(pf)) in ice_init()
4934 ice_set_safe_mode_vlan_cfg(pf); in ice_init()
4937 pcie_print_link_status(pf->pdev); in ice_init()
4940 clear_bit(ICE_DOWN, pf->state); in ice_init()
4941 clear_bit(ICE_SERVICE_DIS, pf->state); in ice_init()
4944 mod_timer(&pf->serv_tmr, round_jiffies(jiffies + pf->serv_tmr_period)); in ice_init()
4949 ice_deinit_pf_sw(pf); in ice_init()
4951 ice_dealloc_vsis(pf); in ice_init()
4953 ice_deinit_dev(pf); in ice_init()
4957 static void ice_deinit(struct ice_pf *pf) in ice_deinit() argument
4959 set_bit(ICE_SERVICE_DIS, pf->state); in ice_deinit()
4960 set_bit(ICE_DOWN, pf->state); in ice_deinit()
4962 ice_deinit_pf_sw(pf); in ice_deinit()
4963 ice_dealloc_vsis(pf); in ice_deinit()
4964 ice_deinit_dev(pf); in ice_deinit()
4968 * ice_load - load pf by init hw and starting VSI
4969 * @pf: pointer to the pf instance
4971 int ice_load(struct ice_pf *pf) in ice_load() argument
4977 err = ice_init_dev(pf); in ice_load()
4981 vsi = ice_get_main_vsi(pf); in ice_load()
4991 err = ice_start_eth(ice_get_main_vsi(pf)); in ice_load()
4996 err = ice_init_rdma(pf); in ice_load()
5000 ice_init_features(pf); in ice_load()
5001 ice_service_task_restart(pf); in ice_load()
5003 clear_bit(ICE_DOWN, pf->state); in ice_load()
5008 ice_vsi_close(ice_get_main_vsi(pf)); in ice_load()
5011 ice_vsi_decfg(ice_get_main_vsi(pf)); in ice_load()
5014 ice_deinit_dev(pf); in ice_load()
5019 * ice_unload - unload pf by stopping VSI and deinit hw
5020 * @pf: pointer to the pf instance
5022 void ice_unload(struct ice_pf *pf) in ice_unload() argument
5024 ice_deinit_features(pf); in ice_unload()
5025 ice_deinit_rdma(pf); in ice_unload()
5027 ice_stop_eth(ice_get_main_vsi(pf)); in ice_unload()
5028 ice_vsi_decfg(ice_get_main_vsi(pf)); in ice_unload()
5030 ice_deinit_dev(pf); in ice_unload()
5044 struct ice_pf *pf; in ice_probe() local
5080 pf = ice_allocate_pf(dev); in ice_probe()
5081 if (!pf) in ice_probe()
5085 pf->aux_idx = -1; in ice_probe()
5096 pf->pdev = pdev; in ice_probe()
5097 pci_set_drvdata(pdev, pf); in ice_probe()
5098 set_bit(ICE_DOWN, pf->state); in ice_probe()
5100 set_bit(ICE_SERVICE_DIS, pf->state); in ice_probe()
5102 hw = &pf->hw; in ice_probe()
5106 hw->back = pf; in ice_probe()
5117 pf->msg_enable = netif_msg_init(debug, ICE_DFLT_NETIF_M); in ice_probe()
5124 err = ice_init(pf); in ice_probe()
5128 err = ice_init_eth(pf); in ice_probe()
5132 err = ice_init_rdma(pf); in ice_probe()
5136 err = ice_init_devlink(pf); in ice_probe()
5140 ice_init_features(pf); in ice_probe()
5145 ice_deinit_rdma(pf); in ice_probe()
5147 ice_deinit_eth(pf); in ice_probe()
5149 ice_deinit(pf); in ice_probe()
5157 * @pf: pointer to the PF struct
5161 static void ice_set_wake(struct ice_pf *pf) in ice_set_wake() argument
5163 struct ice_hw *hw = &pf->hw; in ice_set_wake()
5164 bool wol = pf->wol_ena; in ice_set_wake()
5178 * @pf: pointer to the PF struct
5182 * wake, and that PF reset doesn't undo the LAA.
5184 static void ice_setup_mc_magic_wake(struct ice_pf *pf) in ice_setup_mc_magic_wake() argument
5186 struct device *dev = ice_pf_to_dev(pf); in ice_setup_mc_magic_wake()
5187 struct ice_hw *hw = &pf->hw; in ice_setup_mc_magic_wake()
5193 if (!pf->wol_ena) in ice_setup_mc_magic_wake()
5196 vsi = ice_get_main_vsi(pf); in ice_setup_mc_magic_wake()
5222 struct ice_pf *pf = pci_get_drvdata(pdev); in ice_remove() local
5226 if (!ice_is_reset_in_progress(pf->state)) in ice_remove()
5231 if (test_bit(ICE_FLAG_SRIOV_ENA, pf->flags)) { in ice_remove()
5232 set_bit(ICE_VF_RESETS_DISABLED, pf->state); in ice_remove()
5233 ice_free_vfs(pf); in ice_remove()
5236 ice_service_task_stop(pf); in ice_remove()
5237 ice_aq_cancel_waiting_tasks(pf); in ice_remove()
5238 set_bit(ICE_DOWN, pf->state); in ice_remove()
5240 if (!ice_is_safe_mode(pf)) in ice_remove()
5241 ice_remove_arfs(pf); in ice_remove()
5242 ice_deinit_features(pf); in ice_remove()
5243 ice_deinit_devlink(pf); in ice_remove()
5244 ice_deinit_rdma(pf); in ice_remove()
5245 ice_deinit_eth(pf); in ice_remove()
5246 ice_deinit(pf); in ice_remove()
5248 ice_vsi_release_all(pf); in ice_remove()
5250 ice_setup_mc_magic_wake(pf); in ice_remove()
5251 ice_set_wake(pf); in ice_remove()
5262 struct ice_pf *pf = pci_get_drvdata(pdev); in ice_shutdown() local
5267 pci_wake_from_d3(pdev, pf->wol_ena); in ice_shutdown()
5275 * @pf: board private structure
5279 static void ice_prepare_for_shutdown(struct ice_pf *pf) in ice_prepare_for_shutdown() argument
5281 struct ice_hw *hw = &pf->hw; in ice_prepare_for_shutdown()
5286 ice_vc_notify_reset(pf); in ice_prepare_for_shutdown()
5288 dev_dbg(ice_pf_to_dev(pf), "Tearing down internal switch for shutdown\n"); in ice_prepare_for_shutdown()
5291 ice_pf_dis_all_vsi(pf, false); in ice_prepare_for_shutdown()
5293 ice_for_each_vsi(pf, v) in ice_prepare_for_shutdown()
5294 if (pf->vsi[v]) in ice_prepare_for_shutdown()
5295 pf->vsi[v]->vsi_num = 0; in ice_prepare_for_shutdown()
5302 * @pf: board private structure to reinitialize
5310 static int ice_reinit_interrupt_scheme(struct ice_pf *pf) in ice_reinit_interrupt_scheme() argument
5312 struct device *dev = ice_pf_to_dev(pf); in ice_reinit_interrupt_scheme()
5319 ret = ice_init_interrupt_scheme(pf); in ice_reinit_interrupt_scheme()
5326 ice_for_each_vsi(pf, v) { in ice_reinit_interrupt_scheme()
5327 if (!pf->vsi[v]) in ice_reinit_interrupt_scheme()
5330 ret = ice_vsi_alloc_q_vectors(pf->vsi[v]); in ice_reinit_interrupt_scheme()
5333 ice_vsi_map_rings_to_vectors(pf->vsi[v]); in ice_reinit_interrupt_scheme()
5336 ret = ice_req_irq_msix_misc(pf); in ice_reinit_interrupt_scheme()
5347 if (pf->vsi[v]) in ice_reinit_interrupt_scheme()
5348 ice_vsi_free_q_vectors(pf->vsi[v]); in ice_reinit_interrupt_scheme()
5363 struct ice_pf *pf; in ice_suspend() local
5366 pf = pci_get_drvdata(pdev); in ice_suspend()
5368 if (!ice_pf_state_is_nominal(pf)) { in ice_suspend()
5379 disabled = ice_service_task_stop(pf); in ice_suspend()
5381 ice_deinit_rdma(pf); in ice_suspend()
5384 if (test_and_set_bit(ICE_SUSPENDED, pf->state)) { in ice_suspend()
5386 ice_service_task_restart(pf); in ice_suspend()
5390 if (test_bit(ICE_DOWN, pf->state) || in ice_suspend()
5391 ice_is_reset_in_progress(pf->state)) { in ice_suspend()
5394 ice_service_task_restart(pf); in ice_suspend()
5398 ice_setup_mc_magic_wake(pf); in ice_suspend()
5400 ice_prepare_for_shutdown(pf); in ice_suspend()
5402 ice_set_wake(pf); in ice_suspend()
5409 ice_free_irq_msix_misc(pf); in ice_suspend()
5410 ice_for_each_vsi(pf, v) { in ice_suspend()
5411 if (!pf->vsi[v]) in ice_suspend()
5413 ice_vsi_free_q_vectors(pf->vsi[v]); in ice_suspend()
5415 ice_clear_interrupt_scheme(pf); in ice_suspend()
5418 pci_wake_from_d3(pdev, pf->wol_ena); in ice_suspend()
5431 struct ice_pf *pf; in ice_resume() local
5448 pf = pci_get_drvdata(pdev); in ice_resume()
5449 hw = &pf->hw; in ice_resume()
5451 pf->wakeup_reason = rd32(hw, PFPM_WUS); in ice_resume()
5452 ice_print_wake_reason(pf); in ice_resume()
5457 ret = ice_reinit_interrupt_scheme(pf); in ice_resume()
5461 ret = ice_init_rdma(pf); in ice_resume()
5466 clear_bit(ICE_DOWN, pf->state); in ice_resume()
5467 /* Now perform PF reset and rebuild */ in ice_resume()
5470 clear_bit(ICE_SERVICE_DIS, pf->state); in ice_resume()
5472 if (ice_schedule_reset(pf, reset_type)) in ice_resume()
5475 clear_bit(ICE_SUSPENDED, pf->state); in ice_resume()
5476 ice_service_task_restart(pf); in ice_resume()
5479 mod_timer(&pf->serv_tmr, round_jiffies(jiffies + pf->serv_tmr_period)); in ice_resume()
5496 struct ice_pf *pf = pci_get_drvdata(pdev); in ice_pci_err_detected() local
5498 if (!pf) { in ice_pci_err_detected()
5504 if (!test_bit(ICE_SUSPENDED, pf->state)) { in ice_pci_err_detected()
5505 ice_service_task_stop(pf); in ice_pci_err_detected()
5507 if (!test_bit(ICE_PREPARED_FOR_RESET, pf->state)) { in ice_pci_err_detected()
5508 set_bit(ICE_PFR_REQ, pf->state); in ice_pci_err_detected()
5509 ice_prepare_for_reset(pf, ICE_RESET_PFR); in ice_pci_err_detected()
5525 struct ice_pf *pf = pci_get_drvdata(pdev); in ice_pci_err_slot_reset() local
5542 reg = rd32(&pf->hw, GLGEN_RTRIG); in ice_pci_err_slot_reset()
5561 struct ice_pf *pf = pci_get_drvdata(pdev); in ice_pci_err_resume() local
5563 if (!pf) { in ice_pci_err_resume()
5569 if (test_bit(ICE_SUSPENDED, pf->state)) { in ice_pci_err_resume()
5577 ice_do_reset(pf, ICE_RESET_PFR); in ice_pci_err_resume()
5578 ice_service_task_restart(pf); in ice_pci_err_resume()
5579 mod_timer(&pf->serv_tmr, round_jiffies(jiffies + pf->serv_tmr_period)); in ice_pci_err_resume()
5588 struct ice_pf *pf = pci_get_drvdata(pdev); in ice_pci_err_reset_prepare() local
5590 if (!test_bit(ICE_SUSPENDED, pf->state)) { in ice_pci_err_reset_prepare()
5591 ice_service_task_stop(pf); in ice_pci_err_reset_prepare()
5593 if (!test_bit(ICE_PREPARED_FOR_RESET, pf->state)) { in ice_pci_err_reset_prepare()
5594 set_bit(ICE_PFR_REQ, pf->state); in ice_pci_err_reset_prepare()
5595 ice_prepare_for_reset(pf, ICE_RESET_PFR); in ice_pci_err_reset_prepare()
5739 struct ice_pf *pf = vsi->back; in ice_set_mac_address() local
5740 struct ice_hw *hw = &pf->hw; in ice_set_mac_address()
5752 if (test_bit(ICE_DOWN, pf->state) || in ice_set_mac_address()
5753 ice_is_reset_in_progress(pf->state)) { in ice_set_mac_address()
5759 if (ice_chnl_dmac_fltr_cnt(pf)) { in ice_set_mac_address()
6066 * ice_set_vlan_offload_features - set VLAN offload features for the PF VSI
6067 * @vsi: PF's VSI
6111 * ice_set_vlan_filtering_features - set VLAN filtering features for the PF VSI
6112 * @vsi: PF's VSI
6180 * ice_set_loopback - turn on/off loopback mode on underlying PF
6216 struct ice_pf *pf = vsi->back; in ice_set_features() local
6220 if (ice_is_safe_mode(pf)) { in ice_set_features()
6221 dev_err(ice_pf_to_dev(pf), in ice_set_features()
6227 if (ice_is_reset_in_progress(pf->state)) { in ice_set_features()
6228 dev_err(ice_pf_to_dev(pf), in ice_set_features()
6268 if (!(features & NETIF_F_HW_TC) && ice_is_adq_active(pf)) { in ice_set_features()
6269 dev_err(ice_pf_to_dev(pf), "ADQ is active, can't turn hw_tc_offload off\n"); in ice_set_features()
6276 ena ? set_bit(ICE_FLAG_CLS_FLOWER, pf->flags) : in ice_set_features()
6277 clear_bit(ICE_FLAG_CLS_FLOWER, pf->flags); in ice_set_features()
6287 * ice_vsi_vlan_setup - Setup VLAN offload properties on a PF VSI
6482 struct ice_pf *pf = vsi->back; in ice_up_complete() local
6505 ice_ptp_link_change(pf, pf->hw.pf_id, true); in ice_up_complete()
6514 ice_service_task_schedule(pf); in ice_up_complete()
6597 struct ice_pf *pf = vsi->back; in ice_update_vsi_ring_stats() local
6644 * random value after PF reset. And as we increase the reported stat by in ice_update_vsi_ring_stats()
6648 if (likely(pf->stat_prev_loaded)) { in ice_update_vsi_ring_stats()
6671 struct ice_pf *pf = vsi->back; in ice_update_vsi_stats() local
6674 test_bit(ICE_CFG_BUSY, pf->state)) in ice_update_vsi_stats()
6690 cur_ns->rx_crc_errors = pf->stats.crc_errors; in ice_update_vsi_stats()
6691 cur_ns->rx_errors = pf->stats.crc_errors + in ice_update_vsi_stats()
6692 pf->stats.illegal_bytes + in ice_update_vsi_stats()
6693 pf->stats.rx_len_errors + in ice_update_vsi_stats()
6694 pf->stats.rx_undersize + in ice_update_vsi_stats()
6695 pf->hw_csum_rx_error + in ice_update_vsi_stats()
6696 pf->stats.rx_jabber + in ice_update_vsi_stats()
6697 pf->stats.rx_fragments + in ice_update_vsi_stats()
6698 pf->stats.rx_oversize; in ice_update_vsi_stats()
6699 cur_ns->rx_length_errors = pf->stats.rx_len_errors; in ice_update_vsi_stats()
6701 cur_ns->rx_missed_errors = pf->stats.eth.rx_discards; in ice_update_vsi_stats()
6706 * ice_update_pf_stats - Update PF port stats counters
6707 * @pf: PF whose stats needs to be updated
6709 void ice_update_pf_stats(struct ice_pf *pf) in ice_update_pf_stats() argument
6712 struct ice_hw *hw = &pf->hw; in ice_update_pf_stats()
6717 prev_ps = &pf->stats_prev; in ice_update_pf_stats()
6718 cur_ps = &pf->stats; in ice_update_pf_stats()
6720 if (ice_is_reset_in_progress(pf->state)) in ice_update_pf_stats()
6721 pf->stat_prev_loaded = false; in ice_update_pf_stats()
6723 ice_stat_update40(hw, GLPRT_GORCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6727 ice_stat_update40(hw, GLPRT_UPRCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6731 ice_stat_update40(hw, GLPRT_MPRCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6735 ice_stat_update40(hw, GLPRT_BPRCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6739 ice_stat_update32(hw, PRTRPB_RDPC, pf->stat_prev_loaded, in ice_update_pf_stats()
6743 ice_stat_update40(hw, GLPRT_GOTCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6747 ice_stat_update40(hw, GLPRT_UPTCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6751 ice_stat_update40(hw, GLPRT_MPTCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6755 ice_stat_update40(hw, GLPRT_BPTCL(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6759 ice_stat_update32(hw, GLPRT_TDOLD(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6763 ice_stat_update40(hw, GLPRT_PRC64L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6766 ice_stat_update40(hw, GLPRT_PRC127L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6769 ice_stat_update40(hw, GLPRT_PRC255L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6772 ice_stat_update40(hw, GLPRT_PRC511L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6775 ice_stat_update40(hw, GLPRT_PRC1023L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6778 ice_stat_update40(hw, GLPRT_PRC1522L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6781 ice_stat_update40(hw, GLPRT_PRC9522L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6784 ice_stat_update40(hw, GLPRT_PTC64L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6787 ice_stat_update40(hw, GLPRT_PTC127L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6790 ice_stat_update40(hw, GLPRT_PTC255L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6793 ice_stat_update40(hw, GLPRT_PTC511L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6796 ice_stat_update40(hw, GLPRT_PTC1023L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6799 ice_stat_update40(hw, GLPRT_PTC1522L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6802 ice_stat_update40(hw, GLPRT_PTC9522L(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6809 pf->stat_prev_loaded, &prev_ps->fd_sb_match, in ice_update_pf_stats()
6811 ice_stat_update32(hw, GLPRT_LXONRXC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6814 ice_stat_update32(hw, GLPRT_LXOFFRXC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6817 ice_stat_update32(hw, GLPRT_LXONTXC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6820 ice_stat_update32(hw, GLPRT_LXOFFTXC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6823 ice_update_dcb_stats(pf); in ice_update_pf_stats()
6825 ice_stat_update32(hw, GLPRT_CRCERRS(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6828 ice_stat_update32(hw, GLPRT_ILLERRC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6831 ice_stat_update32(hw, GLPRT_MLFC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6835 ice_stat_update32(hw, GLPRT_MRFC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6839 ice_stat_update32(hw, GLPRT_RLEC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6842 ice_stat_update32(hw, GLPRT_RUC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6845 ice_stat_update32(hw, GLPRT_RFC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6848 ice_stat_update32(hw, GLPRT_ROC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6851 ice_stat_update32(hw, GLPRT_RJC(port), pf->stat_prev_loaded, in ice_update_pf_stats()
6854 cur_ps->fd_sb_status = test_bit(ICE_FLAG_FD_ENA, pf->flags) ? 1 : 0; in ice_update_pf_stats()
6856 pf->stat_prev_loaded = true; in ice_update_pf_stats()
7083 struct ice_pf *pf = vsi->back; in ice_vsi_open_ctrl() local
7087 dev = ice_pf_to_dev(pf); in ice_vsi_open_ctrl()
7139 struct ice_pf *pf = vsi->back; in ice_vsi_open() local
7156 dev_driver_string(ice_pf_to_dev(pf)), vsi->netdev->name); in ice_vsi_open()
7194 * @pf: PF from which all VSIs are being removed
7196 static void ice_vsi_release_all(struct ice_pf *pf) in ice_vsi_release_all() argument
7200 if (!pf->vsi) in ice_vsi_release_all()
7203 ice_for_each_vsi(pf, i) { in ice_vsi_release_all()
7204 if (!pf->vsi[i]) in ice_vsi_release_all()
7207 if (pf->vsi[i]->type == ICE_VSI_CHNL) in ice_vsi_release_all()
7210 err = ice_vsi_release(pf->vsi[i]); in ice_vsi_release_all()
7212 dev_dbg(ice_pf_to_dev(pf), "Failed to release pf->vsi[%d], err %d, vsi_num = %d\n", in ice_vsi_release_all()
7213 i, err, pf->vsi[i]->vsi_num); in ice_vsi_release_all()
7219 * @pf: pointer to the PF instance
7222 * Iterates through the pf->vsi array and rebuilds VSIs of the requested type
7224 static int ice_vsi_rebuild_by_type(struct ice_pf *pf, enum ice_vsi_type type) in ice_vsi_rebuild_by_type() argument
7226 struct device *dev = ice_pf_to_dev(pf); in ice_vsi_rebuild_by_type()
7229 ice_for_each_vsi(pf, i) { in ice_vsi_rebuild_by_type()
7230 struct ice_vsi *vsi = pf->vsi[i]; in ice_vsi_rebuild_by_type()
7244 err = ice_replay_vsi(&pf->hw, vsi->idx); in ice_vsi_rebuild_by_type()
7254 vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx); in ice_vsi_rebuild_by_type()
7272 * ice_update_pf_netdev_link - Update PF netdev link status
7273 * @pf: pointer to the PF instance
7275 static void ice_update_pf_netdev_link(struct ice_pf *pf) in ice_update_pf_netdev_link() argument
7280 ice_for_each_vsi(pf, i) { in ice_update_pf_netdev_link()
7281 struct ice_vsi *vsi = pf->vsi[i]; in ice_update_pf_netdev_link()
7286 ice_get_link_status(pf->vsi[i]->port_info, &link_up); in ice_update_pf_netdev_link()
7288 netif_carrier_on(pf->vsi[i]->netdev); in ice_update_pf_netdev_link()
7289 netif_tx_wake_all_queues(pf->vsi[i]->netdev); in ice_update_pf_netdev_link()
7291 netif_carrier_off(pf->vsi[i]->netdev); in ice_update_pf_netdev_link()
7292 netif_tx_stop_all_queues(pf->vsi[i]->netdev); in ice_update_pf_netdev_link()
7299 * @pf: PF to rebuild
7307 static void ice_rebuild(struct ice_pf *pf, enum ice_reset_req reset_type) in ice_rebuild() argument
7309 struct ice_vsi *vsi = ice_get_main_vsi(pf); in ice_rebuild()
7310 struct device *dev = ice_pf_to_dev(pf); in ice_rebuild()
7311 struct ice_hw *hw = &pf->hw; in ice_rebuild()
7315 if (test_bit(ICE_DOWN, pf->state)) in ice_rebuild()
7318 dev_dbg(dev, "rebuilding PF after reset_type=%d\n", reset_type); in ice_rebuild()
7326 pf->fw_emp_reset_disabled = false; in ice_rebuild()
7338 if (!ice_is_safe_mode(pf)) { in ice_rebuild()
7344 ice_load_pkg(NULL, pf); in ice_rebuild()
7349 dev_err(dev, "clear PF configuration failed %d\n", err); in ice_rebuild()
7375 err = ice_aq_set_port_params(pf->hw.port_info, dvm, NULL); in ice_rebuild()
7384 err = ice_req_irq_msix_misc(pf); in ice_rebuild()
7390 if (test_bit(ICE_FLAG_FD_ENA, pf->flags)) { in ice_rebuild()
7398 /* force guaranteed filter pool for PF */ in ice_rebuild()
7400 /* force shared filter pool for PF */ in ice_rebuild()
7405 if (test_bit(ICE_FLAG_DCB_ENA, pf->flags)) in ice_rebuild()
7406 ice_dcb_rebuild(pf); in ice_rebuild()
7408 /* If the PF previously had enabled PTP, PTP init needs to happen before in ice_rebuild()
7412 if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags)) in ice_rebuild()
7413 ice_ptp_reset(pf); in ice_rebuild()
7415 if (ice_is_feature_supported(pf, ICE_F_GNSS)) in ice_rebuild()
7416 ice_gnss_init(pf); in ice_rebuild()
7418 /* rebuild PF VSI */ in ice_rebuild()
7419 err = ice_vsi_rebuild_by_type(pf, ICE_VSI_PF); in ice_rebuild()
7421 dev_err(dev, "PF VSI rebuild failed: %d\n", err); in ice_rebuild()
7426 if (test_bit(ICE_FLAG_PTP_SUPPORTED, pf->flags)) in ice_rebuild()
7427 ice_ptp_cfg_timestamp(pf, false); in ice_rebuild()
7429 err = ice_vsi_rebuild_by_type(pf, ICE_VSI_SWITCHDEV_CTRL); in ice_rebuild()
7436 err = ice_rebuild_channels(pf); in ice_rebuild()
7445 if (test_bit(ICE_FLAG_FD_ENA, pf->flags)) { in ice_rebuild()
7446 err = ice_vsi_rebuild_by_type(pf, ICE_VSI_CTRL); in ice_rebuild()
7457 ice_fdir_replay_fltrs(pf); in ice_rebuild()
7459 ice_rebuild_arfs(pf); in ice_rebuild()
7465 ice_update_pf_netdev_link(pf); in ice_rebuild()
7468 err = ice_send_version(pf); in ice_rebuild()
7478 clear_bit(ICE_RESET_FAILED, pf->state); in ice_rebuild()
7480 ice_plug_aux_dev(pf); in ice_rebuild()
7481 if (ice_is_feature_supported(pf, ICE_F_SRIOV_LAG)) in ice_rebuild()
7482 ice_lag_rebuild(pf); in ice_rebuild()
7490 set_bit(ICE_RESET_FAILED, pf->state); in ice_rebuild()
7492 /* set this bit in PF state to control service task scheduling */ in ice_rebuild()
7493 set_bit(ICE_NEEDS_RESTART, pf->state); in ice_rebuild()
7508 struct ice_pf *pf = vsi->back; in ice_change_mtu() local
7527 } else if (test_bit(ICE_FLAG_LEGACY_RX, pf->flags)) { in ice_change_mtu()
7537 if (ice_is_reset_in_progress(pf->state)) { in ice_change_mtu()
7557 set_bit(ICE_FLAG_MTU_CHANGED, pf->flags); in ice_change_mtu()
7571 struct ice_pf *pf = np->vsi->back; in ice_eth_ioctl() local
7575 return ice_ptp_get_ts_config(pf, ifr); in ice_eth_ioctl()
7577 return ice_ptp_set_ts_config(pf, ifr); in ice_eth_ioctl()
7748 struct ice_pf *pf = vsi->back; in ice_bridge_getlink() local
7751 bmode = pf->first_sw->bridge_mode; in ice_bridge_getlink()
7809 * hooked up to. Iterates through the PF VSI list and sets the loopback mode (if
7819 struct ice_pf *pf = np->vsi->back; in ice_bridge_setlink() local
7821 struct ice_hw *hw = &pf->hw; in ice_bridge_setlink()
7825 pf_sw = pf->first_sw; in ice_bridge_setlink()
7842 /* Iterates through the PF VSI list and update the loopback in ice_bridge_setlink()
7845 ice_for_each_vsi(pf, v) { in ice_bridge_setlink()
7846 if (!pf->vsi[v]) in ice_bridge_setlink()
7848 err = ice_vsi_update_bridge_mode(pf->vsi[v], mode); in ice_bridge_setlink()
7883 struct ice_pf *pf = vsi->back; in ice_tx_timeout() local
7886 pf->tx_timeout_count++; in ice_tx_timeout()
7892 if (ice_is_pfc_causing_hung_q(pf, txqueue)) { in ice_tx_timeout()
7893 dev_info(ice_pf_to_dev(pf), "Fake Tx hang detected on queue %u, timeout caused by PFC storm\n", in ice_tx_timeout()
7909 if (time_after(jiffies, (pf->tx_timeout_last_recovery + HZ * 20))) in ice_tx_timeout()
7910 pf->tx_timeout_recovery_level = 1; in ice_tx_timeout()
7911 else if (time_before(jiffies, (pf->tx_timeout_last_recovery + in ice_tx_timeout()
7916 struct ice_hw *hw = &pf->hw; in ice_tx_timeout()
7929 pf->tx_timeout_last_recovery = jiffies; in ice_tx_timeout()
7931 pf->tx_timeout_recovery_level, txqueue); in ice_tx_timeout()
7933 switch (pf->tx_timeout_recovery_level) { in ice_tx_timeout()
7935 set_bit(ICE_PFR_REQ, pf->state); in ice_tx_timeout()
7938 set_bit(ICE_CORER_REQ, pf->state); in ice_tx_timeout()
7941 set_bit(ICE_GLOBR_REQ, pf->state); in ice_tx_timeout()
7945 set_bit(ICE_DOWN, pf->state); in ice_tx_timeout()
7947 set_bit(ICE_SERVICE_DIS, pf->state); in ice_tx_timeout()
7951 ice_service_task_schedule(pf); in ice_tx_timeout()
7952 pf->tx_timeout_recovery_level++; in ice_tx_timeout()
8015 struct ice_pf *pf = vsi->back; in ice_validate_mqprio_qopt() local
8030 dev = ice_pf_to_dev(pf); in ice_validate_mqprio_qopt()
8142 * ice_add_vsi_to_fdir - add a VSI to the flow director group for PF
8143 * @pf: ptr to PF device
8146 static int ice_add_vsi_to_fdir(struct ice_pf *pf, struct ice_vsi *vsi) in ice_add_vsi_to_fdir() argument
8148 struct device *dev = ice_pf_to_dev(pf); in ice_add_vsi_to_fdir()
8156 hw = &pf->hw; in ice_add_vsi_to_fdir()
8204 * @pf: ptr to PF device
8210 static int ice_add_channel(struct ice_pf *pf, u16 sw_id, struct ice_channel *ch) in ice_add_channel() argument
8212 struct device *dev = ice_pf_to_dev(pf); in ice_add_channel()
8220 vsi = ice_chnl_vsi_setup(pf, pf->hw.port_info, ch); in ice_add_channel()
8226 ice_add_vsi_to_fdir(pf, vsi); in ice_add_channel()
8318 * @pf: ptr to PF device
8328 ice_setup_hw_channel(struct ice_pf *pf, struct ice_vsi *vsi, in ice_setup_hw_channel() argument
8331 struct device *dev = ice_pf_to_dev(pf); in ice_setup_hw_channel()
8337 ret = ice_add_channel(pf, sw_id, ch); in ice_setup_hw_channel()
8358 * @pf: ptr to PF device
8366 ice_setup_channel(struct ice_pf *pf, struct ice_vsi *vsi, in ice_setup_channel() argument
8369 struct device *dev = ice_pf_to_dev(pf); in ice_setup_channel()
8378 sw_id = pf->first_sw->sw_id; in ice_setup_channel()
8381 ret = ice_setup_hw_channel(pf, vsi, ch, sw_id, ICE_VSI_CHNL); in ice_setup_channel()
8419 struct ice_pf *pf = vsi->back; in ice_create_q_channel() local
8425 dev = ice_pf_to_dev(pf); in ice_create_q_channel()
8437 if (!ice_setup_channel(pf, vsi, ch)) { in ice_create_q_channel()
8462 * @pf: ptr to PF, TC-flower based filter are tracked at PF level
8467 static void ice_rem_all_chnl_fltrs(struct ice_pf *pf) in ice_rem_all_chnl_fltrs() argument
8474 &pf->tc_flower_fltr_list, in ice_rem_all_chnl_fltrs()
8486 status = ice_rem_adv_rule_by_id(&pf->hw, &rule); in ice_rem_all_chnl_fltrs()
8489 dev_dbg(ice_pf_to_dev(pf), "TC flower filter (rule_id %u) does not exist\n", in ice_rem_all_chnl_fltrs()
8492 dev_err(ice_pf_to_dev(pf), "failed to delete TC flower filter, status %d\n", in ice_rem_all_chnl_fltrs()
8502 pf->num_dmac_chnl_fltrs--; in ice_rem_all_chnl_fltrs()
8521 struct ice_pf *pf = vsi->back; in ice_remove_q_channels() local
8526 ice_rem_all_chnl_fltrs(pf); in ice_remove_q_channels()
8530 struct ice_hw *hw = &pf->hw; in ice_remove_q_channels()
8568 ice_fdir_rem_adq_chnl(&pf->hw, ch->ch_vsi->idx); in ice_remove_q_channels()
8573 /* Delete VSI from FW, PF and HW VSI arrays */ in ice_remove_q_channels()
8591 * @pf: ptr to PF
8595 static int ice_rebuild_channels(struct ice_pf *pf) in ice_rebuild_channels() argument
8597 struct device *dev = ice_pf_to_dev(pf); in ice_rebuild_channels()
8605 main_vsi = ice_get_main_vsi(pf); in ice_rebuild_channels()
8609 if (!test_bit(ICE_FLAG_TC_MQPRIO, pf->flags) || in ice_rebuild_channels()
8624 ice_for_each_vsi(pf, i) { in ice_rebuild_channels()
8627 vsi = pf->vsi[i]; in ice_rebuild_channels()
8644 vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx); in ice_rebuild_channels()
8647 err = ice_replay_vsi(&pf->hw, vsi->idx); in ice_rebuild_channels()
8711 struct ice_pf *pf = vsi->back; in ice_create_q_channels() local
8741 dev_err(ice_pf_to_dev(pf), in ice_create_q_channels()
8748 dev_dbg(ice_pf_to_dev(pf), in ice_create_q_channels()
8769 struct ice_pf *pf = vsi->back; in ice_setup_tc_mqprio_qdisc() local
8776 dev = ice_pf_to_dev(pf); in ice_setup_tc_mqprio_qdisc()
8781 clear_bit(ICE_FLAG_TC_MQPRIO, pf->flags); in ice_setup_tc_mqprio_qdisc()
8794 if (pf->hw.port_info->is_custom_tx_enabled) { in ice_setup_tc_mqprio_qdisc()
8798 ice_tear_down_devlink_rate_tree(pf); in ice_setup_tc_mqprio_qdisc()
8807 set_bit(ICE_FLAG_TC_MQPRIO, pf->flags); in ice_setup_tc_mqprio_qdisc()
8813 set_bit(ICE_FLAG_CLS_FLOWER, pf->flags); in ice_setup_tc_mqprio_qdisc()
8829 if (!hw && !test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) in ice_setup_tc_mqprio_qdisc()
8832 if (!hw && !test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) { in ice_setup_tc_mqprio_qdisc()
8833 vsi->req_txq = min_t(int, ice_get_avail_txq_count(pf), in ice_setup_tc_mqprio_qdisc()
8835 vsi->req_rxq = min_t(int, ice_get_avail_rxq_count(pf), in ice_setup_tc_mqprio_qdisc()
8872 clear_bit(ICE_RESET_FAILED, pf->state); in ice_setup_tc_mqprio_qdisc()
8888 if (test_bit(ICE_FLAG_TC_MQPRIO, pf->flags)) { in ice_setup_tc_mqprio_qdisc()
8941 struct ice_pf *pf = np->vsi->back; in ice_setup_tc() local
8952 if (ice_is_eswitch_mode_switchdev(pf)) { in ice_setup_tc()
8957 if (pf->adev) { in ice_setup_tc()
8958 mutex_lock(&pf->adev_mutex); in ice_setup_tc()
8959 device_lock(&pf->adev->dev); in ice_setup_tc()
8961 if (pf->adev->dev.driver) { in ice_setup_tc()
8969 mutex_lock(&pf->tc_mutex); in ice_setup_tc()
8971 mutex_unlock(&pf->tc_mutex); in ice_setup_tc()
8975 device_unlock(&pf->adev->dev); in ice_setup_tc()
8976 mutex_unlock(&pf->adev_mutex); in ice_setup_tc()
9115 struct ice_pf *pf = np->vsi->back; in ice_open() local
9117 if (ice_is_reset_in_progress(pf->state)) { in ice_open()
9138 struct ice_pf *pf = vsi->back; in ice_open_internal() local
9142 if (test_bit(ICE_NEEDS_RESTART, pf->state)) { in ice_open_internal()
9156 ice_check_link_cfg_err(pf, pi->phy.link_info.link_cfg_err); in ice_open_internal()
9160 clear_bit(ICE_FLAG_NO_MEDIA, pf->flags); in ice_open_internal()
9161 if (!test_bit(ICE_PHY_INIT_COMPLETE, pf->state)) { in ice_open_internal()
9177 set_bit(ICE_FLAG_NO_MEDIA, pf->flags); in ice_open_internal()
9206 struct ice_pf *pf = vsi->back; in ice_stop() local
9208 if (ice_is_reset_in_progress(pf->state)) { in ice_stop()