Lines Matching refs:dev
189 static void link_report(struct net_device *dev) in link_report() argument
191 if (!netif_carrier_ok(dev)) in link_report()
192 netdev_info(dev, "link down\n"); in link_report()
197 const struct port_info *p = netdev_priv(dev); in link_report()
223 dev->name, p->link_cfg.speed); in link_report()
227 netdev_info(dev, "link up, %s, full-duplex, %s PAUSE\n", s, in link_report()
234 static void dcb_tx_queue_prio_enable(struct net_device *dev, int enable) in dcb_tx_queue_prio_enable() argument
236 struct port_info *pi = netdev_priv(dev); in dcb_tx_queue_prio_enable()
271 int cxgb4_dcb_enabled(const struct net_device *dev) in cxgb4_dcb_enabled() argument
273 struct port_info *pi = netdev_priv(dev); in cxgb4_dcb_enabled()
285 struct net_device *dev = adapter->port[port_id]; in t4_os_link_changed() local
288 if (netif_running(dev) && link_stat != netif_carrier_ok(dev)) { in t4_os_link_changed()
290 netif_carrier_on(dev); in t4_os_link_changed()
293 if (cxgb4_dcb_enabled(dev)) { in t4_os_link_changed()
294 cxgb4_dcb_reset(dev); in t4_os_link_changed()
295 dcb_tx_queue_prio_enable(dev, false); in t4_os_link_changed()
298 netif_carrier_off(dev); in t4_os_link_changed()
301 link_report(dev); in t4_os_link_changed()
311 struct net_device *dev = adap->port[port_id]; in t4_os_portmod_changed() local
312 struct port_info *pi = netdev_priv(dev); in t4_os_portmod_changed()
315 netdev_info(dev, "port module unplugged\n"); in t4_os_portmod_changed()
317 netdev_info(dev, "%s module inserted\n", mod_str[pi->mod_type]); in t4_os_portmod_changed()
319 netdev_info(dev, "%s: unsupported port module inserted\n", in t4_os_portmod_changed()
320 dev->name); in t4_os_portmod_changed()
322 netdev_info(dev, "%s: unknown port module inserted\n", in t4_os_portmod_changed()
323 dev->name); in t4_os_portmod_changed()
325 netdev_info(dev, "%s: transceiver module error\n", dev->name); in t4_os_portmod_changed()
327 netdev_info(dev, "%s: unknown module type %d inserted\n", in t4_os_portmod_changed()
328 dev->name, pi->mod_type); in t4_os_portmod_changed()
333 pi->link_cfg.redo_l1cfg = netif_running(dev); in t4_os_portmod_changed()
429 static int set_rxmode(struct net_device *dev, int mtu, bool sleep_ok) in set_rxmode() argument
431 struct port_info *pi = netdev_priv(dev); in set_rxmode()
434 __dev_uc_sync(dev, cxgb4_mac_sync, cxgb4_mac_unsync); in set_rxmode()
435 __dev_mc_sync(dev, cxgb4_mac_sync, cxgb4_mac_unsync); in set_rxmode()
438 mtu, (dev->flags & IFF_PROMISC) ? 1 : 0, in set_rxmode()
439 (dev->flags & IFF_ALLMULTI) ? 1 : 0, 1, -1, in set_rxmode()
503 static int link_start(struct net_device *dev) in link_start() argument
505 struct port_info *pi = netdev_priv(dev); in link_start()
514 dev->mtu, -1, -1, -1, in link_start()
515 !!(dev->features & NETIF_F_HW_VLAN_CTAG_RX), true); in link_start()
518 dev->dev_addr, true, &pi->smt_idx); in link_start()
537 struct net_device *dev = adap->port[adap->chan_map[port]]; in dcb_rpl() local
538 int old_dcb_enabled = cxgb4_dcb_enabled(dev); in dcb_rpl()
542 new_dcb_enabled = cxgb4_dcb_enabled(dev); in dcb_rpl()
549 dcb_tx_queue_prio_enable(dev, new_dcb_enabled); in dcb_rpl()
608 struct net_device *dev; in fwevtq_handler() local
611 dev = q->adap->port[q->adap->chan_map[port]]; in fwevtq_handler()
620 cxgb4_dcb_state_fsm(dev, state_input); in fwevtq_handler()
1059 struct net_device *dev = adap->port[i]; in setup_sge_queues() local
1060 struct port_info *pi = netdev_priv(dev); in setup_sge_queues()
1074 "%s-Rx%d", dev->name, j); in setup_sge_queues()
1078 err = t4_sge_alloc_rxq(adap, &q->rspq, false, dev, in setup_sge_queues()
1092 err = t4_sge_alloc_eth_txq(adap, t, dev, in setup_sge_queues()
1093 netdev_get_tx_queue(dev, j), in setup_sge_queues()
1134 static u16 cxgb_select_queue(struct net_device *dev, struct sk_buff *skb, in cxgb_select_queue() argument
1145 if (cxgb4_dcb_enabled(dev) && !is_kdump_kernel()) { in cxgb_select_queue()
1152 netdev_warn(dev, in cxgb_select_queue()
1166 if (dev->num_tc) { in cxgb_select_queue()
1167 struct port_info *pi = netdev2pinfo(dev); in cxgb_select_queue()
1175 txq = netdev_pick_tx(dev, skb, sb_dev); in cxgb_select_queue()
1176 if (xfrm_offload(skb) || is_ptp_enabled(skb, dev) || in cxgb_select_queue()
1190 while (unlikely(txq >= dev->real_num_tx_queues)) in cxgb_select_queue()
1191 txq -= dev->real_num_tx_queues; in cxgb_select_queue()
1196 return netdev_pick_tx(dev, skb, NULL) % dev->real_num_tx_queues; in cxgb_select_queue()
1272 static int cxgb_set_features(struct net_device *dev, netdev_features_t features) in cxgb_set_features() argument
1274 netdev_features_t changed = dev->features ^ features; in cxgb_set_features()
1275 const struct port_info *pi = netdev_priv(dev); in cxgb_set_features()
1285 dev->features = features ^ NETIF_F_HW_VLAN_CTAG_RX; in cxgb_set_features()
1317 static int cxgb4_port_mirror_alloc_queues(struct net_device *dev) in cxgb4_port_mirror_alloc_queues() argument
1319 struct port_info *pi = netdev2pinfo(dev); in cxgb4_port_mirror_alloc_queues()
1320 struct adapter *adap = netdev2adap(dev); in cxgb4_port_mirror_alloc_queues()
1356 "%s-mirrorrxq%d", dev->name, i); in cxgb4_port_mirror_alloc_queues()
1368 dev, msix, &mirror_rxq->fl, in cxgb4_port_mirror_alloc_queues()
1424 static void cxgb4_port_mirror_free_queues(struct net_device *dev) in cxgb4_port_mirror_free_queues() argument
1426 struct port_info *pi = netdev2pinfo(dev); in cxgb4_port_mirror_free_queues()
1427 struct adapter *adap = netdev2adap(dev); in cxgb4_port_mirror_free_queues()
1445 static int cxgb4_port_mirror_start(struct net_device *dev) in cxgb4_port_mirror_start() argument
1447 struct port_info *pi = netdev2pinfo(dev); in cxgb4_port_mirror_start()
1448 struct adapter *adap = netdev2adap(dev); in cxgb4_port_mirror_start()
1460 dev->mtu, (dev->flags & IFF_PROMISC) ? 1 : 0, in cxgb4_port_mirror_start()
1461 (dev->flags & IFF_ALLMULTI) ? 1 : 0, 1, in cxgb4_port_mirror_start()
1462 !!(dev->features & NETIF_F_HW_VLAN_CTAG_RX), true); in cxgb4_port_mirror_start()
1475 dev->dev_addr, true, NULL); in cxgb4_port_mirror_start()
1501 static void cxgb4_port_mirror_stop(struct net_device *dev) in cxgb4_port_mirror_stop() argument
1503 struct port_info *pi = netdev2pinfo(dev); in cxgb4_port_mirror_stop()
1504 struct adapter *adap = netdev2adap(dev); in cxgb4_port_mirror_stop()
1513 int cxgb4_port_mirror_alloc(struct net_device *dev) in cxgb4_port_mirror_alloc() argument
1515 struct port_info *pi = netdev2pinfo(dev); in cxgb4_port_mirror_alloc()
1516 struct adapter *adap = netdev2adap(dev); in cxgb4_port_mirror_alloc()
1536 ret = cxgb4_port_mirror_alloc_queues(dev); in cxgb4_port_mirror_alloc()
1540 ret = cxgb4_port_mirror_start(dev); in cxgb4_port_mirror_alloc()
1549 cxgb4_port_mirror_free_queues(dev); in cxgb4_port_mirror_alloc()
1561 void cxgb4_port_mirror_free(struct net_device *dev) in cxgb4_port_mirror_free() argument
1563 struct port_info *pi = netdev2pinfo(dev); in cxgb4_port_mirror_free()
1564 struct adapter *adap = netdev2adap(dev); in cxgb4_port_mirror_free()
1575 cxgb4_port_mirror_stop(dev); in cxgb4_port_mirror_free()
1576 cxgb4_port_mirror_free_queues(dev); in cxgb4_port_mirror_free()
1922 int cxgb4_create_server(const struct net_device *dev, unsigned int stid, in cxgb4_create_server() argument
1936 adap = netdev2adap(dev); in cxgb4_create_server()
1963 int cxgb4_create_server6(const struct net_device *dev, unsigned int stid, in cxgb4_create_server6() argument
1977 adap = netdev2adap(dev); in cxgb4_create_server6()
1996 int cxgb4_remove_server(const struct net_device *dev, unsigned int stid, in cxgb4_remove_server() argument
2004 adap = netdev2adap(dev); in cxgb4_remove_server()
2120 unsigned int cxgb4_port_chan(const struct net_device *dev) in cxgb4_port_chan() argument
2122 return netdev2pinfo(dev)->tx_chan; in cxgb4_port_chan()
2132 unsigned int cxgb4_port_e2cchan(const struct net_device *dev) in cxgb4_port_e2cchan() argument
2134 return netdev2pinfo(dev)->rx_cchan; in cxgb4_port_e2cchan()
2138 unsigned int cxgb4_dbfifo_count(const struct net_device *dev, int lpfifo) in cxgb4_dbfifo_count() argument
2140 struct adapter *adap = netdev2adap(dev); in cxgb4_dbfifo_count()
2162 unsigned int cxgb4_port_viid(const struct net_device *dev) in cxgb4_port_viid() argument
2164 return netdev2pinfo(dev)->viid; in cxgb4_port_viid()
2174 unsigned int cxgb4_port_idx(const struct net_device *dev) in cxgb4_port_idx() argument
2176 return netdev2pinfo(dev)->port_id; in cxgb4_port_idx()
2191 void cxgb4_iscsi_init(struct net_device *dev, unsigned int tag_mask, in cxgb4_iscsi_init() argument
2194 struct adapter *adap = netdev2adap(dev); in cxgb4_iscsi_init()
2203 int cxgb4_flush_eq_cache(struct net_device *dev) in cxgb4_flush_eq_cache() argument
2205 struct adapter *adap = netdev2adap(dev); in cxgb4_flush_eq_cache()
2229 int cxgb4_sync_txq_pidx(struct net_device *dev, u16 qid, u16 pidx, in cxgb4_sync_txq_pidx() argument
2232 struct adapter *adap = netdev2adap(dev); in cxgb4_sync_txq_pidx()
2262 int cxgb4_read_tpte(struct net_device *dev, u32 stag, __be32 *tpte) in cxgb4_read_tpte() argument
2271 adap = netdev2adap(dev); in cxgb4_read_tpte()
2338 u64 cxgb4_read_sge_timestamp(struct net_device *dev) in cxgb4_read_sge_timestamp() argument
2343 adap = netdev2adap(dev); in cxgb4_read_sge_timestamp()
2351 int cxgb4_bar2_sge_qregs(struct net_device *dev, in cxgb4_bar2_sge_qregs() argument
2358 return t4_bar2_sge_qregs(netdev2adap(dev), in cxgb4_bar2_sge_qregs()
2374 const struct net_device *netdev = neigh->dev; in check_neigh_update()
2378 parent = netdev->dev.parent; in check_neigh_update()
2680 struct net_device *event_dev = ifa->idev->dev; in cxgb4_inet6addr_handler()
2708 parent = event_dev->dev.parent; in cxgb4_inet6addr_handler()
2733 struct net_device *dev; in update_clip() local
2739 dev = adap->port[i]; in update_clip()
2742 if (dev) in update_clip()
2743 ret = cxgb4_update_root_dev_clip(dev); in update_clip()
2839 static int cxgb_open(struct net_device *dev) in cxgb_open() argument
2841 struct port_info *pi = netdev_priv(dev); in cxgb_open()
2845 netif_carrier_off(dev); in cxgb_open()
2860 err = link_start(dev); in cxgb_open()
2866 err = cxgb4_port_mirror_alloc_queues(dev); in cxgb_open()
2870 err = cxgb4_port_mirror_start(dev); in cxgb_open()
2876 netif_tx_start_all_queues(dev); in cxgb_open()
2880 cxgb4_port_mirror_free_queues(dev); in cxgb_open()
2887 static int cxgb_close(struct net_device *dev) in cxgb_close() argument
2889 struct port_info *pi = netdev_priv(dev); in cxgb_close()
2893 netif_tx_stop_all_queues(dev); in cxgb_close()
2894 netif_carrier_off(dev); in cxgb_close()
2898 cxgb4_dcb_reset(dev); in cxgb_close()
2899 dcb_tx_queue_prio_enable(dev, false); in cxgb_close()
2906 cxgb4_port_mirror_stop(dev); in cxgb_close()
2907 cxgb4_port_mirror_free_queues(dev); in cxgb_close()
2914 int cxgb4_create_server_filter(const struct net_device *dev, unsigned int stid, in cxgb4_create_server_filter() argument
2924 adap = netdev2adap(dev); in cxgb4_create_server_filter()
2984 int cxgb4_remove_server_filter(const struct net_device *dev, unsigned int stid, in cxgb4_remove_server_filter() argument
2990 adap = netdev2adap(dev); in cxgb4_remove_server_filter()
3004 static void cxgb_get_stats(struct net_device *dev, in cxgb_get_stats() argument
3008 struct port_info *p = netdev_priv(dev); in cxgb_get_stats()
3016 if (!netif_device_present(dev)) { in cxgb_get_stats()
3054 static int cxgb_ioctl(struct net_device *dev, struct ifreq *req, int cmd) in cxgb_ioctl() argument
3058 struct port_info *pi = netdev_priv(dev); in cxgb_ioctl()
3169 static void cxgb_set_rxmode(struct net_device *dev) in cxgb_set_rxmode() argument
3172 set_rxmode(dev, -1, false); in cxgb_set_rxmode()
3175 static int cxgb_change_mtu(struct net_device *dev, int new_mtu) in cxgb_change_mtu() argument
3177 struct port_info *pi = netdev_priv(dev); in cxgb_change_mtu()
3183 dev->mtu = new_mtu; in cxgb_change_mtu()
3188 static int cxgb4_mgmt_open(struct net_device *dev) in cxgb4_mgmt_open() argument
3193 netif_carrier_off(dev); in cxgb4_mgmt_open()
3233 static int cxgb4_mgmt_set_vf_mac(struct net_device *dev, int vf, u8 *mac) in cxgb4_mgmt_set_vf_mac() argument
3235 struct port_info *pi = netdev_priv(dev); in cxgb4_mgmt_set_vf_mac()
3255 static int cxgb4_mgmt_get_vf_config(struct net_device *dev, in cxgb4_mgmt_get_vf_config() argument
3258 struct port_info *pi = netdev_priv(dev); in cxgb4_mgmt_get_vf_config()
3275 static int cxgb4_mgmt_get_phys_port_id(struct net_device *dev, in cxgb4_mgmt_get_phys_port_id() argument
3278 struct port_info *pi = netdev_priv(dev); in cxgb4_mgmt_get_phys_port_id()
3287 static int cxgb4_mgmt_set_vf_rate(struct net_device *dev, int vf, in cxgb4_mgmt_set_vf_rate() argument
3290 struct port_info *pi = netdev_priv(dev); in cxgb4_mgmt_set_vf_rate()
3388 static int cxgb4_mgmt_set_vf_vlan(struct net_device *dev, int vf, in cxgb4_mgmt_set_vf_vlan() argument
3391 struct port_info *pi = netdev_priv(dev); in cxgb4_mgmt_set_vf_vlan()
3412 static int cxgb4_mgmt_set_vf_link_state(struct net_device *dev, int vf, in cxgb4_mgmt_set_vf_link_state() argument
3415 struct port_info *pi = netdev_priv(dev); in cxgb4_mgmt_set_vf_link_state()
3456 static int cxgb_set_mac_addr(struct net_device *dev, void *p) in cxgb_set_mac_addr() argument
3460 struct port_info *pi = netdev_priv(dev); in cxgb_set_mac_addr()
3470 eth_hw_addr_set(dev, addr->sa_data); in cxgb_set_mac_addr()
3475 static void cxgb_netpoll(struct net_device *dev) in cxgb_netpoll() argument
3477 struct port_info *pi = netdev_priv(dev); in cxgb_netpoll()
3491 static int cxgb_set_tx_maxrate(struct net_device *dev, int index, u32 rate) in cxgb_set_tx_maxrate() argument
3493 struct port_info *pi = netdev_priv(dev); in cxgb_set_tx_maxrate()
3501 if (!can_sched(dev)) in cxgb_set_tx_maxrate()
3515 e = cxgb4_sched_queue_lookup(dev, &qe); in cxgb_set_tx_maxrate()
3539 err = cxgb4_sched_class_unbind(dev, (void *)(&qe), SCHED_QUEUE); in cxgb_set_tx_maxrate()
3562 p.u.params.pktsize = dev->mtu; in cxgb_set_tx_maxrate()
3564 e = cxgb4_sched_class_alloc(dev, &p); in cxgb_set_tx_maxrate()
3573 err = cxgb4_sched_class_bind(dev, (void *)(&qe), SCHED_QUEUE); in cxgb_set_tx_maxrate()
3580 static int cxgb_setup_tc_flower(struct net_device *dev, in cxgb_setup_tc_flower() argument
3585 return cxgb4_tc_flower_replace(dev, cls_flower); in cxgb_setup_tc_flower()
3587 return cxgb4_tc_flower_destroy(dev, cls_flower); in cxgb_setup_tc_flower()
3589 return cxgb4_tc_flower_stats(dev, cls_flower); in cxgb_setup_tc_flower()
3595 static int cxgb_setup_tc_cls_u32(struct net_device *dev, in cxgb_setup_tc_cls_u32() argument
3601 return cxgb4_config_knode(dev, cls_u32); in cxgb_setup_tc_cls_u32()
3603 return cxgb4_delete_knode(dev, cls_u32); in cxgb_setup_tc_cls_u32()
3609 static int cxgb_setup_tc_matchall(struct net_device *dev, in cxgb_setup_tc_matchall() argument
3613 struct adapter *adap = netdev2adap(dev); in cxgb_setup_tc_matchall()
3620 return cxgb4_tc_matchall_replace(dev, cls_matchall, ingress); in cxgb_setup_tc_matchall()
3622 return cxgb4_tc_matchall_destroy(dev, cls_matchall, ingress); in cxgb_setup_tc_matchall()
3625 return cxgb4_tc_matchall_stats(dev, cls_matchall); in cxgb_setup_tc_matchall()
3637 struct net_device *dev = cb_priv; in cxgb_setup_tc_block_ingress_cb() local
3638 struct port_info *pi = netdev2pinfo(dev); in cxgb_setup_tc_block_ingress_cb()
3639 struct adapter *adap = netdev2adap(dev); in cxgb_setup_tc_block_ingress_cb()
3648 if (!tc_cls_can_offload_and_chain0(dev, type_data)) in cxgb_setup_tc_block_ingress_cb()
3653 return cxgb_setup_tc_cls_u32(dev, type_data); in cxgb_setup_tc_block_ingress_cb()
3655 return cxgb_setup_tc_flower(dev, type_data); in cxgb_setup_tc_block_ingress_cb()
3657 return cxgb_setup_tc_matchall(dev, type_data, true); in cxgb_setup_tc_block_ingress_cb()
3666 struct net_device *dev = cb_priv; in cxgb_setup_tc_block_egress_cb() local
3667 struct port_info *pi = netdev2pinfo(dev); in cxgb_setup_tc_block_egress_cb()
3668 struct adapter *adap = netdev2adap(dev); in cxgb_setup_tc_block_egress_cb()
3677 if (!tc_cls_can_offload_and_chain0(dev, type_data)) in cxgb_setup_tc_block_egress_cb()
3682 return cxgb_setup_tc_matchall(dev, type_data, false); in cxgb_setup_tc_block_egress_cb()
3690 static int cxgb_setup_tc_mqprio(struct net_device *dev, in cxgb_setup_tc_mqprio() argument
3693 struct adapter *adap = netdev2adap(dev); in cxgb_setup_tc_mqprio()
3698 return cxgb4_setup_tc_mqprio(dev, mqprio); in cxgb_setup_tc_mqprio()
3703 static int cxgb_setup_tc_block(struct net_device *dev, in cxgb_setup_tc_block() argument
3706 struct port_info *pi = netdev_priv(dev); in cxgb_setup_tc_block()
3720 cb, pi, dev, ingress_only); in cxgb_setup_tc_block()
3723 static int cxgb_setup_tc(struct net_device *dev, enum tc_setup_type type, in cxgb_setup_tc() argument
3728 return cxgb_setup_tc_mqprio(dev, type_data); in cxgb_setup_tc()
3730 return cxgb_setup_tc_block(dev, type_data); in cxgb_setup_tc()
3837 struct net_device *dev, in cxgb_features_check() argument
3840 struct port_info *pi = netdev_priv(dev); in cxgb_features_check()
3854 static netdev_features_t cxgb_fix_features(struct net_device *dev, in cxgb_fix_features() argument
3900 static void cxgb4_mgmt_get_drvinfo(struct net_device *dev, in cxgb4_mgmt_get_drvinfo() argument
3903 struct adapter *adapter = netdev2adap(dev); in cxgb4_mgmt_get_drvinfo()
3935 struct net_device *dev = adap->port[port]; in t4_fatal_err() local
3940 if (!dev) in t4_fatal_err()
3943 netif_tx_stop_all_queues(dev); in t4_fatal_err()
3944 netif_carrier_off(dev); in t4_fatal_err()
4296 dev_err(&adapter->pdev->dev, in adap_init0_tweaks()
5259 dev_warn(&adap->pdev->dev, "could not allocate SRQ, continuing\n"); in adap_init0()
5441 struct net_device *dev = adap->port[i]; in eeh_err_detected() local
5442 if (dev) { in eeh_err_detected()
5443 netif_device_detach(dev); in eeh_err_detected()
5444 netif_carrier_off(dev); in eeh_err_detected()
5474 dev_err(&pdev->dev, "Cannot reenable PCI " in eeh_slot_reset()
5534 struct net_device *dev = adap->port[i]; in eeh_resume() local
5535 if (dev) { in eeh_resume()
5536 if (netif_running(dev)) { in eeh_resume()
5537 link_start(dev); in eeh_resume()
5538 cxgb_set_rxmode(dev); in eeh_resume()
5540 netif_device_attach(dev); in eeh_resume()
6131 static void print_port_info(const struct net_device *dev) in print_port_info() argument
6135 const struct port_info *pi = netdev_priv(dev); in print_port_info()
6160 netdev_info(dev, "Chelsio %s %s\n", adap->params.vpd.id, buf); in print_port_info()
6231 static void cxgb4_mgmt_setup(struct net_device *dev) in cxgb4_mgmt_setup() argument
6233 dev->type = ARPHRD_NONE; in cxgb4_mgmt_setup()
6234 dev->mtu = 0; in cxgb4_mgmt_setup()
6235 dev->hard_header_len = 0; in cxgb4_mgmt_setup()
6236 dev->addr_len = 0; in cxgb4_mgmt_setup()
6237 dev->tx_queue_len = 0; in cxgb4_mgmt_setup()
6238 dev->flags |= IFF_NOARP; in cxgb4_mgmt_setup()
6239 dev->priv_flags |= IFF_NO_QUEUE; in cxgb4_mgmt_setup()
6242 dev->netdev_ops = &cxgb4_mgmt_netdev_ops; in cxgb4_mgmt_setup()
6243 dev->ethtool_ops = &cxgb4_mgmt_ethtool_ops; in cxgb4_mgmt_setup()
6256 dev_warn(&pdev->dev, "Device not initialized\n"); in cxgb4_iov_configure()
6264 dev_err(&pdev->dev, in cxgb4_iov_configure()
6319 …dev_warn(&pdev->dev, "Parent bridge %02x:%02x.%x doesn't support ARI; can't instantiate Virtual Fu… in cxgb4_iov_configure()
6349 SET_NETDEV_DEV(netdev, &pdev->dev); in cxgb4_iov_configure()
6495 struct adapter *adap = netdev2adap(x->xso.dev); in cxgb4_xfrm_add_state()
6516 struct adapter *adap = netdev2adap(x->xso.dev); in cxgb4_xfrm_del_state()
6534 struct adapter *adap = netdev2adap(x->xso.dev); in cxgb4_xfrm_free_state()
6552 struct adapter *adap = netdev2adap(x->xso.dev); in cxgb4_ipsec_offload_ok()
6572 struct adapter *adap = netdev2adap(x->xso.dev); in cxgb4_advance_esn_state()
6615 dev_info(&pdev->dev, "cannot obtain PCI resources\n"); in init_one()
6621 dev_err(&pdev->dev, "cannot enable PCI device\n"); in init_one()
6627 dev_err(&pdev->dev, "cannot map device registers\n"); in init_one()
6648 dev_err(&pdev->dev, "Device %d is not supported\n", device_id); in init_one()
6657 adapter->pdev_dev = &pdev->dev; in init_one()
6683 err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64)); in init_one()
6685 dev_err(&pdev->dev, "no usable DMA configuration\n"); in init_one()
6757 dev_err(&pdev->dev, in init_one()
6765 dev_err(&pdev->dev, "cannot map device bar2 region\n"); in init_one()
6801 SET_NETDEV_DEV(netdev, &pdev->dev); in init_one()
6904 dev_warn(&pdev->dev, "could not allocate SMT, continuing\n"); in init_one()
6910 dev_warn(&pdev->dev, "could not allocate L2T, continuing\n"); in init_one()
6920 dev_warn(&pdev->dev, in init_one()
6930 dev_warn(&pdev->dev, in init_one()
6941 dev_warn(&pdev->dev, in init_one()
6965 dev_warn(&pdev->dev, "could not allocate TID table, " in init_one()
6971 dev_warn(&pdev->dev, in init_one()
6975 dev_warn(&pdev->dev, in init_one()
6979 dev_warn(&pdev->dev, in init_one()
6983 dev_warn(&pdev->dev, in init_one()
6986 dev_warn(&pdev->dev, in init_one()
7044 dev_err(&pdev->dev, "could not register any net devices\n"); in init_one()
7048 dev_warn(&pdev->dev, "only %d net devices registered\n", i); in init_one()