xref: /openbmc/linux/drivers/net/ethernet/ibm/ibmvnic.h (revision d37cf9b63113f13d742713881ce691fc615d8b3b)
1  /* SPDX-License-Identifier: GPL-2.0-or-later */
2  /**************************************************************************/
3  /*                                                                        */
4  /*  IBM System i and System p Virtual NIC Device Driver                   */
5  /*  Copyright (C) 2014 IBM Corp.                                          */
6  /*  Santiago Leon (santi_leon@yahoo.com)                                  */
7  /*  Thomas Falcon (tlfalcon@linux.vnet.ibm.com)                           */
8  /*  John Allen (jallen@linux.vnet.ibm.com)                                */
9  /*                                                                        */
10  /*                                                                        */
11  /* This module contains the implementation of a virtual ethernet device   */
12  /* for use with IBM i/pSeries LPAR Linux.  It utilizes the logical LAN    */
13  /* option of the RS/6000 Platform Architecture to interface with virtual */
14  /* ethernet NICs that are presented to the partition by the hypervisor.   */
15  /*                                                                        */
16  /**************************************************************************/
17  
18  #define IBMVNIC_NAME		"ibmvnic"
19  #define IBMVNIC_DRIVER_VERSION	"1.0.1"
20  #define IBMVNIC_INVALID_MAP	-1
21  #define IBMVNIC_OPEN_FAILED	3
22  
23  /* basic structures plus 100 2k buffers */
24  #define IBMVNIC_IO_ENTITLEMENT_DEFAULT	610305
25  
26  /* Initial module_parameters */
27  #define IBMVNIC_RX_WEIGHT		16
28  /* when changing this, update IBMVNIC_IO_ENTITLEMENT_DEFAULT */
29  #define IBMVNIC_BUFFS_PER_POOL	100
30  #define IBMVNIC_MAX_QUEUES	16
31  #define IBMVNIC_MAX_QUEUE_SZ   4096
32  #define IBMVNIC_MAX_IND_DESCS  16
33  #define IBMVNIC_IND_ARR_SZ	(IBMVNIC_MAX_IND_DESCS * 32)
34  
35  #define IBMVNIC_TSO_BUF_SZ	65536
36  #define IBMVNIC_TSO_BUFS	64
37  #define IBMVNIC_TSO_POOL_MASK	0x80000000
38  
39  /* A VNIC adapter has set of Rx and Tx pools (aka queues). Each Rx/Tx pool
40   * has a set of buffers. The size of each buffer is determined by the MTU.
41   *
42   * Each Rx/Tx pool is also associated with a DMA region that is shared
43   * with the "hardware" (VIOS) and used to send/receive packets. The DMA
44   * region is also referred to as a Long Term Buffer or LTB.
45   *
46   * The size of the DMA region required for an Rx/Tx pool depends on the
47   * number and size (MTU) of the buffers in the pool. At the max levels
48   * of 4096 jumbo frames (MTU=9000) we will need about 9K*4K = 36MB plus
49   * some padding.
50   *
51   * But the size of a single DMA region is limited by MAX_ORDER in the
52   * kernel (about 16MB currently).  To support say 4K Jumbo frames, we
53   * use a set of LTBs (struct ltb_set) per pool.
54   *
55   * IBMVNIC_ONE_LTB_MAX  - max size of each LTB supported by kernel
56   * IBMVNIC_ONE_LTB_SIZE - current max size of each LTB in an ltb_set
57   * (must be <= IBMVNIC_ONE_LTB_MAX)
58   * IBMVNIC_LTB_SET_SIZE - current size of all LTBs in an ltb_set
59   *
60   * Each VNIC can have upto 16 Rx, 16 Tx and 16 TSO pools. The TSO pools
61   * are of fixed length (IBMVNIC_TSO_BUF_SZ * IBMVNIC_TSO_BUFS) of 4MB.
62   *
63   * The Rx and Tx pools can have upto 4096 buffers. The max size of these
64   * buffers is about 9588 (for jumbo frames, including IBMVNIC_BUFFER_HLEN).
65   * So, setting the IBMVNIC_LTB_SET_SIZE for a pool to 4096 * 9588 ~= 38MB.
66   *
67   * There is a trade-off in setting IBMVNIC_ONE_LTB_SIZE. If it is large,
68   * the allocation of the LTB can fail when system is low in memory. If
69   * its too small, we would need several mappings for each of the Rx/
70   * Tx/TSO pools but there is a limit of 255 mappings per vnic in the
71   * VNIC protocol.
72   *
73   * So setting IBMVNIC_ONE_LTB_SIZE to 8MB. With IBMVNIC_LTB_SET_SIZE set
74   * to 38MB, we will need 5 LTBs per Rx and Tx pool and 1 LTB per TSO
75   * pool for the 4MB. Thus the 16 Rx and Tx queues require 32 * 5 = 160
76   * plus 16 for the TSO pools for a total of 176 LTB mappings per VNIC.
77   */
78  #define IBMVNIC_ONE_LTB_MAX	((u32)((1 << MAX_ORDER) * PAGE_SIZE))
79  #define IBMVNIC_ONE_LTB_SIZE	min((u32)(8 << 20), IBMVNIC_ONE_LTB_MAX)
80  #define IBMVNIC_LTB_SET_SIZE	(38 << 20)
81  
82  #define IBMVNIC_BUFFER_HLEN		500
83  #define IBMVNIC_RESET_DELAY 100
84  
85  struct ibmvnic_login_buffer {
86  	__be32 len;
87  	__be32 version;
88  #define INITIAL_VERSION_LB 1
89  	__be32 num_txcomp_subcrqs;
90  	__be32 off_txcomp_subcrqs;
91  	__be32 num_rxcomp_subcrqs;
92  	__be32 off_rxcomp_subcrqs;
93  	__be32 login_rsp_ioba;
94  	__be32 login_rsp_len;
95  	__be32 client_data_offset;
96  	__be32 client_data_len;
97  } __packed __aligned(8);
98  
99  struct ibmvnic_login_rsp_buffer {
100  	__be32 len;
101  	__be32 version;
102  #define INITIAL_VERSION_LRB 1
103  	__be32 num_txsubm_subcrqs;
104  	__be32 off_txsubm_subcrqs;
105  	__be32 num_rxadd_subcrqs;
106  	__be32 off_rxadd_subcrqs;
107  	__be32 off_rxadd_buff_size;
108  	__be32 num_supp_tx_desc;
109  	__be32 off_supp_tx_desc;
110  } __packed __aligned(8);
111  
112  struct ibmvnic_query_ip_offload_buffer {
113  	__be32 len;
114  	__be32 version;
115  #define INITIAL_VERSION_IOB 1
116  	u8 ipv4_chksum;
117  	u8 ipv6_chksum;
118  	u8 tcp_ipv4_chksum;
119  	u8 tcp_ipv6_chksum;
120  	u8 udp_ipv4_chksum;
121  	u8 udp_ipv6_chksum;
122  	u8 large_tx_ipv4;
123  	u8 large_tx_ipv6;
124  	u8 large_rx_ipv4;
125  	u8 large_rx_ipv6;
126  	u8 reserved1[14];
127  	__be16 max_ipv4_header_size;
128  	__be16 max_ipv6_header_size;
129  	__be16 max_tcp_header_size;
130  	__be16 max_udp_header_size;
131  	__be32 max_large_tx_size;
132  	__be32 max_large_rx_size;
133  	u8 reserved2[16];
134  	u8 ipv6_extension_header;
135  #define IPV6_EH_NOT_SUPPORTED	0x00
136  #define IPV6_EH_SUPPORTED_LIM	0x01
137  #define IPV6_EH_SUPPORTED	0xFF
138  	u8 tcp_pseudosum_req;
139  #define TCP_PS_NOT_REQUIRED	0x00
140  #define TCP_PS_REQUIRED		0x01
141  	u8 reserved3[30];
142  	__be16 num_ipv6_ext_headers;
143  	__be32 off_ipv6_ext_headers;
144  	u8 reserved4[154];
145  } __packed __aligned(8);
146  
147  struct ibmvnic_control_ip_offload_buffer {
148  	__be32 len;
149  	__be32 version;
150  #define INITIAL_VERSION_IOB 1
151  	u8 ipv4_chksum;
152  	u8 ipv6_chksum;
153  	u8 tcp_ipv4_chksum;
154  	u8 tcp_ipv6_chksum;
155  	u8 udp_ipv4_chksum;
156  	u8 udp_ipv6_chksum;
157  	u8 large_tx_ipv4;
158  	u8 large_tx_ipv6;
159  	u8 bad_packet_rx;
160  	u8 large_rx_ipv4;
161  	u8 large_rx_ipv6;
162  	u8 reserved4[111];
163  } __packed __aligned(8);
164  
165  struct ibmvnic_fw_component {
166  	u8 name[48];
167  	__be32 trace_buff_size;
168  	u8 correlator;
169  	u8 trace_level;
170  	u8 parent_correlator;
171  	u8 error_check_level;
172  	u8 trace_on;
173  	u8 reserved[7];
174  	u8 description[192];
175  } __packed __aligned(8);
176  
177  struct ibmvnic_fw_trace_entry {
178  	__be32 trace_id;
179  	u8 num_valid_data;
180  	u8 reserved[3];
181  	__be64 pmc_registers;
182  	__be64 timebase;
183  	__be64 trace_data[5];
184  } __packed __aligned(8);
185  
186  struct ibmvnic_statistics {
187  	__be32 version;
188  	__be32 promiscuous;
189  	__be64 rx_packets;
190  	__be64 rx_bytes;
191  	__be64 tx_packets;
192  	__be64 tx_bytes;
193  	__be64 ucast_tx_packets;
194  	__be64 ucast_rx_packets;
195  	__be64 mcast_tx_packets;
196  	__be64 mcast_rx_packets;
197  	__be64 bcast_tx_packets;
198  	__be64 bcast_rx_packets;
199  	__be64 align_errors;
200  	__be64 fcs_errors;
201  	__be64 single_collision_frames;
202  	__be64 multi_collision_frames;
203  	__be64 sqe_test_errors;
204  	__be64 deferred_tx;
205  	__be64 late_collisions;
206  	__be64 excess_collisions;
207  	__be64 internal_mac_tx_errors;
208  	__be64 carrier_sense;
209  	__be64 too_long_frames;
210  	__be64 internal_mac_rx_errors;
211  	u8 reserved[72];
212  } __packed __aligned(8);
213  
214  #define NUM_TX_STATS 3
215  struct ibmvnic_tx_queue_stats {
216  	u64 batched_packets;
217  	u64 direct_packets;
218  	u64 bytes;
219  	u64 dropped_packets;
220  };
221  
222  #define NUM_RX_STATS 3
223  struct ibmvnic_rx_queue_stats {
224  	u64 packets;
225  	u64 bytes;
226  	u64 interrupts;
227  };
228  
229  struct ibmvnic_acl_buffer {
230  	__be32 len;
231  	__be32 version;
232  #define INITIAL_VERSION_IOB 1
233  	u8 mac_acls_restrict;
234  	u8 vlan_acls_restrict;
235  	u8 reserved1[22];
236  	__be32 num_mac_addrs;
237  	__be32 offset_mac_addrs;
238  	__be32 num_vlan_ids;
239  	__be32 offset_vlan_ids;
240  	u8 reserved2[80];
241  } __packed __aligned(8);
242  
243  /* descriptors have been changed, how should this be defined?  1? 4? */
244  
245  #define IBMVNIC_TX_DESC_VERSIONS 3
246  
247  /* is this still needed? */
248  struct ibmvnic_tx_comp_desc {
249  	u8 first;
250  	u8 num_comps;
251  	__be16 rcs[5];
252  	__be32 correlators[5];
253  } __packed __aligned(8);
254  
255  /* some flags that included in v0 descriptor, which is gone
256   * only used for IBMVNIC_TCP_CHKSUM and IBMVNIC_UDP_CHKSUM
257   * and only in some offload_flags variable that doesn't seem
258   * to be used anywhere, can probably be removed?
259   */
260  
261  #define IBMVNIC_TCP_CHKSUM		0x20
262  #define IBMVNIC_UDP_CHKSUM		0x08
263  
264  struct ibmvnic_tx_desc {
265  	u8 first;
266  	u8 type;
267  
268  #define IBMVNIC_TX_DESC 0x10
269  	u8 n_crq_elem;
270  	u8 n_sge;
271  	u8 flags1;
272  #define IBMVNIC_TX_COMP_NEEDED		0x80
273  #define IBMVNIC_TX_CHKSUM_OFFLOAD	0x40
274  #define IBMVNIC_TX_LSO			0x20
275  #define IBMVNIC_TX_PROT_TCP		0x10
276  #define IBMVNIC_TX_PROT_UDP		0x08
277  #define IBMVNIC_TX_PROT_IPV4		0x04
278  #define IBMVNIC_TX_PROT_IPV6		0x02
279  #define IBMVNIC_TX_VLAN_PRESENT		0x01
280  	u8 flags2;
281  #define IBMVNIC_TX_VLAN_INSERT		0x80
282  	__be16 mss;
283  	u8 reserved[4];
284  	__be32 correlator;
285  	__be16 vlan_id;
286  	__be16 dma_reg;
287  	__be32 sge_len;
288  	__be64 ioba;
289  } __packed __aligned(8);
290  
291  struct ibmvnic_hdr_desc {
292  	u8 first;
293  	u8 type;
294  #define IBMVNIC_HDR_DESC		0x11
295  	u8 len;
296  	u8 l2_len;
297  	__be16 l3_len;
298  	u8 l4_len;
299  	u8 flag;
300  	u8 data[24];
301  } __packed __aligned(8);
302  
303  struct ibmvnic_hdr_ext_desc {
304  	u8 first;
305  	u8 type;
306  #define IBMVNIC_HDR_EXT_DESC		0x12
307  	u8 len;
308  	u8 data[29];
309  } __packed __aligned(8);
310  
311  struct ibmvnic_sge_desc {
312  	u8 first;
313  	u8 type;
314  #define IBMVNIC_SGE_DESC		0x30
315  	__be16 sge1_dma_reg;
316  	__be32 sge1_len;
317  	__be64 sge1_ioba;
318  	__be16 reserved;
319  	__be16 sge2_dma_reg;
320  	__be32 sge2_len;
321  	__be64 sge2_ioba;
322  } __packed __aligned(8);
323  
324  struct ibmvnic_rx_comp_desc {
325  	u8 first;
326  	u8 flags;
327  #define IBMVNIC_IP_CHKSUM_GOOD		0x80
328  #define IBMVNIC_TCP_UDP_CHKSUM_GOOD	0x40
329  #define IBMVNIC_END_FRAME			0x20
330  #define IBMVNIC_EXACT_MC			0x10
331  #define IBMVNIC_VLAN_STRIPPED			0x08
332  	__be16 off_frame_data;
333  	__be32 len;
334  	__be64 correlator;
335  	__be16 vlan_tci;
336  	__be16 rc;
337  	u8 reserved[12];
338  } __packed __aligned(8);
339  
340  struct ibmvnic_generic_scrq {
341  	u8 first;
342  	u8 reserved[31];
343  } __packed __aligned(8);
344  
345  struct ibmvnic_rx_buff_add_desc {
346  	u8 first;
347  	u8 reserved[7];
348  	__be64 correlator;
349  	__be32 ioba;
350  	u8 map_id;
351  	__be32 len:24;
352  	u8 reserved2[8];
353  } __packed __aligned(8);
354  
355  struct ibmvnic_rc {
356  	u8 code; /* one of enum ibmvnic_rc_codes */
357  	u8 detailed_data[3];
358  } __packed __aligned(4);
359  
360  struct ibmvnic_generic_crq {
361  	u8 first;
362  	u8 cmd;
363  	u8 params[10];
364  	struct ibmvnic_rc rc;
365  } __packed __aligned(8);
366  
367  struct ibmvnic_version_exchange {
368  	u8 first;
369  	u8 cmd;
370  	__be16 version;
371  #define IBMVNIC_INITIAL_VERSION 1
372  	u8 reserved[8];
373  	struct ibmvnic_rc rc;
374  } __packed __aligned(8);
375  
376  struct ibmvnic_capability {
377  	u8 first;
378  	u8 cmd;
379  	__be16 capability; /* one of ibmvnic_capabilities */
380  	__be64 number;
381  	struct ibmvnic_rc rc;
382  } __packed __aligned(8);
383  
384  struct ibmvnic_login {
385  	u8 first;
386  	u8 cmd;
387  	u8 reserved[6];
388  	__be32 ioba;
389  	__be32 len;
390  } __packed __aligned(8);
391  
392  struct ibmvnic_phys_parms {
393  	u8 first;
394  	u8 cmd;
395  	u8 flags1;
396  #define IBMVNIC_EXTERNAL_LOOPBACK	0x80
397  #define IBMVNIC_INTERNAL_LOOPBACK	0x40
398  #define IBMVNIC_PROMISC		0x20
399  #define IBMVNIC_PHYS_LINK_ACTIVE	0x10
400  #define IBMVNIC_AUTONEG_DUPLEX	0x08
401  #define IBMVNIC_FULL_DUPLEX	0x04
402  #define IBMVNIC_HALF_DUPLEX	0x02
403  #define IBMVNIC_CAN_CHG_PHYS_PARMS	0x01
404  	u8 flags2;
405  #define IBMVNIC_LOGICAL_LNK_ACTIVE 0x80
406  	__be32 speed;
407  #define IBMVNIC_AUTONEG		0x80000000
408  #define IBMVNIC_10MBPS		0x40000000
409  #define IBMVNIC_100MBPS		0x20000000
410  #define IBMVNIC_1GBPS		0x10000000
411  #define IBMVNIC_10GBPS		0x08000000
412  #define IBMVNIC_40GBPS		0x04000000
413  #define IBMVNIC_100GBPS		0x02000000
414  #define IBMVNIC_25GBPS		0x01000000
415  #define IBMVNIC_50GBPS		0x00800000
416  #define IBMVNIC_200GBPS		0x00400000
417  	__be32 mtu;
418  	struct ibmvnic_rc rc;
419  } __packed __aligned(8);
420  
421  struct ibmvnic_logical_link_state {
422  	u8 first;
423  	u8 cmd;
424  	u8 link_state;
425  #define IBMVNIC_LOGICAL_LNK_DN 0x00
426  #define IBMVNIC_LOGICAL_LNK_UP 0x01
427  #define IBMVNIC_LOGICAL_LNK_QUERY 0xff
428  	u8 reserved[9];
429  	struct ibmvnic_rc rc;
430  } __packed __aligned(8);
431  
432  struct ibmvnic_query_ip_offload {
433  	u8 first;
434  	u8 cmd;
435  	u8 reserved[2];
436  	__be32 len;
437  	__be32 ioba;
438  	struct ibmvnic_rc rc;
439  } __packed __aligned(8);
440  
441  struct ibmvnic_control_ip_offload {
442  	u8 first;
443  	u8 cmd;
444  	u8 reserved[2];
445  	__be32 ioba;
446  	__be32 len;
447  	struct ibmvnic_rc rc;
448  } __packed __aligned(8);
449  
450  struct ibmvnic_request_statistics {
451  	u8 first;
452  	u8 cmd;
453  	u8 flags;
454  #define IBMVNIC_PHYSICAL_PORT	0x80
455  	u8 reserved1;
456  	__be32 ioba;
457  	__be32 len;
458  	u8 reserved[4];
459  } __packed __aligned(8);
460  
461  struct ibmvnic_error_indication {
462  	u8 first;
463  	u8 cmd;
464  	u8 flags;
465  #define IBMVNIC_FATAL_ERROR	0x80
466  	u8 reserved1;
467  	__be32 error_id;
468  	__be32 detail_error_sz;
469  	__be16 error_cause;
470  	u8 reserved2[2];
471  } __packed __aligned(8);
472  
473  struct ibmvnic_link_state_indication {
474  	u8 first;
475  	u8 cmd;
476  	u8 reserved1[2];
477  	u8 phys_link_state;
478  	u8 logical_link_state;
479  	u8 reserved2[10];
480  } __packed __aligned(8);
481  
482  struct ibmvnic_change_mac_addr {
483  	u8 first;
484  	u8 cmd;
485  	u8 mac_addr[6];
486  	u8 reserved[4];
487  	struct ibmvnic_rc rc;
488  } __packed __aligned(8);
489  
490  struct ibmvnic_multicast_ctrl {
491  	u8 first;
492  	u8 cmd;
493  	u8 mac_addr[6];
494  	u8 flags;
495  #define IBMVNIC_ENABLE_MC		0x80
496  #define IBMVNIC_DISABLE_MC		0x40
497  #define IBMVNIC_ENABLE_ALL		0x20
498  #define IBMVNIC_DISABLE_ALL	0x10
499  	u8 reserved1;
500  	__be16 reserved2; /* was num_enabled_mc_addr; */
501  	struct ibmvnic_rc rc;
502  } __packed __aligned(8);
503  
504  struct ibmvnic_get_vpd_size {
505  	u8 first;
506  	u8 cmd;
507  	u8 reserved[14];
508  } __packed __aligned(8);
509  
510  struct ibmvnic_get_vpd_size_rsp {
511  	u8 first;
512  	u8 cmd;
513  	u8 reserved[2];
514  	__be64 len;
515  	struct ibmvnic_rc rc;
516  } __packed __aligned(8);
517  
518  struct ibmvnic_get_vpd {
519  	u8 first;
520  	u8 cmd;
521  	u8 reserved1[2];
522  	__be32 ioba;
523  	__be32 len;
524  	u8 reserved[4];
525  } __packed __aligned(8);
526  
527  struct ibmvnic_get_vpd_rsp {
528  	u8 first;
529  	u8 cmd;
530  	u8 reserved[10];
531  	struct ibmvnic_rc rc;
532  } __packed __aligned(8);
533  
534  struct ibmvnic_acl_change_indication {
535  	u8 first;
536  	u8 cmd;
537  	__be16 change_type;
538  #define IBMVNIC_MAC_ACL 0
539  #define IBMVNIC_VLAN_ACL 1
540  	u8 reserved[12];
541  } __packed __aligned(8);
542  
543  struct ibmvnic_acl_query {
544  	u8 first;
545  	u8 cmd;
546  	u8 reserved1[2];
547  	__be32 ioba;
548  	__be32 len;
549  	u8 reserved2[4];
550  } __packed __aligned(8);
551  
552  struct ibmvnic_tune {
553  	u8 first;
554  	u8 cmd;
555  	u8 reserved1[2];
556  	__be32 ioba;
557  	__be32 len;
558  	u8 reserved2[4];
559  } __packed __aligned(8);
560  
561  struct ibmvnic_request_map {
562  	u8 first;
563  	u8 cmd;
564  	u8 reserved1;
565  	u8 map_id;
566  	__be32 ioba;
567  	__be32 len;
568  	u8 reserved2[4];
569  } __packed __aligned(8);
570  
571  struct ibmvnic_request_map_rsp {
572  	u8 first;
573  	u8 cmd;
574  	u8 reserved1;
575  	u8 map_id;
576  	u8 reserved2[8];
577  	struct ibmvnic_rc rc;
578  } __packed __aligned(8);
579  
580  struct ibmvnic_request_unmap {
581  	u8 first;
582  	u8 cmd;
583  	u8 reserved1;
584  	u8 map_id;
585  	u8 reserved2[12];
586  } __packed __aligned(8);
587  
588  struct ibmvnic_request_unmap_rsp {
589  	u8 first;
590  	u8 cmd;
591  	u8 reserved1;
592  	u8 map_id;
593  	u8 reserved2[8];
594  	struct ibmvnic_rc rc;
595  } __packed __aligned(8);
596  
597  struct ibmvnic_query_map {
598  	u8 first;
599  	u8 cmd;
600  	u8 reserved[14];
601  } __packed __aligned(8);
602  
603  struct ibmvnic_query_map_rsp {
604  	u8 first;
605  	u8 cmd;
606  	u8 reserved;
607  	u8 page_size;
608  	__be32 tot_pages;
609  	__be32 free_pages;
610  	struct ibmvnic_rc rc;
611  } __packed __aligned(8);
612  
613  union ibmvnic_crq {
614  	struct ibmvnic_generic_crq generic;
615  	struct ibmvnic_version_exchange version_exchange;
616  	struct ibmvnic_version_exchange version_exchange_rsp;
617  	struct ibmvnic_capability query_capability;
618  	struct ibmvnic_capability query_capability_rsp;
619  	struct ibmvnic_capability request_capability;
620  	struct ibmvnic_capability request_capability_rsp;
621  	struct ibmvnic_login login;
622  	struct ibmvnic_generic_crq login_rsp;
623  	struct ibmvnic_phys_parms query_phys_parms;
624  	struct ibmvnic_phys_parms query_phys_parms_rsp;
625  	struct ibmvnic_phys_parms query_phys_capabilities;
626  	struct ibmvnic_phys_parms query_phys_capabilities_rsp;
627  	struct ibmvnic_phys_parms set_phys_parms;
628  	struct ibmvnic_phys_parms set_phys_parms_rsp;
629  	struct ibmvnic_logical_link_state logical_link_state;
630  	struct ibmvnic_logical_link_state logical_link_state_rsp;
631  	struct ibmvnic_query_ip_offload query_ip_offload;
632  	struct ibmvnic_query_ip_offload query_ip_offload_rsp;
633  	struct ibmvnic_control_ip_offload control_ip_offload;
634  	struct ibmvnic_control_ip_offload control_ip_offload_rsp;
635  	struct ibmvnic_request_statistics request_statistics;
636  	struct ibmvnic_generic_crq request_statistics_rsp;
637  	struct ibmvnic_error_indication error_indication;
638  	struct ibmvnic_link_state_indication link_state_indication;
639  	struct ibmvnic_change_mac_addr change_mac_addr;
640  	struct ibmvnic_change_mac_addr change_mac_addr_rsp;
641  	struct ibmvnic_multicast_ctrl multicast_ctrl;
642  	struct ibmvnic_multicast_ctrl multicast_ctrl_rsp;
643  	struct ibmvnic_get_vpd_size get_vpd_size;
644  	struct ibmvnic_get_vpd_size_rsp get_vpd_size_rsp;
645  	struct ibmvnic_get_vpd get_vpd;
646  	struct ibmvnic_get_vpd_rsp get_vpd_rsp;
647  	struct ibmvnic_acl_change_indication acl_change_indication;
648  	struct ibmvnic_acl_query acl_query;
649  	struct ibmvnic_generic_crq acl_query_rsp;
650  	struct ibmvnic_tune tune;
651  	struct ibmvnic_generic_crq tune_rsp;
652  	struct ibmvnic_request_map request_map;
653  	struct ibmvnic_request_map_rsp request_map_rsp;
654  	struct ibmvnic_request_unmap request_unmap;
655  	struct ibmvnic_request_unmap_rsp request_unmap_rsp;
656  	struct ibmvnic_query_map query_map;
657  	struct ibmvnic_query_map_rsp query_map_rsp;
658  };
659  
660  enum ibmvnic_rc_codes {
661  	SUCCESS = 0,
662  	PARTIALSUCCESS = 1,
663  	PERMISSION = 2,
664  	NOMEMORY = 3,
665  	PARAMETER = 4,
666  	UNKNOWNCOMMAND = 5,
667  	ABORTED = 6,
668  	INVALIDSTATE = 7,
669  	INVALIDIOBA = 8,
670  	INVALIDLENGTH = 9,
671  	UNSUPPORTEDOPTION = 10,
672  };
673  
674  enum ibmvnic_capabilities {
675  	MIN_TX_QUEUES = 1,
676  	MIN_RX_QUEUES = 2,
677  	MIN_RX_ADD_QUEUES = 3,
678  	MAX_TX_QUEUES = 4,
679  	MAX_RX_QUEUES = 5,
680  	MAX_RX_ADD_QUEUES = 6,
681  	REQ_TX_QUEUES = 7,
682  	REQ_RX_QUEUES = 8,
683  	REQ_RX_ADD_QUEUES = 9,
684  	MIN_TX_ENTRIES_PER_SUBCRQ = 10,
685  	MIN_RX_ADD_ENTRIES_PER_SUBCRQ = 11,
686  	MAX_TX_ENTRIES_PER_SUBCRQ = 12,
687  	MAX_RX_ADD_ENTRIES_PER_SUBCRQ = 13,
688  	REQ_TX_ENTRIES_PER_SUBCRQ = 14,
689  	REQ_RX_ADD_ENTRIES_PER_SUBCRQ = 15,
690  	TCP_IP_OFFLOAD = 16,
691  	PROMISC_REQUESTED = 17,
692  	PROMISC_SUPPORTED = 18,
693  	MIN_MTU = 19,
694  	MAX_MTU = 20,
695  	REQ_MTU = 21,
696  	MAX_MULTICAST_FILTERS = 22,
697  	VLAN_HEADER_INSERTION = 23,
698  	RX_VLAN_HEADER_INSERTION = 24,
699  	MAX_TX_SG_ENTRIES = 25,
700  	RX_SG_SUPPORTED = 26,
701  	RX_SG_REQUESTED = 27,
702  	OPT_TX_COMP_SUB_QUEUES = 28,
703  	OPT_RX_COMP_QUEUES = 29,
704  	OPT_RX_BUFADD_Q_PER_RX_COMP_Q = 30,
705  	OPT_TX_ENTRIES_PER_SUBCRQ = 31,
706  	OPT_RXBA_ENTRIES_PER_SUBCRQ = 32,
707  	TX_RX_DESC_REQ = 33,
708  };
709  
710  enum ibmvnic_error_cause {
711  	ADAPTER_PROBLEM = 0,
712  	BUS_PROBLEM = 1,
713  	FW_PROBLEM = 2,
714  	DD_PROBLEM = 3,
715  	EEH_RECOVERY = 4,
716  	FW_UPDATED = 5,
717  	LOW_MEMORY = 6,
718  };
719  
720  enum ibmvnic_commands {
721  	VERSION_EXCHANGE = 0x01,
722  	VERSION_EXCHANGE_RSP = 0x81,
723  	QUERY_CAPABILITY = 0x02,
724  	QUERY_CAPABILITY_RSP = 0x82,
725  	REQUEST_CAPABILITY = 0x03,
726  	REQUEST_CAPABILITY_RSP = 0x83,
727  	LOGIN = 0x04,
728  	LOGIN_RSP = 0x84,
729  	QUERY_PHYS_PARMS = 0x05,
730  	QUERY_PHYS_PARMS_RSP = 0x85,
731  	QUERY_PHYS_CAPABILITIES = 0x06,
732  	QUERY_PHYS_CAPABILITIES_RSP = 0x86,
733  	SET_PHYS_PARMS = 0x07,
734  	SET_PHYS_PARMS_RSP = 0x87,
735  	ERROR_INDICATION = 0x08,
736  	LOGICAL_LINK_STATE = 0x0C,
737  	LOGICAL_LINK_STATE_RSP = 0x8C,
738  	REQUEST_STATISTICS = 0x0D,
739  	REQUEST_STATISTICS_RSP = 0x8D,
740  	COLLECT_FW_TRACE = 0x11,
741  	COLLECT_FW_TRACE_RSP = 0x91,
742  	LINK_STATE_INDICATION = 0x12,
743  	CHANGE_MAC_ADDR = 0x13,
744  	CHANGE_MAC_ADDR_RSP = 0x93,
745  	MULTICAST_CTRL = 0x14,
746  	MULTICAST_CTRL_RSP = 0x94,
747  	GET_VPD_SIZE = 0x15,
748  	GET_VPD_SIZE_RSP = 0x95,
749  	GET_VPD = 0x16,
750  	GET_VPD_RSP = 0x96,
751  	TUNE = 0x17,
752  	TUNE_RSP = 0x97,
753  	QUERY_IP_OFFLOAD = 0x18,
754  	QUERY_IP_OFFLOAD_RSP = 0x98,
755  	CONTROL_IP_OFFLOAD = 0x19,
756  	CONTROL_IP_OFFLOAD_RSP = 0x99,
757  	ACL_CHANGE_INDICATION = 0x1A,
758  	ACL_QUERY = 0x1B,
759  	ACL_QUERY_RSP = 0x9B,
760  	QUERY_MAP = 0x1D,
761  	QUERY_MAP_RSP = 0x9D,
762  	REQUEST_MAP = 0x1E,
763  	REQUEST_MAP_RSP = 0x9E,
764  	REQUEST_UNMAP = 0x1F,
765  	REQUEST_UNMAP_RSP = 0x9F,
766  	VLAN_CTRL = 0x20,
767  	VLAN_CTRL_RSP = 0xA0,
768  };
769  
770  enum ibmvnic_crq_type {
771  	IBMVNIC_CRQ_CMD			= 0x80,
772  	IBMVNIC_CRQ_CMD_RSP		= 0x80,
773  	IBMVNIC_CRQ_INIT_CMD		= 0xC0,
774  	IBMVNIC_CRQ_INIT_RSP		= 0xC0,
775  	IBMVNIC_CRQ_XPORT_EVENT		= 0xFF,
776  };
777  
778  enum ibmvfc_crq_format {
779  	IBMVNIC_CRQ_INIT                 = 0x01,
780  	IBMVNIC_CRQ_INIT_COMPLETE        = 0x02,
781  	IBMVNIC_PARTITION_MIGRATED       = 0x06,
782  	IBMVNIC_DEVICE_FAILOVER          = 0x08,
783  };
784  
785  struct ibmvnic_crq_queue {
786  	union ibmvnic_crq *msgs;
787  	int size, cur;
788  	dma_addr_t msg_token;
789  	/* Used for serialization of msgs, cur */
790  	spinlock_t lock;
791  	bool active;
792  	char name[32];
793  };
794  
795  union sub_crq {
796  	struct ibmvnic_generic_scrq generic;
797  	struct ibmvnic_tx_comp_desc tx_comp;
798  	struct ibmvnic_tx_desc v1;
799  	struct ibmvnic_hdr_desc hdr;
800  	struct ibmvnic_hdr_ext_desc hdr_ext;
801  	struct ibmvnic_sge_desc sge;
802  	struct ibmvnic_rx_comp_desc rx_comp;
803  	struct ibmvnic_rx_buff_add_desc rx_add;
804  };
805  
806  struct ibmvnic_ind_xmit_queue {
807  	union sub_crq *indir_arr;
808  	dma_addr_t indir_dma;
809  	int index;
810  };
811  
812  struct ibmvnic_sub_crq_queue {
813  	union sub_crq *msgs;
814  	int size, cur;
815  	dma_addr_t msg_token;
816  	unsigned long crq_num;
817  	unsigned long hw_irq;
818  	unsigned int irq;
819  	unsigned int pool_index;
820  	int scrq_num;
821  	/* Used for serialization of msgs, cur */
822  	spinlock_t lock;
823  	struct sk_buff *rx_skb_top;
824  	struct ibmvnic_adapter *adapter;
825  	struct ibmvnic_ind_xmit_queue ind_buf;
826  	atomic_t used;
827  	char name[32];
828  	u64 handle;
829  	cpumask_var_t affinity_mask;
830  } ____cacheline_aligned;
831  
832  struct ibmvnic_long_term_buff {
833  	unsigned char *buff;
834  	dma_addr_t addr;
835  	u64 size;
836  	u8 map_id;
837  };
838  
839  struct ibmvnic_ltb_set {
840  	int num_ltbs;
841  	struct ibmvnic_long_term_buff *ltbs;
842  };
843  
844  struct ibmvnic_tx_buff {
845  	struct sk_buff *skb;
846  	int index;
847  	int pool_index;
848  	int num_entries;
849  };
850  
851  struct ibmvnic_tx_pool {
852  	struct ibmvnic_tx_buff *tx_buff;
853  	int *free_map;
854  	int consumer_index;
855  	int producer_index;
856  	struct ibmvnic_ltb_set ltb_set;
857  	int num_buffers;
858  	int buf_size;
859  } ____cacheline_aligned;
860  
861  struct ibmvnic_rx_buff {
862  	struct sk_buff *skb;
863  	dma_addr_t dma;
864  	unsigned char *data;
865  	int size;
866  	int pool_index;
867  };
868  
869  struct ibmvnic_rx_pool {
870  	struct ibmvnic_rx_buff *rx_buff;
871  	int size;			/* # of buffers in the pool */
872  	int index;
873  	int buff_size;
874  	atomic_t available;
875  	int *free_map;
876  	int next_free;
877  	int next_alloc;
878  	int active;
879  	struct ibmvnic_ltb_set ltb_set;
880  } ____cacheline_aligned;
881  
882  struct ibmvnic_vpd {
883  	unsigned char *buff;
884  	dma_addr_t dma_addr;
885  	u64 len;
886  };
887  
888  enum vnic_state {VNIC_PROBING = 1,
889  		 VNIC_PROBED,
890  		 VNIC_OPENING,
891  		 VNIC_OPEN,
892  		 VNIC_CLOSING,
893  		 VNIC_CLOSED,
894  		 VNIC_REMOVING,
895  		 VNIC_REMOVED,
896  		 VNIC_DOWN};
897  
898  enum ibmvnic_reset_reason {VNIC_RESET_FAILOVER = 1,
899  			   VNIC_RESET_MOBILITY,
900  			   VNIC_RESET_FATAL,
901  			   VNIC_RESET_NON_FATAL,
902  			   VNIC_RESET_TIMEOUT,
903  			   VNIC_RESET_CHANGE_PARAM,
904  			   VNIC_RESET_PASSIVE_INIT};
905  
906  struct ibmvnic_rwi {
907  	enum ibmvnic_reset_reason reset_reason;
908  	struct list_head list;
909  };
910  
911  struct ibmvnic_tunables {
912  	u64 rx_queues;
913  	u64 tx_queues;
914  	u64 rx_entries;
915  	u64 tx_entries;
916  	u64 mtu;
917  };
918  
919  struct ibmvnic_adapter {
920  	struct vio_dev *vdev;
921  	struct net_device *netdev;
922  	struct ibmvnic_crq_queue crq;
923  	u8 mac_addr[ETH_ALEN];
924  	struct ibmvnic_query_ip_offload_buffer ip_offload_buf;
925  	dma_addr_t ip_offload_tok;
926  	struct ibmvnic_control_ip_offload_buffer ip_offload_ctrl;
927  	dma_addr_t ip_offload_ctrl_tok;
928  	u32 msg_enable;
929  
930  	/* Vital Product Data (VPD) */
931  	struct ibmvnic_vpd *vpd;
932  	char fw_version[32];
933  
934  	/* Statistics */
935  	struct ibmvnic_statistics stats;
936  	dma_addr_t stats_token;
937  	struct completion stats_done;
938  	int replenish_no_mem;
939  	int replenish_add_buff_success;
940  	int replenish_add_buff_failure;
941  	int replenish_task_cycles;
942  	int tx_send_failed;
943  	int tx_map_failed;
944  
945  	struct ibmvnic_tx_queue_stats *tx_stats_buffers;
946  	struct ibmvnic_rx_queue_stats *rx_stats_buffers;
947  
948  	int phys_link_state;
949  	int logical_link_state;
950  
951  	u32 speed;
952  	u8 duplex;
953  
954  	/* login data */
955  	struct ibmvnic_login_buffer *login_buf;
956  	dma_addr_t login_buf_token;
957  	int login_buf_sz;
958  
959  	struct ibmvnic_login_rsp_buffer *login_rsp_buf;
960  	dma_addr_t login_rsp_buf_token;
961  	int login_rsp_buf_sz;
962  
963  	atomic_t running_cap_crqs;
964  
965  	struct ibmvnic_sub_crq_queue **tx_scrq ____cacheline_aligned;
966  	struct ibmvnic_sub_crq_queue **rx_scrq ____cacheline_aligned;
967  
968  	/* rx structs */
969  	struct napi_struct *napi;
970  	struct ibmvnic_rx_pool *rx_pool;
971  	u64 promisc;
972  
973  	struct ibmvnic_tx_pool *tx_pool;
974  	struct ibmvnic_tx_pool *tso_pool;
975  	struct completion probe_done;
976  	struct completion init_done;
977  	int init_done_rc;
978  
979  	struct completion fw_done;
980  	/* Used for serialization of device commands */
981  	struct mutex fw_lock;
982  	int fw_done_rc;
983  
984  	struct completion reset_done;
985  	int reset_done_rc;
986  	bool wait_for_reset;
987  
988  	/* CPU hotplug instances for online & dead */
989  	struct hlist_node node;
990  	struct hlist_node node_dead;
991  
992  	/* partner capabilities */
993  	u64 min_tx_queues;
994  	u64 min_rx_queues;
995  	u64 min_rx_add_queues;
996  	u64 max_tx_queues;
997  	u64 max_rx_queues;
998  	u64 max_rx_add_queues;
999  	u64 req_tx_queues;
1000  	u64 req_rx_queues;
1001  	u64 req_rx_add_queues;
1002  	u64 min_tx_entries_per_subcrq;
1003  	u64 min_rx_add_entries_per_subcrq;
1004  	u64 max_tx_entries_per_subcrq;
1005  	u64 max_rx_add_entries_per_subcrq;
1006  	u64 req_tx_entries_per_subcrq;
1007  	u64 req_rx_add_entries_per_subcrq;
1008  	u64 tcp_ip_offload;
1009  	u64 promisc_requested;
1010  	u64 promisc_supported;
1011  	u64 min_mtu;
1012  	u64 max_mtu;
1013  	u64 req_mtu;
1014  	u64 prev_mtu;
1015  	u64 max_multicast_filters;
1016  	u64 vlan_header_insertion;
1017  	u64 rx_vlan_header_insertion;
1018  	u64 max_tx_sg_entries;
1019  	u64 rx_sg_supported;
1020  	u64 rx_sg_requested;
1021  	u64 opt_tx_comp_sub_queues;
1022  	u64 opt_rx_comp_queues;
1023  	u64 opt_rx_bufadd_q_per_rx_comp_q;
1024  	u64 opt_tx_entries_per_subcrq;
1025  	u64 opt_rxba_entries_per_subcrq;
1026  	__be64 tx_rx_desc_req;
1027  #define MAX_MAP_ID	255
1028  	DECLARE_BITMAP(map_ids, MAX_MAP_ID);
1029  	u32 num_active_rx_scrqs;
1030  	u32 num_active_rx_pools;
1031  	u32 num_active_rx_napi;
1032  	u32 num_active_tx_scrqs;
1033  	u32 num_active_tx_pools;
1034  
1035  	u32 prev_rx_pool_size;
1036  	u32 prev_tx_pool_size;
1037  	u32 cur_rx_buf_sz;
1038  	u32 prev_rx_buf_sz;
1039  
1040  	struct tasklet_struct tasklet;
1041  	enum vnic_state state;
1042  	/* Used for serialization of state field. When taking both state
1043  	 * and rwi locks, take state lock first.
1044  	 */
1045  	spinlock_t state_lock;
1046  	enum ibmvnic_reset_reason reset_reason;
1047  	struct list_head rwi_list;
1048  	/* Used for serialization of rwi_list. When taking both state
1049  	 * and rwi locks, take state lock first
1050  	 */
1051  	spinlock_t rwi_lock;
1052  	struct work_struct ibmvnic_reset;
1053  	struct delayed_work ibmvnic_delayed_reset;
1054  	unsigned long resetting;
1055  	/* last device reset time */
1056  	unsigned long last_reset_time;
1057  
1058  	bool napi_enabled;
1059  	bool from_passive_init;
1060  	bool login_pending;
1061  	/* protected by rcu */
1062  	bool tx_queues_active;
1063  	bool failover_pending;
1064  	bool force_reset_recovery;
1065  
1066  	struct ibmvnic_tunables desired;
1067  	struct ibmvnic_tunables fallback;
1068  };
1069