1/* SPDX-License-Identifier: GPL-2.0-or-later */
2/**************************************************************************/
3/* */
4/* IBM System i and System p Virtual NIC Device Driver */
5/* Copyright (C) 2014 IBM Corp. */
6/* Santiago Leon (santi_leon@yahoo.com) */
7/* Thomas Falcon (tlfalcon@linux.vnet.ibm.com) */
8/* John Allen (jallen@linux.vnet.ibm.com) */
9/* */
10/* */
11/* This module contains the implementation of a virtual ethernet device */
12/* for use with IBM i/pSeries LPAR Linux. It utilizes the logical LAN */
13/* option of the RS/6000 Platform Architecture to interface with virtual */
14/* ethernet NICs that are presented to the partition by the hypervisor. */
15/* */
16/**************************************************************************/
17
18#define IBMVNIC_NAME "ibmvnic"
19#define IBMVNIC_DRIVER_VERSION "1.0.1"
20#define IBMVNIC_INVALID_MAP -1
21#define IBMVNIC_OPEN_FAILED 3
22
23/* basic structures plus 100 2k buffers */
24#define IBMVNIC_IO_ENTITLEMENT_DEFAULT 610305
25
26/* Initial module_parameters */
27#define IBMVNIC_RX_WEIGHT 16
28/* when changing this, update IBMVNIC_IO_ENTITLEMENT_DEFAULT */
29#define IBMVNIC_BUFFS_PER_POOL 100
30#define IBMVNIC_MAX_QUEUES 16
31#define IBMVNIC_MAX_QUEUE_SZ 4096
32#define IBMVNIC_MAX_IND_DESCS 16
33#define IBMVNIC_IND_ARR_SZ (IBMVNIC_MAX_IND_DESCS * 32)
34
35#define IBMVNIC_TSO_BUF_SZ 65536
36#define IBMVNIC_TSO_BUFS 64
37#define IBMVNIC_TSO_POOL_MASK 0x80000000
38
39/* A VNIC adapter has set of Rx and Tx pools (aka queues). Each Rx/Tx pool
40 * has a set of buffers. The size of each buffer is determined by the MTU.
41 *
42 * Each Rx/Tx pool is also associated with a DMA region that is shared
43 * with the "hardware" (VIOS) and used to send/receive packets. The DMA
44 * region is also referred to as a Long Term Buffer or LTB.
45 *
46 * The size of the DMA region required for an Rx/Tx pool depends on the
47 * number and size (MTU) of the buffers in the pool. At the max levels
48 * of 4096 jumbo frames (MTU=9000) we will need about 9K*4K = 36MB plus
49 * some padding.
50 *
51 * But the size of a single DMA region is limited by MAX_ORDER in the
52 * kernel (about 16MB currently). To support say 4K Jumbo frames, we
53 * use a set of LTBs (struct ltb_set) per pool.
54 *
55 * IBMVNIC_ONE_LTB_MAX - max size of each LTB supported by kernel
56 * IBMVNIC_ONE_LTB_SIZE - current max size of each LTB in an ltb_set
57 * (must be <= IBMVNIC_ONE_LTB_MAX)
58 * IBMVNIC_LTB_SET_SIZE - current size of all LTBs in an ltb_set
59 *
60 * Each VNIC can have upto 16 Rx, 16 Tx and 16 TSO pools. The TSO pools
61 * are of fixed length (IBMVNIC_TSO_BUF_SZ * IBMVNIC_TSO_BUFS) of 4MB.
62 *
63 * The Rx and Tx pools can have upto 4096 buffers. The max size of these
64 * buffers is about 9588 (for jumbo frames, including IBMVNIC_BUFFER_HLEN).
65 * So, setting the IBMVNIC_LTB_SET_SIZE for a pool to 4096 * 9588 ~= 38MB.
66 *
67 * There is a trade-off in setting IBMVNIC_ONE_LTB_SIZE. If it is large,
68 * the allocation of the LTB can fail when system is low in memory. If
69 * its too small, we would need several mappings for each of the Rx/
70 * Tx/TSO pools but there is a limit of 255 mappings per vnic in the
71 * VNIC protocol.
72 *
73 * So setting IBMVNIC_ONE_LTB_SIZE to 8MB. With IBMVNIC_LTB_SET_SIZE set
74 * to 38MB, we will need 5 LTBs per Rx and Tx pool and 1 LTB per TSO
75 * pool for the 4MB. Thus the 16 Rx and Tx queues require 32 * 5 = 160
76 * plus 16 for the TSO pools for a total of 176 LTB mappings per VNIC.
77 */
78#define IBMVNIC_ONE_LTB_MAX ((u32)((1 << MAX_ORDER) * PAGE_SIZE))
79#define IBMVNIC_ONE_LTB_SIZE min((u32)(8 << 20), IBMVNIC_ONE_LTB_MAX)
80#define IBMVNIC_LTB_SET_SIZE (38 << 20)
81
82#define IBMVNIC_BUFFER_HLEN 500
83#define IBMVNIC_RESET_DELAY 100
84
85struct ibmvnic_login_buffer {
86 __be32 len;
87 __be32 version;
88#define INITIAL_VERSION_LB 1
89 __be32 num_txcomp_subcrqs;
90 __be32 off_txcomp_subcrqs;
91 __be32 num_rxcomp_subcrqs;
92 __be32 off_rxcomp_subcrqs;
93 __be32 login_rsp_ioba;
94 __be32 login_rsp_len;
95 __be32 client_data_offset;
96 __be32 client_data_len;
97} __packed __aligned(8);
98
99struct ibmvnic_login_rsp_buffer {
100 __be32 len;
101 __be32 version;
102#define INITIAL_VERSION_LRB 1
103 __be32 num_txsubm_subcrqs;
104 __be32 off_txsubm_subcrqs;
105 __be32 num_rxadd_subcrqs;
106 __be32 off_rxadd_subcrqs;
107 __be32 off_rxadd_buff_size;
108 __be32 num_supp_tx_desc;
109 __be32 off_supp_tx_desc;
110} __packed __aligned(8);
111
112struct ibmvnic_query_ip_offload_buffer {
113 __be32 len;
114 __be32 version;
115#define INITIAL_VERSION_IOB 1
116 u8 ipv4_chksum;
117 u8 ipv6_chksum;
118 u8 tcp_ipv4_chksum;
119 u8 tcp_ipv6_chksum;
120 u8 udp_ipv4_chksum;
121 u8 udp_ipv6_chksum;
122 u8 large_tx_ipv4;
123 u8 large_tx_ipv6;
124 u8 large_rx_ipv4;
125 u8 large_rx_ipv6;
126 u8 reserved1[14];
127 __be16 max_ipv4_header_size;
128 __be16 max_ipv6_header_size;
129 __be16 max_tcp_header_size;
130 __be16 max_udp_header_size;
131 __be32 max_large_tx_size;
132 __be32 max_large_rx_size;
133 u8 reserved2[16];
134 u8 ipv6_extension_header;
135#define IPV6_EH_NOT_SUPPORTED 0x00
136#define IPV6_EH_SUPPORTED_LIM 0x01
137#define IPV6_EH_SUPPORTED 0xFF
138 u8 tcp_pseudosum_req;
139#define TCP_PS_NOT_REQUIRED 0x00
140#define TCP_PS_REQUIRED 0x01
141 u8 reserved3[30];
142 __be16 num_ipv6_ext_headers;
143 __be32 off_ipv6_ext_headers;
144 u8 reserved4[154];
145} __packed __aligned(8);
146
147struct ibmvnic_control_ip_offload_buffer {
148 __be32 len;
149 __be32 version;
150#define INITIAL_VERSION_IOB 1
151 u8 ipv4_chksum;
152 u8 ipv6_chksum;
153 u8 tcp_ipv4_chksum;
154 u8 tcp_ipv6_chksum;
155 u8 udp_ipv4_chksum;
156 u8 udp_ipv6_chksum;
157 u8 large_tx_ipv4;
158 u8 large_tx_ipv6;
159 u8 bad_packet_rx;
160 u8 large_rx_ipv4;
161 u8 large_rx_ipv6;
162 u8 reserved4[111];
163} __packed __aligned(8);
164
165struct ibmvnic_fw_component {
166 u8 name[48];
167 __be32 trace_buff_size;
168 u8 correlator;
169 u8 trace_level;
170 u8 parent_correlator;
171 u8 error_check_level;
172 u8 trace_on;
173 u8 reserved[7];
174 u8 description[192];
175} __packed __aligned(8);
176
177struct ibmvnic_fw_trace_entry {
178 __be32 trace_id;
179 u8 num_valid_data;
180 u8 reserved[3];
181 __be64 pmc_registers;
182 __be64 timebase;
183 __be64 trace_data[5];
184} __packed __aligned(8);
185
186struct ibmvnic_statistics {
187 __be32 version;
188 __be32 promiscuous;
189 __be64 rx_packets;
190 __be64 rx_bytes;
191 __be64 tx_packets;
192 __be64 tx_bytes;
193 __be64 ucast_tx_packets;
194 __be64 ucast_rx_packets;
195 __be64 mcast_tx_packets;
196 __be64 mcast_rx_packets;
197 __be64 bcast_tx_packets;
198 __be64 bcast_rx_packets;
199 __be64 align_errors;
200 __be64 fcs_errors;
201 __be64 single_collision_frames;
202 __be64 multi_collision_frames;
203 __be64 sqe_test_errors;
204 __be64 deferred_tx;
205 __be64 late_collisions;
206 __be64 excess_collisions;
207 __be64 internal_mac_tx_errors;
208 __be64 carrier_sense;
209 __be64 too_long_frames;
210 __be64 internal_mac_rx_errors;
211 u8 reserved[72];
212} __packed __aligned(8);
213
214#define NUM_TX_STATS 3
215struct ibmvnic_tx_queue_stats {
216 u64 packets;
217 u64 bytes;
218 u64 dropped_packets;
219};
220
221#define NUM_RX_STATS 3
222struct ibmvnic_rx_queue_stats {
223 u64 packets;
224 u64 bytes;
225 u64 interrupts;
226};
227
228struct ibmvnic_acl_buffer {
229 __be32 len;
230 __be32 version;
231#define INITIAL_VERSION_IOB 1
232 u8 mac_acls_restrict;
233 u8 vlan_acls_restrict;
234 u8 reserved1[22];
235 __be32 num_mac_addrs;
236 __be32 offset_mac_addrs;
237 __be32 num_vlan_ids;
238 __be32 offset_vlan_ids;
239 u8 reserved2[80];
240} __packed __aligned(8);
241
242/* descriptors have been changed, how should this be defined? 1? 4? */
243
244#define IBMVNIC_TX_DESC_VERSIONS 3
245
246/* is this still needed? */
247struct ibmvnic_tx_comp_desc {
248 u8 first;
249 u8 num_comps;
250 __be16 rcs[5];
251 __be32 correlators[5];
252} __packed __aligned(8);
253
254/* some flags that included in v0 descriptor, which is gone
255 * only used for IBMVNIC_TCP_CHKSUM and IBMVNIC_UDP_CHKSUM
256 * and only in some offload_flags variable that doesn't seem
257 * to be used anywhere, can probably be removed?
258 */
259
260#define IBMVNIC_TCP_CHKSUM 0x20
261#define IBMVNIC_UDP_CHKSUM 0x08
262
263struct ibmvnic_tx_desc {
264 u8 first;
265 u8 type;
266
267#define IBMVNIC_TX_DESC 0x10
268 u8 n_crq_elem;
269 u8 n_sge;
270 u8 flags1;
271#define IBMVNIC_TX_COMP_NEEDED 0x80
272#define IBMVNIC_TX_CHKSUM_OFFLOAD 0x40
273#define IBMVNIC_TX_LSO 0x20
274#define IBMVNIC_TX_PROT_TCP 0x10
275#define IBMVNIC_TX_PROT_UDP 0x08
276#define IBMVNIC_TX_PROT_IPV4 0x04
277#define IBMVNIC_TX_PROT_IPV6 0x02
278#define IBMVNIC_TX_VLAN_PRESENT 0x01
279 u8 flags2;
280#define IBMVNIC_TX_VLAN_INSERT 0x80
281 __be16 mss;
282 u8 reserved[4];
283 __be32 correlator;
284 __be16 vlan_id;
285 __be16 dma_reg;
286 __be32 sge_len;
287 __be64 ioba;
288} __packed __aligned(8);
289
290struct ibmvnic_hdr_desc {
291 u8 first;
292 u8 type;
293#define IBMVNIC_HDR_DESC 0x11
294 u8 len;
295 u8 l2_len;
296 __be16 l3_len;
297 u8 l4_len;
298 u8 flag;
299 u8 data[24];
300} __packed __aligned(8);
301
302struct ibmvnic_hdr_ext_desc {
303 u8 first;
304 u8 type;
305#define IBMVNIC_HDR_EXT_DESC 0x12
306 u8 len;
307 u8 data[29];
308} __packed __aligned(8);
309
310struct ibmvnic_sge_desc {
311 u8 first;
312 u8 type;
313#define IBMVNIC_SGE_DESC 0x30
314 __be16 sge1_dma_reg;
315 __be32 sge1_len;
316 __be64 sge1_ioba;
317 __be16 reserved;
318 __be16 sge2_dma_reg;
319 __be32 sge2_len;
320 __be64 sge2_ioba;
321} __packed __aligned(8);
322
323struct ibmvnic_rx_comp_desc {
324 u8 first;
325 u8 flags;
326#define IBMVNIC_IP_CHKSUM_GOOD 0x80
327#define IBMVNIC_TCP_UDP_CHKSUM_GOOD 0x40
328#define IBMVNIC_END_FRAME 0x20
329#define IBMVNIC_EXACT_MC 0x10
330#define IBMVNIC_VLAN_STRIPPED 0x08
331 __be16 off_frame_data;
332 __be32 len;
333 __be64 correlator;
334 __be16 vlan_tci;
335 __be16 rc;
336 u8 reserved[12];
337} __packed __aligned(8);
338
339struct ibmvnic_generic_scrq {
340 u8 first;
341 u8 reserved[31];
342} __packed __aligned(8);
343
344struct ibmvnic_rx_buff_add_desc {
345 u8 first;
346 u8 reserved[7];
347 __be64 correlator;
348 __be32 ioba;
349 u8 map_id;
350 __be32 len:24;
351 u8 reserved2[8];
352} __packed __aligned(8);
353
354struct ibmvnic_rc {
355 u8 code; /* one of enum ibmvnic_rc_codes */
356 u8 detailed_data[3];
357} __packed __aligned(4);
358
359struct ibmvnic_generic_crq {
360 u8 first;
361 u8 cmd;
362 u8 params[10];
363 struct ibmvnic_rc rc;
364} __packed __aligned(8);
365
366struct ibmvnic_version_exchange {
367 u8 first;
368 u8 cmd;
369 __be16 version;
370#define IBMVNIC_INITIAL_VERSION 1
371 u8 reserved[8];
372 struct ibmvnic_rc rc;
373} __packed __aligned(8);
374
375struct ibmvnic_capability {
376 u8 first;
377 u8 cmd;
378 __be16 capability; /* one of ibmvnic_capabilities */
379 __be64 number;
380 struct ibmvnic_rc rc;
381} __packed __aligned(8);
382
383struct ibmvnic_login {
384 u8 first;
385 u8 cmd;
386 u8 reserved[6];
387 __be32 ioba;
388 __be32 len;
389} __packed __aligned(8);
390
391struct ibmvnic_phys_parms {
392 u8 first;
393 u8 cmd;
394 u8 flags1;
395#define IBMVNIC_EXTERNAL_LOOPBACK 0x80
396#define IBMVNIC_INTERNAL_LOOPBACK 0x40
397#define IBMVNIC_PROMISC 0x20
398#define IBMVNIC_PHYS_LINK_ACTIVE 0x10
399#define IBMVNIC_AUTONEG_DUPLEX 0x08
400#define IBMVNIC_FULL_DUPLEX 0x04
401#define IBMVNIC_HALF_DUPLEX 0x02
402#define IBMVNIC_CAN_CHG_PHYS_PARMS 0x01
403 u8 flags2;
404#define IBMVNIC_LOGICAL_LNK_ACTIVE 0x80
405 __be32 speed;
406#define IBMVNIC_AUTONEG 0x80000000
407#define IBMVNIC_10MBPS 0x40000000
408#define IBMVNIC_100MBPS 0x20000000
409#define IBMVNIC_1GBPS 0x10000000
410#define IBMVNIC_10GBPS 0x08000000
411#define IBMVNIC_40GBPS 0x04000000
412#define IBMVNIC_100GBPS 0x02000000
413#define IBMVNIC_25GBPS 0x01000000
414#define IBMVNIC_50GBPS 0x00800000
415#define IBMVNIC_200GBPS 0x00400000
416 __be32 mtu;
417 struct ibmvnic_rc rc;
418} __packed __aligned(8);
419
420struct ibmvnic_logical_link_state {
421 u8 first;
422 u8 cmd;
423 u8 link_state;
424#define IBMVNIC_LOGICAL_LNK_DN 0x00
425#define IBMVNIC_LOGICAL_LNK_UP 0x01
426#define IBMVNIC_LOGICAL_LNK_QUERY 0xff
427 u8 reserved[9];
428 struct ibmvnic_rc rc;
429} __packed __aligned(8);
430
431struct ibmvnic_query_ip_offload {
432 u8 first;
433 u8 cmd;
434 u8 reserved[2];
435 __be32 len;
436 __be32 ioba;
437 struct ibmvnic_rc rc;
438} __packed __aligned(8);
439
440struct ibmvnic_control_ip_offload {
441 u8 first;
442 u8 cmd;
443 u8 reserved[2];
444 __be32 ioba;
445 __be32 len;
446 struct ibmvnic_rc rc;
447} __packed __aligned(8);
448
449struct ibmvnic_request_statistics {
450 u8 first;
451 u8 cmd;
452 u8 flags;
453#define IBMVNIC_PHYSICAL_PORT 0x80
454 u8 reserved1;
455 __be32 ioba;
456 __be32 len;
457 u8 reserved[4];
458} __packed __aligned(8);
459
460struct ibmvnic_error_indication {
461 u8 first;
462 u8 cmd;
463 u8 flags;
464#define IBMVNIC_FATAL_ERROR 0x80
465 u8 reserved1;
466 __be32 error_id;
467 __be32 detail_error_sz;
468 __be16 error_cause;
469 u8 reserved2[2];
470} __packed __aligned(8);
471
472struct ibmvnic_link_state_indication {
473 u8 first;
474 u8 cmd;
475 u8 reserved1[2];
476 u8 phys_link_state;
477 u8 logical_link_state;
478 u8 reserved2[10];
479} __packed __aligned(8);
480
481struct ibmvnic_change_mac_addr {
482 u8 first;
483 u8 cmd;
484 u8 mac_addr[6];
485 u8 reserved[4];
486 struct ibmvnic_rc rc;
487} __packed __aligned(8);
488
489struct ibmvnic_multicast_ctrl {
490 u8 first;
491 u8 cmd;
492 u8 mac_addr[6];
493 u8 flags;
494#define IBMVNIC_ENABLE_MC 0x80
495#define IBMVNIC_DISABLE_MC 0x40
496#define IBMVNIC_ENABLE_ALL 0x20
497#define IBMVNIC_DISABLE_ALL 0x10
498 u8 reserved1;
499 __be16 reserved2; /* was num_enabled_mc_addr; */
500 struct ibmvnic_rc rc;
501} __packed __aligned(8);
502
503struct ibmvnic_get_vpd_size {
504 u8 first;
505 u8 cmd;
506 u8 reserved[14];
507} __packed __aligned(8);
508
509struct ibmvnic_get_vpd_size_rsp {
510 u8 first;
511 u8 cmd;
512 u8 reserved[2];
513 __be64 len;
514 struct ibmvnic_rc rc;
515} __packed __aligned(8);
516
517struct ibmvnic_get_vpd {
518 u8 first;
519 u8 cmd;
520 u8 reserved1[2];
521 __be32 ioba;
522 __be32 len;
523 u8 reserved[4];
524} __packed __aligned(8);
525
526struct ibmvnic_get_vpd_rsp {
527 u8 first;
528 u8 cmd;
529 u8 reserved[10];
530 struct ibmvnic_rc rc;
531} __packed __aligned(8);
532
533struct ibmvnic_acl_change_indication {
534 u8 first;
535 u8 cmd;
536 __be16 change_type;
537#define IBMVNIC_MAC_ACL 0
538#define IBMVNIC_VLAN_ACL 1
539 u8 reserved[12];
540} __packed __aligned(8);
541
542struct ibmvnic_acl_query {
543 u8 first;
544 u8 cmd;
545 u8 reserved1[2];
546 __be32 ioba;
547 __be32 len;
548 u8 reserved2[4];
549} __packed __aligned(8);
550
551struct ibmvnic_tune {
552 u8 first;
553 u8 cmd;
554 u8 reserved1[2];
555 __be32 ioba;
556 __be32 len;
557 u8 reserved2[4];
558} __packed __aligned(8);
559
560struct ibmvnic_request_map {
561 u8 first;
562 u8 cmd;
563 u8 reserved1;
564 u8 map_id;
565 __be32 ioba;
566 __be32 len;
567 u8 reserved2[4];
568} __packed __aligned(8);
569
570struct ibmvnic_request_map_rsp {
571 u8 first;
572 u8 cmd;
573 u8 reserved1;
574 u8 map_id;
575 u8 reserved2[8];
576 struct ibmvnic_rc rc;
577} __packed __aligned(8);
578
579struct ibmvnic_request_unmap {
580 u8 first;
581 u8 cmd;
582 u8 reserved1;
583 u8 map_id;
584 u8 reserved2[12];
585} __packed __aligned(8);
586
587struct ibmvnic_request_unmap_rsp {
588 u8 first;
589 u8 cmd;
590 u8 reserved1;
591 u8 map_id;
592 u8 reserved2[8];
593 struct ibmvnic_rc rc;
594} __packed __aligned(8);
595
596struct ibmvnic_query_map {
597 u8 first;
598 u8 cmd;
599 u8 reserved[14];
600} __packed __aligned(8);
601
602struct ibmvnic_query_map_rsp {
603 u8 first;
604 u8 cmd;
605 u8 reserved;
606 u8 page_size;
607 __be32 tot_pages;
608 __be32 free_pages;
609 struct ibmvnic_rc rc;
610} __packed __aligned(8);
611
612union ibmvnic_crq {
613 struct ibmvnic_generic_crq generic;
614 struct ibmvnic_version_exchange version_exchange;
615 struct ibmvnic_version_exchange version_exchange_rsp;
616 struct ibmvnic_capability query_capability;
617 struct ibmvnic_capability query_capability_rsp;
618 struct ibmvnic_capability request_capability;
619 struct ibmvnic_capability request_capability_rsp;
620 struct ibmvnic_login login;
621 struct ibmvnic_generic_crq login_rsp;
622 struct ibmvnic_phys_parms query_phys_parms;
623 struct ibmvnic_phys_parms query_phys_parms_rsp;
624 struct ibmvnic_phys_parms query_phys_capabilities;
625 struct ibmvnic_phys_parms query_phys_capabilities_rsp;
626 struct ibmvnic_phys_parms set_phys_parms;
627 struct ibmvnic_phys_parms set_phys_parms_rsp;
628 struct ibmvnic_logical_link_state logical_link_state;
629 struct ibmvnic_logical_link_state logical_link_state_rsp;
630 struct ibmvnic_query_ip_offload query_ip_offload;
631 struct ibmvnic_query_ip_offload query_ip_offload_rsp;
632 struct ibmvnic_control_ip_offload control_ip_offload;
633 struct ibmvnic_control_ip_offload control_ip_offload_rsp;
634 struct ibmvnic_request_statistics request_statistics;
635 struct ibmvnic_generic_crq request_statistics_rsp;
636 struct ibmvnic_error_indication error_indication;
637 struct ibmvnic_link_state_indication link_state_indication;
638 struct ibmvnic_change_mac_addr change_mac_addr;
639 struct ibmvnic_change_mac_addr change_mac_addr_rsp;
640 struct ibmvnic_multicast_ctrl multicast_ctrl;
641 struct ibmvnic_multicast_ctrl multicast_ctrl_rsp;
642 struct ibmvnic_get_vpd_size get_vpd_size;
643 struct ibmvnic_get_vpd_size_rsp get_vpd_size_rsp;
644 struct ibmvnic_get_vpd get_vpd;
645 struct ibmvnic_get_vpd_rsp get_vpd_rsp;
646 struct ibmvnic_acl_change_indication acl_change_indication;
647 struct ibmvnic_acl_query acl_query;
648 struct ibmvnic_generic_crq acl_query_rsp;
649 struct ibmvnic_tune tune;
650 struct ibmvnic_generic_crq tune_rsp;
651 struct ibmvnic_request_map request_map;
652 struct ibmvnic_request_map_rsp request_map_rsp;
653 struct ibmvnic_request_unmap request_unmap;
654 struct ibmvnic_request_unmap_rsp request_unmap_rsp;
655 struct ibmvnic_query_map query_map;
656 struct ibmvnic_query_map_rsp query_map_rsp;
657};
658
659enum ibmvnic_rc_codes {
660 SUCCESS = 0,
661 PARTIALSUCCESS = 1,
662 PERMISSION = 2,
663 NOMEMORY = 3,
664 PARAMETER = 4,
665 UNKNOWNCOMMAND = 5,
666 ABORTED = 6,
667 INVALIDSTATE = 7,
668 INVALIDIOBA = 8,
669 INVALIDLENGTH = 9,
670 UNSUPPORTEDOPTION = 10,
671};
672
673enum ibmvnic_capabilities {
674 MIN_TX_QUEUES = 1,
675 MIN_RX_QUEUES = 2,
676 MIN_RX_ADD_QUEUES = 3,
677 MAX_TX_QUEUES = 4,
678 MAX_RX_QUEUES = 5,
679 MAX_RX_ADD_QUEUES = 6,
680 REQ_TX_QUEUES = 7,
681 REQ_RX_QUEUES = 8,
682 REQ_RX_ADD_QUEUES = 9,
683 MIN_TX_ENTRIES_PER_SUBCRQ = 10,
684 MIN_RX_ADD_ENTRIES_PER_SUBCRQ = 11,
685 MAX_TX_ENTRIES_PER_SUBCRQ = 12,
686 MAX_RX_ADD_ENTRIES_PER_SUBCRQ = 13,
687 REQ_TX_ENTRIES_PER_SUBCRQ = 14,
688 REQ_RX_ADD_ENTRIES_PER_SUBCRQ = 15,
689 TCP_IP_OFFLOAD = 16,
690 PROMISC_REQUESTED = 17,
691 PROMISC_SUPPORTED = 18,
692 MIN_MTU = 19,
693 MAX_MTU = 20,
694 REQ_MTU = 21,
695 MAX_MULTICAST_FILTERS = 22,
696 VLAN_HEADER_INSERTION = 23,
697 RX_VLAN_HEADER_INSERTION = 24,
698 MAX_TX_SG_ENTRIES = 25,
699 RX_SG_SUPPORTED = 26,
700 RX_SG_REQUESTED = 27,
701 OPT_TX_COMP_SUB_QUEUES = 28,
702 OPT_RX_COMP_QUEUES = 29,
703 OPT_RX_BUFADD_Q_PER_RX_COMP_Q = 30,
704 OPT_TX_ENTRIES_PER_SUBCRQ = 31,
705 OPT_RXBA_ENTRIES_PER_SUBCRQ = 32,
706 TX_RX_DESC_REQ = 33,
707};
708
709enum ibmvnic_error_cause {
710 ADAPTER_PROBLEM = 0,
711 BUS_PROBLEM = 1,
712 FW_PROBLEM = 2,
713 DD_PROBLEM = 3,
714 EEH_RECOVERY = 4,
715 FW_UPDATED = 5,
716 LOW_MEMORY = 6,
717};
718
719enum ibmvnic_commands {
720 VERSION_EXCHANGE = 0x01,
721 VERSION_EXCHANGE_RSP = 0x81,
722 QUERY_CAPABILITY = 0x02,
723 QUERY_CAPABILITY_RSP = 0x82,
724 REQUEST_CAPABILITY = 0x03,
725 REQUEST_CAPABILITY_RSP = 0x83,
726 LOGIN = 0x04,
727 LOGIN_RSP = 0x84,
728 QUERY_PHYS_PARMS = 0x05,
729 QUERY_PHYS_PARMS_RSP = 0x85,
730 QUERY_PHYS_CAPABILITIES = 0x06,
731 QUERY_PHYS_CAPABILITIES_RSP = 0x86,
732 SET_PHYS_PARMS = 0x07,
733 SET_PHYS_PARMS_RSP = 0x87,
734 ERROR_INDICATION = 0x08,
735 LOGICAL_LINK_STATE = 0x0C,
736 LOGICAL_LINK_STATE_RSP = 0x8C,
737 REQUEST_STATISTICS = 0x0D,
738 REQUEST_STATISTICS_RSP = 0x8D,
739 COLLECT_FW_TRACE = 0x11,
740 COLLECT_FW_TRACE_RSP = 0x91,
741 LINK_STATE_INDICATION = 0x12,
742 CHANGE_MAC_ADDR = 0x13,
743 CHANGE_MAC_ADDR_RSP = 0x93,
744 MULTICAST_CTRL = 0x14,
745 MULTICAST_CTRL_RSP = 0x94,
746 GET_VPD_SIZE = 0x15,
747 GET_VPD_SIZE_RSP = 0x95,
748 GET_VPD = 0x16,
749 GET_VPD_RSP = 0x96,
750 TUNE = 0x17,
751 TUNE_RSP = 0x97,
752 QUERY_IP_OFFLOAD = 0x18,
753 QUERY_IP_OFFLOAD_RSP = 0x98,
754 CONTROL_IP_OFFLOAD = 0x19,
755 CONTROL_IP_OFFLOAD_RSP = 0x99,
756 ACL_CHANGE_INDICATION = 0x1A,
757 ACL_QUERY = 0x1B,
758 ACL_QUERY_RSP = 0x9B,
759 QUERY_MAP = 0x1D,
760 QUERY_MAP_RSP = 0x9D,
761 REQUEST_MAP = 0x1E,
762 REQUEST_MAP_RSP = 0x9E,
763 REQUEST_UNMAP = 0x1F,
764 REQUEST_UNMAP_RSP = 0x9F,
765 VLAN_CTRL = 0x20,
766 VLAN_CTRL_RSP = 0xA0,
767};
768
769enum ibmvnic_crq_type {
770 IBMVNIC_CRQ_CMD = 0x80,
771 IBMVNIC_CRQ_CMD_RSP = 0x80,
772 IBMVNIC_CRQ_INIT_CMD = 0xC0,
773 IBMVNIC_CRQ_INIT_RSP = 0xC0,
774 IBMVNIC_CRQ_XPORT_EVENT = 0xFF,
775};
776
777enum ibmvfc_crq_format {
778 IBMVNIC_CRQ_INIT = 0x01,
779 IBMVNIC_CRQ_INIT_COMPLETE = 0x02,
780 IBMVNIC_PARTITION_MIGRATED = 0x06,
781 IBMVNIC_DEVICE_FAILOVER = 0x08,
782};
783
784struct ibmvnic_crq_queue {
785 union ibmvnic_crq *msgs;
786 int size, cur;
787 dma_addr_t msg_token;
788 /* Used for serialization of msgs, cur */
789 spinlock_t lock;
790 bool active;
791 char name[32];
792};
793
794union sub_crq {
795 struct ibmvnic_generic_scrq generic;
796 struct ibmvnic_tx_comp_desc tx_comp;
797 struct ibmvnic_tx_desc v1;
798 struct ibmvnic_hdr_desc hdr;
799 struct ibmvnic_hdr_ext_desc hdr_ext;
800 struct ibmvnic_sge_desc sge;
801 struct ibmvnic_rx_comp_desc rx_comp;
802 struct ibmvnic_rx_buff_add_desc rx_add;
803};
804
805struct ibmvnic_ind_xmit_queue {
806 union sub_crq *indir_arr;
807 dma_addr_t indir_dma;
808 int index;
809};
810
811struct ibmvnic_sub_crq_queue {
812 union sub_crq *msgs;
813 int size, cur;
814 dma_addr_t msg_token;
815 unsigned long crq_num;
816 unsigned long hw_irq;
817 unsigned int irq;
818 unsigned int pool_index;
819 int scrq_num;
820 /* Used for serialization of msgs, cur */
821 spinlock_t lock;
822 struct sk_buff *rx_skb_top;
823 struct ibmvnic_adapter *adapter;
824 struct ibmvnic_ind_xmit_queue ind_buf;
825 atomic_t used;
826 char name[32];
827 u64 handle;
828 cpumask_var_t affinity_mask;
829} ____cacheline_aligned;
830
831struct ibmvnic_long_term_buff {
832 unsigned char *buff;
833 dma_addr_t addr;
834 u64 size;
835 u8 map_id;
836};
837
838struct ibmvnic_ltb_set {
839 int num_ltbs;
840 struct ibmvnic_long_term_buff *ltbs;
841};
842
843struct ibmvnic_tx_buff {
844 struct sk_buff *skb;
845 int index;
846 int pool_index;
847 int num_entries;
848};
849
850struct ibmvnic_tx_pool {
851 struct ibmvnic_tx_buff *tx_buff;
852 int *free_map;
853 int consumer_index;
854 int producer_index;
855 struct ibmvnic_ltb_set ltb_set;
856 int num_buffers;
857 int buf_size;
858} ____cacheline_aligned;
859
860struct ibmvnic_rx_buff {
861 struct sk_buff *skb;
862 dma_addr_t dma;
863 unsigned char *data;
864 int size;
865 int pool_index;
866};
867
868struct ibmvnic_rx_pool {
869 struct ibmvnic_rx_buff *rx_buff;
870 int size; /* # of buffers in the pool */
871 int index;
872 int buff_size;
873 atomic_t available;
874 int *free_map;
875 int next_free;
876 int next_alloc;
877 int active;
878 struct ibmvnic_ltb_set ltb_set;
879} ____cacheline_aligned;
880
881struct ibmvnic_vpd {
882 unsigned char *buff;
883 dma_addr_t dma_addr;
884 u64 len;
885};
886
887enum vnic_state {VNIC_PROBING = 1,
888 VNIC_PROBED,
889 VNIC_OPENING,
890 VNIC_OPEN,
891 VNIC_CLOSING,
892 VNIC_CLOSED,
893 VNIC_REMOVING,
894 VNIC_REMOVED,
895 VNIC_DOWN};
896
897enum ibmvnic_reset_reason {VNIC_RESET_FAILOVER = 1,
898 VNIC_RESET_MOBILITY,
899 VNIC_RESET_FATAL,
900 VNIC_RESET_NON_FATAL,
901 VNIC_RESET_TIMEOUT,
902 VNIC_RESET_CHANGE_PARAM,
903 VNIC_RESET_PASSIVE_INIT};
904
905struct ibmvnic_rwi {
906 enum ibmvnic_reset_reason reset_reason;
907 struct list_head list;
908};
909
910struct ibmvnic_tunables {
911 u64 rx_queues;
912 u64 tx_queues;
913 u64 rx_entries;
914 u64 tx_entries;
915 u64 mtu;
916};
917
918struct ibmvnic_adapter {
919 struct vio_dev *vdev;
920 struct net_device *netdev;
921 struct ibmvnic_crq_queue crq;
922 u8 mac_addr[ETH_ALEN];
923 struct ibmvnic_query_ip_offload_buffer ip_offload_buf;
924 dma_addr_t ip_offload_tok;
925 struct ibmvnic_control_ip_offload_buffer ip_offload_ctrl;
926 dma_addr_t ip_offload_ctrl_tok;
927 u32 msg_enable;
928
929 /* Vital Product Data (VPD) */
930 struct ibmvnic_vpd *vpd;
931 char fw_version[32];
932
933 /* Statistics */
934 struct ibmvnic_statistics stats;
935 dma_addr_t stats_token;
936 struct completion stats_done;
937 int replenish_no_mem;
938 int replenish_add_buff_success;
939 int replenish_add_buff_failure;
940 int replenish_task_cycles;
941 int tx_send_failed;
942 int tx_map_failed;
943
944 struct ibmvnic_tx_queue_stats *tx_stats_buffers;
945 struct ibmvnic_rx_queue_stats *rx_stats_buffers;
946
947 int phys_link_state;
948 int logical_link_state;
949
950 u32 speed;
951 u8 duplex;
952
953 /* login data */
954 struct ibmvnic_login_buffer *login_buf;
955 dma_addr_t login_buf_token;
956 int login_buf_sz;
957
958 struct ibmvnic_login_rsp_buffer *login_rsp_buf;
959 dma_addr_t login_rsp_buf_token;
960 int login_rsp_buf_sz;
961
962 atomic_t running_cap_crqs;
963
964 struct ibmvnic_sub_crq_queue **tx_scrq ____cacheline_aligned;
965 struct ibmvnic_sub_crq_queue **rx_scrq ____cacheline_aligned;
966
967 /* rx structs */
968 struct napi_struct *napi;
969 struct ibmvnic_rx_pool *rx_pool;
970 u64 promisc;
971
972 struct ibmvnic_tx_pool *tx_pool;
973 struct ibmvnic_tx_pool *tso_pool;
974 struct completion probe_done;
975 struct completion init_done;
976 int init_done_rc;
977
978 struct completion fw_done;
979 /* Used for serialization of device commands */
980 struct mutex fw_lock;
981 int fw_done_rc;
982
983 struct completion reset_done;
984 int reset_done_rc;
985 bool wait_for_reset;
986
987 /* CPU hotplug instances for online & dead */
988 struct hlist_node node;
989 struct hlist_node node_dead;
990
991 /* partner capabilities */
992 u64 min_tx_queues;
993 u64 min_rx_queues;
994 u64 min_rx_add_queues;
995 u64 max_tx_queues;
996 u64 max_rx_queues;
997 u64 max_rx_add_queues;
998 u64 req_tx_queues;
999 u64 req_rx_queues;
1000 u64 req_rx_add_queues;
1001 u64 min_tx_entries_per_subcrq;
1002 u64 min_rx_add_entries_per_subcrq;
1003 u64 max_tx_entries_per_subcrq;
1004 u64 max_rx_add_entries_per_subcrq;
1005 u64 req_tx_entries_per_subcrq;
1006 u64 req_rx_add_entries_per_subcrq;
1007 u64 tcp_ip_offload;
1008 u64 promisc_requested;
1009 u64 promisc_supported;
1010 u64 min_mtu;
1011 u64 max_mtu;
1012 u64 req_mtu;
1013 u64 prev_mtu;
1014 u64 max_multicast_filters;
1015 u64 vlan_header_insertion;
1016 u64 rx_vlan_header_insertion;
1017 u64 max_tx_sg_entries;
1018 u64 rx_sg_supported;
1019 u64 rx_sg_requested;
1020 u64 opt_tx_comp_sub_queues;
1021 u64 opt_rx_comp_queues;
1022 u64 opt_rx_bufadd_q_per_rx_comp_q;
1023 u64 opt_tx_entries_per_subcrq;
1024 u64 opt_rxba_entries_per_subcrq;
1025 __be64 tx_rx_desc_req;
1026#define MAX_MAP_ID 255
1027 DECLARE_BITMAP(map_ids, MAX_MAP_ID);
1028 u32 num_active_rx_scrqs;
1029 u32 num_active_rx_pools;
1030 u32 num_active_rx_napi;
1031 u32 num_active_tx_scrqs;
1032 u32 num_active_tx_pools;
1033
1034 u32 prev_rx_pool_size;
1035 u32 prev_tx_pool_size;
1036 u32 cur_rx_buf_sz;
1037 u32 prev_rx_buf_sz;
1038
1039 struct tasklet_struct tasklet;
1040 enum vnic_state state;
1041 /* Used for serialization of state field. When taking both state
1042 * and rwi locks, take state lock first.
1043 */
1044 spinlock_t state_lock;
1045 enum ibmvnic_reset_reason reset_reason;
1046 struct list_head rwi_list;
1047 /* Used for serialization of rwi_list. When taking both state
1048 * and rwi locks, take state lock first
1049 */
1050 spinlock_t rwi_lock;
1051 struct work_struct ibmvnic_reset;
1052 struct delayed_work ibmvnic_delayed_reset;
1053 unsigned long resetting;
1054 /* last device reset time */
1055 unsigned long last_reset_time;
1056
1057 bool napi_enabled;
1058 bool from_passive_init;
1059 bool login_pending;
1060 /* protected by rcu */
1061 bool tx_queues_active;
1062 bool failover_pending;
1063 bool force_reset_recovery;
1064
1065 struct ibmvnic_tunables desired;
1066 struct ibmvnic_tunables fallback;
1067};
1068

source code of linux/drivers/net/ethernet/ibm/ibmvnic.h