1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* Atlantic Network Driver 3 * 4 * Copyright (C) 2014-2019 aQuantia Corporation 5 * Copyright (C) 2019-2020 Marvell International Ltd. 6 */ 7 8 /* File aq_ring.h: Declaration of functions for Rx/Tx rings. */ 9 10 #ifndef AQ_RING_H 11 #define AQ_RING_H 12 13 #include "aq_common.h" 14 #include "aq_vec.h" 15 16 #define AQ_XDP_HEADROOM ALIGN(max(NET_SKB_PAD, XDP_PACKET_HEADROOM), 8) 17 #define AQ_XDP_TAILROOM SKB_DATA_ALIGN(sizeof(struct skb_shared_info)) 18 19 struct page; 20 struct aq_nic_cfg_s; 21 22 struct aq_rxpage { 23 struct page *page; 24 dma_addr_t daddr; 25 unsigned int order; 26 unsigned int pg_off; 27 }; 28 29 /* TxC SOP DX EOP 30 * +----------+----------+----------+----------- 31 * 8bytes|len l3,l4 | pa | pa | pa 32 * +----------+----------+----------+----------- 33 * 4/8bytes|len pkt |len pkt | | skb 34 * +----------+----------+----------+----------- 35 * 4/8bytes|is_gso |len,flags |len |len,is_eop 36 * +----------+----------+----------+----------- 37 * 38 * This aq_ring_buff_s doesn't have endianness dependency. 39 * It is __packed for cache line optimizations. 40 */ 41 struct __packed aq_ring_buff_s { 42 union { 43 /* RX/TX */ 44 dma_addr_t pa; 45 /* RX */ 46 struct { 47 u32 rss_hash; 48 u16 next; 49 u8 is_hash_l4; 50 u8 rsvd1; 51 struct aq_rxpage rxdata; 52 u16 vlan_rx_tag; 53 }; 54 /* EOP */ 55 struct { 56 dma_addr_t pa_eop; 57 struct sk_buff *skb; 58 struct xdp_frame *xdpf; 59 }; 60 /* TxC */ 61 struct { 62 u32 mss; 63 u8 len_l2; 64 u8 len_l3; 65 u8 len_l4; 66 u8 is_ipv6:1; 67 u8 rsvd2:7; 68 u32 len_pkt; 69 u16 vlan_tx_tag; 70 }; 71 }; 72 union { 73 struct { 74 u32 len:16; 75 u32 is_ip_cso:1; 76 u32 is_udp_cso:1; 77 u32 is_tcp_cso:1; 78 u32 is_cso_err:1; 79 u32 is_sop:1; 80 u32 is_eop:1; 81 u32 is_gso_tcp:1; 82 u32 is_gso_udp:1; 83 u32 is_mapped:1; 84 u32 is_cleaned:1; 85 u32 is_error:1; 86 u32 is_vlan:1; 87 u32 is_lro:1; 88 u32 rsvd3:3; 89 u16 eop_index; 90 u16 rsvd4; 91 }; 92 u64 flags; 93 }; 94 }; 95 96 struct aq_ring_stats_rx_s { 97 struct u64_stats_sync syncp; /* must be first */ 98 u64 errors; 99 u64 packets; 100 u64 bytes; 101 u64 lro_packets; 102 u64 jumbo_packets; 103 u64 alloc_fails; 104 u64 skb_alloc_fails; 105 u64 polls; 106 u64 pg_losts; 107 u64 pg_flips; 108 u64 pg_reuses; 109 u64 xdp_aborted; 110 u64 xdp_drop; 111 u64 xdp_pass; 112 u64 xdp_tx; 113 u64 xdp_invalid; 114 u64 xdp_redirect; 115 }; 116 117 struct aq_ring_stats_tx_s { 118 struct u64_stats_sync syncp; /* must be first */ 119 u64 errors; 120 u64 packets; 121 u64 bytes; 122 u64 queue_restarts; 123 }; 124 125 union aq_ring_stats_s { 126 struct aq_ring_stats_rx_s rx; 127 struct aq_ring_stats_tx_s tx; 128 }; 129 130 enum atl_ring_type { 131 ATL_RING_TX, 132 ATL_RING_RX, 133 }; 134 135 struct aq_ring_s { 136 struct aq_ring_buff_s *buff_ring; 137 u8 *dx_ring; /* descriptors ring, dma shared mem */ 138 struct aq_nic_s *aq_nic; 139 unsigned int idx; /* for HW layer registers operations */ 140 unsigned int hw_head; 141 unsigned int sw_head; 142 unsigned int sw_tail; 143 unsigned int size; /* descriptors number */ 144 unsigned int dx_size; /* TX or RX descriptor size, */ 145 /* stored here for fater math */ 146 u16 page_order; 147 u16 page_offset; 148 u16 frame_max; 149 u16 tail_size; 150 union aq_ring_stats_s stats; 151 dma_addr_t dx_ring_pa; 152 struct bpf_prog *xdp_prog; 153 enum atl_ring_type ring_type; 154 struct xdp_rxq_info xdp_rxq; 155 }; 156 157 struct aq_ring_param_s { 158 unsigned int vec_idx; 159 unsigned int cpu; 160 cpumask_t affinity_mask; 161 }; 162 163 static inline void *aq_buf_vaddr(struct aq_rxpage *rxpage) 164 { 165 return page_to_virt(rxpage->page) + rxpage->pg_off; 166 } 167 168 static inline dma_addr_t aq_buf_daddr(struct aq_rxpage *rxpage) 169 { 170 return rxpage->daddr + rxpage->pg_off; 171 } 172 173 static inline unsigned int aq_ring_next_dx(struct aq_ring_s *self, 174 unsigned int dx) 175 { 176 return (++dx >= self->size) ? 0U : dx; 177 } 178 179 static inline unsigned int aq_ring_avail_dx(struct aq_ring_s *self) 180 { 181 return (((self->sw_tail >= self->sw_head)) ? 182 (self->size - 1) - self->sw_tail + self->sw_head : 183 self->sw_head - self->sw_tail - 1); 184 } 185 186 struct aq_ring_s *aq_ring_tx_alloc(struct aq_ring_s *self, 187 struct aq_nic_s *aq_nic, 188 unsigned int idx, 189 struct aq_nic_cfg_s *aq_nic_cfg); 190 struct aq_ring_s *aq_ring_rx_alloc(struct aq_ring_s *self, 191 struct aq_nic_s *aq_nic, 192 unsigned int idx, 193 struct aq_nic_cfg_s *aq_nic_cfg); 194 195 int aq_ring_init(struct aq_ring_s *self, const enum atl_ring_type ring_type); 196 void aq_ring_rx_deinit(struct aq_ring_s *self); 197 void aq_ring_free(struct aq_ring_s *self); 198 void aq_ring_update_queue_state(struct aq_ring_s *ring); 199 void aq_ring_queue_wake(struct aq_ring_s *ring); 200 void aq_ring_queue_stop(struct aq_ring_s *ring); 201 bool aq_ring_tx_clean(struct aq_ring_s *self); 202 int aq_xdp_xmit(struct net_device *dev, int num_frames, 203 struct xdp_frame **frames, u32 flags); 204 int aq_ring_rx_clean(struct aq_ring_s *self, 205 struct napi_struct *napi, 206 int *work_done, 207 int budget); 208 int aq_ring_rx_fill(struct aq_ring_s *self); 209 210 struct aq_ring_s *aq_ring_hwts_rx_alloc(struct aq_ring_s *self, 211 struct aq_nic_s *aq_nic, unsigned int idx, 212 unsigned int size, unsigned int dx_size); 213 void aq_ring_hwts_rx_clean(struct aq_ring_s *self, struct aq_nic_s *aq_nic); 214 215 unsigned int aq_ring_fill_stats_data(struct aq_ring_s *self, u64 *data); 216 217 #endif /* AQ_RING_H */ 218