Lines Matching full:xdp
53 return pool->heads[0].xdp.rxq->napi_id; in xsk_pool_get_napi_id()
73 static inline dma_addr_t xsk_buff_xdp_get_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_dma() argument
75 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_xdp_get_dma()
80 static inline dma_addr_t xsk_buff_xdp_get_frame_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_frame_dma() argument
82 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_xdp_get_frame_dma()
98 static inline u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xsk_buff_alloc_batch() argument
100 return xp_alloc_batch(pool, xdp, max); in xsk_buff_alloc_batch()
108 static inline void xsk_buff_free(struct xdp_buff *xdp) in xsk_buff_free() argument
110 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_free()
114 if (likely(!xdp_buff_has_frags(xdp))) in xsk_buff_free()
122 xdp_get_shared_info_from_buff(xdp)->nr_frags = 0; in xsk_buff_free()
127 static inline void xsk_buff_add_frag(struct xdp_buff *xdp) in xsk_buff_add_frag() argument
129 struct xdp_buff_xsk *frag = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_add_frag()
136 struct xdp_buff_xsk *xskb = container_of(first, struct xdp_buff_xsk, xdp); in xsk_buff_get_frag()
144 ret = &frag->xdp; in xsk_buff_get_frag()
152 struct xdp_buff_xsk *xskb = container_of(tail, struct xdp_buff_xsk, xdp); in xsk_buff_del_tail()
159 struct xdp_buff_xsk *xskb = container_of(first, struct xdp_buff_xsk, xdp); in xsk_buff_get_tail()
164 return &frag->xdp; in xsk_buff_get_tail()
167 static inline void xsk_buff_set_size(struct xdp_buff *xdp, u32 size) in xsk_buff_set_size() argument
169 xdp->data = xdp->data_hard_start + XDP_PACKET_HEADROOM; in xsk_buff_set_size()
170 xdp->data_meta = xdp->data; in xsk_buff_set_size()
171 xdp->data_end = xdp->data + size; in xsk_buff_set_size()
172 xdp->flags = 0; in xsk_buff_set_size()
186 static inline void xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp, struct xsk_buff_pool *pool) in xsk_buff_dma_sync_for_cpu() argument
188 struct xdp_buff_xsk *xskb = container_of(xdp, struct xdp_buff_xsk, xdp); in xsk_buff_dma_sync_for_cpu()
287 static inline dma_addr_t xsk_buff_xdp_get_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_dma() argument
292 static inline dma_addr_t xsk_buff_xdp_get_frame_dma(struct xdp_buff *xdp) in xsk_buff_xdp_get_frame_dma() argument
307 static inline u32 xsk_buff_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xsk_buff_alloc_batch() argument
317 static inline void xsk_buff_free(struct xdp_buff *xdp) in xsk_buff_free() argument
321 static inline void xsk_buff_add_frag(struct xdp_buff *xdp) in xsk_buff_add_frag() argument
339 static inline void xsk_buff_set_size(struct xdp_buff *xdp, u32 size) in xsk_buff_set_size() argument
354 static inline void xsk_buff_dma_sync_for_cpu(struct xdp_buff *xdp, struct xsk_buff_pool *pool) in xsk_buff_dma_sync_for_cpu() argument