Lines Matching refs:xsk_buff_pool

11 void xp_add_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs)  in xp_add_xsk()
23 void xp_del_xsk(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_del_xsk()
35 void xp_destroy(struct xsk_buff_pool *pool) in xp_destroy()
45 int xp_alloc_tx_descs(struct xsk_buff_pool *pool, struct xdp_sock *xs) in xp_alloc_tx_descs()
55 struct xsk_buff_pool *xp_create_and_assign_umem(struct xdp_sock *xs, in xp_create_and_assign_umem()
59 struct xsk_buff_pool *pool; in xp_create_and_assign_umem()
117 void xp_set_rxq_info(struct xsk_buff_pool *pool, struct xdp_rxq_info *rxq) in xp_set_rxq_info()
126 static void xp_disable_drv_zc(struct xsk_buff_pool *pool) in xp_disable_drv_zc()
149 int xp_assign_dev(struct xsk_buff_pool *pool, in xp_assign_dev()
228 int xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_sock *umem_xs, in xp_assign_dev_shared()
245 void xp_clear_dev(struct xsk_buff_pool *pool) in xp_clear_dev()
258 struct xsk_buff_pool *pool = container_of(work, struct xsk_buff_pool, in xp_release_deferred()
279 void xp_get_pool(struct xsk_buff_pool *pool) in xp_get_pool()
284 bool xp_put_pool(struct xsk_buff_pool *pool) in xp_put_pool()
298 static struct xsk_dma_map *xp_find_dma_map(struct xsk_buff_pool *pool) in xp_find_dma_map()
359 void xp_dma_unmap(struct xsk_buff_pool *pool, unsigned long attrs) in xp_dma_unmap()
395 static int xp_init_dma_info(struct xsk_buff_pool *pool, struct xsk_dma_map *dma_map) in xp_init_dma_info()
420 int xp_dma_map(struct xsk_buff_pool *pool, struct device *dev, in xp_dma_map()
467 static bool xp_addr_crosses_non_contig_pg(struct xsk_buff_pool *pool, in xp_addr_crosses_non_contig_pg()
473 static bool xp_check_unaligned(struct xsk_buff_pool *pool, u64 *addr) in xp_check_unaligned()
483 static bool xp_check_aligned(struct xsk_buff_pool *pool, u64 *addr) in xp_check_aligned()
489 static struct xdp_buff_xsk *__xp_alloc(struct xsk_buff_pool *pool) in __xp_alloc()
527 struct xdp_buff *xp_alloc(struct xsk_buff_pool *pool) in xp_alloc()
555 static u32 xp_alloc_new_from_fq(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xp_alloc_new_from_fq()
598 static u32 xp_alloc_reused(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 nb_entries) in xp_alloc_reused()
618 u32 xp_alloc_batch(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 max) in xp_alloc_batch()
649 bool xp_can_alloc(struct xsk_buff_pool *pool, u32 count) in xp_can_alloc()
667 void *xp_raw_get_data(struct xsk_buff_pool *pool, u64 addr) in xp_raw_get_data()
674 dma_addr_t xp_raw_get_dma(struct xsk_buff_pool *pool, u64 addr) in xp_raw_get_dma()
690 void xp_dma_sync_for_device_slow(struct xsk_buff_pool *pool, dma_addr_t dma, in xp_dma_sync_for_device_slow()