Lines Matching refs:nfp_map

259 static void nfp_map_bpf_byte_swap(struct nfp_bpf_map *nfp_map, void *value)  in nfp_map_bpf_byte_swap()  argument
264 for (i = 0; i < DIV_ROUND_UP(nfp_map->offmap->map.value_size, 4); i++) in nfp_map_bpf_byte_swap()
265 if (nfp_map->use_map[i].type == NFP_MAP_USE_ATOMIC_CNT) in nfp_map_bpf_byte_swap()
273 nfp_map_bpf_byte_swap_record(struct nfp_bpf_map *nfp_map, void *value) in nfp_map_bpf_byte_swap_record() argument
278 for (i = 0; i < DIV_ROUND_UP(nfp_map->offmap->map.value_size, 4); i++) in nfp_map_bpf_byte_swap_record()
279 if (nfp_map->use_map[i].type == NFP_MAP_UNUSED && in nfp_map_bpf_byte_swap_record()
281 nfp_map->use_map[i].non_zero_update = 1; in nfp_map_bpf_byte_swap_record()
334 struct nfp_bpf_map *nfp_map; in nfp_bpf_map_alloc() local
385 nfp_map = kzalloc(sizeof(*nfp_map) + use_map_size, GFP_USER); in nfp_bpf_map_alloc()
386 if (!nfp_map) in nfp_bpf_map_alloc()
389 offmap->dev_priv = nfp_map; in nfp_bpf_map_alloc()
390 nfp_map->offmap = offmap; in nfp_bpf_map_alloc()
391 nfp_map->bpf = bpf; in nfp_bpf_map_alloc()
392 spin_lock_init(&nfp_map->cache_lock); in nfp_bpf_map_alloc()
396 kfree(nfp_map); in nfp_bpf_map_alloc()
400 nfp_map->tid = res; in nfp_bpf_map_alloc()
404 list_add_tail(&nfp_map->l, &bpf->map_list); in nfp_bpf_map_alloc()
412 struct nfp_bpf_map *nfp_map = offmap->dev_priv; in nfp_bpf_map_free() local
414 nfp_bpf_ctrl_free_map(bpf, nfp_map); in nfp_bpf_map_free()
415 dev_consume_skb_any(nfp_map->cache); in nfp_bpf_map_free()
416 WARN_ON_ONCE(nfp_map->cache_blockers); in nfp_bpf_map_free()
417 list_del_init(&nfp_map->l); in nfp_bpf_map_free()
420 kfree(nfp_map); in nfp_bpf_map_free()