/openbmc/linux/tools/testing/selftests/bpf/ |
H A D | test_lru_map.c | 30 LIBBPF_OPTS(bpf_map_create_opts, opts, .map_flags = map_flags); in create_map() 155 map_flags); in test_lru_sanity0() 159 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_sanity0() 252 map_flags); in test_lru_sanity1() 329 map_flags); in test_lru_sanity2() 435 map_flags); in test_lru_sanity3() 494 map_flags); in test_lru_sanity4() 572 map_flags); in test_lru_sanity5() 620 map_flags); in test_lru_sanity6() 683 map_flags); in test_lru_sanity7() [all …]
|
/openbmc/linux/samples/bpf/ |
H A D | test_lru_dist.c | 205 .map_flags = map_flags, in create_map() 314 map_flags); in test_parallel_lru_dist() 316 if (map_flags & BPF_F_NO_COMMON_LRU) in test_parallel_lru_dist() 341 map_flags); in test_lru_loss0() 345 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_loss0() 393 map_flags); in test_lru_loss1() 397 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_loss1() 470 map_flags); in test_parallel_lru_loss() 473 if (map_flags & BPF_F_NO_COMMON_LRU) in test_parallel_lru_loss() 474 map_fd = create_map(map_type, map_flags, in test_parallel_lru_loss() [all …]
|
H A D | map_perf_test.bpf.c | 36 __uint(map_flags, BPF_F_NO_COMMON_LRU); 44 __uint(map_flags, BPF_F_NUMA_NODE); 70 __uint(map_flags, BPF_F_NO_PREALLOC); 78 __uint(map_flags, BPF_F_NO_PREALLOC); 86 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | tcp_dumpstats_kern.c | 17 __u32 map_flags; member 22 .map_flags = BPF_F_NO_PREALLOC,
|
/openbmc/linux/kernel/bpf/ |
H A D | bpf_local_storage.c | 461 u64 map_flags) in check_flags() argument 467 if (!old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_EXIST) in check_flags() 553 void *value, u64 map_flags, gfp_t gfp_flags) in bpf_local_storage_update() argument 562 if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST) || in bpf_local_storage_update() 564 unlikely((map_flags & BPF_F_LOCK) && in bpf_local_storage_update() 575 err = check_flags(NULL, map_flags); in bpf_local_storage_update() 593 if ((map_flags & BPF_F_LOCK) && !(map_flags & BPF_NOEXIST)) { in bpf_local_storage_update() 600 err = check_flags(old_sdata, map_flags); in bpf_local_storage_update() 631 err = check_flags(old_sdata, map_flags); in bpf_local_storage_update() 635 if (old_sdata && (map_flags & BPF_F_LOCK)) { in bpf_local_storage_update() [all …]
|
H A D | reuseport_array.c | 192 u32 map_flags) in reuseport_array_update_check() argument 194 if (osk && map_flags == BPF_NOEXIST) in reuseport_array_update_check() 197 if (!osk && map_flags == BPF_EXIST) in reuseport_array_update_check() 233 void *value, u64 map_flags) in bpf_fd_reuseport_array_update_elem() argument 243 if (map_flags > BPF_EXIST) in bpf_fd_reuseport_array_update_elem() 273 map_flags); in bpf_fd_reuseport_array_update_elem() 289 err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags); in bpf_fd_reuseport_array_update_elem()
|
H A D | map_in_map.c | 47 inner_map_meta->map_flags = inner_map->map_flags; in bpf_map_meta_alloc() 104 meta0->map_flags == meta1->map_flags && in bpf_map_meta_equal()
|
H A D | arraymap.c | 70 attr->map_flags & BPF_F_PRESERVE_ELEMS) in array_map_alloc_check() 119 if (attr->map_flags & BPF_F_MMAPABLE) { in array_map_alloc() 128 if (attr->map_flags & BPF_F_MMAPABLE) { in array_map_alloc() 214 if (map->map_flags & BPF_F_INNER_MAP) in array_map_gen_lookup() 311 u64 map_flags) in array_map_update_elem() argument 325 if (unlikely(map_flags & BPF_NOEXIST)) in array_map_update_elem() 340 if (map_flags & BPF_F_LOCK) in array_map_update_elem() 350 u64 map_flags) in bpf_percpu_array_update() argument 358 if (unlikely(map_flags > BPF_EXIST)) in bpf_percpu_array_update() 366 if (unlikely(map_flags == BPF_NOEXIST)) in bpf_percpu_array_update() [all …]
|
H A D | hashtab.c | 1079 u64 map_flags) in check_flags() argument 1094 u64 map_flags) in htab_map_update_elem() argument 1197 u64 map_flags) in htab_lru_map_update_elem() argument 1265 void *value, u64 map_flags, in __htab_percpu_map_update_elem() argument 1350 if (map_flags != BPF_EXIST) { in __htab_lru_percpu_map_update_elem() 1390 void *value, u64 map_flags) in htab_percpu_map_update_elem() argument 1693 u64 elem_map_flags, map_flags; in __htab_map_lookup_and_delete_batch() local 1707 map_flags = attr->batch.flags; in __htab_map_lookup_and_delete_batch() 1708 if (map_flags) in __htab_map_lookup_and_delete_batch() 2384 u64 map_flags) in bpf_percpu_hash_update() argument [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | local_storage.c | 28 __uint(map_flags, BPF_F_NO_PREALLOC); 35 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 42 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 49 __uint(map_flags, BPF_F_NO_PREALLOC); 56 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | sockopt_inherit.c | 20 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 27 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 34 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | test_mmap.c | 12 __uint(map_flags, BPF_F_MMAPABLE | BPF_F_RDONLY_PROG); 19 __uint(map_flags, BPF_F_MMAPABLE);
|
H A D | test_btf_map_in_map.c | 46 __uint(map_flags, BPF_F_INNER_MAP); 55 __uint(map_flags, BPF_F_INNER_MAP); 68 __uint(map_flags, BPF_F_INNER_MAP);
|
H A D | test_queue_stack_map.h | 14 __uint(map_flags, 0); 22 __uint(map_flags, 0);
|
H A D | netns_cookie_prog.c | 11 __uint(map_flags, BPF_F_NO_PREALLOC); 18 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | cgrp_ls_recursion.c | 12 __uint(map_flags, BPF_F_NO_PREALLOC); 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | test_map_in_map.c | 11 __uint(map_flags, 0); 19 __uint(map_flags, 0);
|
H A D | task_ls_recursion.c | 18 __uint(map_flags, BPF_F_NO_PREALLOC); 25 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | cgrp_ls_tp_btf.c | 12 __uint(map_flags, BPF_F_NO_PREALLOC); 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
H A D | bench_local_storage_create.c | 20 __uint(map_flags, BPF_F_NO_PREALLOC); 27 __uint(map_flags, BPF_F_NO_PREALLOC);
|
/openbmc/linux/tools/testing/selftests/bpf/map_tests/ |
H A D | map_percpu_stats.c | 262 n_iter, n_real, map_type_to_s(info->type), info->map_flags); in check_expected_number_elements() 331 .map_flags = BPF_F_NO_PREALLOC, in create_hash() 341 .map_flags = BPF_F_NO_PREALLOC, in create_percpu_hash() 357 static int create_lru_hash(__u32 type, __u32 map_flags) in create_lru_hash() argument 361 .map_flags = map_flags, in create_lru_hash() 371 .map_flags = BPF_F_NO_PREALLOC, in create_hash_of_maps()
|
/openbmc/linux/net/xdp/ |
H A D | xskmap.c | 72 attr->map_flags & ~(BPF_F_NUMA_NODE | BPF_F_RDONLY | BPF_F_WRONLY)) in xsk_map_alloc() 162 u64 map_flags) in xsk_map_update_elem() argument 172 if (unlikely(map_flags > BPF_EXIST)) in xsk_map_update_elem() 200 } else if (old_xs && map_flags == BPF_NOEXIST) { in xsk_map_update_elem() 203 } else if (!old_xs && map_flags == BPF_EXIST) { in xsk_map_update_elem()
|
/openbmc/linux/tools/testing/selftests/bpf/benchs/ |
H A D | bench_bpf_hashmap_lookup.c | 22 __u32 map_flags; member 28 .map_flags = 0, 75 args.map_flags = ret; in parse_arg() 169 bpf_map__set_map_flags(ctx.skel->maps.hash_map_bench, args.map_flags); in setup()
|
/openbmc/phosphor-ipmi-flash/tools/ |
H A D | pciaccess.cpp | 66 pciaddr_t size, unsigned map_flags, in pci_device_map_range() argument 69 return ::pci_device_map_range(dev, base, size, map_flags, addr); in pci_device_map_range()
|
H A D | pciaccess.hpp | 49 pciaddr_t size, unsigned map_flags, 77 pciaddr_t size, unsigned map_flags,
|