/openbmc/linux/tools/testing/selftests/bpf/map_tests/ |
H A D | htab_map_batch_ops.c | 28 for (i = 0; i < max_entries; i++) { in map_batch_update() 52 for (i = 0; i < max_entries; i++) { in map_batch_verify() 71 for (i = 0; i < max_entries; i++) { in map_batch_verify() 82 const __u32 max_entries = 10; in __test_map_lookup_and_delete_batch() local 83 value pcpu_values[max_entries]; in __test_map_lookup_and_delete_batch() 98 keys = malloc(max_entries * sizeof(int)); in __test_map_lookup_and_delete_batch() 108 count = max_entries; in __test_map_lookup_and_delete_batch() 126 count = max_entries; in __test_map_lookup_and_delete_batch() 185 while (total < max_entries) { in __test_map_lookup_and_delete_batch() 186 if (max_entries - total < step) in __test_map_lookup_and_delete_batch() [all …]
|
H A D | lpm_trie_map_batch_ops.c | 33 for (i = 0; i < max_entries; i++) { in map_batch_update() 52 for (i = 0; i < max_entries; i++) { in map_batch_verify() 60 for (i = 0; i < max_entries; i++) { in map_batch_verify() 72 const __u32 max_entries = 10; in test_lpm_trie_map_batch_ops() local 82 max_entries, &create_opts); in test_lpm_trie_map_batch_ops() 87 values = malloc(max_entries * sizeof(int)); in test_lpm_trie_map_batch_ops() 88 visited = malloc(max_entries * sizeof(int)); in test_lpm_trie_map_batch_ops() 93 for (step = 1; step < max_entries; step++) { in test_lpm_trie_map_batch_ops() 124 while (total < max_entries) { in test_lpm_trie_map_batch_ops() 125 if (max_entries - total < step) in test_lpm_trie_map_batch_ops() [all …]
|
H A D | array_map_batch_ops.c | 25 for (i = 0; i < max_entries; i++) { in map_batch_update() 46 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify() 47 for (i = 0; i < max_entries; i++) { in map_batch_verify() 64 for (i = 0; i < max_entries; i++) { in map_batch_verify() 74 const __u32 max_entries = 10; in __test_map_lookup_and_update_batch() local 92 keys = calloc(max_entries, sizeof(*keys)); in __test_map_lookup_and_update_batch() 93 values = calloc(max_entries, value_size); in __test_map_lookup_and_update_batch() 94 visited = calloc(max_entries, sizeof(*visited)); in __test_map_lookup_and_update_batch() 100 for (step = 1; step < max_entries; step++) { in __test_map_lookup_and_update_batch() 103 memset(keys, 0, max_entries * sizeof(*keys)); in __test_map_lookup_and_update_batch() [all …]
|
H A D | map_in_map_batch_ops.c | 127 int err, max_entries = OUTER_MAP_ENTRIES; in fetch_and_validate() local 131 fetched_keys = calloc(max_entries, value_size); in fetch_and_validate() 132 fetched_values = calloc(max_entries, value_size); in fetch_and_validate() 138 step_size <= max_entries; in fetch_and_validate() 171 CHECK((total_fetched != max_entries), in fetch_and_validate() 190 __u32 max_entries = OUTER_MAP_ENTRIES; in _map_in_map_batch_ops() local 197 outer_map_keys = calloc(max_entries, value_size); in _map_in_map_batch_ops() 198 inner_map_fds = calloc(max_entries, value_size); in _map_in_map_batch_ops() 207 for (map_index = 0; map_index < max_entries; map_index++) in _map_in_map_batch_ops() 214 inner_map_fds, &max_entries, &opts); in _map_in_map_batch_ops() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_btf_map_in_map.c | 8 __uint(max_entries, 1); 16 __uint(max_entries, 2); 23 __uint(max_entries, 3); 31 __uint(max_entries, 1); 47 __uint(max_entries, 3); 56 __uint(max_entries, 5); 63 __uint(max_entries, 3); 83 __uint(max_entries, 5); 99 __uint(max_entries, 1); 106 __uint(max_entries, 2); [all …]
|
H A D | map_ptr_kern.c | 34 __u32 max_entries; member 44 VERIFY(map->max_entries == max_entries); in check_bpf_map_fields() 56 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr() 67 max_entries)); in check() 101 __uint(max_entries, MAX_ENTRIES); 142 __uint(max_entries, MAX_ENTRIES); 174 __uint(max_entries, MAX_ENTRIES); 191 __uint(max_entries, MAX_ENTRIES); 208 __uint(max_entries, MAX_ENTRIES); 225 __uint(max_entries, MAX_ENTRIES); [all …]
|
H A D | test_unpriv_bpf_disabled.c | 17 __uint(max_entries, 1); 24 __uint(max_entries, 1); 31 __uint(max_entries, 1); 38 __uint(max_entries, 1); 51 __uint(max_entries, 1 << 12); 56 __uint(max_entries, 1);
|
H A D | lsm.c | 16 __uint(max_entries, 1); 23 __uint(max_entries, 1); 30 __uint(max_entries, 1); 37 __uint(max_entries, 1); 44 __uint(max_entries, 1); 51 __uint(max_entries, 1); 58 __uint(max_entries, 1); 65 __uint(max_entries, 1); 75 __uint(max_entries, 1);
|
H A D | verifier_map_ptr_mixing.c | 17 __uint(max_entries, 1); 24 __uint(max_entries, 1); 31 __uint(max_entries, 1); 36 __uint(max_entries, 1); 49 __uint(max_entries, 4); 62 __uint(max_entries, 8);
|
H A D | test_stacktrace_build_id.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 128); 37 __uint(max_entries, 128);
|
H A D | sockmap_verdict_prog.c | 8 __uint(max_entries, 20); 15 __uint(max_entries, 20); 22 __uint(max_entries, 20); 29 __uint(max_entries, 20);
|
H A D | test_stacktrace_map.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 16384); 36 __uint(max_entries, 16384);
|
H A D | test_map_in_map.c | 10 __uint(max_entries, 1); 18 __uint(max_entries, 1); 36 __uint(max_entries, 1); 44 __uint(max_entries, 1);
|
H A D | test_sockmap_pass_prog.c | 7 __uint(max_entries, 20); 14 __uint(max_entries, 20); 21 __uint(max_entries, 20);
|
H A D | test_sockmap_drop_prog.c | 7 __uint(max_entries, 20); 14 __uint(max_entries, 20); 21 __uint(max_entries, 20);
|
/openbmc/linux/drivers/md/persistent-data/ |
H A D | dm-array.c | 26 __le32 max_entries; member 153 uint32_t max_entries, in alloc_ablock() argument 163 (*ab)->max_entries = cpu_to_le32(max_entries); in alloc_ablock() 350 uint32_t max_entries, in insert_new_ablock() argument 407 unsigned int max_entries; member 524 resize->max_entries, in grow_add_tail_block() 765 size -= max_entries; in dm_array_new() 794 entry = index % max_entries; in dm_array_get_value() 813 unsigned int max_entries; in array_set_value() local 826 entry = index % max_entries; in array_set_value() [all …]
|
H A D | dm-btree-remove.c | 197 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local 200 if (max_entries != r_max_entries) { in shift() 205 if (nr_left - count > max_entries) { in shift() 210 if (nr_right + count > max_entries) { in shift() 316 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in delete_center_node() local 319 if (nr_left + shift > max_entries) { in delete_center_node() 357 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in redistribute3() local 363 BUG_ON(target_left > max_entries); in redistribute3() 364 BUG_ON(target_right > max_entries); in redistribute3() 431 if ((left->header.max_entries != center->header.max_entries) || in __rebalance3() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/prog_tests/ |
H A D | btf.c | 72 __u32 max_entries; member 140 .max_entries = 4, 195 .max_entries = 4, 220 .max_entries = 4, 261 .max_entries = 4, 306 .max_entries = 1, 328 .max_entries = 1, 350 .max_entries = 1, 372 .max_entries = 1, 397 .max_entries = 1, [all …]
|
H A D | for_each.c | 13 int i, err, max_entries; in test_hash_map() local 29 max_entries = bpf_map__max_entries(skel->maps.hashmap); in test_hash_map() 30 for (i = 0; i < max_entries; i++) { in test_hash_map() 60 ASSERT_EQ(skel->bss->hashmap_elems, max_entries, "hashmap_elems"); in test_hash_map() 79 __u32 key, num_cpus, max_entries; in test_array_map() local 96 max_entries = bpf_map__max_entries(skel->maps.arraymap); in test_array_map() 97 for (i = 0; i < max_entries; i++) { in test_array_map() 101 if (i != max_entries - 1) in test_array_map()
|
/openbmc/u-boot/cmd/ |
H A D | blkcache.c | 24 stats.max_blocks_per_entry, stats.max_entries); in blkc_show() 31 unsigned blocks_per_entry, max_entries; in blkc_configure() local 36 max_entries = simple_strtoul(argv[2], 0, 0); in blkc_configure() 37 blkcache_configure(blocks_per_entry, max_entries); in blkc_configure() 39 max_entries, blocks_per_entry); in blkc_configure()
|
/openbmc/linux/samples/bpf/ |
H A D | map_perf_test.bpf.c | 21 __uint(max_entries, MAX_ENTRIES); 28 __uint(max_entries, 10000); 35 __uint(max_entries, 10000); 43 __uint(max_entries, MAX_ENTRIES); 50 __uint(max_entries, MAX_NR_CPUS); 62 __uint(max_entries, MAX_ENTRIES); 69 __uint(max_entries, MAX_ENTRIES); 77 __uint(max_entries, MAX_ENTRIES); 85 __uint(max_entries, 10000); 93 __uint(max_entries, MAX_ENTRIES); [all …]
|
H A D | test_map_in_map.bpf.c | 25 __uint(max_entries, MAX_NR_PORTS); 33 __uint(max_entries, 1); 41 __uint(max_entries, 1); 49 __uint(max_entries, 1); 55 __uint(max_entries, MAX_NR_PORTS); 63 __uint(max_entries, 1); 71 __uint(max_entries, 1);
|
/openbmc/linux/kernel/bpf/ |
H A D | arraymap.c | 87 u32 elem_size, index_mask, max_entries; in array_map_alloc() local 94 max_entries = attr->max_entries; in array_map_alloc() 100 mask64 = fls_long(max_entries - 1); in array_map_alloc() 109 max_entries = index_mask + 1; in array_map_alloc() 111 if (max_entries < attr->max_entries) in array_map_alloc() 182 if (map->max_entries != 1) in array_map_direct_value_addr() 198 if (map->max_entries != 1) in array_map_direct_value_meta() 501 map->max_entries != 1) in array_map_check_btf() 545 meta0->max_entries == meta1->max_entries; in array_map_meta_equal() 561 if (info->index >= map->max_entries) in bpf_array_map_seq_start() [all …]
|
/openbmc/linux/net/xdp/ |
H A D | xskmap.c | 70 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc() 76 size = struct_size(m, xsk_map, attr->max_entries); in xsk_map_alloc() 92 return struct_size(m, xsk_map, map->max_entries) + in xsk_map_mem_usage() 110 if (index >= m->map.max_entries) { in xsk_map_get_next_key() 115 if (index == m->map.max_entries - 1) in xsk_map_get_next_key() 127 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); in xsk_map_gen_lookup() 145 if (key >= map->max_entries) in __xsk_map_lookup_elem() 174 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem() 229 if (k >= map->max_entries) in xsk_map_delete_elem() 262 return meta0->max_entries == meta1->max_entries && in xsk_map_meta_equal()
|
/openbmc/linux/fs/ext4/ |
H A D | migrate.c | 103 for (i = 0; i < max_entries; i++) { in update_ind_extent_range() 132 for (i = 0; i < max_entries; i++) { in update_dind_extent_range() 140 lb->curr_block += max_entries; in update_dind_extent_range() 162 for (i = 0; i < max_entries; i++) { in update_tind_extent_range() 170 lb->curr_block += max_entries * max_entries; in update_tind_extent_range() 193 for (i = 0; i < max_entries; i++) { in free_dind_blocks() 232 for (i = 0; i < max_entries; i++) { in free_tind_blocks() 417 unsigned long max_entries; in ext4_ext_migrate() local 510 max_entries = inode->i_sb->s_blocksize >> 2; in ext4_ext_migrate() 526 lb.curr_block += max_entries; in ext4_ext_migrate() [all …]
|