/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_crat.c | 56 .cache_size = 16, 65 .cache_size = 16, 74 .cache_size = 8, 89 .cache_size = 16, 98 .cache_size = 8, 107 .cache_size = 4, 136 .cache_size = 16, 145 .cache_size = 32, 154 .cache_size = 16, 175 .cache_size = 16, [all …]
|
H A D | kfd_crat.h | 167 uint32_t cache_size; member 304 uint32_t cache_size; member
|
/openbmc/qemu/hw/xen/ |
H A D | xen-mapcache.c | 261 hwaddr cache_size = size; in xen_map_cache_unlocked() local 295 cache_size = size + address_offset; in xen_map_cache_unlocked() 296 if (cache_size % MCACHE_BUCKET_SIZE) { in xen_map_cache_unlocked() 297 cache_size += MCACHE_BUCKET_SIZE - (cache_size % MCACHE_BUCKET_SIZE); in xen_map_cache_unlocked() 300 cache_size = MCACHE_BUCKET_SIZE; in xen_map_cache_unlocked() 327 entry->size != cache_size || in xen_map_cache_unlocked() 543 hwaddr test_bit_size, cache_size = size; in xen_replace_cache_entry_unlocked() local 554 cache_size = size + address_offset; in xen_replace_cache_entry_unlocked() 555 if (cache_size % MCACHE_BUCKET_SIZE) { in xen_replace_cache_entry_unlocked() 556 cache_size += MCACHE_BUCKET_SIZE - (cache_size % MCACHE_BUCKET_SIZE); in xen_replace_cache_entry_unlocked() [all …]
|
/openbmc/linux/drivers/infiniband/hw/hfi1/ |
H A D | user_pages.c | 13 static unsigned long cache_size = 256; variable 14 module_param(cache_size, ulong, S_IRUGO | S_IWUSR); 15 MODULE_PARM_DESC(cache_size, "Send and receive side cache size limit (in MB)"); 73 cache_limit_pages = cache_size * (1024 * 1024) / PAGE_SIZE; in hfi1_can_pin_pages()
|
/openbmc/linux/drivers/mtd/ |
H A D | mtdblock.c | 30 unsigned int cache_size; member 88 mtdblk->cache_offset, mtdblk->cache_size); in write_cached_data() 91 mtdblk->cache_size, mtdblk->cache_data); in write_cached_data() 114 unsigned int sect_size = mtdblk->cache_size; in do_cached_write() 162 mtdblk->cache_size = sect_size; in do_cached_write() 184 unsigned int sect_size = mtdblk->cache_size; in do_cached_read() 241 if (unlikely(!mtdblk->cache_data && mtdblk->cache_size)) { in mtdblock_writesect() 273 mtdblk->cache_size = mbd->mtd->erasesize; in mtdblock_open()
|
/openbmc/linux/tools/testing/selftests/resctrl/ |
H A D | cat_test.c | 94 unsigned long cache_size = 0; in cat_perf_miss_val() local 109 ret = get_cache_size(cpu_no, cache_type, &cache_size); in cat_perf_miss_val() 112 ksft_print_msg("Cache size :%lu\n", cache_size); in cat_perf_miss_val() 144 span = cache_size * (count_of_bits - n) / count_of_bits; in cat_perf_miss_val() 166 span = cache_size * n / count_of_bits; in cat_perf_miss_val()
|
H A D | cmt_test.c | 75 unsigned long cache_size = 0; in cmt_resctrl_val() local 89 ret = get_cache_size(cpu_no, "L3", &cache_size); in cmt_resctrl_val() 92 ksft_print_msg("Cache size :%lu\n", cache_size); in cmt_resctrl_val() 113 span = cache_size * n / count_of_bits; in cmt_resctrl_val()
|
H A D | resctrlfs.c | 140 int get_cache_size(int cpu_no, char *cache_type, unsigned long *cache_size) in get_cache_size() argument 173 *cache_size = 0; in get_cache_size() 178 *cache_size = *cache_size * 10 + (cache_str[i] - '0'); in get_cache_size() 182 *cache_size = *cache_size * 1024; in get_cache_size() 186 *cache_size = *cache_size * 1024 * 1024; in get_cache_size()
|
H A D | fill_buf.c | 169 size_t cache_size = span; in run_fill_buf() local 172 ret = fill_cache(cache_size, memflush, op, once); in run_fill_buf()
|
/openbmc/linux/drivers/misc/lkdtm/ |
H A D | usercopy.c | 23 static volatile size_t cache_size = 1024; variable 234 memset(buf, 'B', cache_size); in do_usercopy_slab_whitelist() 237 offset = (cache_size / 4) + unconst; in do_usercopy_slab_whitelist() 238 size = (cache_size / 16) + unconst; in do_usercopy_slab_whitelist() 429 kmem_cache_create_usercopy("lkdtm-usercopy", cache_size, in lkdtm_usercopy_init() 431 cache_size / 4, in lkdtm_usercopy_init() 432 cache_size / 16, in lkdtm_usercopy_init()
|
/openbmc/qemu/linux-user/ |
H A D | linuxload.c | 83 bprm->src.cache_size = retval; in prepare_binprm() 188 if (offset + len <= img->cache_size) { in imgsrc_read() 241 assert(offset < src->cache_size); in imgsrc_mmap() 250 if (offset + len <= src->cache_size) { in imgsrc_mmap() 253 size_t rest = src->cache_size - offset; in imgsrc_mmap()
|
H A D | loader.h | 23 unsigned int cache_size; member
|
/openbmc/linux/arch/sh/kernel/cpu/ |
H A D | proc.c | 68 unsigned int cache_size; in show_cacheinfo() local 70 cache_size = info.ways * info.sets * info.linesz; in show_cacheinfo() 73 type, cache_size >> 10, info.ways); in show_cacheinfo()
|
/openbmc/linux/arch/arm/mm/ |
H A D | cache-uniphier.c | 319 u32 level, cache_size; in __uniphier_cache_init() local 365 if (of_property_read_u32(np, "cache-size", &cache_size) || in __uniphier_cache_init() 366 cache_size == 0 || cache_size % (data->nsets * data->line_size)) { in __uniphier_cache_init() 373 data->way_mask = GENMASK(cache_size / data->nsets / data->line_size - 1, in __uniphier_cache_init()
|
/openbmc/linux/drivers/md/ |
H A D | dm-cache-policy-smq.c | 802 dm_cblock_t cache_size; member 1118 clear_bitset(mq->cache_hit_bits, from_cblock(mq->cache_size)); in end_cache_period() 1138 return from_cblock(mq->cache_size) * p / 100u; in percent_to_target() 1744 mq->cache_size = cache_size; in __smq_create() 1779 if (from_cblock(cache_size)) { in __smq_create() 1780 mq->cache_hit_bits = alloc_bitset(from_cblock(cache_size)); in __smq_create() 1785 clear_bitset(mq->cache_hit_bits, from_cblock(mq->cache_size)); in __smq_create() 1803 if (h_init(&mq->table, &mq->es, from_cblock(cache_size))) in __smq_create() 1844 return __smq_create(cache_size, origin_size, cache_block_size, in smq_create() 1852 return __smq_create(cache_size, origin_size, cache_block_size, in mq_create() [all …]
|
H A D | dm-cache-policy.c | 113 dm_cblock_t cache_size, in dm_cache_policy_create() argument 126 p = type->create(cache_size, origin_size, cache_block_size); in dm_cache_policy_create()
|
H A D | dm-cache-target.c | 333 dm_cblock_t cache_size; member 2321 cache->cache_size, in create_cache_policy() 2364 if (nr_blocks > (1 << 20) && cache->cache_size != size) in set_cache_size() 2370 cache->cache_size = size; in set_cache_size() 2425 dm_block_t cache_size = ca->cache_sectors; in cache_create() local 2428 cache_size = block_div(cache_size, ca->block_size); in cache_create() 2429 set_cache_size(cache, to_cblock(cache_size)); in cache_create() 2901 if (from_cblock(new_size) > from_cblock(cache->cache_size)) { in can_resize() 2957 } else if (csize != cache->cache_size) { in cache_preresume() 3102 (unsigned long long) from_cblock(cache->cache_size), in cache_status() [all …]
|
/openbmc/linux/tools/perf/scripts/python/ |
H A D | arm-cs-trace-disasm.py | 50 cache_size = 64*1024 variable 87 if (len(disasm_cache) > cache_size): 178 global cache_size
|
/openbmc/linux/arch/powerpc/kernel/ |
H A D | cacheinfo.c | 208 static int cache_size(const struct cache *cache, unsigned int *ret) in cache_size() function 211 const __be32 *cache_size; in cache_size() local 215 cache_size = of_get_property(cache->ofnode, propname, NULL); in cache_size() 216 if (!cache_size) in cache_size() 219 *ret = of_read_number(cache_size, 1); in cache_size() 227 if (cache_size(cache, &size)) in cache_size_kb() 292 if (cache_size(cache, &size)) in cache_associativity()
|
/openbmc/linux/arch/microblaze/kernel/cpu/ |
H A D | cache.c | 92 #define CACHE_LOOP_LIMITS(start, end, cache_line_length, cache_size) \ argument 95 if (start < UINT_MAX - cache_size) \ 96 end = min(start + cache_size, end); \ 104 #define CACHE_ALL_LOOP(cache_size, line_length, op) \ argument 106 unsigned int len = cache_size - line_length; \
|
/openbmc/qemu/migration/ |
H A D | page_cache.h | 31 PageCache *cache_init(uint64_t cache_size, size_t page_size, Error **errp);
|
H A D | migration-hmp-cmds.c | 161 info->xbzrle_cache->cache_size); in hmp_info_migrate() 521 uint64_t cache_size; in hmp_migrate_set_parameter() local 637 if (!visit_type_size(v, param, &cache_size, &err)) { in hmp_migrate_set_parameter() 640 if (cache_size > INT64_MAX || (size_t)cache_size != cache_size) { in hmp_migrate_set_parameter() 644 p->xbzrle_cache_size = cache_size; in hmp_migrate_set_parameter()
|
/openbmc/linux/sound/firewire/motu/ |
H A D | amdtp-motu.c | 307 const unsigned int cache_size = cache->size; in cache_event_offsets() local 323 cache_tail = (cache_tail + 1) % cache_size; in cache_event_offsets() 377 const unsigned int cache_size = cache->size; in write_sph() local 387 cache_head = (cache_head + 1) % cache_size; in write_sph()
|
/openbmc/linux/arch/mips/kernel/ |
H A D | pm-cps.c | 197 unsigned cache_size = cache->ways << cache->waybit; in cps_gen_cache_routine() local 209 if (cache_size < 0x8000) in cps_gen_cache_routine() 210 uasm_i_addiu(pp, t1, t0, cache_size); in cps_gen_cache_routine() 212 UASM_i_LA(pp, t1, (long)(CKSEG0 + cache_size)); in cps_gen_cache_routine()
|
/openbmc/linux/net/openvswitch/ |
H A D | flow_table.c | 378 new->cache_size = size; in tbl_mask_cache_alloc() 379 if (new->cache_size > 0) { in tbl_mask_cache_alloc() 381 new->cache_size), in tbl_mask_cache_alloc() 397 if (size == mc->cache_size) in ovs_flow_tbl_masks_cache_resize() 802 if (unlikely(!skb_hash || mc->cache_size == 0)) { in ovs_flow_tbl_lookup_stats() 822 int index = hash & (mc->cache_size - 1); in ovs_flow_tbl_lookup_stats() 949 return READ_ONCE(mc->cache_size); in ovs_flow_tbl_masks_cache_size()
|