/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_active.c | 24 static struct kmem_cache *slab_cache; variable 172 kmem_cache_free(slab_cache, it); in __active_retire() 320 node = kmem_cache_alloc(slab_cache, GFP_ATOMIC); in active_instance() 766 kmem_cache_free(slab_cache, ref->cache); in i915_active_fini() 886 node = kmem_cache_alloc(slab_cache, GFP_KERNEL); in i915_active_acquire_preallocate_barrier() 934 kmem_cache_free(slab_cache, node); in i915_active_acquire_preallocate_barrier() 1198 kmem_cache_destroy(slab_cache); in i915_active_module_exit() 1203 slab_cache = KMEM_CACHE(active_node, SLAB_HWCACHE_ALIGN); in i915_active_module_init() 1204 if (!slab_cache) in i915_active_module_init()
|
/openbmc/linux/mm/ |
H A D | slab.h | 47 struct kmem_cache *slab_cache; member 60 struct kmem_cache *slab_cache; member 112 SLAB_MATCH(compound_head, slab_cache); /* Ensure bit 0 is clear */ 634 return slab->slab_cache; in virt_to_cache()
|
H A D | slab_common.c | 1055 s = slab->slab_cache; in kfree() 1090 skip_orig_size_check(folio_slab(folio)->slab_cache, object); in __ksize() 1093 return slab_ksize(folio_slab(folio)->slab_cache); in __ksize()
|
H A D | slub.c | 1492 if (unlikely(s != slab->slab_cache)) { in free_consistency_checks() 1496 } else if (!slab->slab_cache) { in free_consistency_checks() 2043 slab->slab_cache = s; in allocate_slab() 2100 __free_slab(slab->slab_cache, slab); in rcu_free_slab() 3882 df->s = df->slab->slab_cache; in build_detached_freelist() 4662 struct kmem_cache *s = slab->slab_cache; in __kmem_obj_info() 4764 s = slab->slab_cache; in __check_heap_object() 5018 p->slab_cache = s; in bootstrap() 5022 p->slab_cache = s; in bootstrap()
|
H A D | slab.c | 1402 cachep = slab->slab_cache; in kmem_rcu_free() 2581 slab->slab_cache = cachep; in cache_grow_begin() 3534 cachep = slab->slab_cache; in __kmem_obj_info() 3605 s = folio_slab(folio)->slab_cache; in kmem_cache_free_bulk() 4008 cachep = slab->slab_cache; in __check_heap_object()
|
H A D | memcontrol.c | 2946 off = obj_to_index(slab->slab_cache, slab, p); in mem_cgroup_from_obj_folio()
|
/openbmc/linux/Documentation/translations/zh_CN/mm/ |
H A D | split_page_table_lock.rst | 62 确保架构不使用slab分配器来分配页表:slab使用page->slab_cache来分配其页
|
/openbmc/linux/mm/kasan/ |
H A D | common.c | 294 ____kasan_slab_free(slab->slab_cache, ptr, ip, false, false); in __kasan_slab_free_mempool() 442 return ____kasan_kmalloc(slab->slab_cache, object, size, flags); in __kasan_krealloc()
|
H A D | quarantine.c | 129 return virt_to_slab(qlink)->slab_cache; in qlink_to_cache()
|
H A D | report.c | 495 info->cache = slab->slab_cache; in complete_report_info()
|
H A D | generic.c | 485 cache = slab->slab_cache; in __kasan_record_aux_stack()
|
/openbmc/linux/drivers/md/ |
H A D | dm-bufio.c | 994 struct kmem_cache *slab_cache; member 1182 if (unlikely(c->slab_cache != NULL)) { in alloc_buffer_data() 1184 return kmem_cache_alloc(c->slab_cache, gfp_mask); in alloc_buffer_data() 1207 kmem_cache_free(c->slab_cache, data); in free_buffer_data() 2501 c->slab_cache = kmem_cache_create(slab_name, block_size, align, in dm_bufio_client_create() 2503 if (!c->slab_cache) { in dm_bufio_client_create() 2558 kmem_cache_destroy(c->slab_cache); in dm_bufio_client_create() 2609 kmem_cache_destroy(c->slab_cache); in dm_bufio_client_destroy()
|
H A D | raid5.h | 633 struct kmem_cache *slab_cache; /* for allocating stripes */ member
|
H A D | raid5.c | 2404 sh = alloc_stripe(conf->slab_cache, gfp, conf->pool_size, conf); in grow_one_stripe() 2410 free_stripe(conf->slab_cache, sh); in grow_one_stripe() 2443 conf->slab_cache = sc; in grow_stripes() 2618 free_stripe(conf->slab_cache, osh); in resize_stripes() 2626 kmem_cache_destroy(conf->slab_cache); in resize_stripes() 2656 conf->slab_cache = sc; in resize_stripes() 2714 free_stripe(conf->slab_cache, sh); in drop_one_stripe() 2726 kmem_cache_destroy(conf->slab_cache); in shrink_stripes() 2727 conf->slab_cache = NULL; in shrink_stripes()
|
/openbmc/linux/tools/cgroup/ |
H A D | memcg_slabinfo.py | 199 cache = slab.slab_cache
|
/openbmc/linux/Documentation/mm/ |
H A D | split_page_table_lock.rst | 65 allocation: slab uses page->slab_cache for its pages.
|
/openbmc/linux/mm/kfence/ |
H A D | core.c | 466 slab->slab_cache = cache; in kfence_guarded_alloc()
|