Lines Matching refs:cache

136 void __kasan_unpoison_object_data(struct kmem_cache *cache, void *object)  in __kasan_unpoison_object_data()  argument
138 kasan_unpoison(object, cache->object_size, false); in __kasan_unpoison_object_data()
141 void __kasan_poison_object_data(struct kmem_cache *cache, void *object) in __kasan_poison_object_data() argument
143 kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE), in __kasan_poison_object_data()
161 static inline u8 assign_tag(struct kmem_cache *cache, in assign_tag() argument
171 if (!cache->ctor && !(cache->flags & SLAB_TYPESAFE_BY_RCU)) in assign_tag()
177 return (u8)obj_to_index(cache, virt_to_slab(object), (void *)object); in assign_tag()
187 void * __must_check __kasan_init_slab_obj(struct kmem_cache *cache, in __kasan_init_slab_obj() argument
192 kasan_init_object_meta(cache, object); in __kasan_init_slab_obj()
195 object = set_tag(object, assign_tag(cache, object, true)); in __kasan_init_slab_obj()
200 static inline bool ____kasan_slab_free(struct kmem_cache *cache, void *object, in ____kasan_slab_free() argument
214 if (unlikely(nearest_obj(cache, virt_to_slab(object), object) != in ____kasan_slab_free()
221 if (unlikely(cache->flags & SLAB_TYPESAFE_BY_RCU)) in ____kasan_slab_free()
229 kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE), in ____kasan_slab_free()
236 kasan_save_free_info(cache, tagged_object); in ____kasan_slab_free()
238 return kasan_quarantine_put(cache, object); in ____kasan_slab_free()
241 bool __kasan_slab_free(struct kmem_cache *cache, void *object, in __kasan_slab_free() argument
244 return ____kasan_slab_free(cache, object, ip, true, init); in __kasan_slab_free()
298 void * __must_check __kasan_slab_alloc(struct kmem_cache *cache, in __kasan_slab_alloc() argument
317 tag = assign_tag(cache, object, false); in __kasan_slab_alloc()
324 kasan_unpoison(tagged_object, cache->object_size, init); in __kasan_slab_alloc()
327 if (kasan_stack_collection_enabled() && !is_kmalloc_cache(cache)) in __kasan_slab_alloc()
328 kasan_save_alloc_info(cache, tagged_object, flags); in __kasan_slab_alloc()
333 static inline void *____kasan_kmalloc(struct kmem_cache *cache, in ____kasan_kmalloc() argument
364 redzone_end = round_up((unsigned long)(object + cache->object_size), in ____kasan_kmalloc()
373 if (kasan_stack_collection_enabled() && is_kmalloc_cache(cache)) in ____kasan_kmalloc()
374 kasan_save_alloc_info(cache, (void *)object, flags); in ____kasan_kmalloc()
380 void * __must_check __kasan_kmalloc(struct kmem_cache *cache, const void *object, in __kasan_kmalloc() argument
383 return ____kasan_kmalloc(cache, object, size, flags); in __kasan_kmalloc()