/openbmc/u-boot/fs/yaffs2/ |
H A D | yaffs_allocator.c | 22 * This is basically a simplified slab allocator. 24 * We don't use the Linux slab allocator because slab does not allow 58 struct yaffs_allocator *allocator = in yaffs_deinit_raw_tnodes() local 59 (struct yaffs_allocator *)dev->allocator; in yaffs_deinit_raw_tnodes() 62 if (!allocator) { in yaffs_deinit_raw_tnodes() 67 while (allocator->alloc_tnode_list) { in yaffs_deinit_raw_tnodes() 68 tmp = allocator->alloc_tnode_list->next; in yaffs_deinit_raw_tnodes() 70 kfree(allocator->alloc_tnode_list->tnodes); in yaffs_deinit_raw_tnodes() 71 kfree(allocator->alloc_tnode_list); in yaffs_deinit_raw_tnodes() 72 allocator->alloc_tnode_list = tmp; in yaffs_deinit_raw_tnodes() [all …]
|
/openbmc/qemu/tests/qtest/libqos/ |
H A D | libqos-malloc.c | 215 void alloc_destroy(QGuestAllocator *allocator) in alloc_destroy() argument 222 QTAILQ_FOREACH_SAFE(node, allocator->used, MLIST_ENTNAME, tmp) { in alloc_destroy() 223 if (allocator->opts & (ALLOC_LEAK_WARN | ALLOC_LEAK_ASSERT)) { in alloc_destroy() 228 if (allocator->opts & (ALLOC_LEAK_ASSERT)) { in alloc_destroy() 237 QTAILQ_FOREACH_SAFE(node, allocator->free, MLIST_ENTNAME, tmp) { in alloc_destroy() 238 if ((allocator->opts & mask) == mask) { in alloc_destroy() 239 if ((node->addr != allocator->start) || in alloc_destroy() 240 (node->size != allocator->end - allocator->start)) { in alloc_destroy() 249 g_free(allocator->used); in alloc_destroy() 250 g_free(allocator->free); in alloc_destroy() [all …]
|
H A D | libqos-malloc.h | 39 uint64_t guest_alloc(QGuestAllocator *allocator, size_t size); 40 void guest_free(QGuestAllocator *allocator, uint64_t addr); 43 void alloc_set_flags(QGuestAllocator *allocator, QAllocOpts opts); 48 void alloc_destroy(QGuestAllocator *allocator);
|
/openbmc/linux/rust/alloc/vec/ |
H A D | partial_eq.rs | 3 use crate::alloc::Allocator; 25 __impl_slice_eq1! { [A1: Allocator, A2: Allocator] Vec<T, A1>, Vec<U, A2>, #[stable(feature = "rust… 26 __impl_slice_eq1! { [A: Allocator] Vec<T, A>, &[U], #[stable(feature = "rust1", since = "1.0.0")] } 27 __impl_slice_eq1! { [A: Allocator] Vec<T, A>, &mut [U], #[stable(feature = "rust1", since = "1.0.0"… 28 __impl_slice_eq1! { [A: Allocator] &[T], Vec<U, A>, #[stable(feature = "partialeq_vec_for_ref_slice… 29 __impl_slice_eq1! { [A: Allocator] &mut [T], Vec<U, A>, #[stable(feature = "partialeq_vec_for_ref_s… 30 __impl_slice_eq1! { [A: Allocator] Vec<T, A>, [U], #[stable(feature = "partialeq_vec_for_slice", si… 31 __impl_slice_eq1! { [A: Allocator] [T], Vec<U, A>, #[stable(feature = "partialeq_vec_for_slice", si… 33 __impl_slice_eq1! { [A: Allocator] Cow<'_, [T]>, Vec<U, A> where T: Clone, #[stable(feature = "rust… 38 __impl_slice_eq1! { [A: Allocator, const N: usize] Vec<T, A>, [U; N], #[stable(feature = "rust1", s… [all …]
|
H A D | spec_extend.rs | 3 use crate::alloc::Allocator; 22 impl<T, I, A: Allocator> SpecExtend<T, I> for Vec<T, A> 31 impl<T, I, A: Allocator> TrySpecExtend<T, I> for Vec<T, A> 41 impl<T, I, A: Allocator> SpecExtend<T, I> for Vec<T, A> 50 impl<T, I, A: Allocator> TrySpecExtend<T, I> for Vec<T, A> 60 impl<T, A: Allocator> SpecExtend<T, IntoIter<T>> for Vec<T, A> { 69 impl<T, A: Allocator> TrySpecExtend<T, IntoIter<T>> for Vec<T, A> { 80 impl<'a, T: 'a, I, A: Allocator> SpecExtend<&'a T, I> for Vec<T, A> 90 impl<'a, T: 'a, I, A: Allocator> TrySpecExtend<&'a T, I> for Vec<T, A> 101 impl<'a, T: 'a, A: Allocator> SpecExtend<&'a T, slice::Iter<'a, T>> for Vec<T, A> [all …]
|
H A D | drain.rs | 3 use crate::alloc::{Allocator, Global}; 27 #[unstable(feature = "allocator_api", issue = "32838")] A: Allocator + 'a = Global, 39 impl<T: fmt::Debug, A: Allocator> fmt::Debug for Drain<'_, T, A> { 45 impl<'a, T, A: Allocator> Drain<'a, T, A> { 63 /// Returns a reference to the underlying allocator. 67 pub fn allocator(&self) -> &A { in allocator() method 68 unsafe { self.vec.as_ref().allocator() } in allocator() 142 impl<'a, T, A: Allocator> AsRef<[T]> for Drain<'a, T, A> { 149 unsafe impl<T: Sync, A: Sync + Allocator> Sync for Drain<'_, T, A> {} 151 unsafe impl<T: Send, A: Send + Allocator> Send for Drain<'_, T, A> {} [all …]
|
H A D | into_iter.rs | 5 use crate::alloc::{Allocator, Global}; 37 #[unstable(feature = "allocator_api", issue = "32838")] A: Allocator = Global, 43 // to avoid dropping the allocator twice we need to wrap it into ManuallyDrop 52 impl<T: fmt::Debug, A: Allocator> fmt::Debug for IntoIter<T, A> { 58 impl<T, A: Allocator> IntoIter<T, A> { 93 /// Returns a reference to the underlying allocator. 96 pub fn allocator(&self) -> &A { in allocator() method 148 // Keep our `Drop` impl from dropping the elements and the allocator in into_vecdeque() 174 impl<T, A: Allocator> AsRef<[T]> for IntoIter<T, A> { 181 unsafe impl<T: Send, A: Allocator + Send> Send for IntoIter<T, A> {} [all …]
|
H A D | extract_if.rs | 3 use crate::alloc::{Allocator, Global}; 29 #[unstable(feature = "allocator_api", issue = "32838")] A: Allocator = Global, 44 impl<T, F, A: Allocator> ExtractIf<'_, T, F, A> 48 /// Returns a reference to the underlying allocator. 51 pub fn allocator(&self) -> &A { in allocator() function 52 self.vec.allocator() in allocator() 57 impl<T, F, A: Allocator> Iterator for ExtractIf<'_, T, F, A> 93 impl<T, F, A: Allocator> Drop for ExtractIf<'_, T, F, A>
|
/openbmc/linux/mm/ |
H A D | Kconfig | 131 prompt "Default allocator" 135 Selects the default allocator for the compressed cache for 148 Use the zbud allocator as the default allocator. 154 Use the z3fold allocator as the default allocator. 163 Use the zsmalloc allocator as the default allocator. 175 tristate "2:1 compression allocator (zbud)" 178 A special purpose allocator for storing compressed pages. 185 tristate "3:1 compression allocator (z3fold) (DEPRECATED)" 192 A special purpose allocator for storing compressed pages. 205 prompt "N:1 compression allocator (zsmalloc)" if ZSWAP [all …]
|
/openbmc/linux/fs/nilfs2/ |
H A D | alloc.c | 3 * NILFS dat/inode allocator 23 * @inode: inode of metadata file using this allocator 34 * @inode: inode of metadata file using this allocator 43 * nilfs_palloc_init_blockgroup - initialize private variables for allocator 44 * @inode: inode of metadata file using this allocator 78 * @inode: inode of metadata file using this allocator 93 * @inode: inode of metadata file using this allocator 109 * @inode: inode of metadata file using this allocator 162 * @inode: inode of metadata file using this allocator 234 * nilfs_palloc_delete_block - delete a block on the persistent allocator file [all …]
|
H A D | alloc.h | 3 * Persistent object (dat entry/disk inode) allocator/deallocator 20 * @inode: inode of metadata file using this allocator 40 * nilfs_palloc_req - persistent allocator request and reply 79 * struct nilfs_palloc_cache - persistent object allocator cache
|
/openbmc/linux/rust/alloc/ |
H A D | alloc.rs | 21 // These are the magic symbols to call the global allocator. rustc generates 44 /// The global memory allocator. 46 /// This type implements the [`Allocator`] trait by forwarding calls 47 /// to the allocator registered with the `#[global_allocator]` attribute 60 /// Allocate memory with the global allocator. 63 /// of the allocator registered with the `#[global_allocator]` attribute 67 /// of the [`Global`] type when it and the [`Allocator`] trait become stable. 96 // Make sure we don't accidentally allow omitting the allocator shim in in alloc() 104 /// Deallocate memory with the global allocator. 107 /// of the allocator registered with the `#[global_allocator]` attribute [all …]
|
H A D | boxed.rs | 57 //! For non-zero-sized values, a [`Box`] will use the [`Global`] allocator for 59 //! raw pointer allocated with the [`Global`] allocator, given that the 60 //! [`Layout`] used with the allocator is correct for the type. More precisely, 61 //! a `value: *mut T` that has been allocated with the [`Global`] allocator 65 //! [`Global`] allocator with [`Layout::for_value(&*value)`]. 116 //! free the value with the global allocator. In general, the best practice 118 //! allocator. 172 use crate::alloc::{AllocError, Allocator, Global, Layout}; 201 #[unstable(feature = "allocator_api", issue = "32838")] A: Allocator = Global, 364 impl<T, A: Allocator> Box<T, A> { [all …]
|
H A D | raw_vec.rs | 14 use crate::alloc::{Allocator, Global, Layout}; 43 /// * Uses the excess returned from the allocator to use the largest available capacity. 53 pub(crate) struct RawVec<T, A: Allocator = Global> { 106 impl<T, A: Allocator> RawVec<T, A> { 120 /// Like `new`, but parameterized over the choice of allocator for 128 /// allocator for the returned `RawVec`. 136 /// allocator for the returned `RawVec`. 143 /// of allocator for the returned `RawVec`. 161 /// an allocator could overallocate and return a greater memory block than requested. 236 /// Reconstitutes a `RawVec` from a pointer, capacity, and allocator. [all …]
|
/openbmc/linux/Documentation/core-api/ |
H A D | memory-allocation.rst | 10 or you can directly request pages from the page allocator with 33 zones can be used, how hard the allocator should try to find free 90 useful to understand how hard the page allocator will try to satisfy that 112 **default** page allocator behavior is used. That means that not costly 117 * ``GFP_KERNEL | __GFP_NORETRY`` - overrides the default allocator behavior 122 * ``GFP_KERNEL | __GFP_RETRY_MAYFAIL`` - overrides the default allocator 127 * ``GFP_KERNEL | __GFP_NOFAIL`` - overrides the default allocator behavior 131 Selecting memory allocator 155 request pages from the page allocator. The memory allocated by `vmalloc` 167 cache allocator. The cache should be set up with kmem_cache_create() or
|
H A D | genalloc.rst | 6 implement a new allocator for a specific range of special-purpose memory; 8 driver for that device can certainly write a little allocator to get the 17 Code using this allocator should include <linux/genalloc.h>. The action 28 those used by the page allocator, but it refers to bytes rather than pages. 104 - gen_pool_first_fit is a simple first-fit allocator; this is the default 113 - gen_pool_best_fit, as one would expect, is a simple best-fit allocator.
|
/openbmc/linux/include/drm/ |
H A D | drm_mm.h | 149 * struct drm_mm_node - allocated block in the DRM allocator 151 * This represents an allocated block in a &drm_mm allocator. Except for 183 * struct drm_mm - DRM allocator 185 * DRM range allocator with a few special functions and features geared towards 219 * struct drm_mm_scan - DRM allocator eviction roaster data 264 * drm_mm_initialized - checks whether an allocator is initialized 358 * @mm: &drm_mm allocator to walk 360 * This iterator walks over all nodes in the range allocator. It is implemented 370 * @mm: &drm_mm allocator to walk 372 * This iterator walks over all nodes in the range allocator. It is implemented [all …]
|
/openbmc/linux/Documentation/trace/ |
H A D | events-kmem.rst | 11 - Per-CPU Allocator Activity 55 a simple indicator of page allocator activity. Pages may be allocated from 56 the per-CPU allocator (high performance) or the buddy allocator. 58 If pages are allocated directly from the buddy allocator, the 74 4. Per-CPU Allocator Activity 81 In front of the page allocator is a per-cpu page allocator. It exists only
|
/openbmc/linux/include/trace/events/ |
H A D | xdp.h | 331 __field(const void *, allocator) 338 __entry->allocator = xa->allocator; 341 TP_printk("mem_id=%d mem_type=%s allocator=%p", 344 __entry->allocator 359 __field(const void *, allocator) 368 __entry->allocator = xa->allocator; 373 TP_printk("mem_id=%d mem_type=%s allocator=%p" 377 __entry->allocator,
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_mm.c | 56 * drm_mm provides a simple range allocator. The drivers are free to use the 57 * resource allocator from the linux core if it suits them, the upside of drm_mm 67 * The range allocator also supports reservation of preallocated blocks. This is 71 * after the allocator is initialized, which helps with avoiding looped 93 * some basic allocator dumpers for debugging. 95 * Note that this range allocator is not thread-safe, drivers need to protect 438 * @mm: drm_mm allocator to insert @node into 441 * This functions inserts an already set-up &drm_mm_node into the allocator, 443 * fields must be cleared to 0. This is useful to initialize the allocator with 444 * preallocated objects which must be set-up before the range allocator can be [all …]
|
/openbmc/linux/net/core/ |
H A D | xdp.c | 91 static void mem_allocator_disconnect(void *allocator) in mem_allocator_disconnect() argument 103 if (xa->allocator == allocator) in mem_allocator_disconnect() 232 /* Allocate a cyclic ID that maps to allocator pointer. 246 /* Cyclic allocator, reset next id */ in __mem_id_cyclic_get() 272 void *allocator) in __xdp_reg_mem_model() argument 284 if (!allocator) { in __xdp_reg_mem_model() 311 xdp_alloc->allocator = allocator; in __xdp_reg_mem_model() 313 /* Insert allocator into ID lookup table */ in __xdp_reg_mem_model() 323 page_pool_use_xdp_mem(allocator, mem_allocator_disconnect, mem); in __xdp_reg_mem_model() 335 enum xdp_mem_type type, void *allocator) in xdp_reg_mem_model() argument [all …]
|
/openbmc/linux/Documentation/ABI/testing/ |
H A D | sysfs-fs-ext4 | 5 Controls whether the multiblock allocator should 14 The multiblock allocator will round up allocation 22 The maximum number of extents the multiblock allocator 29 The minimum number of extents the multiblock allocator 88 inode used by the inode allocator in preference to
|
/openbmc/linux/tools/mm/ |
H A D | page_owner_sort.c | 42 int allocator; member 167 return l1->allocator - l2->allocator; in compare_allocator() 422 else if (!strcmp(arg, "allocator") || !strcmp(arg, "ator")) in get_arg_type() 432 int allocator = 0; in get_allocator() local 435 allocator |= ALLOCATOR_CMA; in get_allocator() 437 allocator |= ALLOCATOR_SLAB; in get_allocator() 449 allocator |= ALLOCATOR_VMALLOC; in get_allocator() 451 if (allocator == 0) in get_allocator() 452 allocator = ALLOCATOR_OTHERS; in get_allocator() 453 return allocator; in get_allocator() [all …]
|
/openbmc/linux/Documentation/mm/ |
H A D | page_owner.rst | 33 the page allocator hotpath and if not enabled, then allocation is done 40 most of this code is outside page allocator and its hot path. Building 46 is initialized some time later than that page allocator starts in sparse 177 ator allocator memory allocator for pages 187 ator allocator memory allocator for pages
|
/openbmc/linux/drivers/scsi/cxlflash/ |
H A D | vlun.c | 54 * ba_init() - initializes a block allocator 55 * @ba_lun: Block allocator to initialize. 143 * find_free_range() - locates a free bit within the block allocator 144 * @low: First word in block allocator to start search. 145 * @high: Last word in block allocator to search. 146 * @bali: LUN information structure owning the block allocator to search. 147 * @bit_word: Passes back the word in the block allocator owning the free bit. 179 * ba_alloc() - allocates a block from the block allocator 180 * @ba_lun: Block allocator from which to allocate a block. 232 * @bali: LUN info owning the block allocator. [all …]
|