/openbmc/linux/drivers/acpi/acpica/ |
H A D | uttrack.c | 32 *allocation); 94 struct acpi_debug_mem_block *allocation; in acpi_ut_allocate_and_track() local 105 allocation = in acpi_ut_allocate_and_track() 107 if (!allocation) { in acpi_ut_allocate_and_track() 118 acpi_ut_track_allocation(allocation, size, ACPI_MEM_MALLOC, in acpi_ut_allocate_and_track() 121 acpi_os_free(allocation); in acpi_ut_allocate_and_track() 135 return ((void *)&allocation->user_space); in acpi_ut_allocate_and_track() 157 struct acpi_debug_mem_block *allocation; in acpi_ut_allocate_zeroed_and_track() local 168 allocation = in acpi_ut_allocate_zeroed_and_track() 171 if (!allocation) { in acpi_ut_allocate_zeroed_and_track() [all …]
|
H A D | utalloc.c | 33 void *allocation; in acpi_os_allocate_zeroed() local 37 allocation = acpi_os_allocate(size); in acpi_os_allocate_zeroed() 38 if (allocation) { in acpi_os_allocate_zeroed() 42 memset(allocation, 0, size); in acpi_os_allocate_zeroed() 45 return (allocation); in acpi_os_allocate_zeroed()
|
/openbmc/linux/tools/testing/selftests/resctrl/ |
H A D | mba_test.c | 27 static int runs_per_allocation, allocation = 100; in mba_setup() local 38 if (allocation < ALLOCATION_MIN || allocation > ALLOCATION_MAX) in mba_setup() 41 sprintf(allocation_str, "%d", allocation); in mba_setup() 48 allocation -= ALLOCATION_STEP; in mba_setup() 55 int allocation, runs; in show_mba_info() local 60 for (allocation = 0; allocation < ALLOCATION_MAX / ALLOCATION_STEP; in show_mba_info() 61 allocation++) { in show_mba_info() 71 for (runs = NUM_OF_RUNS * allocation + 1; in show_mba_info() 72 runs < NUM_OF_RUNS * allocation + NUM_OF_RUNS ; runs++) { in show_mba_info() 86 ALLOCATION_MAX - ALLOCATION_STEP * allocation); in show_mba_info()
|
/openbmc/qemu/tests/qemu-iotests/ |
H A D | 175.out | 5 size=1048576, min allocation 9 size=1048576, min allocation 13 size=1048576, max allocation 17 size=1048576, max allocation 21 size=4096, min allocation 25 size=1048576, min allocation
|
/openbmc/linux/mm/ |
H A D | dmapool.c | 57 unsigned int allocation; member 127 if ((dma - page->dma) < pool->allocation) in pool_find_page() 161 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_init_page() 228 size_t allocation; in dma_pool_create() local 245 allocation = max_t(size_t, size, PAGE_SIZE); in dma_pool_create() 248 boundary = allocation; in dma_pool_create() 252 boundary = min(boundary, allocation); in dma_pool_create() 266 retval->allocation = allocation; in dma_pool_create() 306 while (offset + pool->size <= pool->allocation) { in pool_initialise_page() 342 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page() [all …]
|
/openbmc/linux/include/linux/ |
H A D | dmapool.h | 22 size_t size, size_t align, size_t allocation); 34 size_t size, size_t align, size_t allocation); 39 struct device *dev, size_t size, size_t align, size_t allocation) in dma_pool_create() argument 47 struct device *dev, size_t size, size_t align, size_t allocation) in dmam_pool_create() argument
|
/openbmc/linux/Documentation/core-api/ |
H A D | memory-allocation.rst | 7 Linux provides a variety of APIs for memory allocation. You can 14 Most of the memory allocation APIs use GFP flags to express how that 16 pages", the underlying memory allocation function. 18 Diversity of the allocation APIs combined with the numerous GFP flags 26 Of course there are cases when other allocation APIs and different GFP 45 * If the allocation is performed from an atomic context, e.g interrupt 48 ``GFP_NOWAIT`` allocation is likely to fail. Allocations which 51 will be stressed unless allocation succeeds, you may use ``GFP_ATOMIC``. 66 example may be a hardware allocation that maps data directly into 93 * ``GFP_KERNEL & ~__GFP_RECLAIM`` - optimistic allocation without _any_ [all …]
|
H A D | genalloc.rst | 4 There are a number of memory-allocation subsystems in the kernel, each 32 which NUMA node should be used for the allocation of the housekeeping 87 how the allocation functions choose which specific piece of memory to 107 - gen_pool_first_fit_align forces the allocation to have a specific 110 - gen_pool_first_fit_order_align aligns the allocation to the order of the 111 size. A 60-byte allocation will thus be 64-byte aligned, for example. 117 If the indicated memory is not available the allocation fails.
|
/openbmc/linux/Documentation/trace/ |
H A D | events-kmem.rst | 5 The kmem tracing system captures events related to object and page allocation 8 - Slab allocation of small objects of unknown type (kmalloc) 9 - Slab allocation of small objects of known type 10 - Page allocation 17 1. Slab allocation of small objects of unknown type 27 internal fragmented as a result of the allocation pattern. By correlating 29 the allocation sites were. 32 2. Slab allocation of small objects of known type 45 3. Page allocation 54 These four events deal with page allocation and freeing. mm_page_alloc is [all …]
|
/openbmc/linux/Documentation/mm/ |
H A D | page_frags.rst | 11 simple allocation framework for page fragments. This is used by the 17 cache is needed. This provides a central point for the fragment allocation 20 which can be expensive at allocation time. However due to the nature of 23 to be disabled when executing the fragment allocation. 26 allocation. The netdev_alloc_cache is used by callers making use of the 41 avoid calling get_page per allocation.
|
H A D | split_page_table_lock.rst | 30 PTE with pointer to its lock, or returns NULL if allocation failed; 62 must be called on PTE table allocation / freeing. 65 allocation: slab uses page->slab_cache for its pages. 72 allocation and pagetable_pmd_dtor() on freeing. 75 pmd_free_tlb(), but make sure you cover all PMD table allocation / freeing
|
H A D | balance.rst | 12 be that the caller is willing to fail the allocation without incurring the 14 allocation requests that have order-0 fallback options. In such cases, 17 __GFP_IO allocation requests are made to prevent file system deadlocks. 19 In the absence of non sleepable allocation requests, it seems detrimental 46 with a slight change in the allocation routine, it is possible to reduce 74 probably because all allocation requests are coming from intr context 88 watermark[WMARK_HIGH]. When low_on_memory is set, page allocation requests will
|
H A D | page_owner.rst | 10 When allocation happens, information about allocation such as call stack 15 Although we already have tracepoint for tracing page allocation/free, 33 the page allocator hotpath and if not enabled, then allocation is done 35 not affect to allocation performance, especially if the static keys jump 107 -a Sort by memory allocation time. 173 st stacktrace stack trace of the page allocation 186 st stacktrace stack trace of the page allocation
|
/openbmc/qemu/docs/interop/ |
H A D | nbd.rst | 18 QEMU supports the ``base:allocation`` metadata context as defined in the 41 qemu:allocation-depth 43 In the allocation depth context, the entire 32-bit value represents a 49 in addition to the specific ``qemu:allocation-depth`` and 77 ``NBD_CMD_BLOCK_STATUS`` for ``base:allocation`` 85 ``NBD_CMD_BLOCK_STATUS`` for ``qemu:allocation-depth``
|
/openbmc/linux/Documentation/filesystems/ext4/ |
H A D | bigalloc.rst | 15 use clustered allocation, so that each bit in the ext4 block allocation 19 This means that each bit in the block allocation bitmap now addresses 20 256 4k blocks. This shrinks the total size of the block allocation 29 128MiB); however, the minimum allocation unit becomes a cluster, not a
|
/openbmc/qemu/tests/qemu-iotests/tests/ |
H A D | nbd-qemu-allocation.out | 1 QA output created by nbd-qemu-allocation 12 === Check allocation over NBD === 26 base:allocation 27 qemu:allocation-depth
|
/openbmc/linux/fs/dlm/ |
H A D | memory.c | 137 struct dlm_mhandle *dlm_allocate_mhandle(gfp_t allocation) in dlm_allocate_mhandle() argument 139 return kmem_cache_alloc(mhandle_cache, allocation); in dlm_allocate_mhandle() 157 struct dlm_msg *dlm_allocate_msg(gfp_t allocation) in dlm_allocate_msg() argument 159 return kmem_cache_alloc(msg_cache, allocation); in dlm_allocate_msg()
|
H A D | memory.h | 23 struct dlm_mhandle *dlm_allocate_mhandle(gfp_t allocation); 27 struct dlm_msg *dlm_allocate_msg(gfp_t allocation);
|
/openbmc/linux/tools/testing/selftests/cgroup/ |
H A D | test_zswap.c | 121 void *allocation; in no_kmem_bypass_child() local 123 allocation = malloc(values->target_alloc_bytes); in no_kmem_bypass_child() 124 if (!allocation) { in no_kmem_bypass_child() 129 ((char *)allocation)[i] = 'a'; in no_kmem_bypass_child() 132 free(allocation); in no_kmem_bypass_child()
|
/openbmc/linux/Documentation/admin-guide/mm/ |
H A D | numa_memory_policy.rst | 38 use "local allocation" described below. However, during boot 82 A VMA policy will govern the allocation of pages that back 138 support allocation at fault time--a.k.a lazy allocation--so hugetlbfs 140 Although hugetlbfs segments now support lazy allocation, their support 197 closest to the node where the allocation takes place. 200 This mode specifies that the allocation should be attempted 202 allocation fails, the kernel will search other nodes, in order 209 and the policy is interpreted as local allocation. "Local" 210 allocation policy can be viewed as a Preferred policy that 211 starts at the node containing the cpu where the allocation [all …]
|
/openbmc/u-boot/drivers/video/stm32/ |
H A D | Kconfig | 17 int "Maximum horizontal resolution (for memory allocation purposes)" 26 int "Maximum vertical resolution (for memory allocation purposes)" 35 int "Maximum bits per pixel (for memory allocation purposes)"
|
/openbmc/linux/tools/testing/memblock/ |
H A D | README | 17 allocation functionalities of memblock. The main data structure of the boot time 58 allocation functions. Tests for each group are defined in dedicated files, as it 82 Some allocation functions clear the memory in the process, so it is required for 85 points to a block of memory allocated via malloc. For each group of allocation 87 at the end of the test run. The structure of a test runner checking allocation 95 (...allocation checks...)
|
/openbmc/linux/Documentation/admin-guide/ |
H A D | numastat.rst | 12 the preferred node and numa_miss on the node where allocation succeeded. 17 incremented on allocation from a node by CPU on the same node. other_node is 18 similar to numa_miss and is incremented on the node where allocation succeeds 53 preferred node. As a result, such allocation will not increase the numa_foreign
|
/openbmc/linux/drivers/staging/gdm724x/ |
H A D | TODO | 10 - Review use of atomic allocation for tx structs 12 - fix up static tty port allocation to be dynamic
|
/openbmc/openbmc/poky/scripts/pybootchartgui/pybootchartgui/ |
H A D | gui.py | 208 def on_allocation_size_changed(self, widget, allocation): argument 209 self.hadj.page_size = allocation.width 210 self.hadj.page_increment = allocation.width * 0.9 211 self.vadj.page_size = allocation.height 212 self.vadj.page_increment = allocation.height * 0.9 213 self.our_width = allocation.width 216 self.our_height = allocation.height
|