/openbmc/linux/drivers/acpi/acpica/ |
H A D | uttrack.c | 4 * Module Name: uttrack - Memory allocation tracking routines (debug only) 14 * Each memory allocation is tracked via a doubly linked list. Each 32 *allocation); 80 * PARAMETERS: size - Size of the allocation 94 struct acpi_debug_mem_block *allocation; in acpi_ut_allocate_and_track() local 105 allocation = in acpi_ut_allocate_and_track() 107 if (!allocation) { in acpi_ut_allocate_and_track() 109 /* Report allocation error */ in acpi_ut_allocate_and_track() 118 acpi_ut_track_allocation(allocation, size, ACPI_MEM_MALLOC, in acpi_ut_allocate_and_track() 121 acpi_os_free(allocation); in acpi_ut_allocate_and_track() [all …]
|
/openbmc/linux/tools/perf/pmu-events/arch/x86/sapphirerapids/ |
H A D | uncore-cxl.json | 11 "BriefDescription": "Number of Allocation to Mem Rxx AGF 0", 19 "BriefDescription": "Number of Allocation to Cache Req AGF0", 27 "BriefDescription": "Number of Allocation to Cache Rsp AGF", 35 "BriefDescription": "Number of Allocation to Cache Data AGF", 43 "BriefDescription": "Number of Allocation to Cache Rsp AGF", 51 "BriefDescription": "Number of Allocation to Cache Req AGF 1", 59 "BriefDescription": "Number of Allocation to Mem Data AGF", 203 "BriefDescription": "Number of Allocation to Cache Data Packing buffer", 211 "BriefDescription": "Number of Allocation to Cache Req Packing buffer", 219 "BriefDescription": "Number of Allocation to Cache Rsp Packing buffer", [all …]
|
/openbmc/linux/tools/testing/selftests/resctrl/ |
H A D | mba_test.c | 3 * Memory Bandwidth Allocation (MBA) test 23 * For each allocation, run 5 times in order to get average values. 27 static int runs_per_allocation, allocation = 100; in mba_setup() local 38 if (allocation < ALLOCATION_MIN || allocation > ALLOCATION_MAX) in mba_setup() 41 sprintf(allocation_str, "%d", allocation); in mba_setup() 48 allocation -= ALLOCATION_STEP; in mba_setup() 55 int allocation, runs; in show_mba_info() local 60 for (allocation = 0; allocation < ALLOCATION_MAX / ALLOCATION_STEP; in show_mba_info() 61 allocation++) { in show_mba_info() 71 for (runs = NUM_OF_RUNS * allocation + 1; in show_mba_info() [all …]
|
/openbmc/qemu/tests/tcg/aarch64/gdbstub/ |
H A D | test-mte.py | 4 # qMemTag, and QMemTag packets, which are used for manipulating allocation tags. 54 # Test if we can check correctly that the allocation tag for the address 61 report(False, "Logical and allocation tags don't match!") 63 # Test allocation tag 'set and print' commands. Commands on logical 66 # Set the allocation tag for the first granule (16 bytes) of 68 gdb.execute(f"memory-tag set-allocation-tag {ta} 1 04", False, True) 70 # Then set the allocation tag for the second granule to a known 73 gdb.execute(f"memory-tag set-allocation-tag {ta}+16 1 06", False, True) 77 co = gdb.execute(f"memory-tag print-allocation-tag {ta}", False, True) 80 co = gdb.execute(f"memory-tag print-allocation-tag {ta}+16", False, True) [all …]
|
/openbmc/linux/Documentation/core-api/ |
H A D | memory-allocation.rst | 4 Memory Allocation Guide 7 Linux provides a variety of APIs for memory allocation. You can 14 Most of the memory allocation APIs use GFP flags to express how that 16 pages", the underlying memory allocation function. 18 Diversity of the allocation APIs combined with the numerous GFP flags 26 Of course there are cases when other allocation APIs and different GFP 45 * If the allocation is performed from an atomic context, e.g interrupt 48 ``GFP_NOWAIT`` allocation is likely to fail. Allocations which 51 will be stressed unless allocation succeeds, you may use ``GFP_ATOMIC``. 66 example may be a hardware allocation that maps data directly into [all …]
|
/openbmc/linux/mm/ |
H A D | zbud.c | 38 * allocation function, zbud_alloc(), returns an opaque handle to the user, 41 * allocation data and unmap the handle with zbud_unmap() when operations 42 * on the allocation data are complete. 60 * NCHUNKS_ORDER determines the internal allocation granularity, effectively 63 * allocation granularity will be in chunks of size PAGE_SIZE/64. As one chunk 126 /* Converts an allocation size in bytes to size in zbud chunks */ 162 * Add CHUNK_SIZE to the handle if it is the first allocation to jump in encode_handle() 197 * Return: pointer to the new zbud pool or NULL if the metadata allocation 230 * @size: size in bytes of the desired allocation 232 * @handle: handle of the new allocation [all …]
|
H A D | dmapool.c | 57 unsigned int allocation; member 63 struct dma_page { /* cacheable header for 'allocation' bytes */ 127 if ((dma - page->dma) < pool->allocation) in pool_find_page() 161 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_init_page() 221 * Return: a dma allocation pool with the requested characteristics, or 228 size_t allocation; in dma_pool_create() local 245 allocation = max_t(size_t, size, PAGE_SIZE); in dma_pool_create() 248 boundary = allocation; in dma_pool_create() 252 boundary = min(boundary, allocation); in dma_pool_create() 266 retval->allocation = allocation; in dma_pool_create() [all …]
|
/openbmc/linux/Documentation/trace/ |
H A D | events-kmem.rst | 5 The kmem tracing system captures events related to object and page allocation 8 - Slab allocation of small objects of unknown type (kmalloc) 9 - Slab allocation of small objects of known type 10 - Page allocation 17 1. Slab allocation of small objects of unknown type 27 internal fragmented as a result of the allocation pattern. By correlating 29 the allocation sites were. 32 2. Slab allocation of small objects of known type 45 3. Page allocation 54 These four events deal with page allocation and freeing. mm_page_alloc is [all …]
|
/openbmc/qemu/tests/qemu-iotests/ |
H A D | 175.out | 5 size=1048576, min allocation 9 size=1048576, min allocation 13 size=1048576, max allocation 17 size=1048576, max allocation 21 size=4096, min allocation 25 size=1048576, min allocation
|
/openbmc/linux/fs/jfs/ |
H A D | jfs_dmap.c | 19 * SERIALIZATION of the Block Allocation Map. 21 * the working state of the block allocation map is accessed in 24 * 1) allocation and free requests that start at the dmap 28 * 2) allocation requests that start at dmap control page 46 * the working state of the block allocation map also includes read/ 48 * free block count, allocation group level free block counts). 53 * accesses to the persistent state of the block allocation map (limited 139 * FUNCTION: initializate the block allocation map. 252 * FUNCTION: terminate the block allocation map in preparation for 341 * allocation map. [all …]
|
H A D | jfs_imap.h | 21 #define MAXAG 128 /* maximum number of allocation groups */ 23 #define AMAPSIZE 512 /* bytes in the IAG allocation maps */ 39 * inode allocation map: 41 * inode allocation map consists of 43 * . inode allocation group pages (per 4096 inodes) 47 * inode allocation group page (per 4096 inodes of an AG) 51 __le32 iagnum; /* 4: inode allocation group number */ 73 /* allocation bit map: 1 bit per inode (0 - free, 1 - allocated) */ 74 __le32 wmap[EXTSPERIAG]; /* 512: working allocation map */ 75 __le32 pmap[EXTSPERIAG]; /* 512: persistent allocation map */ [all …]
|
H A D | jfs_mount.c | 18 * allocation map inode" (aka fileset inode): 21 * allocation map inode" (aka aggregate inode) where each inode 98 * initialize aggregate inode allocation map in jfs_mount() 106 * open aggregate block allocation map in jfs_mount() 119 * initialize aggregate block allocation map in jfs_mount() 127 * open the secondary aggregate inode allocation map in jfs_mount() 129 * This is a duplicate of the aggregate inode allocation map. in jfs_mount() 149 * initialize secondary aggregate inode allocation map in jfs_mount() 164 * open fileset inode allocation map (aka fileset inode) in jfs_mount() 169 /* open fileset secondary inode allocation map */ in jfs_mount() [all …]
|
/openbmc/linux/Documentation/admin-guide/mm/ |
H A D | numa_memory_policy.rst | 38 use "local allocation" described below. However, during boot 82 A VMA policy will govern the allocation of pages that back 138 support allocation at fault time--a.k.a lazy allocation--so hugetlbfs 140 Although hugetlbfs segments now support lazy allocation, their support 197 closest to the node where the allocation takes place. 200 This mode specifies that the allocation should be attempted 202 allocation fails, the kernel will search other nodes, in order 209 and the policy is interpreted as local allocation. "Local" 210 allocation policy can be viewed as a Preferred policy that 211 starts at the node containing the cpu where the allocation [all …]
|
/openbmc/linux/include/linux/ |
H A D | gfp_types.h | 10 * typedef gfp_t - Memory allocation flags. 14 * the underlying memory allocation function. Not every GFP flag is 96 * pages being in one zone (fair zone allocation policy). 98 * %__GFP_HARDWALL enforces the cpuset memory allocation policy. 100 * %__GFP_THISNODE forces the allocation to be satisfied from the requested 103 * %__GFP_ACCOUNT causes the allocation to be accounted to kmemcg. 123 * the caller guarantees the allocation will allow more memory to be freed 161 * canonical example is THP allocation where a fallback is cheap but 196 * If the allocation does fail, and the caller is in a position to 201 * cannot handle allocation failures. The allocation could block [all …]
|
H A D | dmapool.h | 4 * Allocation pools for DMAable (coherent) memory. 22 size_t size, size_t align, size_t allocation); 34 size_t size, size_t align, size_t allocation); 39 struct device *dev, size_t size, size_t align, size_t allocation) in dma_pool_create() argument 47 struct device *dev, size_t size, size_t align, size_t allocation) in dmam_pool_create() argument
|
/openbmc/linux/arch/x86/include/asm/ |
H A D | hw_irq.h | 63 * irq_alloc_info - X86 specific interrupt allocation info 64 * @type: X86 specific allocation type 65 * @flags: Flags for allocation tweaks 68 * @mask: CPU mask for vector allocation 70 * @data: Allocation specific data 72 * @ioapic: IOAPIC specific allocation data 73 * @uv: UV specific allocation data
|
/openbmc/linux/Documentation/mm/ |
H A D | page_frags.rst | 11 simple allocation framework for page fragments. This is used by the 17 cache is needed. This provides a central point for the fragment allocation 20 which can be expensive at allocation time. However due to the nature of 23 to be disabled when executing the fragment allocation. 26 allocation. The netdev_alloc_cache is used by callers making use of the 41 avoid calling get_page per allocation.
|
/openbmc/linux/sound/core/ |
H A D | pcm_memory.c | 32 MODULE_PARM_DESC(max_alloc_per_card, "Max total allocation bytes per card."); 76 /* the actual allocation size might be bigger than requested, in do_alloc_pages() 82 /* take back on allocation failure */ in do_alloc_pages() 317 * snd_pcm_lib_preallocate_pages - pre-allocation for the given DMA type 321 * @size: the requested pre-allocation size in bytes 322 * @max: the max. allowed pre-allocation size 324 * Do pre-allocation for the given DMA buffer type. 335 …* snd_pcm_lib_preallocate_pages_for_all - pre-allocation for continuous memory type (all substream… 339 * @size: the requested pre-allocation size in bytes 340 * @max: the max. allowed pre-allocation size [all …]
|
/openbmc/linux/fs/ocfs2/ |
H A D | reservations.h | 5 * Allocation reservations function prototypes and structures. 108 * @cstart: start of proposed allocation 109 * @clen: length (in clusters) of proposed allocation 116 * On success, zero is returned and the valid allocation area is set in cstart 129 * @cstart: start of allocation in clusters 130 * @clen: end of allocation in clusters. 132 * Tell the reservation code that bits were used to fulfill allocation in 136 * reservations bitmap. If resv is passed, it's next allocation window will be
|
/openbmc/qemu/docs/interop/ |
H A D | nbd.rst | 18 QEMU supports the ``base:allocation`` metadata context as defined in the 41 qemu:allocation-depth 43 In the allocation depth context, the entire 32-bit value represents a 49 in addition to the specific ``qemu:allocation-depth`` and 77 ``NBD_CMD_BLOCK_STATUS`` for ``base:allocation`` 85 ``NBD_CMD_BLOCK_STATUS`` for ``qemu:allocation-depth``
|
/openbmc/linux/fs/xfs/libxfs/ |
H A D | xfs_alloc.h | 40 xfs_agnumber_t agno; /* allocation group number */ 41 xfs_agblock_t agbno; /* allocation group-relative block # */ 54 char wasdel; /* set if allocation was prev delayed */ 55 char wasfromfl; /* set if allocation is from freelist */ 66 #define XFS_ALLOC_USERDATA (1 << 0)/* allocation is for user data*/ 102 * space matching the requirements in that AG, then the allocation will fail. 108 * viable candidates in the AG, then fail the allocation. 115 * then the allocation fails. 121 * Best effort full filesystem allocation scan. 123 * Locality aware allocation will be attempted in the initial AG, but on failure [all …]
|
H A D | xfs_trans_resv.c | 81 * in the same transaction as an allocation or a free, so we compute them 139 * the allocation btrees: 2 trees * (max depth - 1) * block size 157 * the tree during individual inode allocation. Therefore the finobt 158 * reservation is unconditional for both the inode chunk allocation and 159 * individual inode allocation (modify) cases. 162 * traditional inobt: cover a full finobt shape change plus block allocation. 178 * the allocation btrees: 2 trees * (max depth - 1) * block size 182 * allocation or free and which type of create transaction is in use. An inode 184 * headers (N == 0). An inode chunk allocation requires a chunk sized 186 * reservation is required for allocation o [all...] |
/openbmc/linux/Documentation/filesystems/ext4/ |
H A D | bigalloc.rst | 15 use clustered allocation, so that each bit in the ext4 block allocation 19 This means that each bit in the block allocation bitmap now addresses 20 256 4k blocks. This shrinks the total size of the block allocation 29 128MiB); however, the minimum allocation unit becomes a cluster, not a
|
/openbmc/linux/include/net/page_pool/ |
H A D | types.h | 26 * Fast allocation side cache array/stack 49 * @order: 2^order pages on allocation 75 * struct page_pool_alloc_stats - allocation statistics 79 * @empty: ptr ring is empty, so a slow path allocation was forced 80 * @refill: an allocation which triggered a refill of the cache 143 * Data structure for allocation side 145 * Drivers allocation side usually already perform some kind 147 * require driver to protect allocation side. 161 * association with allocation resource.
|
/openbmc/qemu/tests/qemu-iotests/tests/ |
H A D | block-status-cache | 40 # Pass --allocation-depth to enable the qemu:allocation-depth context, 46 '--allocation-depth', 64 x-dirty-bitmap=qemu:allocation-depth, so we first run a normal `map` 66 qemu:allocation-depth context, and finally another normal `map` to 72 ',x-dirty-bitmap=qemu:allocation-depth' 81 # qemu:allocation-depth maps for want_zero=false. 122 'qemu:allocation-depth')
|