/openbmc/linux/tools/testing/selftests/memory-hotplug/ |
H A D | mem-on-off-test.sh | 25 if ! ls $SYSFS/devices/system/memory/memory* > /dev/null 2>&1; then 30 if ! grep -q 1 $SYSFS/devices/system/memory/memory*/removable; then 43 for memory in $SYSFS/devices/system/memory/memory*; do 63 grep -q online $SYSFS/devices/system/memory/memory$1/state 68 grep -q offline $SYSFS/devices/system/memory/memory$1/state 73 echo online > $SYSFS/devices/system/memory/memory$1/state 78 echo offline > $SYSFS/devices/system/memory/memory$1/state 83 local memory=$1 97 local memory=$1 111 local memory=$1 [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/memory-controllers/fsl/ |
H A D | fsl,ddr.yaml | 7 title: Freescale DDR memory controller 27 - fsl,bsc9132-memory-controller 28 - fsl,mpc8536-memory-controller 29 - fsl,mpc8540-memory-controller 30 - fsl,mpc8541-memory-controller 39 - fsl,p1020-memory-controller 40 - fsl,p1021-memory-controller 41 - fsl,p2020-memory-controller 42 - fsl,qoriq-memory-controller 65 memory-controller@2000 { [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | memory.c | 39 kfree(memory->tags); in nvkm_memory_tags_put() 40 memory->tags = NULL; in nvkm_memory_tags_put() 103 memory->func = func; in nvkm_memory_ctor() 110 struct nvkm_memory *memory = container_of(kref, typeof(*memory), kref); in nvkm_memory_del() local 112 if (memory->func->dtor) in nvkm_memory_del() 113 memory = memory->func->dtor(memory); in nvkm_memory_del() 114 kfree(memory); in nvkm_memory_del() 122 if (memory) { in nvkm_memory_unref() 131 if (memory) in nvkm_memory_ref() 133 return memory; in nvkm_memory_ref() [all …]
|
H A D | firmware.c | 120 struct nvkm_firmware *fw = nvkm_firmware_mem(memory); in nvkm_firmware_mem_map() 122 .memory = &fw->mem.memory, in nvkm_firmware_mem_map() 134 nvkm_firmware_mem_size(struct nvkm_memory *memory) in nvkm_firmware_mem_size() argument 140 nvkm_firmware_mem_addr(struct nvkm_memory *memory) in nvkm_firmware_mem_addr() argument 142 return nvkm_firmware_mem(memory)->phys; in nvkm_firmware_mem_addr() 146 nvkm_firmware_mem_page(struct nvkm_memory *memory) in nvkm_firmware_mem_page() argument 152 nvkm_firmware_mem_target(struct nvkm_memory *memory) in nvkm_firmware_mem_target() argument 154 if (nvkm_firmware_mem(memory)->device->func->tegra) in nvkm_firmware_mem_target() 161 nvkm_firmware_mem_dtor(struct nvkm_memory *memory) in nvkm_firmware_mem_dtor() argument 179 struct nvkm_memory *memory = &fw->mem.memory; in nvkm_firmware_dtor() local [all …]
|
/openbmc/linux/Documentation/admin-guide/mm/ |
H A D | memory-hotplug.rst | 21 downgrading the memory capacity. This dynamic memory resizing, sometimes 67 phase, the memory is visible in memory statistics, such as free and total 140 make use of that memory: the memory block has to be "online". 143 the memory block: the memory block has to be "offlined". 149 memory. 202 memory blocks only. 212 memory blocks; if onlining fails, memory blocks are removed again. 325 however, a memory block might span memory holes. A memory block spanning memory 368 that memory provided by a memory block is managed by 443 memory. [all …]
|
H A D | numaperf.rst | 13 A system supports such heterogeneous memory by grouping each memory type 16 are provided as memory only nodes. While memory only nodes do not provide 19 nodes with local memory and a memory only node for each of compute node:: 31 CPUs or separate memory I/O devices that can initiate memory requests. 42 memory targets. 54 A memory initiator may have multiple memory targets in the same access 97 memory activity. 104 slower performing memory cached by a smaller higher performing memory. The 110 The term "far memory" is used to denote the last level memory in the 118 level memory, so the higher numbered cache level corresponds to memory [all …]
|
H A D | concepts.rst | 5 The memory management in Linux is a complex system that evolved over the 21 the amount of memory that can be installed. The physical memory is not 30 The virtual memory abstracts the details of physical memory from the 35 With virtual memory, each and every memory access uses a virtual 39 memory controller can understand. 67 The address translation requires several memory accesses and memory 103 memory exceeds the maximal addressable size of virtual memory and 143 The `anonymous memory` or `anonymous mappings` represent memory that 160 memory allocated by user space processes etc. 167 reclaimable pages are page cache and anonymous memory. [all …]
|
/openbmc/linux/Documentation/arch/arm64/ |
H A D | kdump.rst | 2 crashkernel memory reservation on arm64 9 reserved memory is needed to pre-load the kdump kernel and boot such 12 That reserved memory for kdump is adapted to be able to minimally 28 Low memory and high memory 31 For kdump reservations, low memory is the memory area under a specific 37 whole system RAM is low memory. Outside of the low memory described 38 above, the rest of system RAM is considered high memory. 63 sufficient for the low memory needs of the kdump kernel. 74 to be allocated from the high memory and low memory respectively. On 80 low memory reservation will be done subsequently. [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | mem.c | 31 struct nvkm_memory memory; member 45 return nvkm_mem(memory)->target; in nvkm_mem_target() 49 nvkm_mem_page(struct nvkm_memory *memory) in nvkm_mem_page() argument 75 .memory = &mem->memory, in nvkm_mem_map_dma() 115 .memory = &mem->memory, in nvkm_mem_map_sgl() 170 *pmemory = &mem->memory; in nvkm_mem_new_host() 227 struct nvkm_memory *memory = NULL; in nvkm_mem_new_type() local 232 argv, argc, &memory); in nvkm_mem_new_type() 235 argv, argc, &memory); in nvkm_mem_new_type() 239 nvkm_memory_unref(&memory); in nvkm_mem_new_type() [all …]
|
/openbmc/linux/Documentation/ABI/testing/ |
H A D | sysfs-devices-memory | 1 What: /sys/devices/system/memory 9 Users: hotplug memory add/remove tools 12 What: /sys/devices/system/memory/memoryX/removable 20 Users: hotplug memory remove tools 40 memory section directory name. 66 Users: hotplug memory remove tools 77 For online memory blocks, it returns in which zone memory 80 and the memory block cannot be offlined. 82 For offline memory blocks, it returns by which zone memory 87 memory block. [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 124 struct nvkm_memory *memory = &iobj->base.memory; in nv50_instobj_kmap() local 128 u64 size = nvkm_memory_size(memory); in nv50_instobj_kmap() 145 nvkm_memory_addr(&eobj->base.memory), in nv50_instobj_kmap() 146 nvkm_memory_size(&eobj->base.memory), in nv50_instobj_kmap() 186 memory = nv50_instobj(memory)->ram; in nv50_instobj_map() 210 iobj->base.memory.ptrs = NULL; in nv50_instobj_release() 285 nv50_instobj_size(struct nvkm_memory *memory) in nv50_instobj_size() argument 291 nv50_instobj_addr(struct nvkm_memory *memory) in nv50_instobj_addr() argument 305 nv50_instobj_release(&iobj->base.memory); in nv50_instobj_bar2() 364 *pmemory = &iobj->base.memory; in nv50_instobj_wrap() [all …]
|
H A D | base.c | 34 struct nvkm_memory *memory = &iobj->memory; in nvkm_instobj_load() local 39 if (!(map = nvkm_kmap(memory))) { in nvkm_instobj_load() 45 nvkm_done(memory); in nvkm_instobj_load() 54 struct nvkm_memory *memory = &iobj->memory; in nvkm_instobj_save() local 63 if (!(map = nvkm_kmap(memory))) { in nvkm_instobj_save() 69 nvkm_done(memory); in nvkm_instobj_save() 109 struct nvkm_memory *memory = NULL; in nvkm_instobj_new() local 120 zero, nvkm_memory_addr(memory), nvkm_memory_size(memory)); in nvkm_instobj_new() 130 nvkm_done(memory); in nvkm_instobj_new() 135 nvkm_memory_unref(&memory); in nvkm_instobj_new() [all …]
|
/openbmc/linux/Documentation/admin-guide/cgroup-v1/ |
H A D | memory.rst | 27 uses of the memory controller. The memory controller can be used to 69 memory.usage_in_bytes show current usage for memory 73 memory.limit_in_bytes set/show limit of memory usage 77 memory.max_usage_in_bytes show max memory usage recorded 86 memory.pressure_level set memory pressure notifications 247 memsw means memory+swap. Usage of memory+swap is limited by 405 # mount -t cgroup none /sys/fs/cgroup/memory -o memory 414 # echo 4M > /sys/fs/cgroup/memory/0/memory.limit_in_bytes 418 # cat /sys/fs/cgroup/memory/0/memory.limit_in_bytes 435 # cat /sys/fs/cgroup/memory/0/memory.usage_in_bytes [all …]
|
/openbmc/phosphor-dbus-interfaces/yaml/xyz/openbmc_project/Inventory/Item/ |
H A D | PersistentMemory.interface.yaml | 2 Implement to provide Persistent memory attributes. 7 The manufacturer ID of this memory module as defined by JEDEC in 16 The manufacturer ID of the memory subsystem controller of this memory 21 The product ID of the memory subsystem controller of this memory 34 Total size of the volatile portion memory in kibibytes (KiB). 42 Total size of the cache portion memory in kibibytes (KiB). 64 Total number of volatile regions this memory can support. 72 Number of unused spare devices available in the memory. 87 description: The current security state of this memory. 94 Memory modes supported by the memory. [all …]
|
/openbmc/linux/drivers/staging/octeon/ |
H A D | ethernet-mem.c | 49 char *memory; in cvm_oct_free_hw_skbuff() local 52 memory = cvmx_fpa_alloc(pool); in cvm_oct_free_hw_skbuff() 53 if (memory) { in cvm_oct_free_hw_skbuff() 59 } while (memory); in cvm_oct_free_hw_skbuff() 79 char *memory; in cvm_oct_fill_hw_memory() local 94 memory = kmalloc(size + 256, GFP_ATOMIC); in cvm_oct_fill_hw_memory() 95 if (unlikely(!memory)) { in cvm_oct_fill_hw_memory() 101 *((char **)fpa - 1) = memory; in cvm_oct_fill_hw_memory() 116 char *memory; in cvm_oct_free_hw_memory() local 124 memory = *((char **)fpa - 1); in cvm_oct_free_hw_memory() [all …]
|
/openbmc/qemu/docs/ |
H A D | memory-hotplug.txt | 1 QEMU memory hotplug 7 Guest support is required for memory hotplug to work. 22 - "slots" is the number of hotpluggable memory slots 29 Creates a guest with 1GB of memory and three hotpluggable memory slots. 31 memory the guest will see after boot is 1GB. The maximum memory the 35 Two monitor commands are used to hotplug memory: 37 - "object_add": creates a memory backend object 44 (qemu) object_add memory-backend-ram,id=mem1,size=1G 82 plugged memory. 84 Two monitor commands are used to hot unplug memory: [all …]
|
/openbmc/linux/Documentation/core-api/ |
H A D | memory-hotplug.rst | 17 to allocate from the new memory. 24 allocate pages from the new memory. 28 longer possible from the memory but some of the memory to be offlined 30 subsystem from the indicated memory block. 34 the memory block that we attempted to offline. 37 Generated after offlining memory is complete. 63 - start_pfn is start_pfn of online/offline memory. 89 When adding/removing memory that uses memory block devices (i.e. ordinary RAM), 94 space once memory has been fully added. And when removing memory, we 100 memory faster than expected: [all …]
|
/openbmc/linux/Documentation/mm/ |
H A D | memory-model.rst | 17 memory models it supports, what the default memory model is and 36 memory. 46 memory to the page allocator. 65 as hot-plug and hot-remove of the physical memory, alternative memory 67 the memory map for larger systems. 98 all the memory sections. 101 initialize the memory sections and the memory maps. 135 allocate memory map on the persistent memory device. 156 subject to its memory ranges being exposed through the sysfs memory 170 events related to device-memory, typically GPU memory. See [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | ram.c | 32 struct nvkm_memory memory; member 41 return nvkm_instobj_wrap(nvkm_vram(memory)->ram->fb->subdev.device, memory, pmemory); in nvkm_vram_kmap() 48 struct nvkm_vram *vram = nvkm_vram(memory); in nvkm_vram_map() 50 .memory = &vram->memory, in nvkm_vram_map() 59 nvkm_vram_size(struct nvkm_memory *memory) in nvkm_vram_size() argument 65 nvkm_vram_addr(struct nvkm_memory *memory) in nvkm_vram_addr() argument 67 struct nvkm_vram *vram = nvkm_vram(memory); in nvkm_vram_addr() 74 nvkm_vram_page(struct nvkm_memory *memory) in nvkm_vram_page() argument 76 return nvkm_vram(memory)->page; in nvkm_vram_page() 86 nvkm_vram_dtor(struct nvkm_memory *memory) in nvkm_vram_dtor() argument [all …]
|
/openbmc/qemu/tests/tcg/aarch64/ |
H A D | Makefile.softmmu-target | 37 memory: CFLAGS+=-DCHECK_UNALIGNED=1 39 memory-sve: memory.c $(LINK_SCRIPT) $(CRT_OBJS) $(MINILIB_OBJS) 42 memory-sve: CFLAGS+=-DCHECK_UNALIGNED=1 -march=armv8.1-a+sve -O3 44 TESTS+=memory-sve 65 .PHONY: memory-record 66 run-memory-record: memory-record memory 71 $(QEMU_OPTS) memory) 73 .PHONY: memory-replay 74 run-memory-replay: memory-replay run-memory-record 79 $(QEMU_OPTS) memory) [all …]
|
/openbmc/linux/Documentation/userspace-api/media/v4l/ |
H A D | dev-mem2mem.rst | 9 A V4L2 memory-to-memory device can compress, decompress, transform, or 10 otherwise convert video data from one format into another format, in memory. 11 Such memory-to-memory devices set the ``V4L2_CAP_VIDEO_M2M`` or 12 ``V4L2_CAP_VIDEO_M2M_MPLANE`` capability. Examples of memory-to-memory 16 A memory-to-memory video node acts just like a normal video node, but it 17 supports both output (sending frames from memory to the hardware) 19 memory) stream I/O. An application will have to setup the stream I/O for 23 Memory-to-memory devices function as a shared resource: you can 32 One of the most common memory-to-memory device is the codec. Codecs 35 See :ref:`codec-controls`. More details on how to use codec memory-to-memory
|
/openbmc/linux/fs/btrfs/tests/ |
H A D | extent-io-tests.c | 727 if (!memory) { in test_eb_mem_ops() 745 memcpy(memory, memory + 16, 16); in test_eb_mem_ops() 751 memcpy(memory, memory + 2048, 16); in test_eb_mem_ops() 756 memcpy(memory, memory + 2048, 2048); in test_eb_mem_ops() 762 memmove(memory + 512, memory + 256, 512); in test_eb_mem_ops() 768 memmove(memory + 2048, memory + 512, 2048); in test_eb_mem_ops() 773 memmove(memory + 512, memory + 2048, 2048); in test_eb_mem_ops() 780 memcpy(memory, memory + 4096 - 128, 256); in test_eb_mem_ops() 786 memcpy(memory + 4096 - 128, memory + 4096 + 128, 256); in test_eb_mem_ops() 792 memmove(memory + 4096 - 128, memory + 4096 - 64, 256); in test_eb_mem_ops() [all …]
|
/openbmc/qemu/tests/multiboot/ |
H A D | mmap.out | 6 Lower memory: 639k 7 Upper memory: 129920k 9 e820 memory map: 24 Lower memory: 639k 25 Upper memory: 104k 27 e820 memory map: 41 Lower memory: 639k 44 e820 memory map: 59 Lower memory: 639k 62 e820 memory map: [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/memory-controllers/ |
H A D | nvidia,tegra210-emc.yaml | 15 sent from the memory controller. 26 - description: external memory clock 36 memory-region: 39 phandle to a reserved memory region describing the table of EMC 42 nvidia,memory-controller: 45 phandle of the memory controller node 52 - nvidia,memory-controller 61 reserved-memory { 72 external-memory-controller@7001b000 { 80 memory-region = <&emc_table>; [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/reserved-memory/ |
H A D | xen,shared-memory.txt | 1 * Xen hypervisor reserved-memory binding 3 Expose one or more memory regions as reserved-memory to the guest 5 to be a shared memory area across multiple virtual machines for 8 For each of these pre-shared memory regions, a range is exposed under 9 the /reserved-memory node as a child node. Each range sub-node is named 13 compatible = "xen,shared-memory-v1" 16 the base guest physical address and size of the shared memory region 20 memory region used for the mapping in the borrower VM. 23 a string that identifies the shared memory region as specified in
|