1# SPDX-License-Identifier: GPL-2.0-only 2 3menu "Memory Management options" 4 5# 6# For some reason microblaze and nios2 hard code SWAP=n. Hopefully we can 7# add proper SWAP support to them, in which case this can be remove. 8# 9config ARCH_NO_SWAP 10 bool 11 12config ZPOOL 13 bool 14 15menuconfig SWAP 16 bool "Support for paging of anonymous memory (swap)" 17 depends on MMU && BLOCK && !ARCH_NO_SWAP 18 default y 19 help 20 This option allows you to choose whether you want to have support 21 for so called swap devices or swap files in your kernel that are 22 used to provide more virtual memory than the actual RAM present 23 in your computer. If unsure say Y. 24 25config ZSWAP 26 bool "Compressed cache for swap pages" 27 depends on SWAP 28 select FRONTSWAP 29 select CRYPTO 30 select ZPOOL 31 help 32 A lightweight compressed cache for swap pages. It takes 33 pages that are in the process of being swapped out and attempts to 34 compress them into a dynamically allocated RAM-based memory pool. 35 This can result in a significant I/O reduction on swap device and, 36 in the case where decompressing from RAM is faster than swap device 37 reads, can also improve workload performance. 38 39config ZSWAP_DEFAULT_ON 40 bool "Enable the compressed cache for swap pages by default" 41 depends on ZSWAP 42 help 43 If selected, the compressed cache for swap pages will be enabled 44 at boot, otherwise it will be disabled. 45 46 The selection made here can be overridden by using the kernel 47 command line 'zswap.enabled=' option. 48 49choice 50 prompt "Default compressor" 51 depends on ZSWAP 52 default ZSWAP_COMPRESSOR_DEFAULT_LZO 53 help 54 Selects the default compression algorithm for the compressed cache 55 for swap pages. 56 57 For an overview what kind of performance can be expected from 58 a particular compression algorithm please refer to the benchmarks 59 available at the following LWN page: 60 https://lwn.net/Articles/751795/ 61 62 If in doubt, select 'LZO'. 63 64 The selection made here can be overridden by using the kernel 65 command line 'zswap.compressor=' option. 66 67config ZSWAP_COMPRESSOR_DEFAULT_DEFLATE 68 bool "Deflate" 69 select CRYPTO_DEFLATE 70 help 71 Use the Deflate algorithm as the default compression algorithm. 72 73config ZSWAP_COMPRESSOR_DEFAULT_LZO 74 bool "LZO" 75 select CRYPTO_LZO 76 help 77 Use the LZO algorithm as the default compression algorithm. 78 79config ZSWAP_COMPRESSOR_DEFAULT_842 80 bool "842" 81 select CRYPTO_842 82 help 83 Use the 842 algorithm as the default compression algorithm. 84 85config ZSWAP_COMPRESSOR_DEFAULT_LZ4 86 bool "LZ4" 87 select CRYPTO_LZ4 88 help 89 Use the LZ4 algorithm as the default compression algorithm. 90 91config ZSWAP_COMPRESSOR_DEFAULT_LZ4HC 92 bool "LZ4HC" 93 select CRYPTO_LZ4HC 94 help 95 Use the LZ4HC algorithm as the default compression algorithm. 96 97config ZSWAP_COMPRESSOR_DEFAULT_ZSTD 98 bool "zstd" 99 select CRYPTO_ZSTD 100 help 101 Use the zstd algorithm as the default compression algorithm. 102endchoice 103 104config ZSWAP_COMPRESSOR_DEFAULT 105 string 106 depends on ZSWAP 107 default "deflate" if ZSWAP_COMPRESSOR_DEFAULT_DEFLATE 108 default "lzo" if ZSWAP_COMPRESSOR_DEFAULT_LZO 109 default "842" if ZSWAP_COMPRESSOR_DEFAULT_842 110 default "lz4" if ZSWAP_COMPRESSOR_DEFAULT_LZ4 111 default "lz4hc" if ZSWAP_COMPRESSOR_DEFAULT_LZ4HC 112 default "zstd" if ZSWAP_COMPRESSOR_DEFAULT_ZSTD 113 default "" 114 115choice 116 prompt "Default allocator" 117 depends on ZSWAP 118 default ZSWAP_ZPOOL_DEFAULT_ZBUD 119 help 120 Selects the default allocator for the compressed cache for 121 swap pages. 122 The default is 'zbud' for compatibility, however please do 123 read the description of each of the allocators below before 124 making a right choice. 125 126 The selection made here can be overridden by using the kernel 127 command line 'zswap.zpool=' option. 128 129config ZSWAP_ZPOOL_DEFAULT_ZBUD 130 bool "zbud" 131 select ZBUD 132 help 133 Use the zbud allocator as the default allocator. 134 135config ZSWAP_ZPOOL_DEFAULT_Z3FOLD 136 bool "z3fold" 137 select Z3FOLD 138 help 139 Use the z3fold allocator as the default allocator. 140 141config ZSWAP_ZPOOL_DEFAULT_ZSMALLOC 142 bool "zsmalloc" 143 select ZSMALLOC 144 help 145 Use the zsmalloc allocator as the default allocator. 146endchoice 147 148config ZSWAP_ZPOOL_DEFAULT 149 string 150 depends on ZSWAP 151 default "zbud" if ZSWAP_ZPOOL_DEFAULT_ZBUD 152 default "z3fold" if ZSWAP_ZPOOL_DEFAULT_Z3FOLD 153 default "zsmalloc" if ZSWAP_ZPOOL_DEFAULT_ZSMALLOC 154 default "" 155 156config ZBUD 157 tristate "2:1 compression allocator (zbud)" 158 depends on ZSWAP 159 help 160 A special purpose allocator for storing compressed pages. 161 It is designed to store up to two compressed pages per physical 162 page. While this design limits storage density, it has simple and 163 deterministic reclaim properties that make it preferable to a higher 164 density approach when reclaim will be used. 165 166config Z3FOLD 167 tristate "3:1 compression allocator (z3fold)" 168 depends on ZSWAP 169 help 170 A special purpose allocator for storing compressed pages. 171 It is designed to store up to three compressed pages per physical 172 page. It is a ZBUD derivative so the simplicity and determinism are 173 still there. 174 175config ZSMALLOC 176 tristate 177 prompt "N:1 compression allocator (zsmalloc)" if ZSWAP 178 depends on MMU 179 help 180 zsmalloc is a slab-based memory allocator designed to store 181 pages of various compression levels efficiently. It achieves 182 the highest storage density with the least amount of fragmentation. 183 184config ZSMALLOC_STAT 185 bool "Export zsmalloc statistics" 186 depends on ZSMALLOC 187 select DEBUG_FS 188 help 189 This option enables code in the zsmalloc to collect various 190 statistics about what's happening in zsmalloc and exports that 191 information to userspace via debugfs. 192 If unsure, say N. 193 194config ZSMALLOC_CHAIN_SIZE 195 int "Maximum number of physical pages per-zspage" 196 default 8 197 range 4 16 198 depends on ZSMALLOC 199 help 200 This option sets the upper limit on the number of physical pages 201 that a zmalloc page (zspage) can consist of. The optimal zspage 202 chain size is calculated for each size class during the 203 initialization of the pool. 204 205 Changing this option can alter the characteristics of size classes, 206 such as the number of pages per zspage and the number of objects 207 per zspage. This can also result in different configurations of 208 the pool, as zsmalloc merges size classes with similar 209 characteristics. 210 211 For more information, see zsmalloc documentation. 212 213menu "SLAB allocator options" 214 215choice 216 prompt "Choose SLAB allocator" 217 default SLUB 218 help 219 This option allows to select a slab allocator. 220 221config SLAB 222 bool "SLAB" 223 depends on !PREEMPT_RT 224 select HAVE_HARDENED_USERCOPY_ALLOCATOR 225 help 226 The regular slab allocator that is established and known to work 227 well in all environments. It organizes cache hot objects in 228 per cpu and per node queues. 229 230config SLUB 231 bool "SLUB (Unqueued Allocator)" 232 select HAVE_HARDENED_USERCOPY_ALLOCATOR 233 help 234 SLUB is a slab allocator that minimizes cache line usage 235 instead of managing queues of cached objects (SLAB approach). 236 Per cpu caching is realized using slabs of objects instead 237 of queues of objects. SLUB can use memory efficiently 238 and has enhanced diagnostics. SLUB is the default choice for 239 a slab allocator. 240 241endchoice 242 243config SLUB_TINY 244 bool "Configure SLUB for minimal memory footprint" 245 depends on SLUB && EXPERT 246 select SLAB_MERGE_DEFAULT 247 help 248 Configures the SLUB allocator in a way to achieve minimal memory 249 footprint, sacrificing scalability, debugging and other features. 250 This is intended only for the smallest system that had used the 251 SLOB allocator and is not recommended for systems with more than 252 16MB RAM. 253 254 If unsure, say N. 255 256config SLAB_MERGE_DEFAULT 257 bool "Allow slab caches to be merged" 258 default y 259 depends on SLAB || SLUB 260 help 261 For reduced kernel memory fragmentation, slab caches can be 262 merged when they share the same size and other characteristics. 263 This carries a risk of kernel heap overflows being able to 264 overwrite objects from merged caches (and more easily control 265 cache layout), which makes such heap attacks easier to exploit 266 by attackers. By keeping caches unmerged, these kinds of exploits 267 can usually only damage objects in the same cache. To disable 268 merging at runtime, "slab_nomerge" can be passed on the kernel 269 command line. 270 271config SLAB_FREELIST_RANDOM 272 bool "Randomize slab freelist" 273 depends on SLAB || (SLUB && !SLUB_TINY) 274 help 275 Randomizes the freelist order used on creating new pages. This 276 security feature reduces the predictability of the kernel slab 277 allocator against heap overflows. 278 279config SLAB_FREELIST_HARDENED 280 bool "Harden slab freelist metadata" 281 depends on SLAB || (SLUB && !SLUB_TINY) 282 help 283 Many kernel heap attacks try to target slab cache metadata and 284 other infrastructure. This options makes minor performance 285 sacrifices to harden the kernel slab allocator against common 286 freelist exploit methods. Some slab implementations have more 287 sanity-checking than others. This option is most effective with 288 CONFIG_SLUB. 289 290config SLUB_STATS 291 default n 292 bool "Enable SLUB performance statistics" 293 depends on SLUB && SYSFS && !SLUB_TINY 294 help 295 SLUB statistics are useful to debug SLUBs allocation behavior in 296 order find ways to optimize the allocator. This should never be 297 enabled for production use since keeping statistics slows down 298 the allocator by a few percentage points. The slabinfo command 299 supports the determination of the most active slabs to figure 300 out which slabs are relevant to a particular load. 301 Try running: slabinfo -DA 302 303config SLUB_CPU_PARTIAL 304 default y 305 depends on SLUB && SMP && !SLUB_TINY 306 bool "SLUB per cpu partial cache" 307 help 308 Per cpu partial caches accelerate objects allocation and freeing 309 that is local to a processor at the price of more indeterminism 310 in the latency of the free. On overflow these caches will be cleared 311 which requires the taking of locks that may cause latency spikes. 312 Typically one would choose no for a realtime system. 313 314endmenu # SLAB allocator options 315 316config SHUFFLE_PAGE_ALLOCATOR 317 bool "Page allocator randomization" 318 default SLAB_FREELIST_RANDOM && ACPI_NUMA 319 help 320 Randomization of the page allocator improves the average 321 utilization of a direct-mapped memory-side-cache. See section 322 5.2.27 Heterogeneous Memory Attribute Table (HMAT) in the ACPI 323 6.2a specification for an example of how a platform advertises 324 the presence of a memory-side-cache. There are also incidental 325 security benefits as it reduces the predictability of page 326 allocations to compliment SLAB_FREELIST_RANDOM, but the 327 default granularity of shuffling on the MAX_ORDER i.e, 10th 328 order of pages is selected based on cache utilization benefits 329 on x86. 330 331 While the randomization improves cache utilization it may 332 negatively impact workloads on platforms without a cache. For 333 this reason, by default, the randomization is enabled only 334 after runtime detection of a direct-mapped memory-side-cache. 335 Otherwise, the randomization may be force enabled with the 336 'page_alloc.shuffle' kernel command line parameter. 337 338 Say Y if unsure. 339 340config COMPAT_BRK 341 bool "Disable heap randomization" 342 default y 343 help 344 Randomizing heap placement makes heap exploits harder, but it 345 also breaks ancient binaries (including anything libc5 based). 346 This option changes the bootup default to heap randomization 347 disabled, and can be overridden at runtime by setting 348 /proc/sys/kernel/randomize_va_space to 2. 349 350 On non-ancient distros (post-2000 ones) N is usually a safe choice. 351 352config MMAP_ALLOW_UNINITIALIZED 353 bool "Allow mmapped anonymous memory to be uninitialized" 354 depends on EXPERT && !MMU 355 default n 356 help 357 Normally, and according to the Linux spec, anonymous memory obtained 358 from mmap() has its contents cleared before it is passed to 359 userspace. Enabling this config option allows you to request that 360 mmap() skip that if it is given an MAP_UNINITIALIZED flag, thus 361 providing a huge performance boost. If this option is not enabled, 362 then the flag will be ignored. 363 364 This is taken advantage of by uClibc's malloc(), and also by 365 ELF-FDPIC binfmt's brk and stack allocator. 366 367 Because of the obvious security issues, this option should only be 368 enabled on embedded devices where you control what is run in 369 userspace. Since that isn't generally a problem on no-MMU systems, 370 it is normally safe to say Y here. 371 372 See Documentation/admin-guide/mm/nommu-mmap.rst for more information. 373 374config SELECT_MEMORY_MODEL 375 def_bool y 376 depends on ARCH_SELECT_MEMORY_MODEL 377 378choice 379 prompt "Memory model" 380 depends on SELECT_MEMORY_MODEL 381 default SPARSEMEM_MANUAL if ARCH_SPARSEMEM_DEFAULT 382 default FLATMEM_MANUAL 383 help 384 This option allows you to change some of the ways that 385 Linux manages its memory internally. Most users will 386 only have one option here selected by the architecture 387 configuration. This is normal. 388 389config FLATMEM_MANUAL 390 bool "Flat Memory" 391 depends on !ARCH_SPARSEMEM_ENABLE || ARCH_FLATMEM_ENABLE 392 help 393 This option is best suited for non-NUMA systems with 394 flat address space. The FLATMEM is the most efficient 395 system in terms of performance and resource consumption 396 and it is the best option for smaller systems. 397 398 For systems that have holes in their physical address 399 spaces and for features like NUMA and memory hotplug, 400 choose "Sparse Memory". 401 402 If unsure, choose this option (Flat Memory) over any other. 403 404config SPARSEMEM_MANUAL 405 bool "Sparse Memory" 406 depends on ARCH_SPARSEMEM_ENABLE 407 help 408 This will be the only option for some systems, including 409 memory hot-plug systems. This is normal. 410 411 This option provides efficient support for systems with 412 holes is their physical address space and allows memory 413 hot-plug and hot-remove. 414 415 If unsure, choose "Flat Memory" over this option. 416 417endchoice 418 419config SPARSEMEM 420 def_bool y 421 depends on (!SELECT_MEMORY_MODEL && ARCH_SPARSEMEM_ENABLE) || SPARSEMEM_MANUAL 422 423config FLATMEM 424 def_bool y 425 depends on !SPARSEMEM || FLATMEM_MANUAL 426 427# 428# SPARSEMEM_EXTREME (which is the default) does some bootmem 429# allocations when sparse_init() is called. If this cannot 430# be done on your architecture, select this option. However, 431# statically allocating the mem_section[] array can potentially 432# consume vast quantities of .bss, so be careful. 433# 434# This option will also potentially produce smaller runtime code 435# with gcc 3.4 and later. 436# 437config SPARSEMEM_STATIC 438 bool 439 440# 441# Architecture platforms which require a two level mem_section in SPARSEMEM 442# must select this option. This is usually for architecture platforms with 443# an extremely sparse physical address space. 444# 445config SPARSEMEM_EXTREME 446 def_bool y 447 depends on SPARSEMEM && !SPARSEMEM_STATIC 448 449config SPARSEMEM_VMEMMAP_ENABLE 450 bool 451 452config SPARSEMEM_VMEMMAP 453 bool "Sparse Memory virtual memmap" 454 depends on SPARSEMEM && SPARSEMEM_VMEMMAP_ENABLE 455 default y 456 help 457 SPARSEMEM_VMEMMAP uses a virtually mapped memmap to optimise 458 pfn_to_page and page_to_pfn operations. This is the most 459 efficient option when sufficient kernel resources are available. 460# 461# Select this config option from the architecture Kconfig, if it is preferred 462# to enable the feature of HugeTLB/dev_dax vmemmap optimization. 463# 464config ARCH_WANT_OPTIMIZE_VMEMMAP 465 bool 466 467config HAVE_MEMBLOCK_PHYS_MAP 468 bool 469 470config HAVE_FAST_GUP 471 depends on MMU 472 bool 473 474# Don't discard allocated memory used to track "memory" and "reserved" memblocks 475# after early boot, so it can still be used to test for validity of memory. 476# Also, memblocks are updated with memory hot(un)plug. 477config ARCH_KEEP_MEMBLOCK 478 bool 479 480# Keep arch NUMA mapping infrastructure post-init. 481config NUMA_KEEP_MEMINFO 482 bool 483 484config MEMORY_ISOLATION 485 bool 486 487# IORESOURCE_SYSTEM_RAM regions in the kernel resource tree that are marked 488# IORESOURCE_EXCLUSIVE cannot be mapped to user space, for example, via 489# /dev/mem. 490config EXCLUSIVE_SYSTEM_RAM 491 def_bool y 492 depends on !DEVMEM || STRICT_DEVMEM 493 494# 495# Only be set on architectures that have completely implemented memory hotplug 496# feature. If you are not sure, don't touch it. 497# 498config HAVE_BOOTMEM_INFO_NODE 499 def_bool n 500 501config ARCH_ENABLE_MEMORY_HOTPLUG 502 bool 503 504config ARCH_ENABLE_MEMORY_HOTREMOVE 505 bool 506 507# eventually, we can have this option just 'select SPARSEMEM' 508menuconfig MEMORY_HOTPLUG 509 bool "Memory hotplug" 510 select MEMORY_ISOLATION 511 depends on SPARSEMEM 512 depends on ARCH_ENABLE_MEMORY_HOTPLUG 513 depends on 64BIT 514 select NUMA_KEEP_MEMINFO if NUMA 515 516if MEMORY_HOTPLUG 517 518config MEMORY_HOTPLUG_DEFAULT_ONLINE 519 bool "Online the newly added memory blocks by default" 520 depends on MEMORY_HOTPLUG 521 help 522 This option sets the default policy setting for memory hotplug 523 onlining policy (/sys/devices/system/memory/auto_online_blocks) which 524 determines what happens to newly added memory regions. Policy setting 525 can always be changed at runtime. 526 See Documentation/admin-guide/mm/memory-hotplug.rst for more information. 527 528 Say Y here if you want all hot-plugged memory blocks to appear in 529 'online' state by default. 530 Say N here if you want the default policy to keep all hot-plugged 531 memory blocks in 'offline' state. 532 533config MEMORY_HOTREMOVE 534 bool "Allow for memory hot remove" 535 select HAVE_BOOTMEM_INFO_NODE if (X86_64 || PPC64) 536 depends on MEMORY_HOTPLUG && ARCH_ENABLE_MEMORY_HOTREMOVE 537 depends on MIGRATION 538 539config MHP_MEMMAP_ON_MEMORY 540 def_bool y 541 depends on MEMORY_HOTPLUG && SPARSEMEM_VMEMMAP 542 depends on ARCH_MHP_MEMMAP_ON_MEMORY_ENABLE 543 544endif # MEMORY_HOTPLUG 545 546# Heavily threaded applications may benefit from splitting the mm-wide 547# page_table_lock, so that faults on different parts of the user address 548# space can be handled with less contention: split it at this NR_CPUS. 549# Default to 4 for wider testing, though 8 might be more appropriate. 550# ARM's adjust_pte (unused if VIPT) depends on mm-wide page_table_lock. 551# PA-RISC 7xxx's spinlock_t would enlarge struct page from 32 to 44 bytes. 552# SPARC32 allocates multiple pte tables within a single page, and therefore 553# a per-page lock leads to problems when multiple tables need to be locked 554# at the same time (e.g. copy_page_range()). 555# DEBUG_SPINLOCK and DEBUG_LOCK_ALLOC spinlock_t also enlarge struct page. 556# 557config SPLIT_PTLOCK_CPUS 558 int 559 default "999999" if !MMU 560 default "999999" if ARM && !CPU_CACHE_VIPT 561 default "999999" if PARISC && !PA20 562 default "999999" if SPARC32 563 default "4" 564 565config ARCH_ENABLE_SPLIT_PMD_PTLOCK 566 bool 567 568# 569# support for memory balloon 570config MEMORY_BALLOON 571 bool 572 573# 574# support for memory balloon compaction 575config BALLOON_COMPACTION 576 bool "Allow for balloon memory compaction/migration" 577 def_bool y 578 depends on COMPACTION && MEMORY_BALLOON 579 help 580 Memory fragmentation introduced by ballooning might reduce 581 significantly the number of 2MB contiguous memory blocks that can be 582 used within a guest, thus imposing performance penalties associated 583 with the reduced number of transparent huge pages that could be used 584 by the guest workload. Allowing the compaction & migration for memory 585 pages enlisted as being part of memory balloon devices avoids the 586 scenario aforementioned and helps improving memory defragmentation. 587 588# 589# support for memory compaction 590config COMPACTION 591 bool "Allow for memory compaction" 592 def_bool y 593 select MIGRATION 594 depends on MMU 595 help 596 Compaction is the only memory management component to form 597 high order (larger physically contiguous) memory blocks 598 reliably. The page allocator relies on compaction heavily and 599 the lack of the feature can lead to unexpected OOM killer 600 invocations for high order memory requests. You shouldn't 601 disable this option unless there really is a strong reason for 602 it and then we would be really interested to hear about that at 603 linux-mm@kvack.org. 604 605config COMPACT_UNEVICTABLE_DEFAULT 606 int 607 depends on COMPACTION 608 default 0 if PREEMPT_RT 609 default 1 610 611# 612# support for free page reporting 613config PAGE_REPORTING 614 bool "Free page reporting" 615 def_bool n 616 help 617 Free page reporting allows for the incremental acquisition of 618 free pages from the buddy allocator for the purpose of reporting 619 those pages to another entity, such as a hypervisor, so that the 620 memory can be freed within the host for other uses. 621 622# 623# support for page migration 624# 625config MIGRATION 626 bool "Page migration" 627 def_bool y 628 depends on (NUMA || ARCH_ENABLE_MEMORY_HOTREMOVE || COMPACTION || CMA) && MMU 629 help 630 Allows the migration of the physical location of pages of processes 631 while the virtual addresses are not changed. This is useful in 632 two situations. The first is on NUMA systems to put pages nearer 633 to the processors accessing. The second is when allocating huge 634 pages as migration can relocate pages to satisfy a huge page 635 allocation instead of reclaiming. 636 637config DEVICE_MIGRATION 638 def_bool MIGRATION && ZONE_DEVICE 639 640config ARCH_ENABLE_HUGEPAGE_MIGRATION 641 bool 642 643config ARCH_ENABLE_THP_MIGRATION 644 bool 645 646config HUGETLB_PAGE_SIZE_VARIABLE 647 def_bool n 648 help 649 Allows the pageblock_order value to be dynamic instead of just standard 650 HUGETLB_PAGE_ORDER when there are multiple HugeTLB page sizes available 651 on a platform. 652 653 Note that the pageblock_order cannot exceed MAX_ORDER and will be 654 clamped down to MAX_ORDER. 655 656config CONTIG_ALLOC 657 def_bool (MEMORY_ISOLATION && COMPACTION) || CMA 658 659config PHYS_ADDR_T_64BIT 660 def_bool 64BIT 661 662config BOUNCE 663 bool "Enable bounce buffers" 664 default y 665 depends on BLOCK && MMU && HIGHMEM 666 help 667 Enable bounce buffers for devices that cannot access the full range of 668 memory available to the CPU. Enabled by default when HIGHMEM is 669 selected, but you may say n to override this. 670 671config MMU_NOTIFIER 672 bool 673 select INTERVAL_TREE 674 675config KSM 676 bool "Enable KSM for page merging" 677 depends on MMU 678 select XXHASH 679 help 680 Enable Kernel Samepage Merging: KSM periodically scans those areas 681 of an application's address space that an app has advised may be 682 mergeable. When it finds pages of identical content, it replaces 683 the many instances by a single page with that content, so 684 saving memory until one or another app needs to modify the content. 685 Recommended for use with KVM, or with other duplicative applications. 686 See Documentation/mm/ksm.rst for more information: KSM is inactive 687 until a program has madvised that an area is MADV_MERGEABLE, and 688 root has set /sys/kernel/mm/ksm/run to 1 (if CONFIG_SYSFS is set). 689 690config DEFAULT_MMAP_MIN_ADDR 691 int "Low address space to protect from user allocation" 692 depends on MMU 693 default 4096 694 help 695 This is the portion of low virtual memory which should be protected 696 from userspace allocation. Keeping a user from writing to low pages 697 can help reduce the impact of kernel NULL pointer bugs. 698 699 For most ia64, ppc64 and x86 users with lots of address space 700 a value of 65536 is reasonable and should cause no problems. 701 On arm and other archs it should not be higher than 32768. 702 Programs which use vm86 functionality or have some need to map 703 this low address space will need CAP_SYS_RAWIO or disable this 704 protection by setting the value to 0. 705 706 This value can be changed after boot using the 707 /proc/sys/vm/mmap_min_addr tunable. 708 709config ARCH_SUPPORTS_MEMORY_FAILURE 710 bool 711 712config MEMORY_FAILURE 713 depends on MMU 714 depends on ARCH_SUPPORTS_MEMORY_FAILURE 715 bool "Enable recovery from hardware memory errors" 716 select MEMORY_ISOLATION 717 select RAS 718 help 719 Enables code to recover from some memory failures on systems 720 with MCA recovery. This allows a system to continue running 721 even when some of its memory has uncorrected errors. This requires 722 special hardware support and typically ECC memory. 723 724config HWPOISON_INJECT 725 tristate "HWPoison pages injector" 726 depends on MEMORY_FAILURE && DEBUG_KERNEL && PROC_FS 727 select PROC_PAGE_MONITOR 728 729config NOMMU_INITIAL_TRIM_EXCESS 730 int "Turn on mmap() excess space trimming before booting" 731 depends on !MMU 732 default 1 733 help 734 The NOMMU mmap() frequently needs to allocate large contiguous chunks 735 of memory on which to store mappings, but it can only ask the system 736 allocator for chunks in 2^N*PAGE_SIZE amounts - which is frequently 737 more than it requires. To deal with this, mmap() is able to trim off 738 the excess and return it to the allocator. 739 740 If trimming is enabled, the excess is trimmed off and returned to the 741 system allocator, which can cause extra fragmentation, particularly 742 if there are a lot of transient processes. 743 744 If trimming is disabled, the excess is kept, but not used, which for 745 long-term mappings means that the space is wasted. 746 747 Trimming can be dynamically controlled through a sysctl option 748 (/proc/sys/vm/nr_trim_pages) which specifies the minimum number of 749 excess pages there must be before trimming should occur, or zero if 750 no trimming is to occur. 751 752 This option specifies the initial value of this option. The default 753 of 1 says that all excess pages should be trimmed. 754 755 See Documentation/admin-guide/mm/nommu-mmap.rst for more information. 756 757config ARCH_WANT_GENERAL_HUGETLB 758 bool 759 760config ARCH_WANTS_THP_SWAP 761 def_bool n 762 763menuconfig TRANSPARENT_HUGEPAGE 764 bool "Transparent Hugepage Support" 765 depends on HAVE_ARCH_TRANSPARENT_HUGEPAGE && !PREEMPT_RT 766 select COMPACTION 767 select XARRAY_MULTI 768 help 769 Transparent Hugepages allows the kernel to use huge pages and 770 huge tlb transparently to the applications whenever possible. 771 This feature can improve computing performance to certain 772 applications by speeding up page faults during memory 773 allocation, by reducing the number of tlb misses and by speeding 774 up the pagetable walking. 775 776 If memory constrained on embedded, you may want to say N. 777 778if TRANSPARENT_HUGEPAGE 779 780choice 781 prompt "Transparent Hugepage Support sysfs defaults" 782 depends on TRANSPARENT_HUGEPAGE 783 default TRANSPARENT_HUGEPAGE_ALWAYS 784 help 785 Selects the sysfs defaults for Transparent Hugepage Support. 786 787 config TRANSPARENT_HUGEPAGE_ALWAYS 788 bool "always" 789 help 790 Enabling Transparent Hugepage always, can increase the 791 memory footprint of applications without a guaranteed 792 benefit but it will work automatically for all applications. 793 794 config TRANSPARENT_HUGEPAGE_MADVISE 795 bool "madvise" 796 help 797 Enabling Transparent Hugepage madvise, will only provide a 798 performance improvement benefit to the applications using 799 madvise(MADV_HUGEPAGE) but it won't risk to increase the 800 memory footprint of applications without a guaranteed 801 benefit. 802endchoice 803 804config THP_SWAP 805 def_bool y 806 depends on TRANSPARENT_HUGEPAGE && ARCH_WANTS_THP_SWAP && SWAP && 64BIT 807 help 808 Swap transparent huge pages in one piece, without splitting. 809 XXX: For now, swap cluster backing transparent huge page 810 will be split after swapout. 811 812 For selection by architectures with reasonable THP sizes. 813 814config READ_ONLY_THP_FOR_FS 815 bool "Read-only THP for filesystems (EXPERIMENTAL)" 816 depends on TRANSPARENT_HUGEPAGE && SHMEM 817 818 help 819 Allow khugepaged to put read-only file-backed pages in THP. 820 821 This is marked experimental because it is a new feature. Write 822 support of file THPs will be developed in the next few release 823 cycles. 824 825endif # TRANSPARENT_HUGEPAGE 826 827# 828# UP and nommu archs use km based percpu allocator 829# 830config NEED_PER_CPU_KM 831 depends on !SMP || !MMU 832 bool 833 default y 834 835config NEED_PER_CPU_EMBED_FIRST_CHUNK 836 bool 837 838config NEED_PER_CPU_PAGE_FIRST_CHUNK 839 bool 840 841config USE_PERCPU_NUMA_NODE_ID 842 bool 843 844config HAVE_SETUP_PER_CPU_AREA 845 bool 846 847config FRONTSWAP 848 bool 849 850config CMA 851 bool "Contiguous Memory Allocator" 852 depends on MMU 853 select MIGRATION 854 select MEMORY_ISOLATION 855 help 856 This enables the Contiguous Memory Allocator which allows other 857 subsystems to allocate big physically-contiguous blocks of memory. 858 CMA reserves a region of memory and allows only movable pages to 859 be allocated from it. This way, the kernel can use the memory for 860 pagecache and when a subsystem requests for contiguous area, the 861 allocated pages are migrated away to serve the contiguous request. 862 863 If unsure, say "n". 864 865config CMA_DEBUG 866 bool "CMA debug messages (DEVELOPMENT)" 867 depends on DEBUG_KERNEL && CMA 868 help 869 Turns on debug messages in CMA. This produces KERN_DEBUG 870 messages for every CMA call as well as various messages while 871 processing calls such as dma_alloc_from_contiguous(). 872 This option does not affect warning and error messages. 873 874config CMA_DEBUGFS 875 bool "CMA debugfs interface" 876 depends on CMA && DEBUG_FS 877 help 878 Turns on the DebugFS interface for CMA. 879 880config CMA_SYSFS 881 bool "CMA information through sysfs interface" 882 depends on CMA && SYSFS 883 help 884 This option exposes some sysfs attributes to get information 885 from CMA. 886 887config CMA_AREAS 888 int "Maximum count of the CMA areas" 889 depends on CMA 890 default 19 if NUMA 891 default 7 892 help 893 CMA allows to create CMA areas for particular purpose, mainly, 894 used as device private area. This parameter sets the maximum 895 number of CMA area in the system. 896 897 If unsure, leave the default value "7" in UMA and "19" in NUMA. 898 899config MEM_SOFT_DIRTY 900 bool "Track memory changes" 901 depends on CHECKPOINT_RESTORE && HAVE_ARCH_SOFT_DIRTY && PROC_FS 902 select PROC_PAGE_MONITOR 903 help 904 This option enables memory changes tracking by introducing a 905 soft-dirty bit on pte-s. This bit it set when someone writes 906 into a page just as regular dirty bit, but unlike the latter 907 it can be cleared by hands. 908 909 See Documentation/admin-guide/mm/soft-dirty.rst for more details. 910 911config GENERIC_EARLY_IOREMAP 912 bool 913 914config STACK_MAX_DEFAULT_SIZE_MB 915 int "Default maximum user stack size for 32-bit processes (MB)" 916 default 100 917 range 8 2048 918 depends on STACK_GROWSUP && (!64BIT || COMPAT) 919 help 920 This is the maximum stack size in Megabytes in the VM layout of 32-bit 921 user processes when the stack grows upwards (currently only on parisc 922 arch) when the RLIMIT_STACK hard limit is unlimited. 923 924 A sane initial value is 100 MB. 925 926config DEFERRED_STRUCT_PAGE_INIT 927 bool "Defer initialisation of struct pages to kthreads" 928 depends on SPARSEMEM 929 depends on !NEED_PER_CPU_KM 930 depends on 64BIT 931 select PADATA 932 help 933 Ordinarily all struct pages are initialised during early boot in a 934 single thread. On very large machines this can take a considerable 935 amount of time. If this option is set, large machines will bring up 936 a subset of memmap at boot and then initialise the rest in parallel. 937 This has a potential performance impact on tasks running early in the 938 lifetime of the system until these kthreads finish the 939 initialisation. 940 941config PAGE_IDLE_FLAG 942 bool 943 select PAGE_EXTENSION if !64BIT 944 help 945 This adds PG_idle and PG_young flags to 'struct page'. PTE Accessed 946 bit writers can set the state of the bit in the flags so that PTE 947 Accessed bit readers may avoid disturbance. 948 949config IDLE_PAGE_TRACKING 950 bool "Enable idle page tracking" 951 depends on SYSFS && MMU 952 select PAGE_IDLE_FLAG 953 help 954 This feature allows to estimate the amount of user pages that have 955 not been touched during a given period of time. This information can 956 be useful to tune memory cgroup limits and/or for job placement 957 within a compute cluster. 958 959 See Documentation/admin-guide/mm/idle_page_tracking.rst for 960 more details. 961 962config ARCH_HAS_CACHE_LINE_SIZE 963 bool 964 965config ARCH_HAS_CURRENT_STACK_POINTER 966 bool 967 help 968 In support of HARDENED_USERCOPY performing stack variable lifetime 969 checking, an architecture-agnostic way to find the stack pointer 970 is needed. Once an architecture defines an unsigned long global 971 register alias named "current_stack_pointer", this config can be 972 selected. 973 974config ARCH_HAS_PTE_DEVMAP 975 bool 976 977config ARCH_HAS_ZONE_DMA_SET 978 bool 979 980config ZONE_DMA 981 bool "Support DMA zone" if ARCH_HAS_ZONE_DMA_SET 982 default y if ARM64 || X86 983 984config ZONE_DMA32 985 bool "Support DMA32 zone" if ARCH_HAS_ZONE_DMA_SET 986 depends on !X86_32 987 default y if ARM64 988 989config ZONE_DEVICE 990 bool "Device memory (pmem, HMM, etc...) hotplug support" 991 depends on MEMORY_HOTPLUG 992 depends on MEMORY_HOTREMOVE 993 depends on SPARSEMEM_VMEMMAP 994 depends on ARCH_HAS_PTE_DEVMAP 995 select XARRAY_MULTI 996 997 help 998 Device memory hotplug support allows for establishing pmem, 999 or other device driver discovered memory regions, in the 1000 memmap. This allows pfn_to_page() lookups of otherwise 1001 "device-physical" addresses which is needed for using a DAX 1002 mapping in an O_DIRECT operation, among other things. 1003 1004 If FS_DAX is enabled, then say Y. 1005 1006# 1007# Helpers to mirror range of the CPU page tables of a process into device page 1008# tables. 1009# 1010config HMM_MIRROR 1011 bool 1012 depends on MMU 1013 1014config GET_FREE_REGION 1015 depends on SPARSEMEM 1016 bool 1017 1018config DEVICE_PRIVATE 1019 bool "Unaddressable device memory (GPU memory, ...)" 1020 depends on ZONE_DEVICE 1021 select GET_FREE_REGION 1022 1023 help 1024 Allows creation of struct pages to represent unaddressable device 1025 memory; i.e., memory that is only accessible from the device (or 1026 group of devices). You likely also want to select HMM_MIRROR. 1027 1028config VMAP_PFN 1029 bool 1030 1031config ARCH_USES_HIGH_VMA_FLAGS 1032 bool 1033config ARCH_HAS_PKEYS 1034 bool 1035 1036config ARCH_USES_PG_ARCH_X 1037 bool 1038 help 1039 Enable the definition of PG_arch_x page flags with x > 1. Only 1040 suitable for 64-bit architectures with CONFIG_FLATMEM or 1041 CONFIG_SPARSEMEM_VMEMMAP enabled, otherwise there may not be 1042 enough room for additional bits in page->flags. 1043 1044config VM_EVENT_COUNTERS 1045 default y 1046 bool "Enable VM event counters for /proc/vmstat" if EXPERT 1047 help 1048 VM event counters are needed for event counts to be shown. 1049 This option allows the disabling of the VM event counters 1050 on EXPERT systems. /proc/vmstat will only show page counts 1051 if VM event counters are disabled. 1052 1053config PERCPU_STATS 1054 bool "Collect percpu memory statistics" 1055 help 1056 This feature collects and exposes statistics via debugfs. The 1057 information includes global and per chunk statistics, which can 1058 be used to help understand percpu memory usage. 1059 1060config GUP_TEST 1061 bool "Enable infrastructure for get_user_pages()-related unit tests" 1062 depends on DEBUG_FS 1063 help 1064 Provides /sys/kernel/debug/gup_test, which in turn provides a way 1065 to make ioctl calls that can launch kernel-based unit tests for 1066 the get_user_pages*() and pin_user_pages*() family of API calls. 1067 1068 These tests include benchmark testing of the _fast variants of 1069 get_user_pages*() and pin_user_pages*(), as well as smoke tests of 1070 the non-_fast variants. 1071 1072 There is also a sub-test that allows running dump_page() on any 1073 of up to eight pages (selected by command line args) within the 1074 range of user-space addresses. These pages are either pinned via 1075 pin_user_pages*(), or pinned via get_user_pages*(), as specified 1076 by other command line arguments. 1077 1078 See tools/testing/selftests/mm/gup_test.c 1079 1080comment "GUP_TEST needs to have DEBUG_FS enabled" 1081 depends on !GUP_TEST && !DEBUG_FS 1082 1083config GUP_GET_PXX_LOW_HIGH 1084 bool 1085 1086config DMAPOOL_TEST 1087 tristate "Enable a module to run time tests on dma_pool" 1088 depends on HAS_DMA 1089 help 1090 Provides a test module that will allocate and free many blocks of 1091 various sizes and report how long it takes. This is intended to 1092 provide a consistent way to measure how changes to the 1093 dma_pool_alloc/free routines affect performance. 1094 1095config ARCH_HAS_PTE_SPECIAL 1096 bool 1097 1098# 1099# Some architectures require a special hugepage directory format that is 1100# required to support multiple hugepage sizes. For example a4fe3ce76 1101# "powerpc/mm: Allow more flexible layouts for hugepage pagetables" 1102# introduced it on powerpc. This allows for a more flexible hugepage 1103# pagetable layouts. 1104# 1105config ARCH_HAS_HUGEPD 1106 bool 1107 1108config MAPPING_DIRTY_HELPERS 1109 bool 1110 1111config KMAP_LOCAL 1112 bool 1113 1114config KMAP_LOCAL_NON_LINEAR_PTE_ARRAY 1115 bool 1116 1117# struct io_mapping based helper. Selected by drivers that need them 1118config IO_MAPPING 1119 bool 1120 1121config SECRETMEM 1122 default y 1123 bool "Enable memfd_secret() system call" if EXPERT 1124 depends on ARCH_HAS_SET_DIRECT_MAP 1125 help 1126 Enable the memfd_secret() system call with the ability to create 1127 memory areas visible only in the context of the owning process and 1128 not mapped to other processes and other kernel page tables. 1129 1130config ANON_VMA_NAME 1131 bool "Anonymous VMA name support" 1132 depends on PROC_FS && ADVISE_SYSCALLS && MMU 1133 1134 help 1135 Allow naming anonymous virtual memory areas. 1136 1137 This feature allows assigning names to virtual memory areas. Assigned 1138 names can be later retrieved from /proc/pid/maps and /proc/pid/smaps 1139 and help identifying individual anonymous memory areas. 1140 Assigning a name to anonymous virtual memory area might prevent that 1141 area from being merged with adjacent virtual memory areas due to the 1142 difference in their name. 1143 1144config USERFAULTFD 1145 bool "Enable userfaultfd() system call" 1146 depends on MMU 1147 help 1148 Enable the userfaultfd() system call that allows to intercept and 1149 handle page faults in userland. 1150 1151config HAVE_ARCH_USERFAULTFD_WP 1152 bool 1153 help 1154 Arch has userfaultfd write protection support 1155 1156config HAVE_ARCH_USERFAULTFD_MINOR 1157 bool 1158 help 1159 Arch has userfaultfd minor fault support 1160 1161config PTE_MARKER_UFFD_WP 1162 bool "Userfaultfd write protection support for shmem/hugetlbfs" 1163 default y 1164 depends on HAVE_ARCH_USERFAULTFD_WP 1165 1166 help 1167 Allows to create marker PTEs for userfaultfd write protection 1168 purposes. It is required to enable userfaultfd write protection on 1169 file-backed memory types like shmem and hugetlbfs. 1170 1171# multi-gen LRU { 1172config LRU_GEN 1173 bool "Multi-Gen LRU" 1174 depends on MMU 1175 # make sure folio->flags has enough spare bits 1176 depends on 64BIT || !SPARSEMEM || SPARSEMEM_VMEMMAP 1177 help 1178 A high performance LRU implementation to overcommit memory. See 1179 Documentation/admin-guide/mm/multigen_lru.rst for details. 1180 1181config LRU_GEN_ENABLED 1182 bool "Enable by default" 1183 depends on LRU_GEN 1184 help 1185 This option enables the multi-gen LRU by default. 1186 1187config LRU_GEN_STATS 1188 bool "Full stats for debugging" 1189 depends on LRU_GEN 1190 help 1191 Do not enable this option unless you plan to look at historical stats 1192 from evicted generations for debugging purpose. 1193 1194 This option has a per-memcg and per-node memory overhead. 1195# } 1196 1197config ARCH_SUPPORTS_PER_VMA_LOCK 1198 def_bool n 1199 1200config PER_VMA_LOCK 1201 def_bool y 1202 depends on ARCH_SUPPORTS_PER_VMA_LOCK && MMU && SMP 1203 help 1204 Allow per-vma locking during page fault handling. 1205 1206 This feature allows locking each virtual memory area separately when 1207 handling page faults instead of taking mmap_lock. 1208 1209source "mm/damon/Kconfig" 1210 1211endmenu 1212